CN109902283A - A kind of information output method and device - Google Patents

A kind of information output method and device Download PDF

Info

Publication number
CN109902283A
CN109902283A CN201810415523.0A CN201810415523A CN109902283A CN 109902283 A CN109902283 A CN 109902283A CN 201810415523 A CN201810415523 A CN 201810415523A CN 109902283 A CN109902283 A CN 109902283A
Authority
CN
China
Prior art keywords
text
term vector
failure
semantic
vector
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201810415523.0A
Other languages
Chinese (zh)
Other versions
CN109902283B (en
Inventor
饶思维
刘裕良
田光见
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Huawei Technologies Co Ltd
Original Assignee
Huawei Technologies Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Huawei Technologies Co Ltd filed Critical Huawei Technologies Co Ltd
Priority to CN201810415523.0A priority Critical patent/CN109902283B/en
Priority to PCT/CN2019/084814 priority patent/WO2019210820A1/en
Publication of CN109902283A publication Critical patent/CN109902283A/en
Application granted granted Critical
Publication of CN109902283B publication Critical patent/CN109902283B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/20Natural language analysis

Abstract

This application discloses a kind of information output method and devices, wherein this method comprises: obtaining failure-description text, failure-description text is used for the failure for describing to occur in network;Pass through the semantic semantic vector for generating model and generating failure-description text;The corresponding semantic vector of related text of a plurality of types of target datas is obtained, the target data is for assisting analysis failure Producing reason;Calculate the correlation of the semantic vector and the semantic vector of the related text of every kind of target data of failure-description text;It determines and exports the first data, first data are that the target data or first data of the correlation maximum of the semantic vector of semantic vector and failure-description text in plurality of target data are the target data that semantic vector is greater than preset threshold with the correlation of the semantic vector of failure-description text in plurality of target data.By implementing the present processes, can accurately find out relevant to failure-description text for assisting the data of analyzing failure cause.

Description

A kind of information output method and device
Technical field
This application involves field of communication technology more particularly to a kind of information output methods and device.
Background technique
When the network equipment breaks down, it will affect normal communication, bring heavy losses to the work and life of people, So the reparation in time of network equipment failure is extremely important.Currently, frontline engineer can be from event when the network equipment breaks down The data that on-site collection is used to assist analyzing failure cause occur for barrier, for example, when one section before and after collection network equipment fault generation In Key Performance Indicator (KPI), equipment alarm, the supplemental characteristics such as device log.And frontline engineer can be existing to failure As being described, failure-description text is obtained.Frontline engineer is by the data such as the KPI of collection and failure-description text with failure work Single form feeds back to O&M department.O&M engineer is according to the failure-description text in fault ticket, by the profession of itself Knowledge selects the supplemental characteristics such as some KPI, equipment alarm, device log from the data that a line is collected manually.Further Ground carries out abnormality detection these data chosen and mutually proves, so that analysis is out of order root because of place, to fault network The reparation of network equipment provides guiding opinion.It is this by manually from supplemental characteristics such as KPI, equipment alarm, device logs In select the fault detection method that supplemental characteristic relevant to failure-description text check analysis, efficiency low velocity is slow, It is unable to satisfy increasingly increased network demand.
In the prior art by searching for the text with failure-description text keyword having the same, and according to the text Associated parameter data progress failure checks analysis.But the high related text that can be used to assist analyzing failure cause of correlation and It may be there is no identical keyword in failure-description text.Therefore, it cannot accurately be found and event by existing mode Barrier description text is associated for assisting the data of analyzing failure cause.
Summary of the invention
This application provides a kind of information output method and device, can automatically and accurately find and failure-description The relevant data for being used to assist analyzing failure cause of text.
In a first aspect, this application provides a kind of information output method, this method comprises: failure-description text is obtained, it should Failure-description text is used for the failure for describing to occur in network;By it is semantic generate model generate failure-description text it is semantic to Amount;The corresponding semantic vector of related text of a plurality of types of target datas is obtained, the target data is for assisting analysis Failure Producing reason;Calculate the semantic vector of the semantic vector of failure-description text and the related text of every kind of target data Correlation;The first data are determined and export, which is semantic vector and failure-description text in every kind of target data The target data of the correlation maximum of semantic vector or first data are semantic vector and failure-description in every kind of target data The correlation of the semantic vector of text is greater than the target data of preset threshold.
The application passes through the semantic vector of the semantic vector of comparison failure-description text and the related text of target data Correlation can accurately find target data associated with failure-description text.For example, failure-description is that " industry is used Family online is slow ", entitled " the downlink bandwidth control for the relative critical index for accident analysis that the application analyzes Packet loss ratio processed ".As can be seen that from it is literal it is both upper it is any can match with associated ingredient, and the application is exactly Both study, which is excavated, by semantic analysis has arrived domain knowledge as " networking speed and packet loss ratio have relationship ", just realize Associated analysis.Therefore, it by implementing method described in first aspect, can automatically and accurately find out and event The relevant data for being used to assist analyzing failure cause of barrier description text.
In a kind of possible embodiment, before obtaining failure-description text, also model can be generated by semanteme and generated The corresponding semantic vector of the related text of a plurality of types of target datas.
And it also can be reserved for the corresponding semantic vector of related text of plurality of target data;Correspondingly, it obtains a variety of The specific embodiment of the corresponding semantic vector of the related text of target data are as follows: obtain the plurality of target data of preservation The corresponding semantic vector of related text.
By implementing the embodiment, the corresponding language of related text of plurality of target data can be pre-generated and saved Adopted vector can directly use the related text of the plurality of target data saved corresponding after receiving failure-description text The semantic vector of semantic vector and failure-description text carries out correlation calculations, thus do not have to receive failure-description text it Afterwards, the corresponding semantic vector of related text of plurality of target data is temporarily generated.As it can be seen that by implementing the embodiment, Be conducive to the semantic vector of the related text of the semantic vector that failure-description text is rapidly calculated and every kind of target data Correlation.
In a kind of possible embodiment, the above-mentioned semantic model that generates is that the corresponding word of text is described according to historical failure Vector matrix training generates, which includes that historical failure describes the corresponding term vector of each word in text, the word Vector is used to indicate the semanteme of word.
The semantic generation model obtained by implementing embodiment training, can more accurately express the language of text Justice.
In a kind of possible embodiment, above-mentioned a plurality of types of target datas include Key Performance Indicator, equipment announcement At least two in police, device log;When above-mentioned target data is Key Performance Indicator, the related text of the target data is The title of Key Performance Indicator;When above-mentioned target data is equipment alarm, the related text of the target data is equipment alarm Mark;When above-mentioned target data is device log, the related text of the target data is the contents fragment of device log.
Second aspect, this application provides a kind of semantic training methods for generating model, this method comprises: obtaining training text This corresponding term vector set, the word in term vector and training text for including in the term vector set correspond, the word to Measure the semanteme for indicating word;It is to be made of at least one term vector that historical failure, which is described text conversion, according to term vector set Term vector matrix;Semantic generation model is obtained according to the training of term vector matrix, which generates model for generating text Semantic vector.
It optionally, can be by the corresponding term vector set of training text after obtaining the corresponding term vector set of training text It is saved, so as to the subsequent term vector using in term vector set.
As it can be seen that method described in second aspect is gradually to model from the semantic of lexical level to the semanteme of sentence surface Model is generated to semanteme, this semantic model training mode that generates is to meet the basic principle of language generation.Therefore, pass through reality The semantic generation model that the training of method described in second aspect obtains is applied, the semanteme of text can be more accurately expressed.
In a kind of possible embodiment, it is by least one that historical failure, which is described text conversion, according to term vector set The specific embodiment of the term vector matrix of a term vector composition are as follows: text is described to historical failure and carries out word segmentation processing, is obtained Historical failure describes the corresponding word sequence being made of at least one word of text;Word sequence is obtained from term vector set includes The corresponding term vector of word;The corresponding term vector of each word for including by word sequence forms term vector matrix.
By implementing the embodiment, historical failure accurately can be described text conversion is by least one term vector group At term vector matrix.
In a kind of possible embodiment, when there is no the corresponding term vectors of word that word sequence includes in term vector set When, generate the corresponding term vector of word that random vector includes as word sequence.
By implementing the embodiment, historical failure accurately can be described text conversion is by least one term vector group At term vector matrix.
In a kind of possible embodiment, the semantic specific embodiment party for generating model is obtained according to the training of term vector matrix Formula are as follows: obtain historical failure and describe the corresponding faulty equipment type of text;According to term vector matrix and class label training classification Model, such distinguishing label include the faulty equipment type;Semantic generation model is obtained according to disaggregated model.
The semantic generation model obtained by implementing embodiment training, can more accurately express the language of text Justice.
In a kind of possible embodiment, according to the specific implementation of term vector matrix and class label train classification models Mode are as follows: term vector matrix and class label input neural network are iterated training, in each repetitive exercise to input The parameter of term vector and neural network in the term vector matrix of neural network is adjusted, to generate the disaggregated model.Pass through The semantic generation model that embodiment training obtains, can more accurately express the semanteme of text.
Optionally, also the term vector in the term vector matrix for using last time repetitive exercise to input can be updated term vector The corresponding term vector of corresponding words in set.By implementing the embodiment, can be retouched according to the historical failure with domain knowledge The term vector in corpus of text amendment term vector set is stated, the word of domain knowledge can be expressed by making the term vector in term vector set more Semantic information.
The third aspect provides a kind of information output apparatus, which can be performed above-mentioned first aspect or On the one hand the method in possible embodiment.The function can also be executed corresponding by hardware realization by hardware Software realization.The hardware or software include one or more units corresponding with above-mentioned function.The unit can be software and/ Or hardware.Based on the same inventive concept, the principle and beneficial effect which solves the problems, such as may refer to above-mentioned In first aspect or the possible embodiment of first aspect and beneficial effect, overlaps will not be repeated.
Fourth aspect provides a kind of model training apparatus, which can be performed above-mentioned second aspect or Method in the two possible embodiments of aspect.The function can also be executed corresponding by hardware realization by hardware Software realization.The hardware or software include one or more units corresponding with above-mentioned function.The unit can be software and/ Or hardware.Based on the same inventive concept, the principle and beneficial effect which solves the problems, such as may refer to above-mentioned In second aspect or the possible embodiment of second aspect and beneficial effect, overlaps will not be repeated.
5th aspect, provides a kind of information output apparatus, which includes: processor, memory, communication Interface;Processor, communication interface are connected with memory;Wherein, communication interface can be transceiver.Communication interface for realizing with Communication between other network elements.Wherein, one or more programs are stored in memory, and processor calling is stored in this and deposits Program in reservoir is to realize the scheme in above-mentioned first aspect or the possible embodiment of first aspect, the information output apparatus The embodiment and beneficial effect solved the problems, such as may refer to above-mentioned first aspect or the possible embodiment of first aspect with And beneficial effect, overlaps will not be repeated.
6th aspect, provides a kind of model training apparatus, which includes: processor, memory, communication Interface;Processor, communication interface are connected with memory;Wherein, communication interface can be transceiver.Communication interface for realizing with Communication between other network elements.Wherein, one or more programs are stored in memory, and processor calling is stored in this and deposits Program in reservoir is to realize the scheme in above-mentioned second aspect or the possible embodiment of second aspect, the model training apparatus The embodiment and beneficial effect solved the problems, such as may refer to above-mentioned second aspect or the possible embodiment of second aspect with And beneficial effect, overlaps will not be repeated.
7th aspect, provides a kind of computer program product, when run on a computer, so that computer executes Above-mentioned first aspect, second aspect, in the possible embodiment of the possible embodiment of first aspect or second aspect Method.
Eighth aspect provides a kind of chip product of information output apparatus, executes above-mentioned first aspect or first aspect Any possible embodiment in method.
9th aspect, provides a kind of chip product of model training apparatus, executes above-mentioned second aspect or second aspect Any possible embodiment in method.
Tenth aspect, has mentioned for a kind of computer readable storage medium, is stored with instruction in computer readable storage medium, When run on a computer, so that computer executes the method for above-mentioned first aspect or the possible embodiment party of first aspect Method in formula.
Tenth on the one hand, has mentioned for a kind of computer readable storage medium, is stored with finger in computer readable storage medium It enables, when run on a computer, so that computer executes the method for above-mentioned second aspect or the possible reality of second aspect Apply the method in mode.
Detailed description of the invention
Fig. 1 is a kind of flow diagram of information output method provided by the embodiments of the present application;
Fig. 2 is a kind of flow diagram of semantic training method for generating model provided by the embodiments of the present application;
Fig. 3 is the schematic diagram for the neural network that a kind of CBOW algorithm provided by the embodiments of the present application uses;
Fig. 4 is a kind of structural schematic diagram of neural network for train classification models provided by the embodiments of the present application;
Fig. 5 is a kind of structural schematic diagram of information output apparatus provided by the embodiments of the present application;
Fig. 6 is a kind of structural schematic diagram of model training apparatus provided by the embodiments of the present application;
Fig. 7 is the structural schematic diagram of another information output apparatus provided by the embodiments of the present application;
Fig. 8 is the structural schematic diagram of another model training apparatus provided by the embodiments of the present application.
Specific embodiment
The application specific embodiment is described in further detail with reference to the accompanying drawing.
The embodiment of the present application provides a kind of information output method and device, can automatically determine and export and failure-description The relevant data for being used to assist analyzing failure cause of text.
Information output method and device provided herein are introduced in detail below.
Referring to Figure 1, Fig. 1 is a kind of flow diagram of information output method provided by the embodiments of the present application.Such as Fig. 1 institute Show, which includes following 101~105 part, in which:
101, information output apparatus obtains failure-description text.
Wherein, failure-description text is the text that phenomenon of the failure is described, i.e., failure-description text is for describing net The failure occurred in network.For example, failure-description text can be " industry user's online is slow ", " online charging server (online Charging system, OCS) communicating interrupt " etc..Failure-description text can be other devices and be sent to information output apparatus 's.For example, phenomenon of the failure is described in frontline engineer, failure-description text is obtained, and collection is used to assist to analyze The data (such as Key Performance Indicator) and failure-description text of failure cause are sent to O&M department in the form of fault ticket Information output apparatus.
102, information output apparatus passes through the semantic semantic vector for generating model and generating failure-description text.
In a kind of possible embodiment, semanteme generation model, which can be, describes the corresponding word of text according to historical failure Vector matrix training generates, which includes that historical failure describes the corresponding term vector of each word in text.
Optionally, for details, reference can be made to the training that semanteme described in Fig. 2 generates model for the semantic training method for generating model Method.That is, semanteme used in information output apparatus, which generates model, can be the language of model training apparatus training in Fig. 2 Justice generates model.The model training apparatus in information output apparatus and Fig. 2 in Fig. 1 can be deployed in same equipment or be deployed in not Same equipment.When the model training apparatus in the information output apparatus and Fig. 2 in Fig. 1 is deployed in different equipment, model Training device has trained the semantic transmittable semantic model that generates after generating model to information output apparatus, thus information output dress Setting can be by the received semantic semantic vector for generating model and generating failure-description text.When in Fig. 1 information output apparatus and When model training apparatus in Fig. 2 is disposed in the same equipment, information output apparatus can obtain language from model training apparatus Justice generates model, so that information output apparatus can pass through the semantic semantic vector for generating model and generating failure-description text.
Certainly the semantic model that generates can not also be generated by the training of mode described in Fig. 2, can also be by other means Training generates, and the embodiment of the present application is without limitation.
In a kind of possible embodiment, information output apparatus generates failure-description text by the semantic model that generates The specific embodiment of semantic vector are as follows:
Failure-description text conversion is term vector matrix according to term vector set by information output apparatus, then by term vector square Battle array input is semantic to generate model, to generate the semantic vector of failure-description text, wherein includes multiple words in the term vector set Vector.Optionally, which can be the model training apparatus in lower Fig. 2 and generates and sends to information output apparatus 's.
Optionally, failure-description text conversion is the specific of term vector matrix according to term vector set by information output apparatus Embodiment are as follows: information output apparatus carries out word segmentation processing to failure-description text, and it is corresponding by extremely to obtain failure-description text The word sequence of few word composition;The corresponding term vector of word that the word sequence includes is obtained from term vector set;By the word order The term vector matrix of the corresponding term vector composition failure-description text of each word that column include.When there is no should in term vector set When the corresponding term vector of the word that word sequence includes, the corresponding term vector of word that random vector includes as the word sequence is generated.
For example, failure-description text includes 4 words, carries out the word sequence that word segmentation processing obtains to failure-description text For " industry ", " user ", " online ", " slow ".Information output apparatus finds " industry " corresponding term vector from term vector set 1, " user " corresponding term vector 2, " online " corresponding term vector 3, do not find " slow " corresponding term vector, then generate random vector Term vector 4 is as " slow " corresponding term vector.Term vector 1~4 is formed the term vector of failure-description text by information output apparatus Matrix.The term vector Input matrix semanteme is generated to the semantic vector that failure-description text is generated in model again.
103, information output apparatus obtains the corresponding semantic vector of related text of a plurality of types of target datas.
Wherein, the target data is for assisting analysis failure Producing reason.Wherein, the execution of 103 parts and 102 parts Sequence can first carry out 102 parts and execute 103 parts again, or can first carry out 103 parts and execute 102 parts again in no particular order.
In a kind of possible embodiment, which includes Key Performance Indicator (KPI), equipment At least two in alarm, device log;When target data is Key Performance Indicator, the related text of the target data is to close The title of key performance indicator;When target data is equipment alarm, the related text of the target data is the mark of equipment alarm; When target data is device log, the related text of the target data is the contents fragment of device log.Wherein, each type Target data be it is multiple.
For example, a plurality of types of target datas include Key Performance Indicator and equipment alarm.Above-mentioned is a plurality of types of Target data is 100 different Key Performance Indicators and 20 different equipment alarms, and 100 Key Performance Indicators are respectively 1~Key Performance Indicator of Key Performance Indicator 100.20 equipment alarms are respectively equipment alarm 1~20.Information output apparatus obtains The corresponding semantic vector of the related text of a plurality of types of target datas taken is that 1~key performance of Key Performance Indicator refers to The corresponding semantic vector of title of mark 100 and the corresponding semantic vector of mark of equipment alarm 1~20.Also It is to say, information output apparatus can obtain 120 semantic vectors.
In a kind of possible embodiment, information output apparatus can pass through semanteme before receiving failure-description text Generate the corresponding semantic vector of related text that model generates a plurality of types of target datas.
Optionally, information output apparatus generates the corresponding semantic vector of related text of a plurality of types of target datas Later, it can be reserved for the corresponding semantic vector of related text of a plurality of types of target datas.Receiving failure-description text After this, so that it may obtain the corresponding semantic vector of related text for saving the plurality of target data, so as to failure-description The semantic vector of text carries out correlation calculations.By implementing the embodiment, it can pre-generate and save plurality of target data The corresponding semantic vector of related text, after receiving failure-description text, can directly with save plurality of target number According to the corresponding semantic vector of related text and failure-description text semantic vector carry out correlation calculations, to not have to After receiving failure-description text, the interim corresponding semantic vector of related text for generating plurality of target data.As it can be seen that By implementing the embodiment, be conducive to rapidly to be calculated the semantic vector of failure-description text and every kind of target data The correlation of the semantic vector of related text.
In a kind of possible embodiment, information output apparatus passes through the semantic correlation for generating model and generating target data The principle and information output apparatus of the corresponding semantic vector of text pass through the semantic semanteme for generating model and generating failure-description text The principle of vector is identical, and this will not be repeated here.
104, the semantic vector of information output apparatus calculating failure-description text and the related text of every kind of target data The correlation of semantic vector.
For example, there is two kinds of target data, respectively 100 different Key Performance Indicators and 20 are not Same equipment alarm, 100 Key Performance Indicators are respectively 1~Key Performance Indicator of Key Performance Indicator 100.20 equipment are accused Alert is respectively equipment alarm 1~20.The semantic vector that information output apparatus calculates failure-description text is key with 100 respectively The correlation of the semantic vector of the related text of energy index, and the semantic vector of calculating failure-description text are set with 20 respectively The correlation of the semantic vector of the related text of standby alarm.Therefore, 120 correlations can be obtained.
In a kind of possible embodiment, the angle of vector can be used as the measurement of correlation, failure-description text The correlation of this semantic vector and the semantic vector of the related text of target data may be expressed as:
Wherein, cos (θ) is the phase of the semantic vector and the semantic vector of the related text of target data of failure-description text Guan Xing, n are the number of dimensions of the semantic vector of the related text of failure-description text and target data, xiFor failure-description text The semantic vector of this i-th dimension, yiThe semantic vector of the related text i-th dimension of target data.
105, information output apparatus is determining and exports the first data.
Wherein, the semantic vector and every kind of mesh in every kind of target data that failure-description text is calculated in information output apparatus After the correlation for marking the semantic vector of the related text of data, information output apparatus is determining and exports the first data, this first Data are the target data of the semantic vector correlation maximum of semantic vector and failure-description text in this every kind target data, or First data are that the semantic vector correlation of semantic vector and failure-description text is greater than default threshold in this every kind target data The target data of value.
For example, the two kinds of target data obtained, respectively 100 different Key Performance Indicators and 20 differences Equipment alarm, 100 Key Performance Indicators are respectively 1~Key Performance Indicator of Key Performance Indicator 100.20 equipment alarms Respectively 1~equipment alarm of equipment alarm 20.The semantic vector of failure-description text refers to 1~key performance of Key Performance Indicator The correlation of the semantic vector of the related text of mark 100 is respectively correlation 1~100.Correlation 1 is maximum correlation, Then information output apparatus exports Key Performance Indicator 1.The semantic vector of failure-description text and 1~equipment alarm of equipment alarm 20 The correlation of semantic vector of related text be respectively correlation 101~120.Correlation 120 is maximum correlation, then believes Cease output device output equipment alarm 20.
For another example, correlation 1 and correlation 2 are the correlation greater than preset threshold, then information output apparatus output is key It can index 1 and Key Performance Indicator 2.Correlation 101 and correlation 102 are the correlation greater than preset threshold, then information exports Device output equipment alarm 1 and equipment alarm 2.
The semantic vector target data bigger with the semantic vector correlation of failure-description text, illustrate the target data with Failure-description text is more related, and user may need to check the target data with analyzing failure cause.For example, failure-description text For " ocs communicating interrupt ", entitled " the ocs communicating interrupt number " of critical index, the semantic vector of the failure-description text Very big with the semantic vector correlation of the title of critical index, user may need to check the critical index to analyze failure Reason.As it can be seen that can be found automatically relevant to failure-description text for assisting by implementing method described in Fig. 1 Help the data of analyzing failure cause.
In the prior art by searching for the text with failure-description text keyword having the same, and according to the text Associated parameter data progress failure checks analysis.But the high related text that can be used to assist analyzing failure cause of correlation and It may be there is no identical keyword in failure-description text.Therefore, it cannot accurately be found and event by existing mode Barrier description text is associated for assisting the data of analyzing failure cause.The embodiment of the present application passes through comparison failure-description text Semantic vector and the semantic vector of the related text of target data correlation, can accurately find and failure-description text This associated target data.For example, failure-description is " industry user online slow ", the embodiment of the present application analyze with its phase Entitled " downlink bandwidth controls packet loss ratio " of the critical index for accident analysis closed.As can be seen that from literal The two do not have it is any can match with associated ingredient, and the application exactly by semantic analysis excavate study arrived " online Speed and packet loss ratio have relationship " as domain knowledge, both just realize associated analysis.
Therefore, it by implementing method described in Fig. 1, can automatically and accurately find and failure-description text It is relevant for assisting the data of analyzing failure cause.
Fig. 2 is referred to, Fig. 2 is a kind of process signal of semantic training method for generating model provided by the embodiments of the present application Figure.As shown in Fig. 2, the training method that the semanteme generates model includes following 201~203 part, in which:
201, model training apparatus obtains the corresponding term vector set of training text.
Wherein, the word in the term vector and training text for including in term vector set corresponds.For example, in training text It then also include 10000 term vectors in term vector set including 10000 words.The term vector is used to indicate the semanteme of word.It can Choosing, after obtaining the corresponding term vector set of training text, it can be reserved for the corresponding term vector set of training text, so as to subsequent Use the term vector in term vector set.
Training text, that is, corpus.In a kind of possible embodiment, training text can be encyclopaedia class text.From encyclopaedia The term vector that the acquistion of class text middle school is arrived has good general semantics.
In a kind of possible embodiment, model training apparatus first pre-processes training text, by sentence cutting Word segmentation processing is carried out to every text again afterwards, the training text after being segmented, and pass through word2vec tool or other tools The corresponding term vector set of training text after obtaining participle.
For example, training text is that " mathematics is one using concepts such as the variation of symbolic language research quantitative structure and spaces Door subject.I likes mathematics ".Model training apparatus is split as two words to by training text, and respectively " mathematics is to utilize symbol One Men Xueke of the concepts such as the variation of speech research quantitative structure and space " and " I likes mathematics ".Again to the two sentences point It carry out not word segmentation processing.Training text after being segmented be " mathematics be changed using symbolic language research quantitative structure and One Men Xueke of the concepts such as space.I likes mathematics ".Model training apparatus is using word2vec tool to the training text after participle This progress traverses sentence by sentence, and traversal terminates just to have obtained the corresponding term vector of each word in training text.Model training apparatus will The term vector set of the corresponding term vector composition of each word in training text is saved.
It is corresponding that model training apparatus can obtain the training text after segmenting by word2vec tool and using CBOW algorithm Term vector set.The thought of CBOW algorithm is to predict current word by given cliction up and down.The mesh of CBOW algorithm training When mark is the context for giving some word, so that the maximum probability that the word occurs.After training, each word is obtained in output layer To a corresponding term vector.Although the idea about modeling of CBOW algorithm is an assorting process, can generate term vector this Byproduct.
For example, Fig. 3 is the schematic diagram for the neural network that CBOW algorithm uses.As shown in figure 3, the neural network is by three-layered node Structure is constituted, respectively input layer, mapping layer and output layer.Wherein, output layer includes constructed good Huffman tree.Huffman A leaf node for tree represents the term vector of a word in training text, the term vector of the corresponding word of each leaf node It is random initializtion.A weight vectors built in each nonleaf node, the dimension of the vector and the term vector of input layer are identical.
Wherein, input layer is the term vector of n-1 word around some word w (t).N is window size.For example, such as Fruit n takes 5, the first two and latter two word that n-1 word around word w (t) is word w (t).Before word w (t) Two and latter two word are respectively w (t-2), w (t-1), w (t+1), w (t+2).It is corresponding, the word of this n-1 word Vector is denoted as v (w (t-2)), v (w (t-1)), v (w (t+1)), v (w (t+2)).Input layer will arrive this n-1 term vector and be transmitted to N-1 term vector is added by mapping layer, mapping layer, i.e., by the corresponding addition of each dimension of n-1 term vector.For example, mapping Layer input is pro (t)=v (w (t-2))+v (w (t-1))+v (w (t+1))+v (w (t+2)).
The vector pro (t) that adduction obtains is input to the root node of Huffman tree by projection layer.Pro (t) is being inputted into root section After point, can calculate root node to each leaf node probability, the training process of model be desirable to obtain by root node to Up to the maximum probability of the corresponding leaf node of w (t), since in the training text of magnanimity, identical context environmental can be multiple Occur, so can constantly correct each weight vectors during traversing training training text, reaches such effect.To training After all words traversal in text is completed, the corresponding term vector of each leaf node of Huffman tree is just each of training text The corresponding term vector of word.Here " all words in training text " include dittograph in training text.
Wherein, when passing through an intermediate node every time from root node to the corresponding leaf node of word w (t), be the equal of into It has gone one time two to classify, classifier can return classifier using softmax.Wherein, the class probability of every subseries are as follows:
Wherein, θiIndicate i-th of weight vectors.Pro (t) is the sum of the term vector of context of w (t), and e is that nature is normal Number.
If containing L intermediate node by the path that root node traverses the corresponding leaf node of word w (t), these nodes On parameter composition parameter vector be [θ1, θ2, θ3..., θL], then the probability of root node to the corresponding leaf node of word w (t) is The product of the probability of two classification, i.e. probability of the root node to the corresponding leaf node of word w (t) every time are as follows:
Wherein, P (w (t) | context (w (t))) is probability of the root node to the corresponding leaf node of word w (t),Indicate that i is incremented by P (context (w (t)), θ one by one from 1 to Li) even multiply quadrature.From root node To other leaf nodes probability calculation method similarly, this will not be repeated here.
202, model training apparatus according to term vector set by historical failure describe text conversion be from least one word to Measure the term vector matrix of composition.
Specifically, it is term vector matrix that a large amount of historical failure can be described text conversion by model training apparatus.Model instruction Practice device and semantic generation model is obtained according to the training of a large amount of term vector matrix.For example, having historical failure to describe text 1~go through Historical failure, can be described 1~historical failure of text and describe text 100 to be respectively converted into term vector by history failure-description text 100 Matrix is to get to 100 term vector matrixes.Model training apparatus obtains semantic generation according to this 100 term vector matrix training Model.
In a kind of possible embodiment, historical failure is described text according to term vector set and turned by model training apparatus The specific embodiment for being changed to the term vector matrix being made of at least one term vector can be with are as follows: model training apparatus to history therefore Barrier description text carries out word segmentation processing, obtains historical failure and describes the corresponding word sequence being made of at least one word of text;From The corresponding term vector of word that the word sequence includes is obtained in term vector set;The corresponding word of each word for including by the word sequence to Amount forms the term vector matrix that the historical failure describes text.When being not present in term vector set, the word that the word sequence includes is corresponding Term vector when, produce the random vector corresponding term vector of word that includes as the word sequence.As it can be seen that by implementing the implementation Mode, historical failure accurately can be described text conversion is the term vector matrix being made of at least one term vector.
For example, it includes 4 words that historical failure, which describes text 1, describes the progress of text 1 word segmentation processing to historical failure and obtains To word order be classified as " industry ", " user ", " online ", " slow ".Model training apparatus finds " industry " from term vector set Corresponding term vector 1, " user " corresponding term vector 2, " online " corresponding term vector 3, do not find " slow " corresponding term vector, then give birth to At random vector term vector 4 as " slow " corresponding term vector.Term vector 1~4 is formed the historical failure by model training apparatus The term vector matrix 1 of text 1 is described.Other historical failures describe the principle and history that text 2~100 is converted to term vector matrix The principle that failure-description text 1 is converted to term vector matrix is identical, and this will not be repeated here.
203, model training apparatus obtains semantic generation model according to the training of term vector matrix.
Specifically, model training apparatus is after obtaining term vector matrix, can by term vector Input matrix neural network into Row training, to obtain semantic generation model.The semanteme generates the semantic vector that model is used to generate text.The semantic vector is used for Indicate the semanteme of text.
As it can be seen that method described in Fig. 2 is to obtain language from the semantic semantic gradually modeling to sentence surface of lexical level Justice generates model, and this semantic model training mode that generates is to meet the basic principle of language generation.Therefore, by implementing to scheme The semantic generation model that the training of method described in 2 obtains, can more accurately express the semanteme of text.
In a kind of possible embodiment, model training apparatus obtains semantic generation model according to the training of term vector matrix Specific embodiment are as follows: model training apparatus obtain historical failure the corresponding faulty equipment type of text is described;Model training For device according to term vector matrix and class label train classification models, such distinguishing label includes the faulty equipment type;Model instruction Practice device and semantic generation model is obtained according to disaggregated model.The semantic generation model obtained by implementing embodiment training, The semanteme of text can more accurately be expressed.
For example, historical failure, which describes the corresponding faulty equipment type of text, can be router, wireline equipment or wirelessly set It is standby etc..For example, historical failure describes the failure that the failure of text description generates for router, then it is corresponding to describe text for historical failure Faulty equipment type be router.Frontline engineer collects the corresponding faulty equipment type of each failure-description text, so By failure-description text, the corresponding faulty equipment type of failure-description text and it is used to assist the data of analyzing failure cause to add afterwards It adds in work order, and work order is sent to O&M terminal and carries out failure reason analysis.Therefore, model training apparatus can be from work order It obtains historical failure and describes the corresponding faulty equipment type of text.
Wherein, the disaggregated model that training obtains is the mould for generating the corresponding faulty equipment type of failure-description text Type.For example, by the corresponding term vector Input matrix disaggregated model of failure-description text 1, the exportable failure-description of the disaggregated model The corresponding faulty equipment type of text 1.
In a kind of possible embodiment, model training apparatus is according to term vector matrix and class label training classification mould The specific embodiment of type are as follows: term vector matrix and class label input neural network are iterated training, in each iteration The parameter of term vector and neural network in the term vector matrix of input neural network is adjusted when training, to generate classification Model.By implementing the embodiment, the disaggregated model that training can be enable to obtain accurately carries out failure-description text Classification.
Optionally, also the term vector in term vector matrix adjusted can be used to update term vector set for model training apparatus The corresponding term vector of middle corresponding words.By implementing the optional mode, can be described according to the historical failure with domain knowledge Corpus of text corrects the term vector in term vector set, makes the term vector in term vector set more and can express the word in failure field Semantic information.
For example, Fig. 4 is a kind of structural schematic diagram of neural network for train classification models.As shown in figure 4, should Neural network includes convolutional layer, pond layer and full articulamentum.The term vector matrix 1 that historical failure describes text 1 includes term vector { w1, w2, w3, w4, w5, w6 }.The dimension of each term vector is 128 dimensions.Model training apparatus obtains term vector matrix 1 Later, term vector matrix 1 is inputted into neural network.As shown in figure 4, there are two convolution kernels for the tool in neural network.Certainly in reality Can also be there are two above convolution kernel in the application of border, the embodiment of the present application is illustrated with two convolution kernels.The left side Convolution kernel 1 includes that term vector carries out convolution two-by-two to term vector matrix 1.For example, w1 and w2, which carries out convolution, obtains C1, w2 and w3 Carry out convolution obtain C2, w3 and w4 carry out convolution obtain C3, w4 and w5 progress convolution obtain C4, w5 and w6 progress convolution obtain C5.The convolution kernel 2 on the right includes that term vector carries out three or three convolution to term vector matrix 1.It is obtained for example, w1, w2 and w3 carry out convolution Convolution is carried out to C6, w2, w3 and w4 and obtains C7, and w3, w4 and w5 carry out convolution and obtain C8, and w4, w5 and w6 carry out convolution and obtain C9.Convolution also carried out to the term vector of other quantity in practical application, the embodiment of the present application with three or three convolution of convolution sum two-by-two into Row illustrates.
As it can be seen that convolution kernel 1 produces characteristic pattern (feature map) C=[C1, C2 ..., C5], convolution kernel 2 is generated One characteristic pattern C=[C6, C7, C8, C9].After model training apparatus obtains the characteristic pattern of each convolution karyogenesis, for every A characteristic pattern, by the operation of maximum pondization choose the maximum value in each dimension as current convolution karyogenesis text feature to Amount.Model training apparatus splices all Text eigenvectors, obtain final historical failure describe text 1 it is semantic to Amount.I.e. as shown in figure 4, model training apparatus chooses maximum value from first dimension of C1~C5, from the 2nd of C1~C5 the Maximum value is chosen in a dimension, chooses maximum value from the 3rd dimension of C1~C5, and so on, until from C1~C5 The 128th dimension in choose arrive maximum value.The maximum value of 128 dimensions of selection is formed convolution kernel by model training apparatus 1 corresponding Text eigenvector 1.Similarly, model training apparatus also obtains the corresponding Text eigenvector 2 of convolution kernel 2.Model instruction Practice device to splice Text eigenvector 1 and Text eigenvector 2, obtains the language that final historical failure describes text 1 Adopted vector.
The semantic vector that obtained historical failure describes text 1 is inputted full articulamentum by model training apparatus, and by history The corresponding faulty equipment type (such as router) of failure-description text 1 is used as class label, inputs full articulamentum.Model training dress It sets and is analyzed in the semantic vector that full articulamentum describes text 1 to historical failure, it is most general that analysis obtains faulty equipment type Rate is interchanger.Due to the faulty equipment class for the maximum probability that the semantic vector that historical failure describes text 1 is analyzed It is not identical that type (i.e. interchanger) and historical failure describe the corresponding class label (i.e. router) of text 1, therefore model training fills Set the faulty equipment type for the maximum probability that record is analyzed by the semantic vector for describing text 1 to historical failure not Correctly.Similarly, historical failure is described the term vector Input matrix neural network of text 2 by model training apparatus according to the above process It is trained, obtains the semantic vector that historical failure describes text 2, and it is right in full articulamentum input historical failure to describe text 2 The faulty equipment type (such as interchanger) answered is used as class label.Model training apparatus describes the semanteme of text 2 to historical failure Vector is analyzed, and it is firewall that analysis, which obtains faulty equipment type maximum probability,.Therefore, model training apparatus record by pair The faulty equipment type for the maximum probability that the semantic vector that historical failure describes text 2 is analyzed is incorrect.Assuming that tool There are 100 historical failures to describe text, remaining 98 historical failure describes text similarly, describes text according to above-mentioned historical failure This 1 mode inputs the training that neural network carries out disaggregated model.The first round is completed describing text 1~100 to historical failure After training, it is assumed that the failure for the maximum probability that the corresponding semantic vector of text 1~50 is analyzed is described according to historical failure Device type is incorrect, and model training apparatus describes the corresponding word of text 1~50 to the parameter and historical failure of neural network Term vector in vector matrix is adjusted.It is again right with the parameter of neologisms vector matrix and neural network after adjustment finishes Historical failure describes text 1~100 and is trained, until describing the corresponding semantic vector of text 1~100 according to historical failure The faulty equipment type for analyzing obtained maximum probability matches with tag along sort, just generates disaggregated model, i.e., is instructed by iteration Practice neural network next life ingredient class model.
Last model training apparatus uses the term vector more neologisms in the term vector matrix of last wheel repetitive exercise input The corresponding term vector of corresponding words in vector set.For example, it is that " networking speed is slow " last wheel changes that historical failure, which describes text 1, It describes 1 equivalent vector matrix of text to historical failure before generation training to be adjusted, by " online " corresponding term vector tune Whole is term vector 1, then corresponding using " online " in 1 substitute vector set of term vector after the completion of last time repetitive exercise Term vector.Historical failure describes to describe text to historical failure before text 2 is " OCS communicating interrupt " last wheel repetitive exercise This 2 equivalent vector matrix is adjusted, and " interruption " corresponding term vector is adjusted to term vector 2, then last wheel iteration After the completion of training, " interruption " corresponding term vector in 2 substitute vector set of term vector is used.Other historical failures description text Originally similarly, this will not be repeated here.
In a kind of possible embodiment, model training apparatus obtains the specific of semantic generation model according to disaggregated model Embodiment are as follows: part more than articulamentum complete in disaggregated model is generated model by model training apparatus.Pass through reality It applies the semantic of embodiment generation and generates model, can accurately generate the semantic vector of text.
The embodiment of the present invention can carry out the division of functional module according to above method example to equipment, for example, can be right The each functional module of each function division is answered, two or more functions can also be integrated in a module.It is above-mentioned Integrated module both can take the form of hardware realization, can also be realized in the form of software function module.It needs to illustrate , be to the division of module in the embodiment of the present invention it is schematical, only a kind of logical function partition in actual implementation may be used To there is other division mode.
Fig. 5 is referred to, Fig. 5 is that the present invention implements a kind of information output apparatus provided.The information output apparatus includes: to obtain Modulus block 501, generation module 502, computing module 503 and output module 504.Wherein:
Module 501 is obtained, for obtaining failure-description text;Generation module 502, for being generated by the semantic model that generates The semantic vector of failure-description text, the failure-description text are used for the failure for describing to occur in network;Module 501 is obtained, also The corresponding semantic vector of related text for obtaining a plurality of types of target datas, the target data is for assisting analysis Failure Producing reason;Computing module 503 is related to every kind of target data for calculating the semantic vector of failure-description text The correlation of the semantic vector of text;Output module 504, for determining and exporting the first data, the first data are every kind of target The target data or the first data of the correlation maximum of the semantic vector of semantic vector and failure-description text are every kind in data Semantic vector is greater than the target data of preset threshold with the correlation of the semantic vector of failure-description text in target data.
In a kind of possible embodiment, generation module 502 is also used to obtaining the acquisition failure-description text of module 501 Before this, pass through the semantic corresponding semantic vector of related text for generating model and generating plurality of target data.
In a kind of possible embodiment, it is that the corresponding term vector of text is described according to historical failure that semanteme, which generates model, Matrix training generates, and term vector matrix includes that historical failure describes the corresponding term vector of each word in text, which uses In the semanteme for indicating word.
In a kind of possible embodiment, a plurality of types of target datas include Key Performance Indicator, equipment alarm, At least two in device log;When target data is Key Performance Indicator, the related text of target data is key performance The title of index;When target data is equipment alarm, the related text of target data is the mark of equipment alarm;Work as number of targets When according to for device log, the related text of target data is the contents fragment of device log.
Fig. 6 is referred to, Fig. 6 is that the present invention implements a kind of model training apparatus provided.The model training apparatus includes obtaining Modulus block 601, conversion module 602 and training module 603, in which:
Module 601 is obtained, for obtaining the corresponding term vector set of training text, the term vector for including in term vector set It is corresponded with the word in training text;Conversion module 602, for historical failure to be described text conversion according to term vector set For the term vector matrix being made of at least one term vector;Training module 603 is also used to obtain language according to the training of term vector matrix Justice generates model, and semanteme generates the semantic vector that model is used to generate text.
In a kind of possible embodiment, conversion module 602 is specifically used for: describing text to historical failure and segments Processing, obtains historical failure and describes the corresponding word sequence being made of at least one word of text;Word is obtained from term vector set The corresponding term vector of the word that sequence includes;The corresponding term vector of each word for including by word sequence forms term vector matrix.
In a kind of possible embodiment, conversion module 602 also particularly useful for: when in term vector set be not present word order When the corresponding term vector of the word that column include, the corresponding term vector of word that random vector includes as word sequence is generated.
In a kind of possible embodiment, training module 603 obtains semantic generation model according to the training of term vector matrix Mode specifically: obtain historical failure the corresponding faulty equipment type of text is described;According to term vector matrix and class label Train classification models, such distinguishing label include the faulty equipment type;Semantic generation model is obtained according to disaggregated model.
In a kind of possible embodiment, training module 603 is according to term vector matrix and class label training classification mould The mode of type specifically: term vector matrix and class label input neural network are iterated training, in each repetitive exercise When in the term vector matrix of input neural network term vector and the parameter of neural network be adjusted, to generate classification mould Type.
Fig. 7 is referred to, Fig. 7 is a kind of structural schematic diagram of information output apparatus disclosed in the embodiment of the present application.Such as Fig. 7 institute Show, which includes processor 701, memory 702 and communication interface 703.Wherein, processor 701, storage Device 702 is connected with communication interface 703.
Wherein, processor 701 can be central processing unit (central processing unit, CPU), general procedure Device, coprocessor, digital signal processor (digital signal processor, DSP), specific integrated circuit (application-specific integrated circuit, ASIC), field programmable gate array (field Programmable gate array, FPGA) or other programmable logic device, transistor logic, hardware component or Person's any combination thereof.The processor 701 is also possible to realize the combination of computing function, such as includes one or more microprocessors Combination, DSP and the combination of microprocessor etc..
Wherein, communication interface 703 is for realizing the communication between other network elements.
Wherein, processor 701 calls the program code stored in memory 702, believes in executable above method embodiment Cease step performed by output device.
Fig. 8 is referred to, Fig. 8 is a kind of structural schematic diagram of model training apparatus disclosed in the embodiment of the present application.Such as Fig. 8 institute Show, which includes processor 801, memory 802 and communication interface 803.Wherein, processor 801, storage Device 802 is connected with communication interface 803.
Wherein, processor 801 can be central processing unit (central processing unit, CPU), general procedure Device, coprocessor, digital signal processor (digital signal processor, DSP), specific integrated circuit (application-specific integrated circuit, ASIC), field programmable gate array (field Programmable gate array, FPGA) or other programmable logic device, transistor logic, hardware component or Person's any combination thereof.The processor 801 is also possible to realize the combination of computing function, such as includes one or more microprocessors Combination, DSP and the combination of microprocessor etc..
Wherein, communication interface 803 is for realizing the communication between other network elements.
Wherein, processor 801 calls the program code stored in memory 802, and mould in above method embodiment can be performed Step performed by type training device.
Based on the same inventive concept, the principle and the application method that each equipment provided in the embodiment of the present application solves the problems, such as Embodiment is similar, therefore the implementation of each equipment may refer to the implementation of method, for succinct description, repeats no more herein.
In the above-described embodiments, it all emphasizes particularly on different fields to the description of each embodiment, there is no the portion being described in detail in some embodiment Point, reference can be made to the related descriptions of other embodiments.
Finally, it should be noted that the above various embodiments is only to illustrate the technical solution of the application, rather than its limitations;To the greatest extent Pipe is described in detail the application referring to foregoing embodiments, those skilled in the art should understand that: its according to So be possible to modify the technical solutions described in the foregoing embodiments, or to some or all of the technical features into Row equivalent replacement;And these are modified or replaceed, each embodiment technology of the application that it does not separate the essence of the corresponding technical solution The range of scheme.

Claims (20)

1. a kind of information output method, which is characterized in that the described method includes:
Failure-description text is obtained, the failure-description text is used for the failure for describing to occur in network;
Pass through the semantic semantic vector for generating model and generating the failure-description text;
The corresponding semantic vector of related text of a plurality of types of target datas is obtained, the target data is divided for assisting Analyse the failure Producing reason;
Calculate the phase of the semantic vector and the semantic vector of the related text of target data described in every kind of the failure-description text Guan Xing;
The first data are determined and export, first data are semantic vector and the failure-description in every kind of target data The target data of the correlation maximum of the semantic vector of text or first data be in every kind of target data semanteme to Measure the target data for being greater than preset threshold with the correlation of the semantic vector of the failure-description text.
2. the method according to claim 1, wherein the method is also before the acquisition failure-description text Include:
Pass through the semantic corresponding semantic vector of related text for generating model and generating a plurality of types of target datas.
3. method according to claim 1 or 2, which is characterized in that the semantic model that generates is retouched according to historical failure State what text corresponding term vector matrix training generated, the term vector matrix include the historical failure describe it is each in text The corresponding term vector of word, the term vector are used to indicate the semanteme of word.
4. method according to any one of claims 1 to 3, which is characterized in that a plurality of types of target packets Include Key Performance Indicator, equipment alarm, at least two in device log;When the target data is the Key Performance Indicator When, the related text of the target data is the title of the Key Performance Indicator;When the target data is equipment announcement When alert, the related text of the target data is the mark of the equipment alarm;When the target data is the device log When, the related text of the target data is the contents fragment of the device log.
5. a kind of semantic training method for generating model, which is characterized in that the described method includes:
The corresponding term vector set of training text is obtained, in the term vector and the training text for including in the term vector set Word correspond, the term vector is used to indicate the semanteme of word;
It is the term vector square being made of at least one term vector that historical failure, which is described text conversion, according to the term vector set Battle array;
Semantic generation model, the semantic semanteme for generating model and being used to generate text are obtained according to term vector matrix training Vector.
6. according to the method described in claim 5, it is characterized in that, described describe historical failure according to the term vector set Text conversion is the term vector matrix being made of at least one term vector, comprising:
Text is described to historical failure and carries out word segmentation processing, obtaining the historical failure, to describe text corresponding by least one word The word sequence of composition;
The corresponding term vector of word that the word sequence includes is obtained from the term vector set;
The corresponding term vector of each word for including by the word sequence forms term vector matrix.
7. according to the method described in claim 6, it is characterized in that, the method also includes:
When the corresponding term vector of the word sequence word that includes is not present in the term vector set, random vector conduct is generated The corresponding term vector of the word that the word sequence includes.
8. according to method described in claim 5~7 any one, which is characterized in that described to be instructed according to the term vector matrix Get semantic generation model, comprising:
It obtains the historical failure and describes the corresponding faulty equipment type of text;
According to the term vector matrix and class label train classification models, the class label includes the faulty equipment class Type;
Semantic generation model is obtained according to the disaggregated model.
9. according to the method described in claim 8, it is characterized in that, described according to the term vector matrix and the class label Train classification models, comprising:
The term vector matrix and class label input neural network are iterated training, in each repetitive exercise pair The parameter of the term vector and the neural network that input in the term vector matrix of the neural network is adjusted, described in generating Disaggregated model.
10. a kind of information output apparatus, which is characterized in that the information output apparatus includes:
Module is obtained, for obtaining failure-description text, the failure-description text is used for the failure for describing to occur in network;
Generation module, for passing through the semantic semantic vector for generating model and generating the failure-description text;
The acquisition module is also used to obtain the corresponding semantic vector of related text of a plurality of types of target datas, institute State target data for assist the analysis failure Producing reason;
Computing module, for calculating the semantic vector of the failure-description text and the related text of every kind of target data The correlation of semantic vector;
Output module, for determining and exporting the first data, first data are semantic vector in every kind of target data It is every kind of mesh with the target data of the correlation maximum of the semantic vector of the failure-description text or first data Mark the target data that semantic vector in data is greater than preset threshold with the correlation of the semantic vector of the failure-description text.
11. device according to claim 10, which is characterized in that
The generation module is also used to before the acquisition module obtains failure-description text, passes through the semantic generation mould The corresponding semantic vector of related text of type generation plurality of target data.
12. device described in 0 or 11 according to claim 1, which is characterized in that the semantic model that generates is according to historical failure Describe what the corresponding term vector matrix training of text generated, the term vector matrix includes that the historical failure describes in text respectively The corresponding term vector of a word, the term vector are used to indicate the semanteme of word.
13. device described in 0~12 any one according to claim 1, which is characterized in that a plurality of types of target datas Including at least two in Key Performance Indicator, equipment alarm, device log;When the target data is that the key performance refers to When mark, the related text of the target data is the title of the Key Performance Indicator;When the target data is the equipment When alarm, the related text of the target data is the mark of the equipment alarm;When the target data is the equipment day When will, the related text of the target data is the contents fragment of the device log.
14. a kind of model training apparatus, which is characterized in that the model training apparatus includes:
Obtain module, for obtaining the corresponding term vector set of training text, the term vector for including in the term vector set with Word in the training text corresponds;
Conversion module is by least one term vector group for historical failure to be described text conversion according to the term vector set At term vector matrix;
The training module is also used to obtain semantic generation model, the semantic generation mould according to term vector matrix training Type is used to generate the semantic vector of text.
15. device according to claim 14, which is characterized in that the conversion module is specifically used for:
Text is described to historical failure and carries out word segmentation processing, obtaining the historical failure, to describe text corresponding by least one word The word sequence of composition;
The corresponding term vector of word that the word sequence includes is obtained from the term vector set;
The corresponding term vector of each word for including by the word sequence forms term vector matrix.
16. device according to claim 15, which is characterized in that the conversion module also particularly useful for:
When the corresponding term vector of the word sequence word that includes is not present in the term vector set, random vector conduct is generated The corresponding term vector of the word that the word sequence includes.
17. device described in 4~16 any one according to claim 1, which is characterized in that the training module is according to institute's predicate Vector matrix training obtains the semantic mode for generating model specifically:
It obtains the historical failure and describes the corresponding faulty equipment type of text;
According to the term vector matrix and the class label train classification models, the class label includes the faulty equipment Type;
Semantic generation model is obtained according to the disaggregated model.
18. device according to claim 17, which is characterized in that the training module is according to the term vector matrix and institute State the mode of class label train classification models specifically:
The term vector matrix and class label input neural network are iterated training, in each repetitive exercise pair The parameter of the term vector and the neural network that input in the term vector matrix of the neural network is adjusted, described in generating Disaggregated model.
19. a kind of computer program product, which is characterized in that when run on a computer, so that computer executes above-mentioned 1 Method described in any one of~9.
20. a kind of computer readable storage medium, which is characterized in that instruction is stored in computer readable storage medium, when it When running on computers so that computer execute above-mentioned any one of 1~9 described in method.
CN201810415523.0A 2018-05-03 2018-05-03 Information output method and device Active CN109902283B (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
CN201810415523.0A CN109902283B (en) 2018-05-03 2018-05-03 Information output method and device
PCT/CN2019/084814 WO2019210820A1 (en) 2018-05-03 2019-04-28 Information output method and apparatus

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201810415523.0A CN109902283B (en) 2018-05-03 2018-05-03 Information output method and device

Publications (2)

Publication Number Publication Date
CN109902283A true CN109902283A (en) 2019-06-18
CN109902283B CN109902283B (en) 2023-06-06

Family

ID=66943185

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201810415523.0A Active CN109902283B (en) 2018-05-03 2018-05-03 Information output method and device

Country Status (2)

Country Link
CN (1) CN109902283B (en)
WO (1) WO2019210820A1 (en)

Cited By (14)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110245233A (en) * 2019-06-19 2019-09-17 北京航空航天大学 A kind of fault determination method and device
CN110378486A (en) * 2019-07-15 2019-10-25 出门问问信息科技有限公司 Internet startup disk method, apparatus, electronic equipment and storage medium
CN110569330A (en) * 2019-07-18 2019-12-13 华瑞新智科技(北京)有限公司 text labeling system, device, equipment and medium based on intelligent word selection
CN111046674A (en) * 2019-12-20 2020-04-21 科大讯飞股份有限公司 Semantic understanding method and device, electronic equipment and storage medium
CN111429155A (en) * 2020-03-25 2020-07-17 中国银行股份有限公司 Bank card dispute processing method and device
CN111651601A (en) * 2020-06-02 2020-09-11 全球能源互联网研究院有限公司 Training method and classification method for fault classification model of power information system
CN111858725A (en) * 2020-04-30 2020-10-30 北京嘀嘀无限科技发展有限公司 Event attribute determination method and system
CN112183994A (en) * 2020-09-23 2021-01-05 南方电网数字电网研究院有限公司 Method and device for evaluating equipment state, computer equipment and storage medium
CN112383421A (en) * 2020-11-03 2021-02-19 中国联合网络通信集团有限公司 Fault positioning method and device
CN112667805A (en) * 2019-09-30 2021-04-16 北京沃东天骏信息技术有限公司 Work order category determination method, device, equipment and medium
CN112925668A (en) * 2021-02-25 2021-06-08 北京百度网讯科技有限公司 Server health evaluation method, device, equipment and storage medium
CN113657022A (en) * 2021-07-15 2021-11-16 华为技术有限公司 Chip fault identification method and related equipment
CN113722494A (en) * 2021-09-10 2021-11-30 中国航空工业集团公司西安飞行自动控制研究所 Equipment fault positioning method based on natural language understanding
CN113657022B (en) * 2021-07-15 2024-05-14 华为技术有限公司 Chip fault recognition method and related equipment

Families Citing this family (27)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110909550B (en) * 2019-11-13 2023-11-03 北京环境特性研究所 Text processing method, text processing device, electronic equipment and readable storage medium
CN111078822A (en) * 2019-11-29 2020-04-28 北京百卓网络技术有限公司 Reader information extraction method and system based on Chinese novel text
CN112988921A (en) * 2019-12-13 2021-06-18 北京四维图新科技股份有限公司 Method and device for identifying map information change
CN111124925B (en) * 2019-12-25 2024-04-05 斑马网络技术有限公司 Scene extraction method, device, equipment and storage medium based on big data
CN111291564B (en) * 2020-03-03 2023-10-31 腾讯科技(深圳)有限公司 Model training method, device and storage medium for word vector acquisition
CN113495949A (en) * 2020-03-18 2021-10-12 北京沃东天骏信息技术有限公司 Method, system, computer system and medium for text recognition
CN111274366A (en) * 2020-03-25 2020-06-12 联想(北京)有限公司 Search recommendation method and device, equipment and storage medium
CN111767721A (en) * 2020-03-26 2020-10-13 北京沃东天骏信息技术有限公司 Information processing method, device and equipment
CN112749553B (en) * 2020-06-05 2023-07-25 腾讯科技(深圳)有限公司 Text information processing method and device for video file and server
CN113779975B (en) * 2020-06-10 2024-03-01 北京猎户星空科技有限公司 Semantic recognition method, device, equipment and medium
CN113822016B (en) * 2020-06-19 2024-03-22 阿里巴巴集团控股有限公司 Text data processing method and device, electronic equipment and readable storage medium
CN112069833B (en) * 2020-09-01 2024-04-30 北京声智科技有限公司 Log analysis method, log analysis device and electronic equipment
CN112507720A (en) * 2020-11-12 2021-03-16 西安交通大学 Graph convolution network root identification method based on causal semantic relation transfer
CN112463378B (en) * 2020-11-27 2023-12-22 北京浪潮数据技术有限公司 Server asset scanning method, system, electronic equipment and storage medium
CN112529104A (en) * 2020-12-23 2021-03-19 东软睿驰汽车技术(沈阳)有限公司 Vehicle fault prediction model generation method, fault prediction method and device
CN112711947B (en) * 2021-01-09 2023-08-22 国网湖北省电力有限公司电力科学研究院 Text vectorization-based fault power failure emergency repair handling reference method
CN112818008A (en) * 2021-01-21 2021-05-18 中广核工程有限公司 Intelligent diagnosis method, system, medium and electronic equipment for nuclear power debugging faults
CN113610112B (en) * 2021-07-09 2024-04-16 中国商用飞机有限责任公司上海飞机设计研究院 Auxiliary decision-making method for aircraft assembly quality defects
CN113591477B (en) * 2021-08-10 2023-09-15 平安银行股份有限公司 Fault positioning method, device, equipment and storage medium based on associated data
CN114036293B (en) * 2021-11-03 2023-06-06 腾讯科技(深圳)有限公司 Data processing method and device and electronic equipment
CN113961708B (en) * 2021-11-10 2024-04-23 北京邮电大学 Power equipment fault tracing method based on multi-level graph convolutional network
CN114265930A (en) * 2021-11-19 2022-04-01 国电南京自动化股份有限公司 Low-voltage user fault report repairing and processing method based on event extraction
CN114218402A (en) * 2021-12-17 2022-03-22 迈创企业管理服务股份有限公司 Method for recommending computer hardware fault replacement part
CN115687031A (en) * 2022-11-15 2023-02-03 北京优特捷信息技术有限公司 Method, device, equipment and medium for generating alarm description text
CN115994217B (en) * 2022-11-29 2024-01-23 南京审计大学 Financial report fraud detection method and system
CN116502058B (en) * 2023-06-28 2023-09-26 长园深瑞能源技术有限公司 AI fault detection analysis method and system applied to charging pile system and cloud platform
CN116738323B (en) * 2023-08-08 2023-10-27 北京全路通信信号研究设计院集团有限公司 Fault diagnosis method, device, equipment and medium for railway signal equipment

Citations (18)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2003173286A (en) * 2001-12-05 2003-06-20 Nippon Telegr & Teleph Corp <Ntt> Method of acquiring sematic information on distributed network
CN101795210A (en) * 2010-01-11 2010-08-04 浪潮通信信息系统有限公司 Method for processing communication network failure
KR101078751B1 (en) * 2011-02-23 2011-11-02 한국과학기술정보연구원 Method and apparatus for detecting association errors of lexical semantic networks
CN102650960A (en) * 2012-03-31 2012-08-29 奇智软件(北京)有限公司 Method and device for eliminating faults of terminal equipment
US20120232905A1 (en) * 2011-03-10 2012-09-13 GM Global Technology Operations LLC Methodology to improve failure prediction accuracy by fusing textual data with reliability model
US20130110496A1 (en) * 2011-10-28 2013-05-02 Sap Ag Calculating Term Similarity Using A Meta-Model Semantic Network
US20130339787A1 (en) * 2012-06-15 2013-12-19 International Business Machines Coporation Systematic failure remediation
CN104361026A (en) * 2014-10-22 2015-02-18 北京航空航天大学 Failure knowledge storage and push method for FMEA (failure mode and effects analysis) process
CN106326346A (en) * 2016-08-06 2017-01-11 上海高欣计算机系统有限公司 Text classification method and terminal device
CN106815252A (en) * 2015-12-01 2017-06-09 阿里巴巴集团控股有限公司 A kind of searching method and equipment
CN106941423A (en) * 2017-04-13 2017-07-11 腾讯科技(深圳)有限公司 Failure cause localization method and device
CN107171819A (en) * 2016-03-07 2017-09-15 北京华为数字技术有限公司 A kind of network fault diagnosis method and device
CN107248927A (en) * 2017-05-02 2017-10-13 华为技术有限公司 Generation method, Fault Locating Method and the device of fault location model
CN107291699A (en) * 2017-07-04 2017-10-24 湖南星汉数智科技有限公司 A kind of sentence semantic similarity computational methods
CN107291693A (en) * 2017-06-15 2017-10-24 广州赫炎大数据科技有限公司 A kind of semantic computation method for improving term vector model
CN107340766A (en) * 2017-07-10 2017-11-10 浙江大学 Power scheduling alarm signal text based on similarity sorts out and method for diagnosing faults
CN107391727A (en) * 2017-08-01 2017-11-24 北京航空航天大学 The method for digging and device of equipment fault sequence pattern
CN107704563A (en) * 2017-09-29 2018-02-16 广州多益网络股份有限公司 A kind of question sentence recommends method and system

Family Cites Families (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103617157B (en) * 2013-12-10 2016-08-17 东北师范大学 Based on semantic Text similarity computing method
CN103744905B (en) * 2013-12-25 2018-03-30 新浪网技术(中国)有限公司 Method for judging rubbish mail and device

Patent Citations (18)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2003173286A (en) * 2001-12-05 2003-06-20 Nippon Telegr & Teleph Corp <Ntt> Method of acquiring sematic information on distributed network
CN101795210A (en) * 2010-01-11 2010-08-04 浪潮通信信息系统有限公司 Method for processing communication network failure
KR101078751B1 (en) * 2011-02-23 2011-11-02 한국과학기술정보연구원 Method and apparatus for detecting association errors of lexical semantic networks
US20120232905A1 (en) * 2011-03-10 2012-09-13 GM Global Technology Operations LLC Methodology to improve failure prediction accuracy by fusing textual data with reliability model
US20130110496A1 (en) * 2011-10-28 2013-05-02 Sap Ag Calculating Term Similarity Using A Meta-Model Semantic Network
CN102650960A (en) * 2012-03-31 2012-08-29 奇智软件(北京)有限公司 Method and device for eliminating faults of terminal equipment
US20130339787A1 (en) * 2012-06-15 2013-12-19 International Business Machines Coporation Systematic failure remediation
CN104361026A (en) * 2014-10-22 2015-02-18 北京航空航天大学 Failure knowledge storage and push method for FMEA (failure mode and effects analysis) process
CN106815252A (en) * 2015-12-01 2017-06-09 阿里巴巴集团控股有限公司 A kind of searching method and equipment
CN107171819A (en) * 2016-03-07 2017-09-15 北京华为数字技术有限公司 A kind of network fault diagnosis method and device
CN106326346A (en) * 2016-08-06 2017-01-11 上海高欣计算机系统有限公司 Text classification method and terminal device
CN106941423A (en) * 2017-04-13 2017-07-11 腾讯科技(深圳)有限公司 Failure cause localization method and device
CN107248927A (en) * 2017-05-02 2017-10-13 华为技术有限公司 Generation method, Fault Locating Method and the device of fault location model
CN107291693A (en) * 2017-06-15 2017-10-24 广州赫炎大数据科技有限公司 A kind of semantic computation method for improving term vector model
CN107291699A (en) * 2017-07-04 2017-10-24 湖南星汉数智科技有限公司 A kind of sentence semantic similarity computational methods
CN107340766A (en) * 2017-07-10 2017-11-10 浙江大学 Power scheduling alarm signal text based on similarity sorts out and method for diagnosing faults
CN107391727A (en) * 2017-08-01 2017-11-24 北京航空航天大学 The method for digging and device of equipment fault sequence pattern
CN107704563A (en) * 2017-09-29 2018-02-16 广州多益网络股份有限公司 A kind of question sentence recommends method and system

Cited By (18)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110245233A (en) * 2019-06-19 2019-09-17 北京航空航天大学 A kind of fault determination method and device
CN110378486A (en) * 2019-07-15 2019-10-25 出门问问信息科技有限公司 Internet startup disk method, apparatus, electronic equipment and storage medium
CN110569330A (en) * 2019-07-18 2019-12-13 华瑞新智科技(北京)有限公司 text labeling system, device, equipment and medium based on intelligent word selection
CN112667805A (en) * 2019-09-30 2021-04-16 北京沃东天骏信息技术有限公司 Work order category determination method, device, equipment and medium
CN112667805B (en) * 2019-09-30 2024-04-09 北京沃东天骏信息技术有限公司 Work order category determining method, device, equipment and medium
CN111046674A (en) * 2019-12-20 2020-04-21 科大讯飞股份有限公司 Semantic understanding method and device, electronic equipment and storage medium
CN111429155A (en) * 2020-03-25 2020-07-17 中国银行股份有限公司 Bank card dispute processing method and device
CN111858725A (en) * 2020-04-30 2020-10-30 北京嘀嘀无限科技发展有限公司 Event attribute determination method and system
CN111651601B (en) * 2020-06-02 2023-04-18 全球能源互联网研究院有限公司 Training method and classification method for fault classification model of power information system
CN111651601A (en) * 2020-06-02 2020-09-11 全球能源互联网研究院有限公司 Training method and classification method for fault classification model of power information system
CN112183994A (en) * 2020-09-23 2021-01-05 南方电网数字电网研究院有限公司 Method and device for evaluating equipment state, computer equipment and storage medium
CN112383421A (en) * 2020-11-03 2021-02-19 中国联合网络通信集团有限公司 Fault positioning method and device
CN112383421B (en) * 2020-11-03 2023-03-24 中国联合网络通信集团有限公司 Fault positioning method and device
CN112925668B (en) * 2021-02-25 2024-04-05 北京百度网讯科技有限公司 Method, device, equipment and storage medium for evaluating server health
CN112925668A (en) * 2021-02-25 2021-06-08 北京百度网讯科技有限公司 Server health evaluation method, device, equipment and storage medium
CN113657022A (en) * 2021-07-15 2021-11-16 华为技术有限公司 Chip fault identification method and related equipment
CN113657022B (en) * 2021-07-15 2024-05-14 华为技术有限公司 Chip fault recognition method and related equipment
CN113722494A (en) * 2021-09-10 2021-11-30 中国航空工业集团公司西安飞行自动控制研究所 Equipment fault positioning method based on natural language understanding

Also Published As

Publication number Publication date
CN109902283B (en) 2023-06-06
WO2019210820A1 (en) 2019-11-07

Similar Documents

Publication Publication Date Title
CN109902283A (en) A kind of information output method and device
Li et al. The emerging graph neural networks for intelligent fault diagnostics and prognostics: A guideline and a benchmark study
Arora Comparative analysis of classification algorithms on different datasets using WEKA
JP2021182430A (en) Method for training semantic analysis model, device, electronic apparatus, and storage medium
CN111858123A (en) Fault root cause analysis method and device based on directed graph network
CN109614495B (en) Related company mining method combining knowledge graph and text information
CN110516757A (en) A kind of transformer fault detection method and relevant apparatus
CN109471793A (en) A kind of webpage automatic test defect positioning method based on deep learning
CN113312447A (en) Semi-supervised log anomaly detection method based on probability label estimation
CN112217674A (en) Alarm root cause identification method based on causal network mining and graph attention network
Matas Comparing Network Centrality Measures as Tools for Identifying Key Concepts in Complex Networks: A Case of Wikipedia.
CN111709244A (en) Deep learning method for identifying causal relationship of contradictory dispute events
CN106874963A (en) A kind of Fault Diagnosis Method for Distribution Networks and system based on big data technology
Khoshraftar et al. Dynamic graph embedding via lstm history tracking
Pavlenko et al. Criterion of cyber-physical systems sustainability
CN115328782A (en) Semi-supervised software defect prediction method based on graph representation learning and knowledge distillation
CN113268370A (en) Root cause alarm analysis method, system, equipment and storage medium
CN112015890B (en) Method and device for generating movie script abstract
Qi et al. Solving asymmetric decision problems with influence diagrams
Fu et al. Nonlinear dynamic measurement method of software reliability based on data mining
CN113378178A (en) Deep learning-based graph confidence learning software vulnerability detection method
CN110908919B (en) Response test system based on artificial intelligence and application thereof
CN112507720A (en) Graph convolution network root identification method based on causal semantic relation transfer
Pokorádi Methodolody of Advanced Graph Model-based Vehicle Systems' Analysis
CN116578336A (en) Software clone detection method based on plagiarism-detector countermeasure

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant