CN107818076A - For the semantic processes of natural language - Google Patents

For the semantic processes of natural language Download PDF

Info

Publication number
CN107818076A
CN107818076A CN201610818984.3A CN201610818984A CN107818076A CN 107818076 A CN107818076 A CN 107818076A CN 201610818984 A CN201610818984 A CN 201610818984A CN 107818076 A CN107818076 A CN 107818076A
Authority
CN
China
Prior art keywords
project
quantization means
semantic
group
item sets
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201610818984.3A
Other languages
Chinese (zh)
Other versions
CN107818076B (en
Inventor
秦涛
刘铁岩
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Microsoft Technology Licensing LLC
Original Assignee
Microsoft Technology Licensing LLC
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Microsoft Technology Licensing LLC filed Critical Microsoft Technology Licensing LLC
Priority to CN201610818984.3A priority Critical patent/CN107818076B/en
Publication of CN107818076A publication Critical patent/CN107818076A/en
Application granted granted Critical
Publication of CN107818076B publication Critical patent/CN107818076B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/30Semantic analysis
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/044Recurrent networks, e.g. Hopfield networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Health & Medical Sciences (AREA)
  • Health & Medical Sciences (AREA)
  • Artificial Intelligence (AREA)
  • General Physics & Mathematics (AREA)
  • Computational Linguistics (AREA)
  • General Engineering & Computer Science (AREA)
  • Biomedical Technology (AREA)
  • Evolutionary Computation (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • Data Mining & Analysis (AREA)
  • Biophysics (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Machine Translation (AREA)

Abstract

In embodiment of the disclosure, it is proposed that a kind of method and apparatus of semantic processes for natural language.After acquisition includes the Item Sets of multiple projects, it is determined that an each quantization means of the project on one group of semantic dimension and another quantization means in another group of dimension, then generate the semantic values of each project using the quantization means in this two groups of dimensions.In accordance with an embodiment of the present disclosure, semantic values can be used in identifying project concentration disparity items between semantic dependency, and each quantization means can be shared by multiple projects in Item Sets.Therefore, same quantization means are shared by multiple projects, the disclosure can not only effectively reduce the size of semantic model, and can significantly improve the speed of semantic processes during the semantic processes for natural language.

Description

For the semantic processes of natural language
Background technology
Natural language processing refers to the technology using computer disposal human language, and it enables a computer to understand the mankind Language.Computer is trained by the corpus for manually marking or not marking, generates the semantic expressiveness of natural language.Natural language Speech processing is an important directions in machine learning field, and it can apply to semantic analysis, information retrieval, machine translation, language Speech modeling and chat robots etc..
Recognition with Recurrent Neural Network (Recurrent Neural Networks, RNN) is a kind of god of node orientation connection cyclization Through network, its internal state can be expressed as dynamic time sequence data.Different from the acyclic neutral net of in general, RNN can profit The list entries of arbitrary sequence is handled with its internal memory.RNN can be remembered to information above and is applied to current In the calculating of output, i.e., the node between hidden layer in neutral net has connection, and the input of hidden layer is not only including defeated Entering the output of layer also includes the output of last moment hidden layer.Therefore, RNN is suitable for use in the word or text of prediction subsequent time Using such as text generation, machine translation, speech recognition and iamge description.However, current RNN semantic processes process needs Take a large amount of storage spaces and processing speed is slow.
The content of the invention
Inventors noted that can obtain a large amount of corpus data from network, these data are easy to collect.Base Most semantic scenes can be covered in the semantic model that a large amount of corpus data are trained, it is thus possible to effectively applied In being applied to actual natural language processing.Based on this understanding, each project (example is represented different from the use of one group of multi-C vector Such as word) semantic values conventional method, embodiment of the disclosure represents item using two groups or more multi-C vector Purpose semantic values.By means of this two groups or more multi-C vector, the model size and processing speed of semantic model can Optimised, this is all markedly different from any known arrangement in operation principle and mechanism.
For example, in accordance with an embodiment of the present disclosure, the Item Sets for including multiple projects can be obtained.Then, using two with On subvector represent the semantic vector of each project together, wherein semantic vector can be used in identifying project the difference of concentration Semantic dependency between project, and each subvector can be shared by multiple projects in Item Sets respectively.Therefore, pass through So that multiple projects share same subvector, embodiment of the disclosure is during the semantic processes for natural language, not only The size of semantic model can be effectively reduced, and the speed of semantic processes can be significantly improved.
It is their specific realities below in order to which simplified form introduces the selection to concept to provide Summary Applying in mode to be further described.Summary is not intended to identify the key feature or principal character of the disclosure, is also not intended to Limit the scope of the present disclosure.
Brief description of the drawings
With reference to accompanying drawing and refer to described further below, above and other feature, advantage and the aspect of each embodiment of the disclosure It will be apparent.In the accompanying drawings, same or analogous accompanying drawing mark represents same or analogous element, wherein:
Fig. 1 is shown in which that the block diagram of computing system/server of one or more other embodiments of the present disclosure can be implemented;
Fig. 2 shows the flow chart of the method for generative semantics value in accordance with an embodiment of the present disclosure;
Fig. 3 A and 3B show the exemplary plot of the table for Item Sets in accordance with an embodiment of the present disclosure;
Fig. 4 shows the flow chart for being used for the method by multiple allocations of items in table in accordance with an embodiment of the present disclosure;
Fig. 5 shows the exemplary plot of some rows in the table for Item Sets in accordance with an embodiment of the present disclosure;
Fig. 6 shows the flow chart for being used to determine the method for associated project in accordance with an embodiment of the present disclosure;And
Fig. 7 shows the exemplary plot of the prediction process of the semantic model based on RNN in accordance with an embodiment of the present disclosure.
In all of the figs, same or similar reference numeral represents same or similar element.
Embodiment
Embodiment of the disclosure is more fully described below with reference to accompanying drawings.Although some of the disclosure are shown in accompanying drawing Embodiment, it should be understood that, the disclosure can be realized by various forms, and should not be construed as being limited to this In the embodiment that illustrates, it is in order to more thorough and be fully understood by the disclosure conversely to provide these embodiments.It should be understood that It is being given for example only property of the drawings and Examples effect of the disclosure, is not intended to limit the protection domain of the disclosure.
Usually, the semantic modeling process for the natural language being described herein is considered a kind of machine learning Journey.Machine learning is that a kind of automatically analyzed from data obtains rule, and the algorithm that assimilated equations are predicted to unknown data. Term as used herein " semantic model " refers to that foundation is known with associated priori such as the syntaxes, grammer, morphology of specific languages Know the model established, it is determined for the semantic relevance between word.Term as used herein " training process " or " learning process " refers to the process of optimize semantic model using corpus or training data.For example, semantic model can pass through Training learns gradually to optimize the semantic values of each project, so as to improve the accuracy of semantic model.Above and below the disclosure Wen Zhong, to discuss convenient purpose, term " training " or " study " are interchangeable." vector " used herein is also referred to as For " embedded vector (embedding vector) ", it is used for the Semantic mapping of each project into hyperspace to form language Adopted quantization means.
Terms used herein " comprising " and its deformation are that opening includes, i.e., " include but is not limited to ".Term "based" It is " being based at least partially on ".Term " one embodiment " expression " at least one embodiment ";Term " another embodiment " represents " at least one further embodiment ".The related definition of other terms provides in will be described below.
Traditionally, in machine-learning process, semantic model is generally using corpus in Item Sets (such as vocabulary) Project (such as word or other character strings) be trained to determine the semantic values of each project, the semantic values are generally by vector Represent.However, when the number of the project in corpus is too big, the size of semantic model will also become very huge.For example, work as When training the semantic model based on RNN, when multiple projects in Item Sets be present, it is necessary to generate a term vector (one-hot Vector), its dimension is equal to the number of the project in Item Sets, and each project is generally also provided with respective embedded vector.For example, When the number of the project in Item Sets reaches tens million of ranks, the size of semantic model is up to tens GB (GB), The size of the memory of existing computing device (such as GPU and mobile computing device include mobile phone and tablet personal computer) is far more than, So that existing computing device to semantic model according to can not be trained.Therefore, traditional Semantic Modeling Method can cause language The size of adopted model is excessively huge.
Traditional method is directed to each project in vocabulary, generates a high dimension vector for representing its semantic values. For example, table 1 shows traditional vector representation for project, vector x is generated for project " January "1, for project " February " generates vector x2Deng wherein vector x1And x2Etc. being semantic expressiveness of the corresponding project in hyperspace.It is known , semantic association degree that the distance between vector can be between expression project.That is, the distance between vector is shorter, table Relevance is higher between showing two projects, for example, due to " January " and " one " semantic relevance be more than " January " and The semantic relevance of " two ", therefore, vector x1With x5The distance between be less than vector x1With x6The distance between.
Table 1
Vector Project
x1 January
x2 February
x5 one
x6 two
In addition, in the semantic model based on RNN, in the next item after predicting current project, it is necessary to computational item Mesh concentrates the probability of each project.However, when the number of the project in Item Sets reaches tens million of ranks, the place of semantic model Reason speed will also become very slow.Even if use current most fast processor, it is also possible to need to take up to for tens years To complete the training process of so huge Item Sets.Therefore, when traditional Semantic Modeling Method needs to consume very long training Between.
Therefore, the present disclosure proposes a kind of Semantic Modeling Method based on multigroup dimension.In accordance with an embodiment of the present disclosure, make The semantic values of each project are generated with the quantization means in two groups or more dimension so that the quantization table in every group of dimension Showing can be shared by multiple projects in Item Sets.As an example, table 2 show in accordance with an embodiment of the present disclosure be directed to project Vector representation, each project represents that such as project " January " is by vector x by two vector joints1And y1Joint expression, item Mesh " February " is by vector x1And y2Joint expression, wherein vector x1And x2For the quantization means in first group of dimension, and it is vectorial y1And y2For the quantization means in second group of dimension.
Table 2
Vector Project
(x1, y1) January
(x1, y2) February
(x2, y1) one
(x2, y2) two
In this way, share same quantization means by multiple projects, for example, project " January " and " February " shares the vector x in first group of dimension1, the disclosure is during the semantic modeling for natural language, not only The size of semantic model can be effectively reduced, and the processing speed of semantic model can be significantly improved.
Illustrate the general principle of the disclosure and some sample implementations below with reference to Fig. 1 to Fig. 7.Fig. 1 shows it In can implement one or more other embodiments of the present disclosure computing system/server 100 block diagram.It should be appreciated that shown in Fig. 1 What the computing system/server 100 gone out was merely exemplary, without should form to the function of embodiment described herein and Any restrictions of scope.
As shown in figure 1, computing system/server 100 is the form of universal computing device.Computing system/server 100 Component can include but is not limited to one or more processors or processing unit 110, memory 120, storage device 130, one Or multiple communication units 140, one or more input equipments 150 and one or more output equipments 160.Processing unit 110 It can be reality or virtual processor and various processing can be performed according to the program stored in memory 120.In many places Manage device system in, multiple processing unit for parallel execution computer executable instructions, with improve computing system/server 100 and Row disposal ability.
Computing system/server 100 generally includes multiple computer-readable storage mediums.Such medium can calculate system It is the addressable any medium that can be obtained of system/server 100, including but not limited to volatibility and non-volatile media, removable Unload and non-dismountable medium.Memory 120 can be volatile memory (such as register, cache, random access storage Device (RAM)), nonvolatile memory is (for example, read-only storage (ROM), Electrically Erasable Read Only Memory (EEPROM), flash memory) or their certain combination.Storage device 130 can be detachable or non-removable medium, and can So that including machine readable media, such as flash drive, disk or any other medium, it can be used in storage information And/or data (such as data 170) and it can be accessed in computing system/server 100.
Computing system/server 100 may further include other detachable/non-dismountable, volatile, nonvolatile Storage medium.Although not shown in FIG. 1, can provide for from detachable, non-volatile magnetic disk being read out or writing Disk drive and the disc drives for being read out or writing from detachable, anonvolatile optical disk.In such cases, each Driving can be connected to bus (not shown) by one or more data media interfaces.Memory 120 can include one or Multiple program products 122, it has one or more program module collections, and these program modules are configured as performing and retouched herein The function for the various embodiments stated.
Communication unit 140 is realized and communicated by communication media with other computing device.Additionally, computing system/ The function of the component of server 100 can realize that these computing machines can with single computing cluster or multiple computing machines By being communicated.Therefore, computing system/server 100 can use and other one or more servers, net The logic of network personal computer (PC) or another general networking node is connected to be operated in networked environment.
Input equipment 150 can be one or more various input equipments, such as mouse, keyboard, trackball etc..Output is set Standby 160 can be one or more output equipments, such as display, loudspeaker, printer etc..Computing system/server 100 is also It can be communicated as desired by communication unit 140 with one or more external equipment (not shown), external equipment is such as Storage device, display device etc., make it that user is led to the equipment that computing system/server 100 interacts with one or more Letter, or with causing any equipment of other computing device communications of computing system/server 100 with one or more (for example, net Card, modem etc.) communicated.Such communication can perform via input/output (I/O) interface (not shown).
As shown in figure 1, being stored with data 170 in storage device 130, it includes training data 172 (such as corpus), meter Calculating systems/servers 100 can be trained using training data, so as to export training result 180 by output equipment 160. In logic, training result 180 can be considered as one or more tables, and the table is including each project in training data in multigroup dimension On quantization means (such as vector).Of course, it should be noted that it is merely exemplary in the form of a table to export training result 180 , it is not intended to the scope of the present disclosure is limited in any way.Any other additional or alternative data format is all feasible.
Some example embodiments based on the generation table 180 of training data 172 have been described below in detail.Fig. 2 shows basis The flow chart of the method 200 for generative semantics value of embodiment of the disclosure.It should be appreciated that method 200 can be by with reference to figure 1 Described processing unit 110 performs.
202, the Item Sets for including multiple projects are obtained.For example, can be from storage device 130, communication unit 140, defeated Enter the grade Network Capture corpus of equipment 150 as training data.It is stored in corpus in the actual use of language and truly occurs The linguistic data crossed, it generally includes the data set of a large amount of texts by arrangement.In corpus, the project quilt that each occurs As a word example (token).In certain embodiments, deduplication operation can be carried out to word example, and by the word example after duplicate removal As multiple projects in Item Sets.In certain embodiments, Item Sets can include:Word, such as English " one ", " January ", Chinese " year ", " hour ", etc.;Phrase, such as " day-and-a-half ";And/or some other character string, For example, numeral, network address, digital alphabet combination etc..
204, each project for concentration of identifying project is on one group of semantic dimension (being referred to as " first group of semantic dimension ") Quantization means (being referred to as " the first quantization means ") and the quantization on another group of semantic dimension (being referred to as " second group of semantic dimension ") Represent and (be referred to as " the second quantization means ").
In certain embodiments, quantization means may be implemented as vector.For the sake of for convenience of description, hereafter will mainly it describe Such embodiment.Of course it is to be understood that what this was merely exemplary, it is not intended to limit the scope of the present disclosure in any way.Appoint What his appropriate data format is also feasible.In addition, for ease of distinguishing, vector corresponding with quantization means is referred to as " son Vector ".For example, the first quantization means may be implemented as the first subvector x, and the second quantization means may be implemented as second Subvector y.
Term " one group of semantic dimension " as used herein represents the hyperspace of certain dimension.Correspondingly, the quantization of project Represent to refer to the semantic quantization means of the semantic values of project in hyperspace, every one-dimensional expression project in semantic dimension exists Sub- semantic values in the dimension.Alternatively, first group of semantic dimension and second group of semantic dimension can have identical dimension, example Such as 1000 dimensions or 1024 dimensions, and first group of semantic dimension can be the row in table, and second group of semantic dimension can be in table Row.For example, it may be determined that 1024 the first subvector x tieed up of each project and the second subvector of 1024 dimensions in Item Sets y.Further below by with reference to the example embodiment that the first subvector and the second subvector are determined in the description of figure 4 action 204.
206, at least generated based on the first quantization means and the second quantization means each project semantic values (such as to Amount), such as the semantic values of each project can be expressed as to vectorial (x, y).In accordance with an embodiment of the present disclosure, the vector generated It can be used in the semantic dependency identified project between the sundry item in Item Sets.As described above, for example, between vector Distance can be between expression project the distance between semantic association degree, vector it is shorter, represent relevance between two projects It is higher, that is to say, that to be quantized by the semantic relation between vectorial project.
For example, the semanteme that " image " is similar with " picture " expression, then the vector distance between them is relatively short.Example again Such as, because the relation of " Beijing " and " China " is similar to " Washington " and " U.S. " relation, therefore, for " Beijing " vector with The distance between vector for " China " is substantially equal to for the vector in " Washington " and between the vector in " U.S. " Distance.
In addition, as shown in Table 2 above, the first subvector x and/or the second subvector y are total to by multiple projects in Item Sets Enjoy, that is, multiple projects share the first subvector x, and multiple projects share the second subvector y.
Although it should be appreciated that only show to represent the semantic values of each project by two subvectors in method 200, but Represent that the semantic values of each project are also possible by more than two subvectors.For example, three subvectors can be passed through (x, y, z) represents the semantic values of each project, and wherein z represents another quantization table of each project on the 3rd group of semantic dimension Show.That is, the protection domain of the disclosure is not limited by subvector number.The concentration in addition, although method 200 is identified project Each item object vector, but the vector in a subset for the concentration that can also only identify project.
In certain embodiments, the subvector and second group of language of first group of semantic dimension can be represented using the mode of table The subvector of adopted dimension, the representation of other certain data formats is also feasible.With reference to figure 3A, show according to the disclosure Embodiment the table 300 for Item Sets.The example embodiment of generation table 300 is described in detail below with reference to Fig. 4.Herein In described embodiment, the vector x in table 300 is usedrAnd vector xcJoint represents a project, thus vector xrAnd xc Subvector is collectively referred to as, or is known respectively as row vector and column vector.
As shown in Figure 3A, by all items composition in Item Sets into table 300, each position correspondence in table 300 is in one Individual project.For example, in table 300, the vectorial subvector by the first row of word " January "With the subvector of first rowThe two joint expression, the vectorial subvector by the first row of word " February "With the subvector of secondary seriesThe two Joint represents.As can be seen here, the subvector of the first rowAt least shared by word " January " and " February ".Namely Say, subvector of all items in table 300 in i-th row on first group of semantic dimension isAnd all items in jth row Subvector of the mesh on second group of semantic dimension beTherefore, the word in table 300 in ith row and jth column is by subvectorWithThe two joint represents.
Fig. 3 B show the table 350 for Item Sets in accordance with an embodiment of the present disclosure.As shown in Figure 3 B, in table 350, The word of ith row and jth column opening position by the i-th row subvectorWith the subvector of jth rowThe two joint represent, i.e., to AmountWhereinWithIt is higher-dimension (such as 1024 dimensions) vector.For example, in accordance with an embodiment of the present disclosure, word The vector of " January " can be expressed asAnd the vector of word " February " can be expressed as
Traditionally, it is used to represent semantic values, it is necessary to generate one for each project for the Item Sets with V project Vector, then need to amount to V it is vectorial.However, as described by method 200 and Fig. 3 A-3B, in accordance with an embodiment of the present disclosure Method needs only at leastIndividual row vector andIndividual column vector, i.e., altogetherIndividual vector.In some cases, in table When row and column in 300 is unequal, vectorial number is slightly larger thanBut when V value is ten million rank, the disclosure to Amount number is far smaller than the number of vectors of traditional method.Therefore, the disclosure can effectively reduce the number of vector, so as to effectively Reduce the size of semantic model, for example, the disclosure can reduce the size of semantic model from traditional 80 GB (GB) To 70 Mbytes (MB).
Fig. 4 shows the flow for being used for the method 400 by multiple allocations of items in table in accordance with an embodiment of the present disclosure Figure.The description of method 400 is distributed project in table (such as table 300) exactly based on training data, and determines often to go in table With the vector value of each column.It should be appreciated that method 400 can be the son of action 204 in the method 200 above with reference to described by Fig. 2 Action, and can be as being performed with reference to the processing unit 110 described by figure 1.
402, by multiple project organizations in Item Sets into table so that the project being in a line in table 300 has There is identical row vector, and the project in same row has identical column vector.In some embodiments, can there will be phase Same a line that the project of same prefix is initially distributed in table, and the project with identical suffix is initially distributed Same row in table.In certain embodiments, for English word or character string, prefix represents English word or character string Front portion, and the rear portion of postfix notation English word or character string, for example, by word " react ", " real " and " return " is distributed in same a line, and word " billion ", " million " and " trillion " is distributed in same row. Next, in 404-408, dispensing position of the training dataset come adjusted iterm in table 300 is utilized.
Specifically, 404, based on dispensing position renewal vector.For example, being based on existing dispensing position, language material is utilized Storehouse is trained to all row vectors in table 300 and all column vectors, until row vector and/or column vector become to restrain.Example Such as, similar project " image " and the vector value of " picture " can be updated so that their vector space distance is shorter.Therefore, All or part of row vector x in table 300rWith column vector xcValue can be updated, for example, row vectorValue from (0.025,0.365,0.263 ...) be updated to (0.012,0.242,0.347 ...).
406, based on the vector adjustment dispensing position after renewal.In certain embodiments, based on being had determined in 404 Table 300 in all row vectors and column vector value, dispensing position of the corpus to all items in table 300 can be utilized It is adjusted, make it that the loss function of all items is minimum.Loss function is that one kind of optimization aim is used to learn or train The form of model, loss function can represent the measurement of the overall loss for model optimization, wherein loss can represent to miss The various factors such as error in classification.Distribution position can be adjusted by minimizing the negative log-likelihood function of next word in sequence Put, for example, in the context with T word, overall negative log-likelihood function NLL can be represented by formula (1).
Wherein NLL can be defined asWherein NLLwRepresent to be directed to particular words w Negative log-likelihood function.
, can be by NLL for the ease of descriptionwL (w, r (w), c (w)) is expressed as, wherein (r (w), c (w)) represents word w The position of row and column in table 300.L can be utilizedr(w, r (w)) and carry out lc(w, c (w)) represents l (w, r (w), c (w)) word Language w row loss and word w row lose.Therefore, word w negative log-likelihood function NLLwIt can be expressed as again:
Wherein SwRepresent the set of all positions of the word w in corpus.
It is then possible to positions of the word w in table 300 is adjusted, make it that loss function NLL is minimum.For example, it is assumed that by word Language w is from the new position (i, j) that home position (r (w), c (w)) is moved in table 300.It can be assumed the distribution position of other words Put constant and calculate row loss l respectivelyr(w, i) and lcThe loss of (w, j) row, is then defined as l according to formula (2) by l (w, i, j)r (w,i)+lc(w,j).Then, an accepted way of doing sth (3) can be converted by minimizing loss function.
Wherein a (w, i, j)=1 represents that word w is dispensed on the new position (i, j) in table 300, and wherein SrAnd Sc The set of the row and column in table is represented respectively.
Therefore, standard minimal weight perfect matching problem can be equal to the problem of above-mentioned minimum loss function, thus Such as can be solved by minimum cost maximum-flow algorithm (MCMF).It is any being currently known or exploitation in the future MCMF methods can be used in combination with embodiments of the invention, the scope of the present invention not limited to this.Therefore, by MCMF methods, Dispensing position of each project in table can be adjusted using corpus so that the dispensing position of all items in Item Sets It is more accurate.
408, judge whether to meet the condition of convergence.For example, the number that the condition of convergence can be iteration has reached predetermined Number.Alternatively or additionally, the condition of convergence can the time of iteration reach the scheduled time.In another example the condition of convergence Can be that loss function starts to restrain, i.e., the rate of change of loss function is less than Threshold variety.If not meeting the condition of convergence, Then repetitive operation 404-408, until meeting the condition of convergence.If having met the condition of convergence, 410, store multiple projects and exist The value of all row vectors and column vector in dispensing position and table 300 in table 300 uses for follow-up.
By the above-mentioned means, being trained by corpus to all row vectors and column vector in table, and pass through Corpus adjusts dispensing position of all items in table, enables to all items in Item Sets to be distributed in table most Suitable position.Therefore, the value of all row vectors and column vector that train in accordance with an embodiment of the present disclosure is also very accurate, has It ensure that to effect the accuracy of semantic model.
Fig. 5 shows the exemplary plot of some rows in the table of Item Sets in accordance with an embodiment of the present disclosure.For example, passing through Several times after iteration adjustment, there is semantic relevance, or tool between the project in same a line and/or same row in table 300 There is semantic and grammar association.For example, in the 832nd row 510, the project of distribution is place name;In the 887th row 520, distribution Project is network address;In the 889th row 530, the project of distribution is the expression about concept of time.Therefore, according to the reality of the disclosure The Semantic Modeling Method for applying example can effectively between discovery project semantic relevance, or semanteme between discovery project and Grammar association, and adjust dispensing position of each project in table according to semantic.Therefore, in accordance with an embodiment of the present disclosure The accuracy of semantic model can not only be ensured, and the size of semantic model can be reduced.
Fig. 6 shows the flow chart for being used to determine the method 600 of associated project in accordance with an embodiment of the present disclosure.Root According to method 600, it may be determined that another project (also known as " Section 2 (also known as " first item ") associated with a project Mesh ").It should be appreciated that method 600 can perform after the action 206 in the method 200 above with reference to described by Fig. 2, and Can be as being performed with reference to the processing unit 110 described by figure 1.
602, at least based on the second quantization means, it is determined that on first group of semantic dimension, associated with project Three quantization means.For example it is assumed that first item is word " January ", then the first quantization means are subvectorAnd the second amount Change is expressed as subvectorThen, based on subvectorWith the current hidden layer state vector in RNN, in all rows of table 300 It is determined that with the subvector of the maximally related row of first item (such as the row vector of the 2nd row).Then 604, at least based on the 3rd Quantization means, it is determined that the 4th on second group of semantic dimension, associated with project quantization means.For example, based in 602 The row vector of determinationWith the current hidden layer state vector in RNN, determined in all row of table 300 most related to first item Row subvector (such as the 1st row column vector).Although it should be appreciated that first determined in above-mentioned example another purpose row to Then amount determines another purpose column vector again, but can also first determine another purpose column vector and then determine another again Purpose row vector.
606, according to the 3rd quantization means and the 4th quantization means, another project is determined.For example, it is based on row vectorWith Column vectorThe project that another project is the second row first row in table 300, i.e. word " one " can be determined.In some embodiments In, first item can be the current term in sentence, then the semantic vector based on current term, and can predict to appear in Next word after in sentence, current term.
Traditionally, for the Item Sets with V project, it is determined that during another project maximally related with first item, pass Then system method selects it needs to be determined that the correlation of each project and first item in Item Sets according to all correlations Another project, thus need V*t0Total time, wherein t0It is to determine the time of each project and the correlation of first item.So And the V project organization is existed in method in accordance with an embodiment of the present disclosureIn the form of row, and first from A line, Ran Houcong are selected in rowThe row of selection one in row, thus need only toTotal time.In some cases, When row and column in table 300 is unequal, determine that another purpose total time is slightly larger thanBut the value in V is ten million During rank, processing time in accordance with an embodiment of the present disclosure is far smaller than the processing time of traditional method, such as will can locate Reason time from traditional decades are reduced to several days or even several hours.Therefore, the disclosure can effectively improve semantic model Processing speed.
Fig. 7 shows the exemplary plot of the prediction process based on RNN in accordance with an embodiment of the present disclosure.For example, described by Fig. 7 The prediction process based on RNN can be used for the word of the data prediction current location t based on upper position t-1 in sentence. Hidden layer state vector can be divided into hidden layer state row vector hrWith hidden layer state column vector hc, wtRepresent the word of t positions.Make n The dimension of column vector is inputted for the line of input vector sum in RNN semantic models, and it is the hidden layer shape in RNN semantic models to make m The column vector of the dimension, wherein position t-1 of state vector∈ sets of real numbers Rn, position t row vector∈ sets of real numbers Rn, position T-1 hidden layer state row vector∈ sets of real numbers Rm.Therefore, can be determined by formula (4) position t-1 hidden layer status Bar to AmountWith position t hidden layer state row vector
Wherein W, U and b represent affine transformation parameter, wherein W ∈ sets of real numbers Rm*n, U ∈ sets of real numbers Rm*m, and b ∈ are real Manifold Rm, f represent neutral net in nonlinear activation function (such as sigmoid functions).In addition, row vector and column vector are equal From input vector matrix Xc, Xr∈ sets of real numbers
It is then possible to row probability Ps of the word w in position t is determined by formula (5)r(wt) and word w position t row probability Pc(wt)。
Wherein r (w) represents word w line index, and c (w) represents word w column index,∈ sets of real numbers RmIt is Yr∈ is real ManifoldI-th vector, and∈ sets of real numbers RmIt is Yc∈ sets of real numbersI-th of vector, and SrWith ScThe set of the row and column in table is represented respectively.
Therefore, it can determine that next word appears in the probability of every row by formula (5), and it is then determined that next word goes out The probability of present each column, so that it is determined that the word of the row position of the row of maximum probability and maximum probability is as maximum probability Next word.
As can be seen here, the column vector based on position t-1With position t-1 hidden layer state row vectorsPosition can be determined Put t-1 hidden layer state column vectorsIt is then based onRow probability Ps of the word w in position t can be determinedr(wt), that is, extremely Few column vector based on previous wordThe row vector of current term can be determined.In addition, the row vector by position tWith Position t-1 hidden layer state column vectorsPosition t hidden layer state row vectors can be determinedIt is then based onWord can be determined Row probability Ps of the w in position tc(wt), that is, at least row vector based on current termTo determine the column vector of current term. Therefore, for the Item Sets with V project, the general of word is determined by the calculating respectively of every row probability and each column probability Rate, the next item down object time can will be predicted from V*t0It is reduced toIt is effectively improved the processing of RNN semantic models Speed.
Method and function described herein can be performed by one or more hardware logic components at least in part. Such as, but not limited to, the exemplary types for the hardware logic component that can be used include field programmable gate array (FPGA), specially With integrated circuit (ASIC), Application Specific Standard Product (ASSP), on-chip system (SOC), CPLD (CPLD) etc..
For implement disclosed method program code can using any combinations of one or more programming languages come Write.These program codes can be supplied to the place of all-purpose computer, special-purpose computer or other programmable data processing units Manage device or controller so that program code when processor or controller execution when by making defined in flow chart and/or block diagram Function/operation is carried out.Program code can be performed completely on machine, partly performed on machine, as stand alone software Performed on machine packet portion and partly perform or performed completely on remote machine or server on the remote machine.
In the context of present disclosure, machine readable media can be tangible medium, and it can include or store The program for using or being used in combination with instruction execution system, device or equipment for instruction execution system, device or equipment.Machine Device computer-readable recording medium can be machine-readable signal medium or machine-readable storage medium.Machine readable media can include but unlimited In electronics, magnetic, optical, electromagnetism, infrared or semiconductor system, device or equipment, or the above appoints What appropriate combination.The more specific example of machinable medium can include being electrically connected, be portable based on one or more lines Formula computer disks, hard disk, random access memory (RAM), read-only storage (ROM), Erasable Programmable Read Only Memory EPROM (EPROM or flash memory), optical fiber, portable compact disk read-only storage (CD-ROM), optical storage device, magnetic storage are set Standby or the above any appropriate combination.
Although in addition, depicting each operation using certain order, this should be understood to requirement so operation with shown The certain order that goes out performs in sequential order, or requires that the operation of all diagrams should be performed to obtain desired result. Under certain environment, multitask and parallel processing are probably favourable.Similarly, although containing some tools in being discussed above Body realizes details, but these are not construed as the limitation to the scope of the present disclosure.In the context individually realized Some features of description can also be realized in single realization in combination.On the contrary, described in the context of single realization Various features can also be realized in multiple realizations individually or in any suitable subcombination.
It is listed below some sample implementations of the disclosure.
The disclosure may be implemented as a kind of electronic equipment, and the electronic equipment includes processing unit and memory, described Memory is coupled to the processing unit and is stored with instruction, and the instruction performs following when being performed by the processing unit Action:Obtaining includes the Item Sets of multiple projects;Determine first of the project in the Item Sets on first group of semantic dimension Quantization means and the second quantization means on second group of semantic dimension;And at least based on first quantization means and described Second quantization means generate the semantic values of the project, wherein the semantic values can be used in determining the project and the item Semantic dependency between the sundry item that mesh is concentrated, at least one in first quantization means and second quantization means It is individual to be shared by the project and at least one other project in the Item Sets.
In certain embodiments, wherein the semantic values are by vector representation, first quantization means are by the first subvector Represent, and second quantization means are represented by the second subvector.
In certain embodiments, the action also includes:The semantic values based on the project, determine the Item Sets In another project associated with the project.
In certain embodiments, wherein the project is the first word in sentence, another project is the sentence It is middle to appear in the second word after first word, and determine that another project includes:Based on described first The semantic values of word predict second word.
In certain embodiments, wherein determining that another project includes:Second quantization means are at least based on, it is determined that 3rd quantization means on first group of semantic dimension, associated with the project;At least quantify based on the described 3rd Represent, it is determined that the 4th on second group of semantic dimension, associated with project quantization means;And according to described 3rd quantization means and the 4th quantization means determine another project.
In certain embodiments, wherein determining that first quantization means and second quantization means include:By described in The multiple project organization in Item Sets is into table so that the project being in a line in the table has at described first group Identical quantization means on semantic dimension, and the project in same row has the identical on second group of semantic dimension Quantization means;And the dispensing position of the project in the table is adjusted using training dataset.
In certain embodiments, wherein the multiple project organization in the Item Sets is included into table:There to be phase The same a line of the allocation of items of same prefix in the table, the prefix represent the front portion of the project;And will tool There is the same row of the allocation of items of identical suffix in the table, the rear portion of project described in the postfix notation.
In certain embodiments, wherein adjusting the dispensing position bag of the project in the table using training dataset Include:Iteration performs following operation at least once, until the condition of convergence is satisfied, the condition of convergence and iteration time, iteration time Number or the Parameters variation of training pattern are relevant:Based on the dispensing position, first quantization means and second amount are updated Change and represent;And based on first quantization means after renewal and second quantization means, adjust the dispensing position.
The disclosure may be implemented as a kind of computer implemented method, and methods described includes:Acquisition includes multiple projects Item Sets;Determine first quantization means of the project in the Item Sets on first group of semantic dimension and semantic at second group The second quantization means in dimension;And at least generated based on first quantization means and second quantization means described The semantic values of project, the semantic values can be used in determining the semanteme between the sundry item in the project and the Item Sets Correlation, it is at least one by the project and the Item Sets in first quantization means and second quantization means At least one other project share.
In certain embodiments, wherein the semantic values are by vector representation, first quantization means are by the first subvector Represent, and second quantization means are represented by the second subvector.
In certain embodiments, this method also includes:The semantic values based on the project, are determined in the Item Sets Another project associated with the project.
In certain embodiments, wherein the project is the first word in sentence, another project is the sentence It is middle to appear in the second word after first word, and determine that another project includes:Based on described first The semantic values of word predict second word.
In certain embodiments, wherein determining that another project includes:Second quantization means are at least based on, it is determined that 3rd quantization means on first group of semantic dimension, associated with the project;At least quantify based on the described 3rd Represent, it is determined that the 4th on second group of semantic dimension, associated with project quantization means;And according to described 3rd quantization means and the 4th quantization means determine another project.
In certain embodiments, wherein determining that first quantization means and second quantization means include:By described in The multiple project organization in Item Sets is into table so that the project being in a line in the table has at described first group Identical quantization means on semantic dimension, and the project in same row has the identical on second group of semantic dimension Quantization means;And the dispensing position of the project in the table is adjusted using training dataset.
In certain embodiments, wherein the multiple project organization in the Item Sets is included into table:There to be phase The same a line of the allocation of items of same prefix in the table, the prefix represent the front portion of the project;And will tool There is the same row of the allocation of items of identical suffix in the table, the rear portion of project described in the postfix notation.
In certain embodiments, wherein adjusting the dispensing position bag of the project in the table using training dataset Include:Iteration performs following operation at least once, until the condition of convergence is satisfied, the condition of convergence and iteration time, iteration time Number or the Parameters variation of training pattern are relevant:Based on the dispensing position, first quantization means and second amount are updated Change and represent;And based on first quantization means after renewal and second quantization means, adjust the dispensing position.
The disclosure may be implemented as a kind of computer program product, and it is stored in non-transitory, computer storage medium simultaneously And including machine-executable instruction, when the machine-executable instruction is run in a device so that the equipment:Obtain including more The Item Sets of individual project;Determine first quantization means of the project in the Item Sets on first group of semantic dimension and second The second quantization means on group semantic dimension;And at least it is based on first quantization means and the second quantization means next life Into the semantic values of the project, the semantic values can be used in determining between the sundry item in the project and the Item Sets Semantic dependency, it is at least one by the project and the item in first quantization means and second quantization means At least one other project that mesh is concentrated is shared.
In certain embodiments, wherein the semantic values are by vector representation, first quantization means are by the first subvector Represent, and second quantization means are represented by the second subvector.
In certain embodiments, wherein also causing the equipment when machine-executable instruction is run in a device:Base In the semantic values of the project, another project associated with the project in the Item Sets is determined.
In certain embodiments, wherein the project is the first word in sentence, another project is the sentence It is middle to appear in the second word after first word, and determine that another project includes:Based on described first The semantic values of word predict second word.
In certain embodiments, wherein determining that another project includes:Second quantization means are at least based on, it is determined that 3rd quantization means on first group of semantic dimension, associated with the project;At least quantify based on the described 3rd Represent, it is determined that the 4th on second group of semantic dimension, associated with project quantization means;And according to described 3rd quantization means and the 4th quantization means determine another project.
In certain embodiments, wherein determining that first quantization means and second quantization means include:By described in The multiple project organization in Item Sets is into table so that the project being in a line in the table has at described first group Identical quantization means on semantic dimension, and the project in same row has the identical on second group of semantic dimension Quantization means;And the dispensing position of the project in the table is adjusted using training dataset.
In certain embodiments, wherein the multiple project organization in the Item Sets is included into table:There to be phase The same a line of the allocation of items of same prefix in the table, the prefix represent the front portion of the project;And will tool There is the same row of the allocation of items of identical suffix in the table, the rear portion of project described in the postfix notation.
In certain embodiments, wherein adjusting the dispensing position bag of the project in the table using training dataset Include:Iteration performs following operation at least once, until the condition of convergence is satisfied, the condition of convergence and iteration time, iteration time Number or the Parameters variation of training pattern are relevant:Based on the dispensing position, first quantization means and second amount are updated Change and represent;And based on first quantization means after renewal and second quantization means, adjust the dispensing position.
, should although describing the disclosure using the language specific to architectural feature and/or method logical action When understanding that the theme defined in appended claims is not necessarily limited to special characteristic described above or action.On on the contrary, Special characteristic and action described by face are only the exemplary forms for realizing claims.

Claims (20)

1. a kind of electronic equipment, including:
Processing unit;
Memory, coupled to the processing unit and instruction is stored with, the instruction is held when being performed by the processing unit Row is following to be acted:
Obtaining includes the Item Sets of multiple projects;
Determine first quantization means of the project in the Item Sets on first group of semantic dimension and in second group of semantic dimension On the second quantization means;And
At least the semantic values of the project, the semanteme are generated based on first quantization means and second quantization means Value can be used in determining the semantic dependency between the sundry item in the project and the Item Sets, and described first quantifies table Show and be total to at least one in second quantization means by the project and at least one other project in the Item Sets Enjoy.
2. equipment according to claim 1, wherein the semantic values, by vector representation, first quantization means are by first Subvector is represented, and second quantization means are represented by the second subvector.
3. equipment according to claim 1, the action also includes:
The semantic values based on the project, determine another project associated with the project in the Item Sets.
4. equipment according to claim 3, wherein the project is the first word in sentence, another project is institute The second word after first word will be appeared in by stating in sentence, and determine that another project includes:
Second word is predicted based on the semantic values of first word.
5. equipment according to claim 3, wherein determining that another project includes:
Second quantization means are at least based on, it is determined that on first group of semantic dimension, associated with the project 3rd quantization means;
The 3rd quantization means are at least based on, it is determined that on second group of semantic dimension, associated with the project 4th quantization means;And
Another project is determined according to the 3rd quantization means and the 4th quantization means.
6. equipment according to claim 1, wherein determining that first quantization means and second quantization means include:
By the multiple project organization in the Item Sets into table so that the project being in a line in the table has Identical quantization means on first group of semantic dimension, and the project in same row has in second group of semantic dimensions Identical quantization means on degree;And
The dispensing position of the project in the table is adjusted using training dataset.
7. equipment according to claim 6, wherein the multiple project organization in the Item Sets is included into table:
By the same a line of the allocation of items with identical prefix in the table, the prefix represents the previous portion of the project Point;And
By the same row of the allocation of items with identical suffix in the table, the latter portion of project described in the postfix notation Point.
8. equipment according to claim 6, wherein adjusting point of the project in the table using training dataset Include with position:
Iteration performs following operation at least once, until the condition of convergence is satisfied, the condition of convergence and at least one following phase Close:The Parameters variation of iteration time, iterations and training pattern:
Based on the dispensing position, first quantization means and second quantization means are updated;And
Based on first quantization means after renewal and second quantization means, the dispensing position is adjusted.
9. a kind of computer implemented method, including:
Obtaining includes the Item Sets of multiple projects;
Determine first quantization means of the project in the Item Sets on first group of semantic dimension and in second group of semantic dimension On the second quantization means;And
At least the semantic values of the project, the semanteme are generated based on first quantization means and second quantization means Value can be used in determining the semantic dependency between the sundry item in the project and the Item Sets, and described first quantifies table Show and be total to at least one in second quantization means by the project and at least one other project in the Item Sets Enjoy.
10. according to the method for claim 9, wherein the semantic values are by vector representation, first quantization means are by the One subvector is represented, and second quantization means are represented by the second subvector.
11. the method according to claim 11, in addition to:
The semantic values based on the project, determine another project associated with the project in the Item Sets.
12. according to the method for claim 11, wherein the project is the first word in sentence, another project is The second word after first word will be appeared in the sentence, and determine that another project includes:
Second word is predicted based on the semantic values of first word.
13. according to the method for claim 11, wherein determining that another project includes:
Second quantization means are at least based on, it is determined that on first group of semantic dimension, associated with the project 3rd quantization means;
The 3rd quantization means are at least based on, it is determined that on second group of semantic dimension, associated with the project 4th quantization means;And
Another project is determined according to the 3rd quantization means and the 4th quantization means.
14. according to the method for claim 9, wherein determining first quantization means and the second quantization means bag Include:
By the multiple project organization in the Item Sets into table so that the project being in a line in the table has Identical quantization means on first group of semantic dimension, and the project in same row has in second group of semantic dimensions Identical quantization means on degree;And
The dispensing position of the project in the table is adjusted using training dataset.
15. according to the method for claim 14, wherein the multiple project organization in the Item Sets is included into table:
By the same a line of the allocation of items with identical prefix in the table, the prefix represents the previous portion of the project Point;And
By the same row of the allocation of items with identical suffix in the table, the latter portion of project described in the postfix notation Point.
16. according to the method for claim 14, wherein adjusting the project using training dataset in the table Dispensing position includes:
Iteration performs following operation at least once, until the condition of convergence is satisfied, the condition of convergence and at least one following phase Close:The Parameters variation of iteration time, iterations and training pattern:
Based on the dispensing position, first quantization means and second quantization means are updated;And
Based on first quantization means after renewal and second quantization means, the dispensing position is adjusted.
17. a kind of computer program product, the computer program product is stored in non-transitory, computer storage medium simultaneously And cause the equipment when running in a device including machine-executable instruction, the machine-executable instruction:
Obtaining includes the Item Sets of multiple projects;
Determine first quantization means of the project in the Item Sets on first group of semantic dimension and in second group of semantic dimension On the second quantization means;And
At least the semantic values of the project, the semanteme are generated based on first quantization means and second quantization means Value can be used in determining the semantic dependency between the sundry item in the project and the Item Sets, and described first quantifies table Show and be total to at least one in second quantization means by the project and at least one other project in the Item Sets Enjoy.
18. computer program product according to claim 17, wherein the machine-executable instruction is run in a device When also cause the equipment:
Second quantization means are at least based on, it is determined that on first group of semantic dimension, associated with the project 3rd quantization means;
The 3rd quantization means are at least based on, it is determined that on second group of semantic dimension, associated with the project 4th quantization means;And
It is associated with the project in the Item Sets to determine according to the 3rd quantization means and the 4th quantization means Another project.
19. computer program product according to claim 17, wherein determining first quantization means and described second Quantization means include:
By the multiple project organization in the Item Sets into table so that the project being in a line in the table has Identical quantization means on first group of semantic dimension, and the project in same row has in second group of semantic dimensions Identical quantization means on degree;And
The dispensing position of the project in the table is adjusted using training dataset.
20. computer program product according to claim 19, exist wherein adjusting the project using training dataset Dispensing position in the table includes:
Iteration performs following operation at least once, until the condition of convergence is satisfied, the condition of convergence and at least one following phase Close:The Parameters variation of iteration time, iterations and training pattern:
Based on the dispensing position, first quantization means and second quantization means are updated;And
Based on first quantization means after renewal and second quantization means, the dispensing position is adjusted.
CN201610818984.3A 2016-09-12 2016-09-12 Semantic processing for natural language Active CN107818076B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201610818984.3A CN107818076B (en) 2016-09-12 2016-09-12 Semantic processing for natural language

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201610818984.3A CN107818076B (en) 2016-09-12 2016-09-12 Semantic processing for natural language

Publications (2)

Publication Number Publication Date
CN107818076A true CN107818076A (en) 2018-03-20
CN107818076B CN107818076B (en) 2021-11-12

Family

ID=61601158

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201610818984.3A Active CN107818076B (en) 2016-09-12 2016-09-12 Semantic processing for natural language

Country Status (1)

Country Link
CN (1) CN107818076B (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110807207A (en) * 2019-10-30 2020-02-18 腾讯科技(深圳)有限公司 Data processing method and device, electronic equipment and storage medium

Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN1310825A (en) * 1998-06-23 2001-08-29 微软公司 Methods and apparatus for classifying text and for building a text classifier
CN101796511A (en) * 2007-08-31 2010-08-04 微软公司 Identification of semantic relationships within reported speech
CN102646091A (en) * 2011-02-22 2012-08-22 日电(中国)有限公司 Dependence relationship labeling method, device and system
US20130151441A1 (en) * 2011-12-13 2013-06-13 Xerox Corporation Multi-task learning using bayesian model with enforced sparsity and leveraging of task correlations
CN103955714A (en) * 2014-04-09 2014-07-30 中国科学院信息工程研究所 Navy detection model construction method and system and navy detection method
US20150074027A1 (en) * 2013-09-06 2015-03-12 Microsoft Corporation Deep Structured Semantic Model Produced Using Click-Through Data
CN104765728A (en) * 2014-01-08 2015-07-08 富士通株式会社 Method and device for training neural network and method for determining sparse feature vector
CN105051721A (en) * 2013-01-29 2015-11-11 微软技术许可有限责任公司 Translating natural language descriptions to programs in a domain-specific language for spreadsheets

Patent Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN1310825A (en) * 1998-06-23 2001-08-29 微软公司 Methods and apparatus for classifying text and for building a text classifier
CN101796511A (en) * 2007-08-31 2010-08-04 微软公司 Identification of semantic relationships within reported speech
CN102646091A (en) * 2011-02-22 2012-08-22 日电(中国)有限公司 Dependence relationship labeling method, device and system
US20130151441A1 (en) * 2011-12-13 2013-06-13 Xerox Corporation Multi-task learning using bayesian model with enforced sparsity and leveraging of task correlations
CN105051721A (en) * 2013-01-29 2015-11-11 微软技术许可有限责任公司 Translating natural language descriptions to programs in a domain-specific language for spreadsheets
US20150074027A1 (en) * 2013-09-06 2015-03-12 Microsoft Corporation Deep Structured Semantic Model Produced Using Click-Through Data
CN104765728A (en) * 2014-01-08 2015-07-08 富士通株式会社 Method and device for training neural network and method for determining sparse feature vector
CN103955714A (en) * 2014-04-09 2014-07-30 中国科学院信息工程研究所 Navy detection model construction method and system and navy detection method

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
TOMAS MIKOLOV ET AL.: "Recurrent neural network based language model", 《INTERSPEECH 2010》 *
WENLIN CHEN ET AL.: "Strategies for Training Large Vocabulary Neural Language Models", 《ARXIV:1512.04906V1》 *
陈恩红 等: "单词嵌入——自然语言的连续空间表示", 《数据采集与处理》 *

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110807207A (en) * 2019-10-30 2020-02-18 腾讯科技(深圳)有限公司 Data processing method and device, electronic equipment and storage medium

Also Published As

Publication number Publication date
CN107818076B (en) 2021-11-12

Similar Documents

Publication Publication Date Title
CN108363790B (en) Method, device, equipment and storage medium for evaluating comments
CN104615767B (en) Training method, search processing method and the device of searching order model
CN106663124A (en) Generating and using a knowledge-enhanced model
CN108628825A (en) Text message Similarity Match Method, device, computer equipment and storage medium
CN105404632B (en) System and method for carrying out serialized annotation on biomedical text based on deep neural network
CN111191002B (en) Neural code searching method and device based on hierarchical embedding
CN110036399A (en) Neural Network Data input system
Kreutzer et al. Bandit structured prediction for neural sequence-to-sequence learning
CN109829162A (en) A kind of text segmenting method and device
Chien et al. Bayesian sparse topic model
US11373043B2 (en) Technique for generating and utilizing virtual fingerprint representing text data
CN113961666B (en) Keyword recognition method, apparatus, device, medium, and computer program product
CN113656547A (en) Text matching method, device, equipment and storage medium
CN112434134A (en) Search model training method and device, terminal equipment and storage medium
de Souza et al. A deep learning approach for sentiment analysis applied to hotel’s reviews
CN115329075A (en) Text classification method based on distributed machine learning
US20220284172A1 (en) Machine learning technologies for structuring unstructured data
CN113723077B (en) Sentence vector generation method and device based on bidirectional characterization model and computer equipment
CN112667797B (en) Question-answer matching method, system and storage medium for self-adaptive transfer learning
CN107818076A (en) For the semantic processes of natural language
WO2020215581A1 (en) Chinese coding method and apparatus based on bidirectional long short-term memory network model
CN112989801B (en) Sequence labeling method, device and equipment
CN115906861A (en) Statement emotion analysis method and device based on interaction aspect information fusion
CN109871414A (en) Biomedical entity relationship classification method based on the context vector kernel of graph
CN115983269A (en) Intelligent community data named entity identification method, terminal and computer medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant