CN115408498B - Data dynamic identification method based on natural language - Google Patents
Data dynamic identification method based on natural language Download PDFInfo
- Publication number
- CN115408498B CN115408498B CN202211359030.2A CN202211359030A CN115408498B CN 115408498 B CN115408498 B CN 115408498B CN 202211359030 A CN202211359030 A CN 202211359030A CN 115408498 B CN115408498 B CN 115408498B
- Authority
- CN
- China
- Prior art keywords
- preset
- data
- sample
- splicing
- label
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
- 238000000034 method Methods 0.000 title claims abstract description 55
- 238000012795 verification Methods 0.000 claims abstract description 25
- 238000012549 training Methods 0.000 claims abstract description 20
- 238000002474 experimental method Methods 0.000 claims abstract description 7
- 238000009499 grossing Methods 0.000 claims description 24
- 230000008569 process Effects 0.000 claims description 16
- 238000009826 distribution Methods 0.000 claims description 14
- 238000004422 calculation algorithm Methods 0.000 claims description 13
- 238000000605 extraction Methods 0.000 claims description 7
- 238000004364 calculation method Methods 0.000 claims description 6
- 238000013507 mapping Methods 0.000 claims description 4
- 238000012986 modification Methods 0.000 claims description 4
- 230000004048 modification Effects 0.000 claims description 4
- 238000005070 sampling Methods 0.000 claims description 3
- 238000011161 development Methods 0.000 description 3
- 230000018109 developmental process Effects 0.000 description 3
- 230000006870 function Effects 0.000 description 3
- 230000000694 effects Effects 0.000 description 2
- 239000000284 extract Substances 0.000 description 2
- 238000000926 separation method Methods 0.000 description 2
- 230000009286 beneficial effect Effects 0.000 description 1
- 238000005259 measurement Methods 0.000 description 1
- 238000003058 natural language processing Methods 0.000 description 1
- 238000004088 simulation Methods 0.000 description 1
- 238000006467 substitution reaction Methods 0.000 description 1
- 238000010200 validation analysis Methods 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/30—Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
- G06F16/33—Querying
- G06F16/3331—Query processing
- G06F16/334—Query execution
- G06F16/3344—Query execution using natural language analysis
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/30—Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
- G06F16/35—Clustering; Classification
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F40/00—Handling natural language data
- G06F40/10—Text processing
- G06F40/12—Use of codes for handling textual entities
- G06F40/126—Character encoding
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F40/00—Handling natural language data
- G06F40/30—Semantic analysis
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Computational Linguistics (AREA)
- Artificial Intelligence (AREA)
- Audiology, Speech & Language Pathology (AREA)
- General Health & Medical Sciences (AREA)
- Health & Medical Sciences (AREA)
- Data Mining & Analysis (AREA)
- Databases & Information Systems (AREA)
- Machine Translation (AREA)
Abstract
The application discloses a data dynamic identification method based on natural language, mainly relates to the technical field of data dynamic identification, and is used for solving the problems of general applicability and high uncertainty of the existing model performance. The method comprises the following steps: determining semantic tag data corresponding to the sample data; generating an experiment set; splitting the data into a training data set and a verification data set; importing sample data in a training data set into a preset encoder; splicing into sample splicing data; importing semantic label data in a training data set into a preset generator, and further splicing into label splicing data; determining the cost value of the distance between the sample splicing data and the label splicing data to obtain a trained preset discriminator; obtaining a trained preset encoder and a preset generator; obtaining verification sample splicing data; obtaining verification label splicing data; and completing the matching of the data. According to the method, the fitting degree of the model and the data is improved, and the accuracy is prompted.
Description
Technical Field
The application relates to the technical field of dynamic data identification, in particular to a dynamic data identification method based on natural language.
Background
The text classification task in the natural language processing field is widely applied to various fields and industries, the application scene range is wide, and various resources (service platforms, hardware, technical frameworks, data and the like) which can be utilized in the development and implementation process are various.
The existing method for dynamically identifying data comprises the following steps: and (3) adopting a general structure, taking a BERT or similar model [ CLS ] layer as the input of a classifier, and carrying out classification task training. Can meet certain industrial requirements under normal conditions, and has simple realization and short development period.
However, as data asset security management becomes more and more standard, a certain bottleneck exists in the performance level in a real service scene through a general architecture, and the universality and uncertainty of the corresponding model performance are large.
Disclosure of Invention
In view of the above-mentioned shortcomings in the prior art, the present invention provides a dynamic data identification method based on natural language, so as to solve the above-mentioned technical problems.
The application provides a data dynamic identification method based on natural language, which comprises the following steps: acquiring a sample set, and determining semantic tag data corresponding to each sample data in the sample set; generating an experiment set based on the sample data, the semantic label data and the mapping relation between the sample data and the semantic label data; dividing the experiment set into a training data set and a verification data set; importing sample data in a training data set into a preset encoder; acquiring a plurality of sample sub-splicing data from a hidden layer of a preset encoder based on a preset sample splicing data acquisition position, and further splicing the sample sub-splicing data into sample splicing data; importing semantic label data and preset reference dimension data in a training data set into a preset generator; acquiring a plurality of label sub-splicing data from a hidden layer of a preset generator based on a preset label splicing data acquisition position, and splicing the label sub-splicing data into label splicing data; determining a distance cost value between the sample splicing data and the label splicing data based on a preset distribution distance equation; importing the distance cost value, the preset learning rate, the preset smoothing constant and the initial discriminator weight value into a preset optimizer to complete weight updating of the preset discriminator so as to obtain a trained preset discriminator; importing a preset learning rate, a preset smoothing constant, an initial encoder weight value and a trained preset discriminator into a preset optimizer to finish updating of the preset encoder weight value; importing a preset learning rate, a preset smoothing constant, an initial encoder weight value and a trained preset discriminator into a preset optimizer to finish updating of a preset generator weight value; to obtain a trained preset encoder and a preset generator; obtaining verification sample splicing data corresponding to sample data in a verification data set based on a trained preset encoder; obtaining verification label splicing data based on the trained preset generator or label splicing data; and matching the verification sample splicing data with the verification label splicing data based on a trained preset discriminator or a preset matching degree calculation formula.
Further, the hidden layer of the preset encoder and the hidden layer of the preset generator both comprise a text semantic coding network layer and a label semantic coding network layer.
Further, determining semantic tag data corresponding to each sample data in the sample set specifically includes: acquiring a semantic tag set through a preset semantic tag interface; the semantic tag set comprises semantic tag data; or, semantic tag data corresponding to each sample data is obtained through a preset keyword/subject term extraction algorithm; or analyzing the part of speech of the sample data through a preset sample part of speech analysis algorithm to obtain preset attribute words corresponding to the sample data, and splicing the preset attribute words into semantic tag data; or when the preset associated data set corresponding to the sample set is obtained, extracting the keywords/subject terms corresponding to the preset associated data set through a keyword/subject term extraction algorithm and a preset sample part-of-speech analysis algorithm to obtain semantic tag data.
Further, acquiring a sample set specifically includes: and acquiring real service data or replacing open source service data or artificial sample data as a sample set through a preset sample uploading process.
Further, before determining a distance cost value between the sample stitching data and the label stitching data based on a preset inter-distribution distance equation, the method further includes: replacing joint distribution in the Wasserstein-distance method with an encoder, replacing edge distribution with a generator, and replacing sampling with sample splicing data and label splicing data; obtaining a preset distance cost value calculation formula:wherein D () is the output result of the preset discriminator,the data is spliced for the samples and,data is spliced for the label.
Further, importing the distance cost value, the preset learning rate, the preset smoothing constant and the initial discriminator weight value into a preset optimizer, and completing weight update of the preset discriminator, which specifically comprises: updating the formula by the preset discriminator weight:updating the weight value of the preset discriminator; wherein,to preset the weight values of the encoders generated in the updating process,in order to be the cost value of the distance,in order to preset the learning rate,andis a preset smoothing constant; when the weight value is larger than c or smaller than-c, through a preset clipping formula:performing gradient clipping on the weight value of the preset discriminator; where c is the clipping threshold.
Further, leading the preset learning rate, the preset smoothing constant, the initial encoder weight value and the trained preset discriminator into the preset optimizer, completing the updating of the preset encoder weight value, and specifically comprising: updating the formula by presetting the encoder weight:updating the weight value of a preset encoder; wherein,to preset the weight values of the encoders generated in the updating process,the data is spliced for the samples and,in order to preset the learning rate,andis a preset smoothing constant.
Further, importing the preset learning rate, the preset smoothing constant, the initial encoder weight value and the trained preset discriminator into the preset optimizer, and completing updating of the weight value of the preset generator, specifically including: updating the formula by presetting the weight of the generator:updating the weight value of the preset generator; wherein,for the weight value of the preset generator generated in the updating process,the data is spliced for the label and,in order to preset the learning rate,andis a preset smoothing constant.
Further, after obtaining the trained preset encoder and preset generator, the method further comprises: and modifying the semantic label data by presetting a semantic label modification interface.
As can be appreciated by those skilled in the art, the present invention has at least the following beneficial effects:
(1) And through presetting the sample splicing data acquisition position and presetting the label splicing data acquisition position, the sample splicing data and the label splicing data which are formed by splicing different depth hidden layers of the related structure are acquired. Because the hidden layer corresponding to the acquisition position is preset, relevant technicians can select proper hidden layer input data or output data for flexible splicing according to the network structure characteristics of the relevant technicians (a preset encoder, a preset generator and the like).
(2) Due to reasons such as data asset privacy and safety, when effective real service scene data cannot be obtained during model training, relevant open source data can be obtained through a preset interface, semantic tag data with associated requirements can be obtained, a simulation training effect separated from the real service scene data can be achieved, risks caused by problems such as data privacy disclosure are avoided, and compared with a traditional scheme, the effect is improved to a certain extent.
(3) Finally, the range and definition of the semantic tag data can be modified through the preset semantic tag modification interface, so that the classification function can be dynamically adjusted to a certain degree, more diversified requirements are met, and better support is provided for user customization.
Drawings
Some embodiments of the disclosure are described below with reference to the accompanying drawings, in which:
fig. 1 is a flowchart of a dynamic data identification method based on natural language according to an embodiment of the present application.
Detailed Description
It should be understood by those skilled in the art that the embodiments described below are only preferred embodiments of the present disclosure, and do not mean that the present disclosure can be implemented only by the preferred embodiments, which are merely intended to explain the technical principles of the present disclosure and not to limit the scope of the present disclosure. All other embodiments that can be derived by one of ordinary skill in the art from the preferred embodiments provided by the disclosure and that fall within the scope of the disclosure are intended to be encompassed by the present disclosure without any inventive step.
It should also be noted that the terms "comprises," "comprising," or any other variation thereof, are intended to cover a non-exclusive inclusion, such that a process, method, article, or apparatus that comprises a list of elements does not include only those elements but may include other elements not expressly listed or inherent to such process, method, article, or apparatus. Without further limitation, an element defined by the phrases "comprising a," "8230," "8230," or "comprising" does not exclude the presence of additional identical elements in the process, method, article, or apparatus comprising the element.
The technical solutions proposed in the embodiments of the present application are described in detail below with reference to the accompanying drawings.
Fig. 1 further provides a data dynamic recognition method based on natural language for the embodiment of the present application, and as shown in fig. 1, the method provided by the embodiment of the present application mainly includes the following steps:
and 110, acquiring a sample set, and determining semantic tag data corresponding to each sample data in the sample set.
It should be noted that the sample set may be real service data, and in the case of lack of support of the real service data, or the real service scene data cannot directly participate in training due to privacy and security, the sample set may also be substitute open-source service data rich in similar semantics. The replacement open source service data may be replaced for similar near semantic meaning or supplemented by other open source data sets. In the event that no suitable substitute open source business data can be found, approximately 50 samples (artificial sample data) can be artificially made for each tag as a set of samples. The specific contents of the real service data, the alternative open source service data and the artificial sample data can be determined by those skilled in the art according to actual conditions.
The method for obtaining the sample set may be obtaining through a preset sample uploading process.
The method for determining semantic tag data corresponding to each sample data in the sample set may specifically be: (1) Acquiring a semantic tag set uploaded by an operator through a preset semantic tag interface; wherein the semantic tag set comprises semantic tag data; (2) And importing the sample set into a preset keyword/subject term extraction algorithm to obtain semantic tag data corresponding to each sample data. The preset keyword/subject term extraction algorithm can be any algorithm with a keyword/subject term extraction function, such as a TF-IDF algorithm; (3) The method has the advantages that the labels obtained by the method are more generalized and are suitable for individual requirements, or under the condition that original data are not labeled, data related to samples can be searched by combining other business data of companies, and all semantic label data can be obtained through business logic and algorithm association.
Thus, the acquisition of the sample set, the acquisition of the semantic tag data and the mapping of the sample data and the semantic tag data are completed.
Step 120, generating an experiment set based on the sample data, the semantic tag data and the mapping relation between the sample data and the semantic tag data; and splitting the experiment set into a training data set and a verification data set.
Step 130, importing sample data in the training data set into a preset encoder; and acquiring a plurality of sample sub-splicing data from a hidden layer of a preset encoder based on a preset sample splicing data acquisition position, and splicing the sample sub-splicing data into sample splicing data.
It should be noted that the hidden layer of the preset encoder at least includes a text semantic coding network and a tag semantic coding network. And presetting a sample splicing data acquisition position for defining which hidden layer corresponding to the text semantic coding network and the label semantic coding network extracts hidden layer input data or output data as sample sub-splicing data.And splicing all the obtained sample sub-splicing data into sample splicing data. For example,for the output of the step 1 text semantic coding network,outputting the labeled semantic coding network in the step 1;the data is stitched for the sample.
Step 140, importing semantic label data and preset reference dimension data in the training data set into a preset generator; and acquiring a plurality of label sub-splicing data from the hidden layer of the preset generator based on the preset label splicing data acquisition position, and splicing the label sub-splicing data into label splicing data.
It should be noted that the hidden layer of the preset generator at least includes a text semantic coding network and a label semantic coding network. And presetting a tag splicing data acquisition position, and defining which hidden layer corresponding to the text semantic coding network and the tag semantic coding network extracts hidden layer input data or output data as tag sub-splicing data. And then splicing all the obtained label sub-splicing data into label splicing data. For example,the output of the network is generated for the tag semantics in step 1,generating the output of the network for the text semantics in the step 1;data is spliced for the label. In addition, the preset reference dimension data may be a null value or supervisory information related to the tag, and the specific content thereof may be determined by those skilled in the art.
In addition, since the sample data may correspond to a plurality of semantic tag data, one or more semantic tag data may exist inside the semantic tag data. The separation method of the semantic tag data can be any feasible method, for example, the separator "|" is used for separation.
Based on the step 130 and the step 140, those skilled in the art can understand that the sample stitching data and the label stitching data formed by stitching different depth hidden layers of the related structure are obtained by presetting the sample stitching data obtaining position and the label stitching data obtaining position. Because the hidden layer corresponding to the acquisition position is preset, relevant technicians can select proper hidden layer input data or output data for flexible splicing according to the network structure characteristics of the relevant technicians (a preset encoder, a preset generator and the like).
Thus, the acquisition of the sample splicing data and the label splicing data is completed.
Step 150, determining a distance cost value between the sample splicing data and the label splicing data based on a preset inter-distribution distance equation; and importing the distance cost value, the preset learning rate, the preset smoothing constant and the initial discriminator weight value into a preset optimizer, and finishing weight updating of the preset discriminator to obtain the trained preset discriminator.
It should be noted that the preset inter-distribution distance equation is any feasible measurement formula capable of calculating the difference between the two output distributions.
The method for obtaining the preset inter-distribution distance equation may specifically be: and (3) adopting a Wasserstein-distance formula as an object equation, jointly distributing and edge distribution corresponding to an encoder and a generator, and sampling (X, Y) corresponding to sample splicing data and label splicing data.
According to the formula:and obtaining the preset distance cost value. Wherein D () presets the discriminator output result,the data is spliced for the samples and,data is spliced for the label.
In addition, the above theoretical method based on Wasserstein-distance may be replaced by any theory capable of calculating the difference between two output distributions, such as Kullback-leibler divergence theory, jensen-Shannon divergence theory, and the like.
The distance cost value, the preset learning rate, the preset smoothing constant and the initial discriminator weight value are led into a preset optimizer, and the weight update of the preset discriminator is completed, specifically, the formula is updated by the preset discriminator weight:updating the weight value of the preset discriminator; wherein,to preset the weight values of the encoders generated in the updating process,in order to be the cost value of the distance,in order to preset the learning rate,andis a preset smoothing constant; when the weight value is larger than c or smaller than-c, through a preset clipping formula:performing gradient clipping on a preset discriminator weight value; where c is the clipping threshold.
Thus, the preset encoder training is completed.
Step 160, importing a preset learning rate, a preset smoothing constant, an initial encoder weight value and a trained preset discriminator into a preset optimizer to complete updating of the preset encoder weight value; importing a preset learning rate, a preset smoothing constant, an initial encoder weight value and a trained preset discriminator into a preset optimizer to finish updating of a preset generator weight value; to obtain a trained preset encoder and preset generator.
Wherein, will predetermine the learning rate, predetermine smooth constant, initial encoder weighted value, the leading-in optimizer of predetermineeing of the good ware of distinguishing of training, accomplish the update of predetermineeing the encoder weighted value, specifically can be for, through predetermineeing encoder weight update formula:updating the weight value of a preset encoder; wherein,to preset the weight values of the encoders generated in the updating process,the data is spliced for the samples and,in order to preset the learning rate,andis a preset smoothing constant.
Wherein, leading-in the optimizer of predetermineeing preset learning rate, predetermineeing smooth constant, initial encoder weighted value, the good ware of predetermineeing of training, the update of the completion preset generater weighted value specifically is, through predetermineeing generater weight update formula:to update the tableSetting a weight value of a generator; wherein,to update the weight values of the preset generators generated in the process,the data is spliced for the label and,in order to preset the learning rate,andis a preset smoothing constant.
Thus, the training of the preset encoder and the preset generator is completed.
In addition, after obtaining the trained preset encoder and preset generator, the application can further: and modifying the semantic label data by presetting a semantic label modification interface. And then can adjust classification function to a certain extent developments, satisfy more diversified demand, provide better support for user-defined.
Step 170, obtaining verification sample splicing data corresponding to sample data in the verification data set based on the trained preset encoder; obtaining verification label splicing data based on the trained preset generator or label splicing data; and completing the matching of the verification sample splicing data and the verification label splicing data based on a trained preset discriminator or a preset matching degree calculation formula.
It should be noted that, based on the tag splicing data, the method for obtaining the verification tag splicing data includes: and directly using the label splicing data as verification label splicing data. The preset matching degree calculation formula is any existing formula capable of calculating the matching degree of the verification sample splicing data and the verification label splicing data.
Wherein the validation tag splice data is generated online with a generator. When a new label appears and output data needs to be modified, the model does not need to be retrained, and semantic label data with higher customization degree can be provided on line.
So far, the technical solutions of the present disclosure have been described in connection with the foregoing embodiments, but it is easily understood by those skilled in the art that the scope of the present disclosure is not limited to only these specific embodiments. The technical solutions in the above embodiments can be split and combined, and equivalent changes or substitutions can be made on related technical features by those skilled in the art without departing from the technical principles of the present disclosure, and any changes, equivalents, improvements, and the like made within the technical concept and/or technical principles of the present disclosure will fall within the protection scope of the present disclosure.
Claims (9)
1. A dynamic data identification method based on natural language is characterized in that the method comprises the following steps:
acquiring a sample set, and determining semantic tag data corresponding to each sample data in the sample set;
generating an experiment set based on the sample data, the semantic tag data and the mapping relation between the sample data and the semantic tag data; splitting the experiment set into a training data set and a verification data set;
importing sample data in a training data set into a preset encoder; acquiring a plurality of sample sub-splicing data from a hidden layer of a preset encoder based on a preset sample splicing data acquisition position, and splicing the sample sub-splicing data into sample splicing data;
importing semantic label data and preset reference dimension data in a training data set into a preset generator; acquiring a plurality of label sub-splicing data from a hidden layer of a preset generator based on a preset label splicing data acquisition position, and splicing the label sub-splicing data into label splicing data;
determining a distance cost value between the sample splicing data and the label splicing data based on a preset inter-distribution distance equation; importing the distance cost value, the preset learning rate, the preset smoothing constant and the initial discriminator weight value into a preset optimizer to complete weight updating of the preset discriminator so as to obtain a trained preset discriminator;
importing the preset learning rate, the preset smoothing constant, the initial encoder weight value and the trained preset discriminator into the preset optimizer to finish the updating of the preset encoder weight value; importing the preset learning rate, the preset smoothing constant, the initial encoder weight value and the trained preset discriminator into the preset optimizer to finish updating of the weight value of the preset generator; so as to obtain a trained preset encoder and a preset generator;
obtaining verification sample splicing data corresponding to sample data in a verification data set based on a trained preset encoder; obtaining verification label splicing data based on the trained preset generator or label splicing data; and matching the verification sample splicing data with the verification label splicing data based on a trained preset discriminator or a preset matching degree calculation formula.
2. The dynamic natural language based data recognition method of claim 1,
the hidden layer of the preset encoder and the hidden layer of the preset generator both comprise a text semantic coding network layer and a label semantic coding network layer.
3. The dynamic data identification method based on natural language according to claim 1, wherein determining semantic tag data corresponding to each sample data in the sample set specifically includes:
acquiring a semantic tag set through a preset semantic tag interface; wherein the semantic tag set comprises semantic tag data; or,
obtaining semantic tag data corresponding to each sample data through a preset keyword/subject term extraction algorithm;
or analyzing the part of speech of the sample data through a preset sample part of speech analysis algorithm to obtain preset attribute words corresponding to the sample data, and splicing the preset attribute words into semantic tag data;
or when a preset associated data set corresponding to the sample set is obtained, extracting keywords/subject terms corresponding to the preset associated data set through a keyword/subject term extraction algorithm and a preset sample part-of-speech analysis algorithm to obtain semantic tag data.
4. The dynamic data identification method based on natural language according to claim 1, wherein the obtaining of the sample set specifically includes:
and acquiring real service data or replacing open source service data or artificial sample data as a sample set through a preset sample uploading process.
5. The dynamic natural language-based data recognition method of claim 1, wherein before determining the cost value of the distance between the sample-stitched data and the label-stitched data based on a preset inter-distribution distance equation, the method further comprises:
replacing joint distribution in the Wasserstein-distance method with an encoder, replacing edge distribution with a generator, and replacing sampling with sample splicing data and label splicing data;
6. The dynamic data recognition method according to claim 1, wherein the step of importing the distance cost value, the preset learning rate, the preset smoothing constant, and the initial discriminator weight value into a preset optimizer to complete updating of the weight of the preset discriminator includes:
updating the formula by the preset discriminator weight:updating the weight value of the preset discriminator; wherein,for the weight values of the preset encoders generated in the updating process,in order to be the cost value of the distance,in order to preset the learning rate,andis a preset smoothing constant; when the weight value is larger than c or smaller than-c, through a preset clipping formula:performing gradient clipping on the weight value of the preset discriminator; where c is the clipping threshold.
7. The dynamic data recognition method according to claim 1, wherein the updating of the preset encoder weight value is completed by importing the preset learning rate, the preset smoothing constant, the initial encoder weight value, and the trained preset discriminator into the preset optimizer, and specifically includes:
updating the formula by presetting the encoder weight:updating the weight value of a preset encoder; wherein,to preset the weight values of the encoders generated in the updating process,the data is spliced for the samples and,in order to preset the learning rate,andd () is the preset discriminator output result, j represents the value from 1.
8. The dynamic data recognition method according to claim 1, wherein the step of importing the preset learning rate, the preset smoothing constant, the initial encoder weight value, and the trained preset discriminator into the preset optimizer to complete updating of the preset generator weight value includes:
updating the formula by presetting the weight of the generator:updating the weight value of the preset generator; wherein,to update the weight values of the preset generators generated in the process,the data is spliced for the label and,in order to preset the learning rate,andd () is a preset discriminator output result, and m represents the number of computations of D (), for a preset smoothing constant.
9. The dynamic natural language based data recognition method of claim 1, wherein after obtaining the trained predictive coder and predictive generator, the method further comprises:
and modifying the semantic label data by presetting a semantic label modification interface.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202211359030.2A CN115408498B (en) | 2022-11-02 | 2022-11-02 | Data dynamic identification method based on natural language |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202211359030.2A CN115408498B (en) | 2022-11-02 | 2022-11-02 | Data dynamic identification method based on natural language |
Publications (2)
Publication Number | Publication Date |
---|---|
CN115408498A CN115408498A (en) | 2022-11-29 |
CN115408498B true CN115408498B (en) | 2023-03-24 |
Family
ID=84169251
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202211359030.2A Active CN115408498B (en) | 2022-11-02 | 2022-11-02 | Data dynamic identification method based on natural language |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN115408498B (en) |
Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN113936217A (en) * | 2021-10-25 | 2022-01-14 | 华中师范大学 | Priori semantic knowledge guided high-resolution remote sensing image weakly supervised building change detection method |
CN115049936A (en) * | 2022-08-12 | 2022-09-13 | 武汉大学 | High-resolution remote sensing image-oriented boundary enhancement type semantic segmentation method |
Family Cites Families (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111583276B (en) * | 2020-05-06 | 2022-04-19 | 西安电子科技大学 | CGAN-based space target ISAR image component segmentation method |
CN111582175B (en) * | 2020-05-09 | 2023-07-21 | 中南大学 | High-resolution remote sensing image semantic segmentation method for sharing multi-scale countermeasure features |
CN112784965B (en) * | 2021-01-28 | 2022-07-29 | 广西大学 | Large-scale multi-element time series data anomaly detection method oriented to cloud environment |
CN114973062B (en) * | 2022-04-25 | 2024-08-20 | 西安电子科技大学 | Multimode emotion analysis method based on Transformer |
CN115035418A (en) * | 2022-06-15 | 2022-09-09 | 杭州电子科技大学 | Remote sensing image semantic segmentation method and system based on improved deep LabV3+ network |
-
2022
- 2022-11-02 CN CN202211359030.2A patent/CN115408498B/en active Active
Patent Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN113936217A (en) * | 2021-10-25 | 2022-01-14 | 华中师范大学 | Priori semantic knowledge guided high-resolution remote sensing image weakly supervised building change detection method |
CN115049936A (en) * | 2022-08-12 | 2022-09-13 | 武汉大学 | High-resolution remote sensing image-oriented boundary enhancement type semantic segmentation method |
Also Published As
Publication number | Publication date |
---|---|
CN115408498A (en) | 2022-11-29 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN110008349B (en) | Computer-implemented method and apparatus for event risk assessment | |
CN110147445A (en) | Intension recognizing method, device, equipment and storage medium based on text classification | |
CN110489555A (en) | A kind of language model pre-training method of combination class word information | |
CN104503998B (en) | For the kind identification method and device of user query sentence | |
CN111428504B (en) | Event extraction method and device | |
CN111091009B (en) | Document association auditing method based on semantic analysis | |
CN104462064A (en) | Method and system for prompting content input in information communication of mobile terminals | |
CN113986660A (en) | Matching method, device, equipment and storage medium of system adjustment strategy | |
CN111177402A (en) | Evaluation method and device based on word segmentation processing, computer equipment and storage medium | |
CN112671985A (en) | Agent quality inspection method, device, equipment and storage medium based on deep learning | |
CN112257425A (en) | Power data analysis method and system based on data classification model | |
CN110110087A (en) | A kind of Feature Engineering method for Law Text classification based on two classifiers | |
CN114661951A (en) | Video processing method and device, computer equipment and storage medium | |
CN114491034A (en) | Text classification method and intelligent device | |
CN113868422A (en) | Multi-label inspection work order problem traceability identification method and device | |
CN114064893A (en) | Abnormal data auditing method, device, equipment and storage medium | |
CN115408498B (en) | Data dynamic identification method based on natural language | |
CN117933253A (en) | Data processing method, device, electronic equipment and computer readable medium | |
CN110852082B (en) | Synonym determination method and device | |
CN105653619B (en) | The update method and device in correct log library in intelligent Answer System | |
CN116663547A (en) | Sample generation method and device | |
CN110851572A (en) | Session labeling method and device, storage medium and electronic equipment | |
CN114429140A (en) | Case cause identification method and system for causal inference based on related graph information | |
CN113449506A (en) | Data detection method, device and equipment and readable storage medium | |
Dong et al. | End-to-end topic classification without asr |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |