CN114265764A - Deep learning operator test data generation method based on weighted sampling - Google Patents
Deep learning operator test data generation method based on weighted sampling Download PDFInfo
- Publication number
- CN114265764A CN114265764A CN202111471772.XA CN202111471772A CN114265764A CN 114265764 A CN114265764 A CN 114265764A CN 202111471772 A CN202111471772 A CN 202111471772A CN 114265764 A CN114265764 A CN 114265764A
- Authority
- CN
- China
- Prior art keywords
- variation
- mutation
- sampling
- test data
- group
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
- 238000000034 method Methods 0.000 title claims abstract description 134
- 238000012360 testing method Methods 0.000 title claims abstract description 66
- 238000005070 sampling Methods 0.000 title claims abstract description 49
- 238000013135 deep learning Methods 0.000 title claims abstract description 26
- 230000035772 mutation Effects 0.000 claims abstract description 46
- 238000012614 Monte-Carlo sampling Methods 0.000 claims abstract description 6
- 239000011159 matrix material Substances 0.000 claims description 13
- 238000012216 screening Methods 0.000 claims description 13
- 238000002474 experimental method Methods 0.000 claims description 4
- 238000013100 final test Methods 0.000 claims description 4
- 230000001960 triggered effect Effects 0.000 claims description 4
- 238000004458 analytical method Methods 0.000 claims description 3
- 238000012217 deletion Methods 0.000 claims description 3
- 230000037430 deletion Effects 0.000 claims description 3
- 230000000694 effects Effects 0.000 claims description 3
- 230000001172 regenerating effect Effects 0.000 claims description 2
- 230000009286 beneficial effect Effects 0.000 claims 1
- 238000010606 normalization Methods 0.000 claims 1
- 238000013507 mapping Methods 0.000 abstract description 2
- 238000004422 calculation algorithm Methods 0.000 description 9
- 238000010586 diagram Methods 0.000 description 6
- 238000013136 deep learning model Methods 0.000 description 3
- 238000001514 detection method Methods 0.000 description 3
- 238000010801 machine learning Methods 0.000 description 3
- 238000013473 artificial intelligence Methods 0.000 description 2
- 210000004556 brain Anatomy 0.000 description 2
- 238000004364 calculation method Methods 0.000 description 2
- 238000013461 design Methods 0.000 description 2
- 238000005516 engineering process Methods 0.000 description 2
- 230000006870 function Effects 0.000 description 2
- 238000012986 modification Methods 0.000 description 2
- 230000004048 modification Effects 0.000 description 2
- 238000012549 training Methods 0.000 description 2
- 238000012935 Averaging Methods 0.000 description 1
- 238000000342 Monte Carlo simulation Methods 0.000 description 1
- 230000005856 abnormality Effects 0.000 description 1
- 238000013528 artificial neural network Methods 0.000 description 1
- 230000008030 elimination Effects 0.000 description 1
- 238000003379 elimination reaction Methods 0.000 description 1
- 238000011156 evaluation Methods 0.000 description 1
- 238000004519 manufacturing process Methods 0.000 description 1
- 238000007620 mathematical function Methods 0.000 description 1
- 238000005259 measurement Methods 0.000 description 1
- 238000003058 natural language processing Methods 0.000 description 1
- 238000011160 research Methods 0.000 description 1
- 238000012163 sequencing technique Methods 0.000 description 1
- 238000010200 validation analysis Methods 0.000 description 1
Images
Landscapes
- Measuring Or Testing Involving Enzymes Or Micro-Organisms (AREA)
Abstract
A deep learning operator test data generation method based on weighted sampling is characterized in that a group of mutation methods designed in advance are regarded as mapping to different data value taking domains, Monte Carlo sampling is carried out in advance, so that the effectiveness of each mutation method is obtained, sampling is carried out by taking the effectiveness as the weight, and finally, deep learning operator test data are generated more effectively while data diversity is considered. The data variation method combination can be flexibly added, deleted and changed, and the basic variation method provided by the method comprises the following steps: byte, noise, and inversion variations. The byte variation refers to the variation of byte operation on the floating point number binary code of the test data, and comprises the steps of adding, deleting, negating, shifting and randomly resetting each byte; noise variation refers to the random use of various types of noise applied to test data; the inverse variation is a variation in which the pseudo-reciprocal of test data is calculated using the test data as a denominator. The above variations are based on different basic ideas, taking into account both effectiveness and diversity.
Description
Technical Field
The invention belongs to the field of deep learning test, in particular to the field of testing accuracy of deep learning operators. Users of various deep learning models wish to test whether the deep learning operators on which the models depend have accuracy problems by some method.
Background
With the gradual maturity of deep learning technology and the emerging demand of intelligent application, various deep learning models are proposed and widely applied to various scenes in various fields. However, people tend to pay attention to the correctness and completeness of the overall function or flow of a model when researching, using and testing a deep learning model, and ignore the correctness of a specific operator depended on the back of the model. In fact, the deep learning operator often causes a series of accuracy problems, which cause serious damage and influence on the model, and therefore, an accuracy test for the deep learning operator is necessary.
Different from the traditional test, the test case generation of the deep learning operator accuracy test is difficult, and people cannot accurately know the specific characteristics of the effective test case, so that the test case cannot be accurately designed, and the effective test case can be searched only in a random sampling mode. However, pure uniform random sampling has many problems, which are often extremely inefficient and do not give enough attention to some important value intervals, such as: floating point numbers less than 1. By taking the idea of the conventional mutation test as a reference, mutating the randomly generated test data can effectively improve the probability of the problem of the test triggering accuracy, but the problem still exists in that it is not possible to determine which mutation method is used at which time. The present invention solves the above problems well by not only providing a more efficient set of basic variation methods and allowing for flexible variation, but also by means of monte carlo methods and weighted sampling, a solution is proposed to determine how to use which variation method at what time.
Disclosure of Invention
The invention aims to solve the problems that: the test case generation of the deep learning operator accuracy test is difficult to accurately design, and the generation efficiency is low. People cannot accurately know the specific characteristics of the effective test cases, so that the test cases cannot be precisely designed, and the effective test cases can be searched only in a random sampling mode. However, pure uniform random sampling has many problems, which are often extremely inefficient and do not give enough attention to some important value intervals, such as: floating point numbers less than 1. Aiming at the problems, the invention designs a series of test data mutation methods, and evaluates the effectiveness of the methods through Monte Carlo sampling, so that a proper mutation method and corresponding weights are selected to carry out weighted mutation data sampling, and the generation efficiency, effectiveness and diversity of test cases of the deep learning operator accuracy test are greatly improved.
The technical scheme of the invention is as follows: a deep learning operator test data generation method based on weighted sampling is characterized in that a group of data mutation methods which are designed in advance are regarded as mapping to different data value taking domains, and Monte Carlo sampling is carried out in advance, so that the effectiveness of each mutation method is obtained, sampling is carried out with weights by taking the effectiveness as the weights, and finally the purpose of more effectively generating deep learning operator test data under the condition of considering data diversity is achieved. The data mutation method combination can be flexibly added, deleted and modified according to the needs, and the basic data mutation method provided by the method comprises the following steps: byte variation, noise variation, and inversion variation. The byte variation refers to the variation of byte operation on floating point number binary codes of the test data, the operation comprises the addition, deletion, negation, shift and random reset of each byte of the binary codes, and the floating 32 and floating 16 format codes are supported; noise variance refers to uniformly randomly applying gaussian noise or uniform noise to test data; the inversion variation is the variation of the pseudo reciprocal of the test data in a value domain calculated by taking the test data as a denominator. The above variations are based on different basic ideas, such as: the boundary data is more likely to cause abnormality, the uniform sampling is disadvantageous for a number less than 1, and the like, and both effectiveness and diversity are taken into consideration. The method comprises the following steps:
1) the variation method is defined as follows: the invention predefines several variation methods as basic variation groups, the specific variation methods are stated in 1.1, and the variation methods can be added, reduced and modified as required during actual use to define the variation method group specific to the practice.
1.1) the basic variation method group comprises the following variation methods: byte variation, noise variation, and inversion variation. The byte variation refers to the variation of byte operation on floating point number binary codes of the test data, the operation comprises the addition, deletion, negation, shift and random reset of each byte of the binary codes, and the floating 32 and floating 16 format codes are supported; noise variance refers to uniformly randomly applying gaussian noise or uniform noise to test data; the inversion variation is to calculate the variation of the pseudo reciprocal of the test data in a value domain by taking the test data as a denominator;
2) random sampling is carried out, and the accuracy problem triggering success rate after variation of each variation method is obtained;
2.1) obtaining a seed random test case by using uniform random sampling;
2.2) applying each variation method to the seed random test case respectively and testing the variation effect, and counting the accuracy problem triggering success rate of each variation method after variation after multiple iterations;
3) and (3) effectiveness analysis: according to the basic idea of the Monte Carlo sampling method, the triggering success rate of the post-mutation accuracy problem of each mutation method is regarded as the effectiveness of the mutation method under the operator, and the sampling weight matrix is obtained by normalizing the triggering success rate;
4) screening by a variation method: screening the variation methods in the variation method group according to needs;
4.1) mutation method elimination: for some special cases, such as: the method has a success rate of 0 or a variation method which greatly reduces the success rate, a large number of variation methods which have a success rate almost equal to that of a random method, application scenes do not require diversity and a large number of samples are urgently needed with high success rate, and the like, and a variation method group needs to be further screened;
4.2) the screening and rejecting operations mainly comprise: screening out the mutation method with high effectiveness of pre-K in the mutation group to form a new mutation group, and rejecting the mutation method with low effectiveness of pre-K in the mutation group to form a new mutation group, and if necessary, redefining the mutation group and repeating the whole process
4.3) matrix modification: modifying the validity matrix according to the screening, rejecting or modifying result of the variation method group, and regenerating a weight matrix for sampling with weight;
5) sampling with weight: according to the weight matrix obtained in step 4.2, sampling with weight and checking whether the accuracy problem is successfully triggered;
5.1) sequencing the weight matrix from large to small, simultaneously recording the index of each variation method, then uniformly sampling in the interval from 0 to the sum of the weights, and finding out the corresponding variation method, wherein the sum of the weights of the method and the previous variation methods is greater than or equal to the random value, and the sum of the weights of the previous variation methods is less than the random value;
5.2) randomly sampling to obtain seed test data, applying a mutation method obtained by sampling to the data to obtain final test data, and inputting the final test data into an operator to check whether the accuracy problem is successfully triggered;
the invention is characterized in that:
1. the method is a novel method, and the deep learning operator test data generation method giving consideration to data generation efficiency, effectiveness and diversity is realized by applying various variation methods to randomly generated seed test cases and respectively evaluating the effectiveness of the seed test cases;
2. a basic variation method group which is proved to be effective through a large amount of experiments and an effectiveness reference thereof are provided, and a method user is allowed to flexibly add, delete and modify variation methods according to needs;
3. the success rate of accuracy triggering after variation obtained by Monte Carlo sampling is regarded as the effectiveness measurement of the variation method, and a novel variation method evaluation method with theoretical basis is provided;
based on the three points, the method can effectively solve the problem of low test case generation efficiency of the deep learning operator accuracy test, remarkably improves the test case generation efficiency, effectiveness and diversity of the deep learning operator accuracy test, and better provides support for the subsequent deep learning operator accuracy test.
Drawings
FIG. 1 is a general architecture diagram of the present invention
FIG. 2 is a diagram of the variation method definition sub-process architecture of the present invention
FIG. 3 is a diagram of the Monte Carlo stochastic sampling subprocess architecture of the present invention
FIG. 4 is a diagram of the validation analysis subprocess architecture of the present invention
FIG. 5 is a diagram showing the framework of the screening subprocess of the mutation method of the present invention
FIG. 6 is a diagram of the weighted sampling sub-process architecture of the present invention
Detailed Description
The key technology related by the invention is that some existing deep learning operators are utilized to sample and generate test data based on a weighted mutation method, accuracy problem detection is carried out, the generation and the mutation of a random tensor are mainly realized through NumPy, and the accuracy problem detection and the deep learning operators mainly relate to TensorFlow, PyTorch, MNN framework and MRE/MARE algorithm.
1. Tensor generation and variation
In the invention, the production and variation of the random tensor are mainly carried out through a NumPy library, the NumPy is an extended program library of Python language, supports a large number of dimensional array and matrix operations, and provides a large number of mathematical function libraries aiming at array operations, including functions of linear algebra, random number generation and the like.
2. Deep learning operator
The invention relates to an operator under three frames, which are respectively: tensorflow, PyTorch, and MNN. TensorFlow is a symbolic mathematical system developed and maintained by Google Brain (Google Brain) of Google artificial intelligence team and based on data flow programming, and is widely applied to programming realization of various machine learning algorithms, and the predecessor of the symbolic mathematical system is a neural network algorithm library DistBelef of Google. PyTorch is an open-source Python machine learning library introduced by Facebook Artificial Intelligence research institute (FAIR) and is commonly used in various machine learning applications such as natural language processing based on Torch. MNN is an efficient, lightweight deep learning framework developed by ali that supports deep model reasoning and training, especially with the performance of reasoning and training at the end-side leading up in the industry. Currently, MNN has been widely used in more than 20 apps of acriba, moto, mao, youku, etc.
The deep learning operator related by the invention is realized under three frames respectively as follows:
(1)TensorFlow:tf.nn.bias_add、tf.nn.avg_pool、tf.nn.max_pool、tf.nn.softmax、tf.nn.sigmoid、tf.nn.tanh、tf.nn.relu、tf.nn.conv2d、tf.nn.reduce_mean、tf.matmul、tf.nn.reduce_max、tf.keras.layers.BatchNormalization。
(2)PyTorch:torch.add、F.avg_pool2d、F.max_pool2d、F.softmax、torch.sigmoid、torch.tanh、torch.nn.functional.relu、torch.nn.Conv2d、torch.mean、torch.matmul、torch.max、torch.nn.BatchNorm2d。
(3)MNN:MNN.expr.bias_add、MNN.expr.avg_pool、MNN.expr.max_pool、MNN.expr.softmax、MNN.expr.sigmoid、MNN.expr.tanh、MNN.expr.relu、MNN.nn.conv、MNN.expr.reduce_mean、MNN.expr.matmul、MNN.expr.reduce_max、MNN.nn.batch_norm。
3. accuracy problem detection
The algorithm for detecting the accuracy problem mainly comprises an MRE algorithm and a MARE algorithm. The MRE and MARE algorithms are defined as follows:
let the computation results of TensorFlow, PyTorch, and MNN be f for a given operator and a given input, respectivelyt,fp,fmThe variance between these results is VarTM(Tensorflow and MNN), VarTP(Tensorflow and PyTorch) and VarMP(MNN and PyTorch), the benchmark result f of the operator calculationbThe calculation method is as follows: if min (Var)TP,VarTM,VarMP)=VarTP,then fb=(ft+fp)/2;min(VarTP,VarTM,VarMP)=VarTM,then fb=(ft+fm)/2;if(VarTP,VarTM,VarMP)=VarMP,thenfb=(fm+fp)/2。
MRE and MARE are defined based on the error between the actual computation result and the benchmark result, respectively, as:
in actual use, whether the accuracy problem occurs is detected by comparing the actual calculated values of the MRE and the MARE with a preset threshold value.
4. Examples of the invention
The following uses specific examples to illustrate the steps of the present invention and to show the results.
The experimental environment is as follows: tensorflow 2.0, PyTorch 1.8.1, MNN 1.1.4, the graphics card is GeForce GTX 1080 Ti.
The overall process of the invention is shown in fig. 1, and the specific implementation steps are as follows:
1) defining a variation method group, wherein a basic variation method group is directly used in an experiment, the method group comprises 34 variation methods of 3 types, the random sampling iteration frequency is set to be 20000 times, and the sampling frequency with weight is set to be 15000 times;
2) and respectively carrying out experiments under MRE and MARE by using 36 operators under 3 frames, presetting two groups of different MRE and MARE algorithm threshold values beta and gamma for the operators, and respectively ensuring that the problem triggering rate of the accuracy of random sampling is below 15% and between 30% and 50%. The accuracy problem triggering rate of all operators under different conditions is obtained in the stage;
3) and (3) regarding the accuracy problem triggering rate as the effectiveness of the mutation method, converting the accuracy problem triggering rate into normalized weight matrixes, carrying out weighted sampling according to the weight matrixes after proper screening and modification, and testing after each sampling to detect whether the generated data successfully triggers the accuracy problem of the operator.
4) The experimental results generated by the operator test data under each frame are shown in table 1. The data are obtained by averaging the results of 4 conditions of two different threshold values of MRE and MARE, and the success rate of sampling performed by using the method is obviously higher than that of a random sampling algorithm under various operators of various frames.
Table 1 Generation of experimental results of operator test data under frames
Claims (6)
1. A deep learning operator test data generation method based on weighted sampling is characterized in that a variation method group is defined based on a basic variation method group which is defined in advance, Monte Carlo random sampling is carried out on each method in advance, so that the effectiveness of each variation method is obtained, the variation groups are further evaluated and refined, weighted sampling is finally carried out, and the purpose of more effectively generating deep learning operator test data under the condition of considering data diversity is finally achieved. The method comprises the following steps:
1) the variation method is defined as follows: the method predefines a plurality of variation methods as basic variation groups, and can add, reduce and modify the variation methods as required during actual use to define the variation method group exclusive for the practice, wherein the variation methods all belong to tensor variation methods and are different from the traditional scalar variation;
2) and (3) carrying out Monte Carlo random sampling, and counting to obtain the post-mutation accuracy problem trigger success rate of each mutation method: compared with the traditional scalar variation test, tensor variation is more complex, the effect is more difficult to predict and directly evaluate, and the most intuitive and reasonable method is to approximate the success rate of each variation method through Monte Carlo random sampling;
3) and (3) effectiveness analysis: according to the basic idea of the Monte Carlo sampling method, the triggering success rate of the post-mutation accuracy problem of each mutation method is regarded as the effectiveness of the mutation method under the operator, and the sampling weight matrix is obtained by normalizing the triggering success rate;
4) screening by a variation method: screening the variation methods in the variation method group according to needs;
5) sampling with weight: and according to the weight matrix obtained by the step 4.2, sampling with the weight and checking whether the accuracy problem is successfully triggered.
2. The variation method definition sub-process of claim 1, wherein:
1) a basic set of variant methods is predefined, the set of methods comprising: byte variation, noise variation, and inversion variation. The byte variation refers to the variation of byte operation on floating point number binary codes of the test data, the operation comprises the addition, deletion, negation, shift and random reset of each byte of the binary codes, and the floating 32 and floating 16 format codes are supported; noise variance refers to uniformly randomly applying gaussian noise or uniform noise to test data; the inversion variation is to calculate the variation of the pseudo reciprocal of the test data in a value domain by taking the test data as a denominator;
2) the above-described predefined basic variation methods are based on different basic ideas, such as: the boundary data are easy to cause abnormity, uniform sampling is not beneficial to the number less than 1, and the like, a large number of experiments prove that both effectiveness and diversity are considered, the target is tensor variation, and the variation operation is more complex than scalar variation and has more statistical characteristics;
3) based on the basic mutation method group, the basic mutation method group can be amplified, reduced and modified according to actual needs to form a new mutation method group, but the mutation methods are required to be ensured to be suitable for tensor and simultaneously suitable for float32 and float16 format coding.
3. The monte carlo random sampling sub-process of claim 1, wherein:
1) obtaining a seed random test case by using uniform random sampling;
2) each variation method is applied to the seed random test case respectively, the variation effect is tested, after multiple iterations, the accuracy problem triggering success rate after variation of each variation method is counted, all the variation methods are guaranteed to be treated fairly in the random sampling process, and the accuracy problem triggering success rate after variation is equal to the number of samples successfully triggering the accuracy problem/the total number of samples after variation.
4. The validity analysis subprocess of claim 1, wherein:
1) the triggering success rate of the post-mutation accuracy problem of each mutation method is regarded as the effectiveness of the mutation method under the operator, and the effectiveness of the mutation method is further analyzed;
2) and (4) regarding the effectiveness of each mutation method under a certain operator as a proportion expected to be used in weighted sampling, and obtaining a weight matrix of the mutation method group through normalization.
5. The variation method screening subprocess according to claim 1, characterized in that:
1) for some special cases: the method has a success rate of 0 or a variation method which greatly reduces the success rate, a large number of variation methods which have a success rate almost equal to that of a random method, application scenes do not require diversity and a large number of samples are urgently needed with high success rate, and the like, and a variation method group needs to be further screened;
2) the screening and removing operation mainly comprises the following steps: screening out the mutation method with high effectiveness of the front K in the mutation group to form a new mutation group, and rejecting the mutation method with low effectiveness of the front K in the mutation group to form a new mutation group, and if necessary, redefining the mutation group and repeating the whole process;
3) and modifying the validity matrix according to the screening, rejecting or modifying result of the variation method group, and regenerating a weight matrix for weighted sampling.
6. The weighted sampling sub-process of claim 1, wherein:
1) sorting the weight matrix from large to small, simultaneously recording the index of each variation method, uniformly sampling in the interval from 0 to the sum of the weights, and finding out the corresponding variation method, wherein the sum of the weights of the method and the previous variation methods is greater than or equal to the random value, and the sum of the weights of the previous variation methods is less than the random value;
2) and randomly sampling to obtain seed test data, applying a mutation method obtained by sampling to the data to obtain final test data, and inputting the final test data into an operator to check whether the accuracy problem is successfully triggered.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202111471772.XA CN114265764A (en) | 2021-11-30 | 2021-11-30 | Deep learning operator test data generation method based on weighted sampling |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202111471772.XA CN114265764A (en) | 2021-11-30 | 2021-11-30 | Deep learning operator test data generation method based on weighted sampling |
Publications (1)
Publication Number | Publication Date |
---|---|
CN114265764A true CN114265764A (en) | 2022-04-01 |
Family
ID=80826256
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202111471772.XA Pending CN114265764A (en) | 2021-11-30 | 2021-11-30 | Deep learning operator test data generation method based on weighted sampling |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN114265764A (en) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN115687158A (en) * | 2022-12-28 | 2023-02-03 | 北京邮电大学 | Protocol software testing method and device |
-
2021
- 2021-11-30 CN CN202111471772.XA patent/CN114265764A/en active Pending
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN115687158A (en) * | 2022-12-28 | 2023-02-03 | 北京邮电大学 | Protocol software testing method and device |
CN115687158B (en) * | 2022-12-28 | 2023-03-10 | 北京邮电大学 | Protocol software testing method and device |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN108647520B (en) | Intelligent fuzzy test method and system based on vulnerability learning | |
CN112508243B (en) | Training method and device for multi-fault prediction network model of power information system | |
CN110232280B (en) | Software security vulnerability detection method based on tree structure convolutional neural network | |
CN106022517A (en) | Risk prediction method and device based on nucleus limit learning machine | |
CN110956309A (en) | Flow activity prediction method based on CRF and LSTM | |
CN110162972B (en) | UAF vulnerability detection method based on statement joint coding deep neural network | |
CN112735541A (en) | Sewage treatment water quality prediction method based on simple circulation unit neural network | |
Masabo et al. | Big data: deep learning for detecting malware | |
CN114265764A (en) | Deep learning operator test data generation method based on weighted sampling | |
CN115114166A (en) | Ether house intelligent contract security testing method | |
Zhang et al. | Generating test data for both paths coverage and faults detection using genetic algorithms: multi-path case | |
Wu et al. | Spp-cnn: An efficient framework for network robustness prediction | |
Shaochuan | Bayesian multiple changepoint detection for stochastic models in continuous time | |
CN113159441A (en) | Prediction method and device for implementation condition of banking business project | |
CN117195044A (en) | Fault analysis method, device, computer equipment, storage medium and program product | |
Jin et al. | An approach to constructing effective training data for a classification model to evaluate the reliability of a passive safety system | |
Miranda et al. | How noisy data affects geometric semantic genetic programming | |
Zheng et al. | Software Defect Prediction Model Based on Improved Deep Forest and AutoEncoder by Forest. | |
CN115242457A (en) | Method and device for detecting log data, electronic equipment and storage medium | |
Obert et al. | Efficient binary static code data flow analysis using unsupervised learning | |
Liu et al. | MC/DC test data generation algorithm based on whale genetic algorithm | |
Mohan et al. | White-Box and black-box reliability modeling framework: integration through analytical model and user profile validation via deep learning—a practitioner’s approach | |
Singh et al. | Improving event log quality using autoencoders and performing quantitative analysis with conformance checking | |
Juneja et al. | Optimization of dejong function using ga under different selection algorithms | |
CN117056203B (en) | Numerical expression type metamorphic relation selection method based on complexity |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination |