WO2021164228A1 - Method and system for selecting augmentation strategy for image data - Google Patents

Method and system for selecting augmentation strategy for image data Download PDF

Info

Publication number
WO2021164228A1
WO2021164228A1 PCT/CN2020/111666 CN2020111666W WO2021164228A1 WO 2021164228 A1 WO2021164228 A1 WO 2021164228A1 CN 2020111666 W CN2020111666 W CN 2020111666W WO 2021164228 A1 WO2021164228 A1 WO 2021164228A1
Authority
WO
WIPO (PCT)
Prior art keywords
strategy
classification
sample
classification model
trained
Prior art date
Application number
PCT/CN2020/111666
Other languages
French (fr)
Chinese (zh)
Inventor
王俊
高鹏
谢国彤
杨苏辉
Original Assignee
平安科技(深圳)有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 平安科技(深圳)有限公司 filed Critical 平安科技(深圳)有限公司
Publication of WO2021164228A1 publication Critical patent/WO2021164228A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02TCLIMATE CHANGE MITIGATION TECHNOLOGIES RELATED TO TRANSPORTATION
    • Y02T10/00Road transport of goods or passengers
    • Y02T10/10Internal combustion engine [ICE] based vehicles
    • Y02T10/40Engine management systems

Definitions

  • This application relates to the field of artificial intelligence technology, and in particular to a method and system for selecting an augmentation strategy for image data.
  • the inventor realizes that there are various augmentation strategies currently available, and they perform differently when facing different data sets, and it is difficult to determine which augmentation strategy is most effective for the current type of image data set.
  • the embodiments of the present application provide a method and device for selecting an augmentation strategy for image data to solve the problem in the prior art that it is difficult to determine which augmentation strategy is most effective for the current type of image data set.
  • an augmentation strategy selection method for image data includes: selecting a plurality of pending strategy subsets from an augmentation strategy set to a preset sample training set Perform sample augmentation to obtain multiple augmented sample training sets, wherein each of the pending strategy subsets is composed of at least one augmentation strategy in the augmentation strategy set; each of the augmented strategy subsets is used Training the initialized classification model on the sample training set to obtain multiple trained classification models; inputting a preset sample verification set into each of the trained classification models to obtain the classification accuracy corresponding to the trained classification model; A Bayesian optimization algorithm is used to determine the optimal strategy subset from a plurality of pending strategy subsets based on the classification accuracy corresponding to each of the trained classification models.
  • an augmentation strategy selection system for image data including an augmenter, a classification model, and a controller;
  • the augmenter is used to select multiple undetermined strategy subsets from the augmentation strategy set to perform sample augmentation on the preset sample training set to obtain multiple augmented sample training sets, wherein each of the The undetermined strategy subset is composed of at least one augmentation strategy in the augmentation strategy set;
  • the classification model is used to train an initialized classification model using each of the augmented sample training sets to obtain a plurality of trained classification models; and input a preset sample verification set into each of the trained samples A classification model to obtain the classification accuracy corresponding to the trained classification model;
  • the controller is configured to use a Bayesian optimization algorithm to determine an optimal strategy subset from a plurality of pending strategy subsets based on the classification accuracy corresponding to each of the trained classification models.
  • a computer-readable storage medium includes a stored program, and when the program runs, the device where the storage medium is located is controlled to perform the following steps:
  • a Bayesian optimization algorithm is used to determine the optimal strategy subset from a plurality of pending strategy subsets based on the classification accuracy corresponding to each of the trained classification models.
  • a computer device including a memory, a processor, and a computer program stored in the memory and running on the processor, and the processor executes all When the computer program is described, the following steps are implemented:
  • a Bayesian optimization algorithm is used to determine the optimal strategy subset from a plurality of pending strategy subsets based on the classification accuracy corresponding to each of the trained classification models.
  • each augmented sample training set is used to train the initialized classification model to obtain multiple trained classification models, and use samples to verify Validate the trained classification model, and then according to the classification accuracy of the classification model and the Bayesian optimization algorithm to obtain a suitable augmentation strategy that meets this type of sample, which can improve the efficiency of augmentation strategy selection.
  • FIG. 1 is a flowchart of an optional method for selecting an augmentation strategy for image data according to an embodiment of the present application
  • FIG. 2 is a schematic diagram of an optional image data augmentation strategy selection system provided by an embodiment of the present application.
  • FIG. 3 is a functional block diagram of an optional controller provided by an embodiment of the present application.
  • Fig. 4 is a schematic diagram of an optional computer device provided by an embodiment of the present application.
  • first, second, third, etc. may be used to describe terminals in the embodiments of the present application, these terminals should not be limited to these terms. These terms are only used to distinguish terminals from each other.
  • the first terminal may also be referred to as the second terminal, and similarly, the second terminal may also be referred to as the first terminal.
  • the word “if” as used herein can be interpreted as “when” or “when” or “in response to determination” or “in response to detection”.
  • the phrase “if determined” or “if detected (statement or event)” can be interpreted as “when determined” or “in response to determination” or “when detected (statement or event) )” or “in response to detection (statement or event)”.
  • Fig. 1 is a flowchart of a method for selecting an augmentation strategy for image data according to an embodiment of the present application. As shown in Fig. 1, the method includes:
  • Step S01 Select multiple pending strategy subsets from the augmented strategy set to perform sample augmentation on the preset sample training set to obtain multiple augmented sample training sets, where each pending strategy subset is augmented by At least one augmentation strategy in the strategy set;
  • Step S02 Use each augmented sample training set to train an initialized classification model to obtain multiple trained classification models
  • Step S03 input the preset sample verification set into each trained classification model, and obtain the classification accuracy corresponding to the trained classification model;
  • step S04 the Bayesian optimization algorithm is used to determine the optimal strategy subset from the multiple pending strategy subsets based on the classification accuracy corresponding to each trained classification model.
  • the samples in the sample training set are graphic data samples.
  • each augmented sample training set is used to train the initialized classification model to obtain multiple trained classification models, and use samples to verify Validate the trained classification model, and then according to the classification accuracy of the classification model and the Bayesian optimization algorithm to obtain a suitable augmentation strategy that meets this type of sample, which can improve the efficiency of augmentation strategy selection.
  • Step S01 Select multiple pending strategy subsets from the augmented strategy set to perform sample augmentation on the preset sample training set to obtain multiple augmented sample training sets, where each pending strategy subset is augmented by At least one augmentation strategy in the strategy set;
  • the samples in the sample training set are medical image samples of the same type, such as lung images and stomach images.
  • Each training sample has a label.
  • a training sample with a positive label is a lung image marked as having pneumonia symptoms
  • a training sample with a negative label is a lung image without pneumonia symptoms.
  • the training sample is a 512*512 medical image sample.
  • augmentation strategies include rotation transformation, flip transformation, zoom transformation, translation transformation, scale transformation, region cropping, adding noise, segmented affine, random masking, boundary detection, contrast transformation, color dithering, random mixing, and composite overlay.
  • the augmentation strategy is, for example, flipped transformation.
  • Scale the image is enlarged or reduced according to the preset scale factor, or the scale space is constructed by filtering the image with the preset scale factor to change the size or blur degree of the image content;
  • Crop Crop the region of interest of the picture
  • Piecewise Affine Place a regular point grid on the image, and move these points and the surrounding image area according to the number of samples in the normal distribution;
  • Random concealment The loss of information is realized in a rectangular area with a selectable area and a random location to achieve conversion. The loss of information in all channels produces black rectangular blocks, and the loss of information in some channels produces color noise;
  • Edge Detect Detect all edges in the image, mark them as black and white images, and then superimpose the result with the original image;
  • Contrast transformation In the HSV color space of the image, change the saturation S and V brightness components, keep the hue H unchanged, and perform exponential calculations on the S and V components of each pixel (the exponent factor is between 0.25 and 4). Time), increase the light changes;
  • Color jitter Randomly change the exposure, saturation and hue of the image to form pictures under different lighting and colors, as much as possible to make the model use different light conditions as small as possible Situation
  • Random mixing (Mix up): a data augmentation method based on the principle of neighborhood risk minimization, using linear interpolation to obtain new sample data;
  • Sample Pairing Two images are randomly selected after being processed by the basic data augmentation operation and then superimposed into a new sample in the form of pixel averaging.
  • the label of the new template is one of the original sample labels.
  • any three augmentation strategies are randomly selected from the above 14 augmentation strategies to form a undetermined strategy subset, that is, a undetermined strategy subset includes 3 augmentation strategies, and each augmentation strategy includes 3 A strategy parameter, namely strategy type ( ⁇ ), probability value ( ⁇ ), amplitude ( ⁇ ).
  • a subset of pending strategies can be represented in the form of a numerical matrix:
  • each row represents an augmentation strategy.
  • Step S02 Use each augmented sample training set to train an initialized classification model to obtain multiple trained classification models.
  • the classification model is a convolutional neural network model, which is composed of a convolutional neural network and a fully connected network, and its specific structure includes at least a convolutional network layer, a pooling layer, and a fully connected network layer.
  • the specific steps during training include:
  • the convolutional neural network uses the convolutional neural network to extract the feature map of each sample in the augmented sample training set of the input classification model; according to the feature map, classify and predict a corresponding sample in the augmented sample training set to obtain the classification result; obtain The classification result set and the loss function of the mean square error of the label set of all samples in the sample training set; the convolutional neural network is optimized through backpropagation, so that the value of the loss function converges, and the optimized and trained classification model is obtained.
  • the initial convolutional neural network performs feature extraction on labeled samples and performs preset rounds of training, so that the convolutional neural network layer can effectively extract more generalized features (such as edges, textures, etc.).
  • the accuracy of the model can be improved after the gradient is continuously reduced, so that the value of the loss function converges to the minimum, and the weight and bias of the convolutional layer and the fully connected layer are automatically adjusted to make the classification model the best optimization.
  • the classification model may also be a long- and short-term neural network model, a random forest model, a support vector machine model, a maximum entropy model, etc., which are not limited here.
  • Step S03 Input the preset sample verification set into each trained classification model to obtain the classification accuracy corresponding to the trained classification model.
  • the samples in the preset sample validation set are also labeled.
  • a training sample with a positive label is a lung image marked as having pneumonia symptoms
  • a training sample with a negative label is marked as having no pneumonia symptoms.
  • a preset sample verification set is used to verify the trained classification model.
  • the sample verification set corresponding to each classification model is different, which can achieve better model generalization performance and effectively solve the problem of overfitting that may be introduced by sample augmentation.
  • the method further includes:
  • the sample size ratio in the sample training set and the sample verification set may be 2:8, 4:6, 6:4, 8:2, etc. Understandably, each time a sample is drawn, 50% of the samples in the sample verification set are randomly selected to form a verification subset. In other embodiments, the ratio of random selection may be 30%, 40%, 60%, and so on.
  • a cross-validation method is used to validate the classification model.
  • the cross-validation method is either a ten-fold cross-validation method or a five-fold cross-validation method.
  • a five-fold cross-validation method is adopted. Specifically, multiple training samples are randomly divided into 10 parts, and 2 of them are taken as the cross-validation set each time, and the remaining 8 parts are used as the training set. When training, first use 8 of them to train the initialized classification model, and then classify and label the 2 cross-validation sets, so as to repeat the training and verification process 5 times, each time the selected cross-validation set is different, until all The training samples of are all classified and labeled again.
  • Step S03 specifically includes:
  • Step S031 input the preset sample verification set into each trained classification model
  • Step S032 Obtain the training accuracy and verification accuracy of the classification model output
  • Step S033 judging whether the classification model fits well according to the training accuracy and the verification accuracy
  • Step S034 Determine the well-fitted classification model as the trained classification model, and use the verification accuracy of the trained classification model as the classification accuracy of the classification model.
  • the training rounds of the classification model can be preset.
  • the training round is 100 trainings.
  • the sample validation set is input into the classification model to obtain the output of the classification model.
  • the training accuracy and verification accuracy of the classification model are judged to determine whether the training classification model fits well. Specifically, when (training accuracy-verification accuracy)/verification accuracy ⁇ 10%, then the classification model is considered The fit is good.
  • the verification accuracy of a well-fitted classification model is used as the classification accuracy.
  • step S04 the Bayesian optimization algorithm is used to determine the optimal strategy subset from the multiple pending strategy subsets based on the classification accuracy corresponding to each trained classification model.
  • the undetermined strategy subset (numerical matrix) is used as the x value of the sample point, and the classification accuracy is used as the y value of the sample point to form multiple sample points, and A regression model of the Gaussian process is constructed based on multiple sample points, and the objective function is learned and fitted to find a subset of strategies that promote the objective function to the global optimal value.
  • Step S04 specifically includes:
  • the optimal strategy subset is determined from multiple pending strategy subsets. Among them, the classification model trained on the sample training set after augmenting the optimal strategy subset has the highest classification accuracy.
  • the optimal strategy subset is determined from a plurality of pending strategy subsets based on the classification accuracy and the Bayesian optimization algorithm. In other embodiments, other algorithms may also be used for selection, which is not limited here.
  • Bayesian optimization algorithm learns and fits the acquisition function to find the objective function
  • the strategy parameter that f(x) promotes to the global optimal value.
  • every Bayesian optimization iteration uses a new sample point to test the objective function f(x)
  • use this information to update the prior distribution of the objective function f(x)
  • use the Bayesian optimization algorithm to test the posterior The distribution gives the sample points where the global maximum value is most likely to occur.
  • the covariance is only related to x, and has nothing to do with y.
  • the posterior probability distribution of f n+1 can be estimated through the first n sample points : P(f n+1
  • the probability of improvement (POI) is used as the acquisition function.
  • the get function is:
  • f(x) is the objective function value of x
  • x is the verification accuracy
  • f(X+) is the optimal objective function value of x so far
  • ⁇ (x) and ⁇ (x) are obtained by Gaussian process respectively
  • the mean and variance of the objective function are the posterior distribution of f(x)
  • ⁇ ( ⁇ ) represents the normal cumulative distribution function.
  • is the trade-off coefficient. If there is no such coefficient, the POI function will tend to take a point around X+ and converge to a position close to f(X+), that is, it tends to develop rather than explore, so this item is added to make a trade-off. By constantly trying new x, the next largest point should be larger or at least equal to it.
  • the next sample is between the intersection f(X+) and the confidence region.
  • f(X+) we can assume that the samples below f(X+) can be discarded, because we only need to search for the parameter that makes the objective function take the maximum value. So through the iterative process, the observation area is reduced until the optimal solution is searched, so that the POI(X) is maximized.
  • the embodiment of the present application provides an augmentation strategy selection system for image data.
  • the system includes an augmenter 10, a classification model 20, and a controller 30;
  • the augmenter 10 is used to select multiple undetermined strategy subsets from the augmentation strategy set to perform sample augmentation on the preset sample training set to obtain multiple augmented sample training sets, where each undetermined strategy sub-set
  • the set consists of at least one augmentation strategy in the augmentation strategy set.
  • the augmentation strategy set includes rotation transformation, flip transformation, zoom transformation, translation transformation, scale transformation, region cropping, noise addition, segmented affine, random masking, boundary detection, contrast transformation, color dithering, random mixing and compounding Overlay.
  • the augmentation strategy is, for example, flip transformation.
  • any of the three augmentation strategies mentioned above are randomly selected to form a subset of undetermined strategies.
  • Each augmentation strategy includes three strategy parameters, which are strategy type ( ⁇ ), probability value ( ⁇ ), and amplitude ( ⁇ ). ). Then a subset of pending strategies can be represented in the form of a numerical matrix:
  • each row represents an augmentation strategy.
  • the classification model 20 includes a training unit 210 and a verification unit 220.
  • the training unit 210 is configured to use each augmented sample training set to train an initialized classification model to obtain multiple trained classification models;
  • the verification unit 220 is configured to input a preset sample validation set into each trained The classification model obtains the classification accuracy corresponding to the trained classification model.
  • the classification model is a convolutional neural network model, which is composed of a convolutional neural network and a fully connected network, and its specific structure includes at least a convolutional network layer, a pooling layer, and a fully connected network layer.
  • the training unit 210 includes an extraction subunit, a classification subunit, a first acquisition subunit, and an optimization subunit.
  • the extraction subunit is used to use the convolutional neural network to extract the feature map of each sample in the augmented sample training set of the input classification model; the classification subunit is used to compare the augmented sample training set according to the feature map Perform classification prediction corresponding to a sample to obtain the classification result; obtain the subunit, which is used to obtain the loss function of the mean square error between the classification result set and the label set of all samples in the sample training set; the optimization subunit is used to pair The convolutional neural network is optimized to make the value of the loss function converge to obtain the optimized and trained classification model.
  • the initial convolutional neural network performs feature extraction on labeled samples and performs preset rounds of training, so that the convolutional neural network layer can effectively extract more generalized features (such as edges, textures, etc.).
  • the accuracy of the model can be improved after the gradient is continuously reduced, so that the value of the loss function converges to the minimum, and the weight and bias of the convolutional layer and the fully connected layer are automatically adjusted to make the classification model the best optimization.
  • the samples in the preset sample validation set are also labeled.
  • a training sample with a positive label is a lung image marked as having pneumonia symptoms
  • a training sample with a negative label is marked as having no pneumonia symptoms.
  • a preset sample validation set is used to verify the trained classification model.
  • the sample validation set corresponding to each classification model is different, which can achieve better model generalization performance and effectively solve the problem of overfitting that may be introduced by sample augmentation.
  • the verification unit 220 includes an input subunit, a second acquisition subunit, a judgment subunit, and a determination subunit.
  • the input subunit is used to input the preset sample verification set into each trained classification model
  • the second acquisition subunit is used to acquire the training accuracy and verification accuracy of the classification model output
  • the judgment subunit is used to judge whether the classification model fits well according to the training accuracy and verification accuracy
  • the determination subunit is used to determine a well-fitted classification model as a trained classification model, and use the verification accuracy of the trained classification model as the classification accuracy of the classification model.
  • the training rounds of the classification model can be preset.
  • the training round is 100 trainings.
  • the sample validation set is input into the classification model to obtain the output of the classification model.
  • the training accuracy and verification accuracy of the classification model are judged to determine whether the training classification model fits well. Specifically, when (training accuracy-verification accuracy)/verification accuracy ⁇ 10%, then the classification model is considered The fit is good.
  • the verification accuracy of a well-fitted classification model is used as the classification accuracy.
  • the system also includes a database 40 and a processing module 50.
  • the database 40 is used to store the sample training set and the sample verification set.
  • the processing module 50 is configured to randomly select a plurality of verification subsets from a preset sample verification set; input the plurality of verification subsets into each trained classification model respectively.
  • the sample size ratio in the sample training set and the sample verification set may be 2:8, 4:6, 6:4, 8:2, etc. Understandably, each time a sample is drawn, 50% of the samples in the sample verification set are randomly selected to form a verification subset. In other embodiments, the ratio of random selection may be 30%, 40%, 60%, and so on.
  • a cross-validation method is used to validate the classification model.
  • the cross-validation method is either a ten-fold cross-validation method or a five-fold cross-validation method.
  • a five-fold cross-validation method is adopted. Specifically, multiple training samples are randomly divided into 10 parts, and 2 of them are taken as the cross-validation set each time, and the remaining 8 parts are used as the training set. When training, first use 8 of them to train the initialized classification model, and then classify and label the 2 cross-validation sets, so as to repeat the training and verification process 5 times, each time the selected cross-validation set is different, until all The training samples of are all classified and labeled again.
  • the controller 30 is configured to use a Bayesian optimization algorithm to determine an optimal strategy subset from a plurality of pending strategy subsets based on the classification accuracy corresponding to each trained classification model.
  • the controller 30 determines an optimal strategy subset from a plurality of pending strategy subsets based on the classification accuracy and using a Bayesian optimization algorithm. In other embodiments, other algorithms may also be used for selection, which is not limited here.
  • the controller 30 includes a construction unit 310, a first determination unit 320, and a second determination unit 330.
  • the constructing unit 310 is configured to construct a regression model of the Gaussian process based on a plurality of sample points, where each sample point includes the classification accuracy of the trained classification model and the undetermined strategy subset used to train the classification model;
  • the first determining unit 320 is configured to determine the acquisition function of the Bayesian optimization algorithm according to the regression model
  • the second determining unit 330 is used to determine the optimal strategy subset from a plurality of pending strategy subsets through the maximum optimization of the acquisition function, wherein the optimal strategy subset is used to augment the classification model trained on the sample training set.
  • the classification accuracy is the highest.
  • Bayesian optimization algorithm learns and fits the acquisition function to find the objective function
  • the strategy parameter that f(x) promotes to the global optimal value.
  • every Bayesian optimization iteration uses a new sample point to test the objective function f(x)
  • use this information to update the prior distribution of the objective function f(x)
  • use the Bayesian optimization algorithm to test the posterior The distribution gives the sample points where the global maximum value is most likely to occur.
  • the covariance is only related to x, and has nothing to do with y.
  • the posterior probability distribution of f n+1 can be estimated through the first n sample points : P(f n+1
  • the probability of improvement (POI) is used as the acquisition function.
  • the get function is:
  • f(x) is the objective function value of x
  • x is the verification accuracy
  • f(X+) is the optimal objective function value of x so far
  • ⁇ (x) and ⁇ (x) are obtained by Gaussian process respectively
  • the mean and variance of the objective function are the posterior distribution of f(x)
  • ⁇ ( ⁇ ) represents the normal cumulative distribution function.
  • is the trade-off coefficient. If there is no such coefficient, the POI function will tend to take a point around X+ and converge to a position close to f(X+), that is, it tends to develop rather than explore, so this item is added to make a trade-off. By constantly trying new x, the next largest point should be larger or at least equal to it.
  • the next sample is between the intersection f(X+) and the confidence region.
  • f(X+) we can assume that the samples below f(X+) can be discarded, because we only need to search for the parameter that makes the objective function take the maximum value. So through the iterative process, the observation area is reduced until the optimal solution is searched, so that the POI(X) is maximized.
  • the controller 30 selects the optimal augmentation strategy
  • the controller 30 is also used to output the optimal augmentation strategy to the augmenter 10, and the augmenter 10 confirms the optimal augmentation strategy as a preset sample The augmentation strategy of the training set. Understandably, after the augmenter 10 obtains the optimal augmentation strategy, every time the augmenter performs sample augmentation, it will use the optimal augmentation strategy output by the controller for sample augmentation.
  • the embodiment of the present application provides a computer-readable storage medium.
  • the computer-readable storage medium may be non-volatile or volatile.
  • the storage medium includes a stored program, and when the program is running, the device where the storage medium is located is controlled to perform the following steps:
  • each undetermined strategy subset is from the augmented strategy set At least one augmentation strategy composition; use each augmented sample training set to train the initialized classification model to obtain multiple trained classification models; input the preset sample validation set into each trained classification model to obtain training The classification accuracy corresponding to a good classification model; the Bayesian optimization algorithm is used to determine the optimal strategy subset from multiple pending strategy subsets based on the classification accuracy corresponding to each trained classification model.
  • the device where the storage medium is located is controlled to execute the step of using a Bayesian optimization algorithm to determine the optimal strategy subset from a plurality of pending strategy subsets based on the classification accuracy corresponding to each trained classification model, include:
  • the Bayesian optimization algorithm is determined according to the regression model Acquisition function: Through the maximum optimization of the acquisition function, the optimal strategy subset is determined from multiple pending strategy subsets. Among them, the classification model trained on the sample training set after augmenting the optimal strategy subset has the highest classification accuracy.
  • the device where the storage medium is located is controlled to execute the input of a preset sample verification set into each trained classification model to obtain the classification accuracy corresponding to the trained classification model, including:
  • Input the preset sample validation set into each trained classification model obtain the training accuracy and verification accuracy of the output of the classification model; judge whether the classification model fits well according to the training accuracy and verification accuracy; determine the well-fitted classification model as The trained classification model, and the verification accuracy of the trained classification model is used as the classification accuracy of the classification model.
  • the device where the storage medium is located is controlled to execute a classification model trained and initialized with each augmented sample training set to obtain multiple trained classification models, including: extracting input using a convolutional neural network The feature map of each sample in the augmented sample training set of the classification model; according to the feature map, classify and predict a corresponding sample in the augmented sample training set to obtain the classification result; obtain the classification result set and the sample training set The loss function of the mean square error of the label set of all samples; the convolutional neural network is optimized by backpropagation, so that the value of the loss function converges, and the optimized training classification model is obtained.
  • controlling the device where the storage medium is located before executing the input of the preset sample verification set into each trained classification model to obtain the classification accuracy corresponding to the trained classification model also includes: The sample validation set of the sample validation set randomly selects multiple validation subsets; input multiple validation subsets into each trained classification model.
  • Fig. 4 is a schematic diagram of a computer device provided by an embodiment of the present application.
  • the computer device 100 of this embodiment includes a processor 101, a memory 102, and a computer program 103 stored in the memory 102 and running on the processor 101.
  • the processor 101 executes the computer program 103 when the computer program 103 is executed.
  • the method of selecting the augmentation strategy of the image data in the example is not repeated here to avoid repetition.
  • the computer program is executed by the processor 101, the function of each model/unit in the image data augmentation strategy selection system in the embodiment is realized. In order to avoid repetition, it will not be repeated here.
  • the computer device 100 may be a computing device such as a desktop computer, a notebook, a palmtop computer, and a cloud server.
  • the computer device may include, but is not limited to, a processor 101 and a memory 102.
  • FIG. 3 is only an example of the computer device 100 and does not constitute a limitation on the computer device 100. It may include more or less components than shown, or a combination of certain components, or different components.
  • computer equipment may also include input and output devices, network access devices, buses, and so on.
  • the so-called processor 101 may be a central processing unit (Central Processing Unit, CPU), other general-purpose processors, digital signal processors (Digital Signal Processor, DSP), application specific integrated circuits (ASIC), Field-Programmable Gate Array (FPGA) or other programmable logic devices, discrete gates or transistor logic devices, discrete hardware components, etc.
  • the general-purpose processor may be a microprocessor or the processor may also be any conventional processor or the like.
  • the memory 102 may be an internal storage unit of the computer device 100, such as a hard disk or memory of the computer device 100.
  • the memory 102 may also be an external storage device of the computer device 100, such as a plug-in hard disk equipped on the computer device 100, a smart memory card (Smart Media Card, SMC), a Secure Digital (SD) card, and a flash memory card (Flash). Card) and so on.
  • the memory 102 may also include both an internal storage unit of the computer device 100 and an external storage device.
  • the memory 102 is used to store computer programs and other programs and data required by the computer equipment.
  • the memory 102 can also be used to temporarily store data that has been output or will be output.
  • the disclosed system, device, and method can be implemented in other ways.
  • the device embodiments described above are only illustrative.
  • the division of the units is only a logical function division, and there may be other divisions in actual implementation, for example, multiple units or components may be combined. Or it can be integrated into another system, or some features can be ignored or not implemented.
  • the displayed or discussed mutual coupling or direct coupling or communication connection may be indirect coupling or communication connection through some interfaces, devices or units, and may be in electrical, mechanical or other forms.
  • the units described as separate components may or may not be physically separated, and the components displayed as units may or may not be physical units, that is, they may be located in one place, or they may be distributed on multiple network units. Some or all of the units may be selected according to actual needs to achieve the objectives of the solutions of the embodiments.
  • the functional units in the various embodiments of the present application may be integrated into one processing unit, or each unit may exist alone physically, or two or more units may be integrated into one unit.
  • the above-mentioned integrated unit may be implemented in the form of hardware, or may be implemented in the form of hardware plus software functional units.
  • the above-mentioned integrated unit implemented in the form of a software functional unit may be stored in a computer readable storage medium.
  • the above-mentioned software functional unit is stored in a storage medium and includes several instructions to make a computer device (which may be a personal computer, a server, or a network device, etc.) or a processor (Processor) execute the method described in each embodiment of the present application. Part of the steps.
  • the aforementioned storage media include: U disk, mobile hard disk, read-only memory (Read-Only Memory, ROM), random access memory (Random Access Memory, RAM), magnetic disks or optical disks and other media that can store program codes. .

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Artificial Intelligence (AREA)
  • General Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • Evolutionary Computation (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Computational Linguistics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Health & Medical Sciences (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • Evolutionary Biology (AREA)
  • General Health & Medical Sciences (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Image Analysis (AREA)

Abstract

The present application relates to the field of artificial intelligence, and provided in the embodiments thereof are a method and system for selecting an augmentation strategy for image data. The present application relates to the technical field of artificial intelligence. The method comprises: from within an augmentation strategy set, selecting a plurality of strategy subsets to be determined to perform sample augmentation on a preset sample training set so as to obtain a plurality of augmented sample training sets; training an initialized classification model by using each augmented sample training set so as to obtain a plurality of trained classification models; inputting a preset sample validation set into each trained classification model to obtain classification accuracy degrees corresponding to the trained classification models; and determining an optimal strategy subset from among the plurality of strategy subsets by using a Bayesian optimization algorithm on the basis of the classification accuracy degrees corresponding to each trained classification model. The technical solution provided in the embodiments of the present application is able to solve the problem of difficulty in determining which augmentation strategy is most effective for the current type of image sample.

Description

一种图像数据的增广策略选取方法及系统Method and system for selecting augmentation strategy of image data
本申请要求于2020年02月20日提交中国专利局、申请号为202010095784.6,发明名称为“一种图像数据的增广策略选取方法及系统”的中国专利申请的优先权,其全部内容通过引用结合在本申请中。This application claims the priority of a Chinese patent application filed with the Chinese Patent Office on February 20, 2020, the application number is 202010095784.6, and the invention title is "A method and system for selecting an augmentation strategy for image data", the entire content of which is incorporated by reference Incorporated in this application.
技术领域Technical field
本申请涉及人工智能技术领域,尤其涉及一种图像数据的增广策略选取方法及系统。This application relates to the field of artificial intelligence technology, and in particular to a method and system for selecting an augmentation strategy for image data.
背景技术Background technique
深度学习在计算机视觉领域的成功一定程度归功于拥有大量带标记的训练数据,因为模型的性能通常会随着训练数据质量、多样性和数量的增加而相应提高。然而,要收集足够多的高质量数据来训练模型以使其具有良好的性能,往往非常困难和成本昂贵。The success of deep learning in the field of computer vision is due to a large amount of labeled training data, because the performance of the model usually increases with the increase in the quality, diversity and quantity of training data. However, it is often very difficult and expensive to collect enough high-quality data to train the model to have good performance.
目前常用一些数据增广策略去增加数据量,用以训练计算机视觉模型,如平移、旋转和翻转等通过随机“扩充”来增加训练样本的数量和多样性。At present, some data augmentation strategies are commonly used to increase the amount of data for training computer vision models, such as translation, rotation, and flipping through random "expansion" to increase the number and diversity of training samples.
发明人意识到,目前现有的增广策略各式各样,在面对不同的数据集的时候表现不一,难以确定哪种增广策略对当前类型的图像数据集最有效。The inventor realizes that there are various augmentation strategies currently available, and they perform differently when facing different data sets, and it is difficult to determine which augmentation strategy is most effective for the current type of image data set.
发明内容Summary of the invention
有鉴于此,本申请实施例提供了一种图像数据的增广策略选取方法及装置,用以解决现有技术中难以确定哪种增广策略对当前类型的图像数据集最有效的问题。In view of this, the embodiments of the present application provide a method and device for selecting an augmentation strategy for image data to solve the problem in the prior art that it is difficult to determine which augmentation strategy is most effective for the current type of image data set.
为了实现上述目的,根据本申请的一个方面,提供了一种图像数据的增广策略选取方法,所述方法包括:从增广策略集合中选取多个待定策略子集对预设的样本训练集进行样本增广,得到多个增广后的样本训练集,其中,每个所述待定策略子集由所述增广策略集合中至少一个增广策略组成;利用每个所述增广后的样本训练集训练初始化的分类模型,得到多个训练后的分类模型;将预设的样本验证集输入每个所述训练后的分类模型,得到训练好的所述分类模型对应的分类准确度;利用贝叶斯优化算法基于每个所述训练好的分类模型对应的分类准确度,从多个待定策略子集中确定最优策略子集。In order to achieve the above objective, according to one aspect of the present application, an augmentation strategy selection method for image data is provided. The method includes: selecting a plurality of pending strategy subsets from an augmentation strategy set to a preset sample training set Perform sample augmentation to obtain multiple augmented sample training sets, wherein each of the pending strategy subsets is composed of at least one augmentation strategy in the augmentation strategy set; each of the augmented strategy subsets is used Training the initialized classification model on the sample training set to obtain multiple trained classification models; inputting a preset sample verification set into each of the trained classification models to obtain the classification accuracy corresponding to the trained classification model; A Bayesian optimization algorithm is used to determine the optimal strategy subset from a plurality of pending strategy subsets based on the classification accuracy corresponding to each of the trained classification models.
为了实现上述目的,根据本申请的一个方面,提供了一种图像数据的增广策略选取系统,所述系统包括增广器、分类模型及控制器;In order to achieve the foregoing objective, according to one aspect of the present application, an augmentation strategy selection system for image data is provided, the system including an augmenter, a classification model, and a controller;
所述增广器,用于从增广策略集合中选取多个待定策略子集对预设的样本训练集进行样本增广,得到多个增广后的样本训练集,其中,每个所述待定策略子集由所述增广策略集合中至少一个增广策略组成;The augmenter is used to select multiple undetermined strategy subsets from the augmentation strategy set to perform sample augmentation on the preset sample training set to obtain multiple augmented sample training sets, wherein each of the The undetermined strategy subset is composed of at least one augmentation strategy in the augmentation strategy set;
所述分类模型,用于利用每个所述增广后的样本训练集训练初始化的分类模型,得到多个训练后的分类模型;并将预设的样本验证集输入每个所述训练后的分类模型,得到训练好的所述分类模型对应的分类准确度;The classification model is used to train an initialized classification model using each of the augmented sample training sets to obtain a plurality of trained classification models; and input a preset sample verification set into each of the trained samples A classification model to obtain the classification accuracy corresponding to the trained classification model;
所述控制器,用于利用贝叶斯优化算法基于每个所述训练好的分类模型对应的分类准确度,从多个待定策略子集中确定最优策略子集。The controller is configured to use a Bayesian optimization algorithm to determine an optimal strategy subset from a plurality of pending strategy subsets based on the classification accuracy corresponding to each of the trained classification models.
为了实现上述目的,根据本申请的一个方面,提供了一种计算机可读存储介质,所 述存储介质包括存储的程序,在所述程序运行时控制所述存储介质所在设备执行以下步骤:In order to achieve the above objective, according to one aspect of the present application, a computer-readable storage medium is provided, the storage medium includes a stored program, and when the program runs, the device where the storage medium is located is controlled to perform the following steps:
从增广策略集合中选取多个待定策略子集对预设的样本训练集进行样本增广,得到多个增广后的样本训练集,其中,每个所述待定策略子集由所述增广策略集合中至少一个增广策略组成;Select multiple undetermined strategy subsets from the augmented strategy set to perform sample augmentation on the preset sample training set to obtain multiple augmented sample training sets, wherein each of the undetermined strategy subsets is augmented by the At least one augmentation strategy in the broad strategy set;
利用每个所述增广后的样本训练集训练初始化的分类模型,得到多个训练后的分类模型;Training an initialized classification model using each of the augmented sample training sets to obtain multiple trained classification models;
将预设的样本验证集输入每个所述训练后的分类模型,得到训练好的所述分类模型对应的分类准确度;Input a preset sample verification set into each of the trained classification models to obtain the classification accuracy corresponding to the trained classification model;
利用贝叶斯优化算法基于每个所述训练好的分类模型对应的分类准确度,从多个待定策略子集中确定最优策略子集。A Bayesian optimization algorithm is used to determine the optimal strategy subset from a plurality of pending strategy subsets based on the classification accuracy corresponding to each of the trained classification models.
为了实现上述目的,根据本申请的一个方面,提供了一种计算机设备,包括存储器、处理器以及存储在所述存储器中并可在所述处理器上运行的计算机程序,所述处理器执行所述计算机程序时实现以下步骤:In order to achieve the above objective, according to one aspect of the present application, a computer device is provided, including a memory, a processor, and a computer program stored in the memory and running on the processor, and the processor executes all When the computer program is described, the following steps are implemented:
从增广策略集合中选取多个待定策略子集对预设的样本训练集进行样本增广,得到多个增广后的样本训练集,其中,每个所述待定策略子集由所述增广策略集合中至少一个增广策略组成;Select multiple undetermined strategy subsets from the augmented strategy set to perform sample augmentation on the preset sample training set to obtain multiple augmented sample training sets, wherein each of the undetermined strategy subsets is augmented by the At least one augmentation strategy in the broad strategy set;
利用每个所述增广后的样本训练集训练初始化的分类模型,得到多个训练后的分类模型;Training an initialized classification model using each of the augmented sample training sets to obtain multiple trained classification models;
将预设的样本验证集输入每个所述训练后的分类模型,得到训练好的所述分类模型对应的分类准确度;Input a preset sample verification set into each of the trained classification models to obtain the classification accuracy corresponding to the trained classification model;
利用贝叶斯优化算法基于每个所述训练好的分类模型对应的分类准确度,从多个待定策略子集中确定最优策略子集。A Bayesian optimization algorithm is used to determine the optimal strategy subset from a plurality of pending strategy subsets based on the classification accuracy corresponding to each of the trained classification models.
在本方案中,通过利用不同的增广策略对同类样本分别进行样本增广,从而利用每个增广后的样本训练集训练初始化的分类模型,得到多个训练后的分类模型,利用样本验证集验证训练后的分类模型,然后根据分类模型的分类准确度及贝叶斯优化算法来获取符合该类样本的合适的增广策略,能够提高增广策略选取效率。In this scheme, by using different augmentation strategies to perform sample augmentation on similar samples, each augmented sample training set is used to train the initialized classification model to obtain multiple trained classification models, and use samples to verify Validate the trained classification model, and then according to the classification accuracy of the classification model and the Bayesian optimization algorithm to obtain a suitable augmentation strategy that meets this type of sample, which can improve the efficiency of augmentation strategy selection.
附图说明Description of the drawings
图1是本申请实施例提供的一种可选的图像数据的增广策略选取方法的流程图;FIG. 1 is a flowchart of an optional method for selecting an augmentation strategy for image data according to an embodiment of the present application;
图2是本申请实施例提供的一种可选的图像数据的增广策略选取系统的示意图;2 is a schematic diagram of an optional image data augmentation strategy selection system provided by an embodiment of the present application;
图3是本申请实施例提供的一种可选的控制器的功能框图;FIG. 3 is a functional block diagram of an optional controller provided by an embodiment of the present application;
图4是本申请实施例提供的一种可选的计算机设备的示意图。Fig. 4 is a schematic diagram of an optional computer device provided by an embodiment of the present application.
具体实施方式Detailed ways
为了更好的理解本申请的技术方案,下面结合附图对本申请实施例进行详细描述。In order to better understand the technical solutions of the present application, the following describes the embodiments of the present application in detail with reference to the accompanying drawings.
应当明确,所描述的实施例仅仅是本申请一部分实施例,而不是全部的实施例。基于本申请中的实施例,本领域普通技术人员在没有作出创造性劳动前提下所获得的所有其它 实施例,都属于本申请保护的范围。It should be clear that the described embodiments are only a part of the embodiments of the present application, rather than all of the embodiments. Based on the embodiments in this application, all other embodiments obtained by those of ordinary skill in the art without creative work shall fall within the protection scope of this application.
在本申请实施例中使用的术语是仅仅出于描述特定实施例的目的,而非旨在限制本申请。在本申请实施例和所附权利要求书中所使用的单数形式的“一种”、“所述”和“该”也旨在包括多数形式,除非上下文清楚地表示其他含义。The terms used in the embodiments of the present application are only for the purpose of describing specific embodiments, and are not intended to limit the present application. The singular forms of "a", "the" and "the" used in the embodiments of the present application and the appended claims are also intended to include plural forms, unless the context clearly indicates other meanings.
应当理解,本文中使用的术语“和/或”仅仅是一种描述关联对象的关联关系,表示可以存在三种关系,例如,A和/或B,可以表示:单独存在A,同时存在A和B,单独存在B这三种情况。另外,本文中字符“/”,一般表示前后关联对象是一种“或”的关系。It should be understood that the term "and/or" used in this text is only an association relationship describing the associated objects, indicating that there can be three types of relationships, for example, A and/or B can mean that A alone exists, and both A and A exist at the same time. B, there are three cases of B alone. In addition, the character "/" in this text generally indicates that the associated objects before and after are in an "or" relationship.
应当理解,尽管在本申请实施例中可能采用术语第一、第二、第三等来描述终端,但这些终端不应限于这些术语。这些术语仅用来将终端彼此区分开。例如,在不脱离本申请实施例范围的情况下,第一终端也可以被称为第二终端,类似地,第二终端也可以被称为第一终端。It should be understood that although the terms first, second, third, etc. may be used to describe terminals in the embodiments of the present application, these terminals should not be limited to these terms. These terms are only used to distinguish terminals from each other. For example, without departing from the scope of the embodiments of the present application, the first terminal may also be referred to as the second terminal, and similarly, the second terminal may also be referred to as the first terminal.
取决于语境,如在此所使用的词语“如果”可以被解释成为“在……时”或“当……时”或“响应于确定”或“响应于检测”。类似地,取决于语境,短语“如果确定”或“如果检测(陈述的条件或事件)”可以被解释成为“当确定时”或“响应于确定”或“当检测(陈述的条件或事件)时”或“响应于检测(陈述的条件或事件)”。Depending on the context, the word "if" as used herein can be interpreted as "when" or "when" or "in response to determination" or "in response to detection". Similarly, depending on the context, the phrase "if determined" or "if detected (statement or event)" can be interpreted as "when determined" or "in response to determination" or "when detected (statement or event) )" or "in response to detection (statement or event)".
图1是根据本申请实施例的一种图像数据的增广策略选取方法的流程图,如图1所示,该方法包括:Fig. 1 is a flowchart of a method for selecting an augmentation strategy for image data according to an embodiment of the present application. As shown in Fig. 1, the method includes:
步骤S01,从增广策略集合中选取多个待定策略子集对预设的样本训练集进行样本增广,得到多个增广后的样本训练集,其中,每个待定策略子集由增广策略集合中至少一个增广策略组成;Step S01: Select multiple pending strategy subsets from the augmented strategy set to perform sample augmentation on the preset sample training set to obtain multiple augmented sample training sets, where each pending strategy subset is augmented by At least one augmentation strategy in the strategy set;
步骤S02,利用每个增广后的样本训练集训练初始化的分类模型,得到多个训练后的分类模型;Step S02: Use each augmented sample training set to train an initialized classification model to obtain multiple trained classification models;
步骤S03,将预设的样本验证集输入每个训练后的分类模型,得到训练好的分类模型对应的分类准确度;Step S03, input the preset sample verification set into each trained classification model, and obtain the classification accuracy corresponding to the trained classification model;
步骤S04,利用贝叶斯优化算法基于每个训练好的分类模型对应的分类准确度,从多个待定策略子集中确定最优策略子集。In step S04, the Bayesian optimization algorithm is used to determine the optimal strategy subset from the multiple pending strategy subsets based on the classification accuracy corresponding to each trained classification model.
其中,样本训练集中的样本为图形数据样本。Among them, the samples in the sample training set are graphic data samples.
在本方案中,通过利用不同的增广策略对同类样本分别进行样本增广,从而利用每个增广后的样本训练集训练初始化的分类模型,得到多个训练后的分类模型,利用样本验证集验证训练后的分类模型,然后根据分类模型的分类准确度及贝叶斯优化算法来获取符合该类样本的合适的增广策略,能够提高增广策略选取效率。In this scheme, by using different augmentation strategies to perform sample augmentation on similar samples, each augmented sample training set is used to train the initialized classification model to obtain multiple trained classification models, and use samples to verify Validate the trained classification model, and then according to the classification accuracy of the classification model and the Bayesian optimization algorithm to obtain a suitable augmentation strategy that meets this type of sample, which can improve the efficiency of augmentation strategy selection.
下面对本实施例提供的图像数据的增广策略选取方法的具体技术方案进行详细的说明。The specific technical solution of the method for selecting the augmentation strategy of image data provided by this embodiment will be described in detail below.
步骤S01,从增广策略集合中选取多个待定策略子集对预设的样本训练集进行样本增广,得到多个增广后的样本训练集,其中,每个待定策略子集由增广策略集合中至少一个增广策略组成;Step S01: Select multiple pending strategy subsets from the augmented strategy set to perform sample augmentation on the preset sample training set to obtain multiple augmented sample training sets, where each pending strategy subset is augmented by At least one augmentation strategy in the strategy set;
在本实施方式中,样本训练集中的样本是同类型的医学影像样本,例如肺部影像、胃部影像等。每个训练样本都设有标签,例如带有正标签的训练样本,即为标记为有肺炎症状的肺部影像,带有负标签的训练样本,标记为没有肺炎症状的肺部影像。示例性的,训练样本为512*512的医学影像样本。In this embodiment, the samples in the sample training set are medical image samples of the same type, such as lung images and stomach images. Each training sample has a label. For example, a training sample with a positive label is a lung image marked as having pneumonia symptoms, and a training sample with a negative label is a lung image without pneumonia symptoms. Exemplarily, the training sample is a 512*512 medical image sample.
其中,增广策略包括旋转变换、翻转变换、缩放变换、平移变换、尺度变换、区域裁剪、添加噪声、分段仿射、随机掩盖、边界检测、对比度变换、颜色抖动、随机混合及复合叠加。增广策略例如为翻转变换。Among them, augmentation strategies include rotation transformation, flip transformation, zoom transformation, translation transformation, scale transformation, region cropping, adding noise, segmented affine, random masking, boundary detection, contrast transformation, color dithering, random mixing, and composite overlay. The augmentation strategy is, for example, flipped transformation.
1)旋转变换(Rotation):随机旋转图像预设的角度,改变图像内容的朝向;1) Rotation: randomly rotate the image at a preset angle, changing the orientation of the image content;
2)翻转变换(Flip):沿着水平或者垂直方向翻转图像;2) Flip transform (Flip): Flip the image along the horizontal or vertical direction;
3)缩放变换(Zoom):按照预设的比例放大或者缩小图像;3) Zoom: zoom in or zoom out the image according to the preset ratio;
4)平移变换(Shift):在图像平面上对图像以预设方式进行平移;4) Translation transformation (Shift): Translate the image in a preset manner on the image plane;
5)尺度变换(Scale):对图像按照预设的尺度因子,进行放大或缩小,或者利用预设的尺度因子对图像滤波构造尺度空间,改变图像内容的大小或模糊程度;5) Scale: the image is enlarged or reduced according to the preset scale factor, or the scale space is constructed by filtering the image with the preset scale factor to change the size or blur degree of the image content;
6)区域裁剪(Crop):裁剪图片的感兴趣区域;6) Crop: Crop the region of interest of the picture;
7)添加噪声(Noise):在原来的图片上随机叠加一些噪声;7) Add noise (Noise): randomly superimpose some noise on the original picture;
8)分段仿射(Piecewise Affine):在图像上放置一个规则的点网格,根据正态分布的样本数量移动这些点及周围的图像区域;8) Piecewise Affine: Place a regular point grid on the image, and move these points and the surrounding image area according to the number of samples in the normal distribution;
9)随机掩盖(Dropout):在面积大小可选定、位置随机的矩形区域上丢失信息实现转换,所有通道的信息丢失产生黑色矩形块,部分通道的信息丢失产生彩色噪声;9) Random concealment (Dropout): The loss of information is realized in a rectangular area with a selectable area and a random location to achieve conversion. The loss of information in all channels produces black rectangular blocks, and the loss of information in some channels produces color noise;
10)边界检测(Edge Detect):检测图像中的所有边缘,将它们标记为黑白图像,再将结果与原始图像叠加;10) Edge Detect: Detect all edges in the image, mark them as black and white images, and then superimpose the result with the original image;
11)对比度变换(Contrast):在图像的HSV颜色空间,改变饱和度S和V亮度分量,保持色调H不变,对每个像素的S和V分量进行指数运算(指数因子在0.25到4之间),增加光照变化;11) Contrast transformation (Contrast): In the HSV color space of the image, change the saturation S and V brightness components, keep the hue H unchanged, and perform exponential calculations on the S and V components of each pixel (the exponent factor is between 0.25 and 4). Time), increase the light changes;
12)颜色抖动(Color jitter):对图像的曝光度(exposure)、饱和度(saturation)和色调(hue)进行随机变化形成不同光照及颜色下的图片,尽可能使得模型能够使用不同光照条件小的情形;12) Color jitter: Randomly change the exposure, saturation and hue of the image to form pictures under different lighting and colors, as much as possible to make the model use different light conditions as small as possible Situation
13)随机混合(Mix up):基于邻域风险最小化原则的数据增广方法,使用线性插值得到新样本数据;13) Random mixing (Mix up): a data augmentation method based on the principle of neighborhood risk minimization, using linear interpolation to obtain new sample data;
14)复合叠加(Sample Pairing):随机抽取两张图片分别经过基础数据增广操作处理后经像素取平均值的形式叠加合成一个新的样本,新样板的标签为原样本标签中的一种。14) Sample Pairing: Two images are randomly selected after being processed by the basic data augmentation operation and then superimposed into a new sample in the form of pixel averaging. The label of the new template is one of the original sample labels.
在本实施方式中,从上述14种增广策略中随机抽取上述任意3种增广策略组成一个待定策略子集,即一个待定策略子集包括3种增广策略,每种增广策略包括3个策略参数,分别为策略类型(μ)、概率值(α)、幅度(β)。那么一个待定策略子集可以用数值矩阵形式来表示:In this embodiment, any three augmentation strategies are randomly selected from the above 14 augmentation strategies to form a undetermined strategy subset, that is, a undetermined strategy subset includes 3 augmentation strategies, and each augmentation strategy includes 3 A strategy parameter, namely strategy type (μ), probability value (α), amplitude (β). Then a subset of pending strategies can be represented in the form of a numerical matrix:
Figure PCTCN2020111666-appb-000001
Figure PCTCN2020111666-appb-000001
其中,每一行表示一个增广策略。利用数值矩阵来表示待定策略子集,提高计算效率。Among them, each row represents an augmentation strategy. Use a numerical matrix to represent a subset of undetermined strategies to improve computational efficiency.
步骤S02,利用每个增广后的样本训练集训练初始化的分类模型,得到多个训练后的分类模型。Step S02: Use each augmented sample training set to train an initialized classification model to obtain multiple trained classification models.
在本实施方式中,分类模型为卷积神经网络模型,由卷积神经网络和全连接网络组成,其具体构成至少包括卷积网络层、池化层和全连接网络层。训练时具体步骤包括:In this embodiment, the classification model is a convolutional neural network model, which is composed of a convolutional neural network and a fully connected network, and its specific structure includes at least a convolutional network layer, a pooling layer, and a fully connected network layer. The specific steps during training include:
利用卷积神经网络提取输入分类模型的增广后的样本训练集中的每个样本的特征图;根据特征图,对增广后的样本训练集中的对应一个样本进行分类预测,得到分类结果;获取分类结果集合与样本训练集中的所有样本的标签集合的均方误差的损失函数;通过反向传播对卷积神经网络进行优化,以使得损失函数的值收敛,得到优化训练后的分类模型。Use the convolutional neural network to extract the feature map of each sample in the augmented sample training set of the input classification model; according to the feature map, classify and predict a corresponding sample in the augmented sample training set to obtain the classification result; obtain The classification result set and the loss function of the mean square error of the label set of all samples in the sample training set; the convolutional neural network is optimized through backpropagation, so that the value of the loss function converges, and the optimized and trained classification model is obtained.
在本实施方式中,分类结果有两种,分别是肺炎和非肺炎。初始的卷积神经网络对带有标签的样本进行特征提取,并进行预设轮次的训练,使得卷积神经网络层能够有效提取更泛化的特征(例如边缘、纹理等)。进行反向传播时,不断地梯度下降之后,模型的精度才能得到提高,使得损失函数的值收敛至最小,其中会自动调整卷积层与全连接层的权重和偏置,从而使得分类模型最优化。In this embodiment, there are two classification results, namely pneumonia and non-pneumonia. The initial convolutional neural network performs feature extraction on labeled samples and performs preset rounds of training, so that the convolutional neural network layer can effectively extract more generalized features (such as edges, textures, etc.). When performing backpropagation, the accuracy of the model can be improved after the gradient is continuously reduced, so that the value of the loss function converges to the minimum, and the weight and bias of the convolutional layer and the fully connected layer are automatically adjusted to make the classification model the best optimization.
在其他实施方式中,分类模型还可以是长短时神经网络模型、随机森林模型、支持向量机模型、最大熵模型等等,在此不做限定。In other embodiments, the classification model may also be a long- and short-term neural network model, a random forest model, a support vector machine model, a maximum entropy model, etc., which are not limited here.
步骤S03,将预设的样本验证集输入每个训练后的分类模型,得到训练好的分类模型对应的分类准确度。Step S03: Input the preset sample verification set into each trained classification model to obtain the classification accuracy corresponding to the trained classification model.
具体地,预设的样本验证集中的样本也设有标签,例如带有正标签的训练样本,即为标记为有肺炎症状的肺部影像,带有负标签的训练样本,标记为没有肺炎症状的肺部影像。采用预设的样本验证集对训练后的分类模型进行验证,每个分类模型对应的样本验证集不同,能够实现更好的模型泛化性能,有效解决样本增广可能引入的过度拟合问题。Specifically, the samples in the preset sample validation set are also labeled. For example, a training sample with a positive label is a lung image marked as having pneumonia symptoms, and a training sample with a negative label is marked as having no pneumonia symptoms. Of the lungs. A preset sample verification set is used to verify the trained classification model. The sample verification set corresponding to each classification model is different, which can achieve better model generalization performance and effectively solve the problem of overfitting that may be introduced by sample augmentation.
在步骤S03之前,方法还包括:Before step S03, the method further includes:
从预设的样本验证集随机抽取多个验证子集;Randomly select multiple verification subsets from the preset sample verification set;
将多个验证子集分别输入每个训练后的分类模型。Input multiple validation subsets into each trained classification model.
在本实施方式中,采用随机抽取方式,样本训练集与样本验证集中的样本量比例可以是2:8,4:6,6:4,8:2等。可以理解地,每次抽取时,随机抽取样本验证集中50%的样本组成验证子集。在其他实施方式中,随机抽取的比例可以是30%、40%、60%等等。In this embodiment, a random sampling method is adopted, and the sample size ratio in the sample training set and the sample verification set may be 2:8, 4:6, 6:4, 8:2, etc. Understandably, each time a sample is drawn, 50% of the samples in the sample verification set are randomly selected to form a verification subset. In other embodiments, the ratio of random selection may be 30%, 40%, 60%, and so on.
在另一种实施方式中,采用交叉验证方法对分类模型进行验证。交叉验证方法为十折交叉验证方法或五折交叉验证方法中的任意一种。例如采用五折交叉验证方法,具体地,将多个训练样本随机分成10份,每次取其中2份作为交叉验证集,其余8份作为训练集。训 练时,先用其中的8份对初始化后的分类模型进行训练,然后对2份交叉验证集进行分类标注,以此重复训练及验证过程5次,每次选取的交叉验证集不同,直至所有的训练样本都被分类标注一遍。In another embodiment, a cross-validation method is used to validate the classification model. The cross-validation method is either a ten-fold cross-validation method or a five-fold cross-validation method. For example, a five-fold cross-validation method is adopted. Specifically, multiple training samples are randomly divided into 10 parts, and 2 of them are taken as the cross-validation set each time, and the remaining 8 parts are used as the training set. When training, first use 8 of them to train the initialized classification model, and then classify and label the 2 cross-validation sets, so as to repeat the training and verification process 5 times, each time the selected cross-validation set is different, until all The training samples of are all classified and labeled again.
步骤S03,具体包括:Step S03 specifically includes:
步骤S031,将预设的样本验证集输入每个训练后的分类模型;Step S031, input the preset sample verification set into each trained classification model;
步骤S032,获取分类模型输出的训练精度及验证精度;Step S032: Obtain the training accuracy and verification accuracy of the classification model output;
步骤S033,根据训练精度和验证精度判断分类模型是否拟合良好;Step S033, judging whether the classification model fits well according to the training accuracy and the verification accuracy;
步骤S034,将拟合良好的分类模型确定为训练好的分类模型,并将训练好的分类模型的验证精度作为分类模型的分类准确度。Step S034: Determine the well-fitted classification model as the trained classification model, and use the verification accuracy of the trained classification model as the classification accuracy of the classification model.
其中,每个分类模型的训练过程中,分类模型的训练轮次可以预先设定,例如训练轮次为训练100次,在100次训练后,再将样本验证集输入分类模型,得到分类模型输出的训练精度和验证精度,并对分类模型进行拟合判断,以确定训练后的分类模型是否拟合良好,具体地,当(训练精度-验证精度)/验证精度≤10%,那么认为分类模型拟合良好。在本实施方式中,将拟合良好的分类模型的验证精度作为分类准确度。Among them, in the training process of each classification model, the training rounds of the classification model can be preset. For example, the training round is 100 trainings. After 100 trainings, the sample validation set is input into the classification model to obtain the output of the classification model. The training accuracy and verification accuracy of the classification model are judged to determine whether the training classification model fits well. Specifically, when (training accuracy-verification accuracy)/verification accuracy ≤ 10%, then the classification model is considered The fit is good. In this embodiment, the verification accuracy of a well-fitted classification model is used as the classification accuracy.
步骤S04,利用贝叶斯优化算法基于每个训练好的分类模型对应的分类准确度,从多个待定策略子集中确定最优策略子集。In step S04, the Bayesian optimization algorithm is used to determine the optimal strategy subset from the multiple pending strategy subsets based on the classification accuracy corresponding to each trained classification model.
采用贝叶斯优化算法才寻找最优策略子集时,将待定策略子集(数值矩阵)作为样本点的x值,将分类准确度作为样本点的y值,从而组成多个样本点,并基于多个样本点构建高斯过程的回归模型,通过对目标函数进行学习拟合,找到使目标函数向全局最优值提升的策略子集。When the Bayesian optimization algorithm is used to find the optimal strategy subset, the undetermined strategy subset (numerical matrix) is used as the x value of the sample point, and the classification accuracy is used as the y value of the sample point to form multiple sample points, and A regression model of the Gaussian process is constructed based on multiple sample points, and the objective function is learned and fitted to find a subset of strategies that promote the objective function to the global optimal value.
步骤S04具体包括:Step S04 specifically includes:
基于多个样本点构建高斯过程的回归模型,其中,每个样本点包括训练好的分类模型的分类准确度及训练分类模型所采用的待定策略子集;Construct a regression model of the Gaussian process based on multiple sample points, where each sample point includes the classification accuracy of the trained classification model and the undetermined strategy subset used to train the classification model;
根据回归模型确定贝叶斯优化算法的获取函数;Determine the acquisition function of the Bayesian optimization algorithm according to the regression model;
通过对获取函数的最大优化,从多个待定策略子集中确定最优策略子集,其中,利用最优策略子集增广后样本训练集训练得到的分类模型的分类准确度最高。Through the maximum optimization of the acquisition function, the optimal strategy subset is determined from multiple pending strategy subsets. Among them, the classification model trained on the sample training set after augmenting the optimal strategy subset has the highest classification accuracy.
在本实施方式中,基于分类准确度并利用贝叶斯优化算法从多个待定策略子集中确定最优策略子集。在其他实施方式中,也可以采用其他算法来选取,在此不做限定。In this embodiment, the optimal strategy subset is determined from a plurality of pending strategy subsets based on the classification accuracy and the Bayesian optimization algorithm. In other embodiments, other algorithms may also be used for selection, which is not limited here.
可以理解地,我们认为y与x=(μ,α,β)之间存在某种函数关系,即y=f(x)贝叶斯优化算法通过对获取函数进行学习拟合,找到使目标函数f(x)向全局最优值提升的策略参数。每一次贝叶斯优化迭代使用新的样本点来测试目标函数f(x)时,利用这个信息来更新目标函数f(x)的先验分布,最后,利用贝叶斯优化算法测试由后验分布给出的全局最值最可能出现的位置的样本点。Understandably, we believe that there is a certain functional relationship between y and x=(μ,α,β), that is, y=f(x) Bayesian optimization algorithm learns and fits the acquisition function to find the objective function The strategy parameter that f(x) promotes to the global optimal value. When every Bayesian optimization iteration uses a new sample point to test the objective function f(x), use this information to update the prior distribution of the objective function f(x), and finally, use the Bayesian optimization algorithm to test the posterior The distribution gives the sample points where the global maximum value is most likely to occur.
在本实施方式中,贝叶斯优化迭代的过程中,通过获取函数指导我们去选择样本点,不断修正GP高斯过程曲线去逼近目标函数f(x),当获取函数最大的时候说明选择的样本点最优,相当于我们就搜索到了令目标函数f(x)最大的最优策略子集。In this embodiment, during the Bayesian optimization iteration process, we guide us to select sample points through the acquisition function, and continuously modify the GP Gaussian process curve to approximate the objective function f(x). When the acquisition function is the largest, the selected sample is explained The point is optimal, which means that we have searched for the optimal strategy subset that maximizes the objective function f(x).
由于f(x)形式无法显性求取,我们用高斯过程来逼近,Since the form of f(x) cannot be obtained explicitly, we use Gaussian process to approximate,
即f(x)~GP(m(x),k(x,x′)),其中m(x)代表样本点f(x)的数学期望E(f(x)),在贝叶斯优化中通常取0,k(x,x')为核函数,描述的是x的协方差。That is, f(x)~GP(m(x), k(x,x′)), where m(x) represents the mathematical expectation E(f(x)) of the sample point f(x), in Bayesian optimization Usually 0, k(x, x') is the kernel function, which describes the covariance of x.
对于每个x都有一个对应的高斯分布,而对于一组{x 1,x 2...x n},假设y值服从联合正态分布,其均值为0,协方差为: For each x there is a corresponding Gaussian distribution, and for a set of {x 1 , x 2 ... x n }, assuming that the y value obeys the joint normal distribution, its mean value is 0, and the covariance is:
Figure PCTCN2020111666-appb-000002
其中,协方差只与x有关,和y无关。
Figure PCTCN2020111666-appb-000002
Among them, the covariance is only related to x, and has nothing to do with y.
对于一个新的样本点x n+1,联合高斯分布为: For a new sample point x n+1 , the joint Gaussian distribution is:
Figure PCTCN2020111666-appb-000003
Figure PCTCN2020111666-appb-000003
因此可以通过前n个样本点估计出f n+1的后验概率分布:P(f n+1|D 1:t,x t+1)~N(μ n(x),σ n 2(x)),其中,μ n(x)=k TK -1f 1:n;σ n 2(x)=k(x n+1,x n+1)-k TK -1k; Therefore, the posterior probability distribution of f n+1 can be estimated through the first n sample points : P(f n+1 |D 1:t ,x t+1 )~N(μ n (x),σ n 2 ( x)), where μ n (x)=k T K -1 f 1:n ; σ n 2 (x)=k(x n+1 ,x n+1 )-k T K -1 k;
在本实施方式中,采用改进概率(Probability of Improvement,POI)作为获取函数。In this embodiment, the probability of improvement (POI) is used as the acquisition function.
获取函数为:
Figure PCTCN2020111666-appb-000004
The get function is:
Figure PCTCN2020111666-appb-000004
其中,f(x)为x的目标函数值,x为验证精度,f(X+)为到目前为止最优的x的目标函数值,μ(x),σ(x)分别是高斯过程所得到的目标函数的均值和方差,即f(x)的后验分布,Φ(·)表示的是正态累计分布函数。ξ为trade-off系数,如果没有该系数,POI函数会倾向于取在X+周围的点,收敛到接近f(X+)附近的位置,即倾向于开发而不是探索,因此加入该项进行权衡。通过不断尝试新的x,下一个最大点应该要比它大或至少与之相等。因此,下一个采样在交叉点f(X+)和置信域之间,我们能假定在f(X+)点以下的样本是可以丢弃的,因为我们只需要搜索令目标函数取极大值的参数,于是通过迭代这一过程缩小了观察区域,直到搜索到最优解,使得POI(X)最大。Among them, f(x) is the objective function value of x, x is the verification accuracy, f(X+) is the optimal objective function value of x so far, μ(x) and σ(x) are obtained by Gaussian process respectively The mean and variance of the objective function are the posterior distribution of f(x), and Φ(·) represents the normal cumulative distribution function. ξ is the trade-off coefficient. If there is no such coefficient, the POI function will tend to take a point around X+ and converge to a position close to f(X+), that is, it tends to develop rather than explore, so this item is added to make a trade-off. By constantly trying new x, the next largest point should be larger or at least equal to it. Therefore, the next sample is between the intersection f(X+) and the confidence region. We can assume that the samples below f(X+) can be discarded, because we only need to search for the parameter that makes the objective function take the maximum value. So through the iterative process, the observation area is reduced until the optimal solution is searched, so that the POI(X) is maximized.
本申请实施例提供了一种图像数据的增广策略选取系统,如图2所示,系统包括增广器10、分类模型20及控制器30;The embodiment of the present application provides an augmentation strategy selection system for image data. As shown in FIG. 2, the system includes an augmenter 10, a classification model 20, and a controller 30;
增广器10,用于从增广策略集合中选取多个待定策略子集对预设的样本训练集进行样本增广,得到多个增广后的样本训练集,其中,每个待定策略子集由增广策略集合中至少 一个增广策略组成。具体地,增广策略集合包括旋转变换、翻转变换、缩放变换、平移变换、尺度变换、区域裁剪、噪声添加、分段仿射、随机掩盖、边界检测、对比度变换、颜色抖动、随机混合及复合叠加。其中,增广策略例如为翻转变换。The augmenter 10 is used to select multiple undetermined strategy subsets from the augmentation strategy set to perform sample augmentation on the preset sample training set to obtain multiple augmented sample training sets, where each undetermined strategy sub-set The set consists of at least one augmentation strategy in the augmentation strategy set. Specifically, the augmentation strategy set includes rotation transformation, flip transformation, zoom transformation, translation transformation, scale transformation, region cropping, noise addition, segmented affine, random masking, boundary detection, contrast transformation, color dithering, random mixing and compounding Overlay. Among them, the augmentation strategy is, for example, flip transformation.
在本实施方式中,随机抽取上述任意3种增广策略组成一个待定策略子集,每种增广策略包括3个策略参数,分别为策略类型(μ)、概率值(α)、幅度(β)。那么一个待定策略子集可以用数值矩阵形式来表示:In this embodiment, any of the three augmentation strategies mentioned above are randomly selected to form a subset of undetermined strategies. Each augmentation strategy includes three strategy parameters, which are strategy type (μ), probability value (α), and amplitude (β). ). Then a subset of pending strategies can be represented in the form of a numerical matrix:
Figure PCTCN2020111666-appb-000005
Figure PCTCN2020111666-appb-000005
其中,每一行表示一个增广策略。利用数值矩阵来表示待定策略子集,提高计算效率。Among them, each row represents an augmentation strategy. Use a numerical matrix to represent a subset of undetermined strategies to improve computational efficiency.
分类模型20包括训练单元210及验证单元220。训练单元210,用于利用每个增广后的样本训练集训练初始化的分类模型,得到多个训练后的分类模型;验证单元220,用于将预设的样本验证集输入每个训练后的分类模型,得到训练好的分类模型对应的分类准确度。The classification model 20 includes a training unit 210 and a verification unit 220. The training unit 210 is configured to use each augmented sample training set to train an initialized classification model to obtain multiple trained classification models; the verification unit 220 is configured to input a preset sample validation set into each trained The classification model obtains the classification accuracy corresponding to the trained classification model.
在本实施方式中,分类模型为卷积神经网络模型,由卷积神经网络和全连接网络组成,其具体构成至少包括卷积网络层、池化层和全连接网络层。In this embodiment, the classification model is a convolutional neural network model, which is composed of a convolutional neural network and a fully connected network, and its specific structure includes at least a convolutional network layer, a pooling layer, and a fully connected network layer.
训练单元210包括提取子单元、分类子单元、第一获取子单元及优化子单元。The training unit 210 includes an extraction subunit, a classification subunit, a first acquisition subunit, and an optimization subunit.
提取子单元,用于利用卷积神经网络提取输入分类模型的增广后的样本训练集中的每个样本的特征图;分类子单元,用于根据特征图,对增广后的样本训练集中的对应一个样本进行分类预测,得到分类结果;获取子单元,用于获取分类结果集合与样本训练集中的所有样本的标签集合的均方误差的损失函数;优化子单元,用于通过反向传播对卷积神经网络进行优化,以使得损失函数的值收敛,得到优化训练后的分类模型。The extraction subunit is used to use the convolutional neural network to extract the feature map of each sample in the augmented sample training set of the input classification model; the classification subunit is used to compare the augmented sample training set according to the feature map Perform classification prediction corresponding to a sample to obtain the classification result; obtain the subunit, which is used to obtain the loss function of the mean square error between the classification result set and the label set of all samples in the sample training set; the optimization subunit is used to pair The convolutional neural network is optimized to make the value of the loss function converge to obtain the optimized and trained classification model.
在本实施方式中,分类结果有两种,分别是肺炎和非肺炎。初始的卷积神经网络对带有标签的样本进行特征提取,并进行预设轮次的训练,使得卷积神经网络层能够有效提取更泛化的特征(例如边缘、纹理等)。进行反向传播时,不断地梯度下降之后,模型的精度才能得到提高,使得损失函数的值收敛至最小,其中会自动调整卷积层与全连接层的权重和偏置,从而使得分类模型最优化。In this embodiment, there are two classification results, namely pneumonia and non-pneumonia. The initial convolutional neural network performs feature extraction on labeled samples and performs preset rounds of training, so that the convolutional neural network layer can effectively extract more generalized features (such as edges, textures, etc.). When performing backpropagation, the accuracy of the model can be improved after the gradient is continuously reduced, so that the value of the loss function converges to the minimum, and the weight and bias of the convolutional layer and the fully connected layer are automatically adjusted to make the classification model the best optimization.
具体地,预设的样本验证集中的样本也设有标签,例如带有正标签的训练样本,即为标记为有肺炎症状的肺部影像,带有负标签的训练样本,标记为没有肺炎症状的肺部影像。采用预设的样本验证集对训练后的分类模型进行验证,每个分类模型对应的样本验证集不同,能够实现更好的模型泛化性能,有效解决样本增广可能引入的过度拟合问题。Specifically, the samples in the preset sample validation set are also labeled. For example, a training sample with a positive label is a lung image marked as having pneumonia symptoms, and a training sample with a negative label is marked as having no pneumonia symptoms. Of the lungs. A preset sample validation set is used to verify the trained classification model. The sample validation set corresponding to each classification model is different, which can achieve better model generalization performance and effectively solve the problem of overfitting that may be introduced by sample augmentation.
验证单元220包括输入子单元、第二获取子单元、判断子单元及确定子单元。The verification unit 220 includes an input subunit, a second acquisition subunit, a judgment subunit, and a determination subunit.
输入子单元,用于将预设的样本验证集输入每个训练后的分类模型;The input subunit is used to input the preset sample verification set into each trained classification model;
第二获取子单元,用于获取分类模型输出的训练精度及验证精度;The second acquisition subunit is used to acquire the training accuracy and verification accuracy of the classification model output;
判断子单元,用于根据训练精度和验证精度判断分类模型是否拟合良好;The judgment subunit is used to judge whether the classification model fits well according to the training accuracy and verification accuracy;
确定子单元,用于将拟合良好的分类模型确定为训练好的分类模型,并将训练好的分 类模型的验证精度作为分类模型的分类准确度。The determination subunit is used to determine a well-fitted classification model as a trained classification model, and use the verification accuracy of the trained classification model as the classification accuracy of the classification model.
其中,每个分类模型的训练过程中,分类模型的训练轮次可以预先设定,例如训练轮次为训练100次,在100次训练后,再将样本验证集输入分类模型,得到分类模型输出的训练精度和验证精度,并对分类模型进行拟合判断,以确定训练后的分类模型是否拟合良好,具体地,当(训练精度-验证精度)/验证精度≤10%,那么认为分类模型拟合良好。在本实施方式中,将拟合良好的分类模型的验证精度作为分类准确度。Among them, in the training process of each classification model, the training rounds of the classification model can be preset. For example, the training round is 100 trainings. After 100 trainings, the sample validation set is input into the classification model to obtain the output of the classification model. The training accuracy and verification accuracy of the classification model are judged to determine whether the training classification model fits well. Specifically, when (training accuracy-verification accuracy)/verification accuracy ≤ 10%, then the classification model is considered The fit is good. In this embodiment, the verification accuracy of a well-fitted classification model is used as the classification accuracy.
系统还包括数据库40及处理模块50,数据库40用于存储样本训练集及样本验证集。The system also includes a database 40 and a processing module 50. The database 40 is used to store the sample training set and the sample verification set.
处理模块50用于从预设的样本验证集随机抽取多个验证子集;将多个验证子集分别输入每个训练后的分类模型。The processing module 50 is configured to randomly select a plurality of verification subsets from a preset sample verification set; input the plurality of verification subsets into each trained classification model respectively.
在本实施方式中,采用随机抽取方式,样本训练集与样本验证集中的样本量比例可以是2:8,4:6,6:4,8:2等。可以理解地,每次抽取时,随机抽取样本验证集中50%的样本组成验证子集。在其他实施方式中,随机抽取的比例可以是30%、40%、60%等等。In this embodiment, a random sampling method is adopted, and the sample size ratio in the sample training set and the sample verification set may be 2:8, 4:6, 6:4, 8:2, etc. Understandably, each time a sample is drawn, 50% of the samples in the sample verification set are randomly selected to form a verification subset. In other embodiments, the ratio of random selection may be 30%, 40%, 60%, and so on.
在另一种实施方式中,采用交叉验证方法对分类模型进行验证。交叉验证方法为十折交叉验证方法或五折交叉验证方法中的任意一种。例如采用五折交叉验证方法,具体地,将多个训练样本随机分成10份,每次取其中2份作为交叉验证集,其余8份作为训练集。训练时,先用其中的8份对初始化后的分类模型进行训练,然后对2份交叉验证集进行分类标注,以此重复训练及验证过程5次,每次选取的交叉验证集不同,直至所有的训练样本都被分类标注一遍。In another embodiment, a cross-validation method is used to validate the classification model. The cross-validation method is either a ten-fold cross-validation method or a five-fold cross-validation method. For example, a five-fold cross-validation method is adopted. Specifically, multiple training samples are randomly divided into 10 parts, and 2 of them are taken as the cross-validation set each time, and the remaining 8 parts are used as the training set. When training, first use 8 of them to train the initialized classification model, and then classify and label the 2 cross-validation sets, so as to repeat the training and verification process 5 times, each time the selected cross-validation set is different, until all The training samples of are all classified and labeled again.
控制器30,用于利用贝叶斯优化算法基于每个训练好的分类模型对应的分类准确度,从多个待定策略子集中确定最优策略子集。The controller 30 is configured to use a Bayesian optimization algorithm to determine an optimal strategy subset from a plurality of pending strategy subsets based on the classification accuracy corresponding to each trained classification model.
在本实施方式中,控制器30基于分类准确度并利用贝叶斯优化算法从多个待定策略子集中确定最优策略子集。在其他实施方式中,也可以采用其他算法来选取,在此不做限定。In this embodiment, the controller 30 determines an optimal strategy subset from a plurality of pending strategy subsets based on the classification accuracy and using a Bayesian optimization algorithm. In other embodiments, other algorithms may also be used for selection, which is not limited here.
请参阅图3,可选地,控制器30包括构建单元310、第一确定单元320、第二确定单元330。Please refer to FIG. 3. Optionally, the controller 30 includes a construction unit 310, a first determination unit 320, and a second determination unit 330.
构建单元310,用于基于多个样本点构建高斯过程的回归模型,其中,每个样本点包括训练好的分类模型的分类准确度及训练分类模型所采用的待定策略子集;The constructing unit 310 is configured to construct a regression model of the Gaussian process based on a plurality of sample points, where each sample point includes the classification accuracy of the trained classification model and the undetermined strategy subset used to train the classification model;
第一确定单元320,用于根据回归模型确定贝叶斯优化算法的获取函数;The first determining unit 320 is configured to determine the acquisition function of the Bayesian optimization algorithm according to the regression model;
第二确定单元330,用于通过对获取函数的最大优化,从多个待定策略子集中确定最优策略子集,其中,利用最优策略子集增广后样本训练集训练得到的分类模型的分类准确度最高。The second determining unit 330 is used to determine the optimal strategy subset from a plurality of pending strategy subsets through the maximum optimization of the acquisition function, wherein the optimal strategy subset is used to augment the classification model trained on the sample training set. The classification accuracy is the highest.
可以理解地,我们认为y与x=(μ,α,β)之间存在某种函数关系,即y=f(x)贝叶斯优化算法通过对获取函数进行学习拟合,找到使目标函数f(x)向全局最优值提升的策略参数。每一次贝叶斯优化迭代使用新的样本点来测试目标函数f(x)时,利用这个信息来更新目标函数f(x)的先验分布,最后,利用贝叶斯优化算法测试由后验分布给出的全局最值最可能出现的位置的样本点。Understandably, we believe that there is a certain functional relationship between y and x=(μ,α,β), that is, y=f(x) Bayesian optimization algorithm learns and fits the acquisition function to find the objective function The strategy parameter that f(x) promotes to the global optimal value. When every Bayesian optimization iteration uses a new sample point to test the objective function f(x), use this information to update the prior distribution of the objective function f(x), and finally, use the Bayesian optimization algorithm to test the posterior The distribution gives the sample points where the global maximum value is most likely to occur.
在本实施方式中,贝叶斯优化迭代的过程中,通过获取函数指导我们去选择样本点,不断修正GP高斯过程曲线去逼近目标函数f(x),当获取函数最大的时候说明选择的样本 点最优,相当于我们就搜索到了令目标函数f(x)最大的最优策略子集。In this embodiment, during the Bayesian optimization iteration process, we guide us to select sample points through the acquisition function, and continuously modify the GP Gaussian process curve to approximate the objective function f(x). When the acquisition function is the largest, the selected sample is explained The point is optimal, which means that we have searched for the optimal strategy subset that maximizes the objective function f(x).
由于f(x)形式无法显性求取,我们用高斯过程来逼近,Since the form of f(x) cannot be obtained explicitly, we use Gaussian process to approximate,
即f(x)~GP(m(x),k(x,x′)),其中m(x)代表样本点f(x)的数学期望E(f(x)),在贝叶斯优化中通常取0,k(x,x')为核函数,描述的是x的协方差。That is, f(x)~GP(m(x), k(x,x′)), where m(x) represents the mathematical expectation E(f(x)) of the sample point f(x), in Bayesian optimization Usually 0, k(x, x') is the kernel function, which describes the covariance of x.
对于每个x都有一个对应的高斯分布,而对于一组{x 1,x 2...x n},假设y值服从联合正态分布,其均值为0,协方差为: For each x there is a corresponding Gaussian distribution, and for a set of {x 1 , x 2 ... x n }, assuming that the y value obeys the joint normal distribution, its mean value is 0, and the covariance is:
Figure PCTCN2020111666-appb-000006
其中,协方差只与x有关,和y无关。
Figure PCTCN2020111666-appb-000006
Among them, the covariance is only related to x, and has nothing to do with y.
对于一个新的样本点x n+1,联合高斯分布为: For a new sample point x n+1 , the joint Gaussian distribution is:
Figure PCTCN2020111666-appb-000007
Figure PCTCN2020111666-appb-000007
因此可以通过前n个样本点估计出f n+1的后验概率分布:P(f n+1|D 1:t,x t+1)~N(μ n(x),σ n 2(x)),其中,μ n(x)=k TK -1f 1:n;σ n 2(x)=k(x n+1,x n+1)-k TK -1k; Therefore, the posterior probability distribution of f n+1 can be estimated through the first n sample points : P(f n+1 |D 1:t ,x t+1 )~N(μ n (x),σ n 2 ( x)), where μ n (x)=k T K -1 f 1:n ; σ n 2 (x)=k(x n+1 ,x n+1 )-k T K -1 k;
在本实施方式中,采用改进概率(Probability of Improvement,POI)作为获取函数。In this embodiment, the probability of improvement (POI) is used as the acquisition function.
获取函数为:
Figure PCTCN2020111666-appb-000008
The get function is:
Figure PCTCN2020111666-appb-000008
其中,f(x)为x的目标函数值,x为验证精度,f(X+)为到目前为止最优的x的目标函数值,μ(x),σ(x)分别是高斯过程所得到的目标函数的均值和方差,即f(x)的后验分布,Φ(·)表示的是正态累计分布函数。ξ为trade-off系数,如果没有该系数,POI函数会倾向于取在X+周围的点,收敛到接近f(X+)附近的位置,即倾向于开发而不是探索,因此加入该项进行权衡。通过不断尝试新的x,下一个最大点应该要比它大或至少与之相等。因此,下一个采样在交叉点f(X+)和置信域之间,我们能假定在f(X+)点以下的样本是可以丢弃的,因为我们只需要搜索令目标函数取极大值的参数,于是通过迭代这一过程缩小了观察区域,直到搜索到最优解,使得POI(X)最大。Among them, f(x) is the objective function value of x, x is the verification accuracy, f(X+) is the optimal objective function value of x so far, μ(x) and σ(x) are obtained by Gaussian process respectively The mean and variance of the objective function are the posterior distribution of f(x), and Φ(·) represents the normal cumulative distribution function. ξ is the trade-off coefficient. If there is no such coefficient, the POI function will tend to take a point around X+ and converge to a position close to f(X+), that is, it tends to develop rather than explore, so this item is added to make a trade-off. By constantly trying new x, the next largest point should be larger or at least equal to it. Therefore, the next sample is between the intersection f(X+) and the confidence region. We can assume that the samples below f(X+) can be discarded, because we only need to search for the parameter that makes the objective function take the maximum value. So through the iterative process, the observation area is reduced until the optimal solution is searched, so that the POI(X) is maximized.
进一步地,在控制器30选取出最优增广策略后,控制器30还用于输出最优增广策略至增广器10,增广器10将最优增广策略确认为预设的样本训练集的增广策略。可以理解地,在增广器10获取最优增广策略后,增广器每次进行样本增广时,都将采用控制器输出的最优增广策略来进行样本增广。Further, after the controller 30 selects the optimal augmentation strategy, the controller 30 is also used to output the optimal augmentation strategy to the augmenter 10, and the augmenter 10 confirms the optimal augmentation strategy as a preset sample The augmentation strategy of the training set. Understandably, after the augmenter 10 obtains the optimal augmentation strategy, every time the augmenter performs sample augmentation, it will use the optimal augmentation strategy output by the controller for sample augmentation.
本申请实施例提供了一种计算机可读存储介质。其中,所述计算机可读存储介质可以是非易失性,也可以是易失性的。其中,存储介质包括存储的程序,在程序运行时控制存储介质所在设备执行以下步骤:The embodiment of the present application provides a computer-readable storage medium. Wherein, the computer-readable storage medium may be non-volatile or volatile. Wherein, the storage medium includes a stored program, and when the program is running, the device where the storage medium is located is controlled to perform the following steps:
从增广策略集合中选取多个待定策略子集对预设的样本训练集进行样本增广,得到多个增广后的样本训练集,其中,每个待定策略子集由增广策略集合中至少一个增广策略组成;利用每个增广后的样本训练集训练初始化的分类模型,得到多个训练后的分类模型;将预设的样本验证集输入每个训练后的分类模型,得到训练好的分类模型对应的分类准确度;利用贝叶斯优化算法基于每个训练好的分类模型对应的分类准确度,从多个待定策略子集中确定最优策略子集。Select multiple undetermined strategy subsets from the augmented strategy set to perform sample augmentation on the preset sample training set to obtain multiple augmented sample training sets, where each undetermined strategy subset is from the augmented strategy set At least one augmentation strategy composition; use each augmented sample training set to train the initialized classification model to obtain multiple trained classification models; input the preset sample validation set into each trained classification model to obtain training The classification accuracy corresponding to a good classification model; the Bayesian optimization algorithm is used to determine the optimal strategy subset from multiple pending strategy subsets based on the classification accuracy corresponding to each trained classification model.
可选地,在程序运行时控制存储介质所在设备执行利用贝叶斯优化算法基于每个训练好的分类模型对应的分类准确度,从多个待定策略子集中确定最优策略子集的步骤,包括:Optionally, when the program is running, the device where the storage medium is located is controlled to execute the step of using a Bayesian optimization algorithm to determine the optimal strategy subset from a plurality of pending strategy subsets based on the classification accuracy corresponding to each trained classification model, include:
基于多个样本点构建高斯过程的回归模型,其中,每个样本点包括训练好的分类模型的分类准确度及训练分类模型所采用的待定策略子集;根据回归模型确定贝叶斯优化算法的获取函数;通过对获取函数的最大优化,从多个待定策略子集中确定最优策略子集,其中,利用最优策略子集增广后样本训练集训练得到的分类模型的分类准确度最高。Construct a regression model of the Gaussian process based on multiple sample points, where each sample point includes the classification accuracy of the trained classification model and the undetermined strategy subset used to train the classification model; the Bayesian optimization algorithm is determined according to the regression model Acquisition function: Through the maximum optimization of the acquisition function, the optimal strategy subset is determined from multiple pending strategy subsets. Among them, the classification model trained on the sample training set after augmenting the optimal strategy subset has the highest classification accuracy.
可选地,在程序运行时控制存储介质所在设备执行将预设的样本验证集输入每个训练后的分类模型,得到训练好的分类模型对应的分类准确度,包括:Optionally, when the program is running, the device where the storage medium is located is controlled to execute the input of a preset sample verification set into each trained classification model to obtain the classification accuracy corresponding to the trained classification model, including:
将预设的样本验证集输入每个训练后的分类模型;获取分类模型输出的训练精度及验证精度;根据训练精度和验证精度判断分类模型是否拟合良好;将拟合良好的分类模型确定为训练好的分类模型,并将训练好的分类模型的验证精度作为分类模型的分类准确度。Input the preset sample validation set into each trained classification model; obtain the training accuracy and verification accuracy of the output of the classification model; judge whether the classification model fits well according to the training accuracy and verification accuracy; determine the well-fitted classification model as The trained classification model, and the verification accuracy of the trained classification model is used as the classification accuracy of the classification model.
可选地,在程序运行时控制存储介质所在设备执行利用每个增广后的样本训练集训练初始化的分类模型,得到多个训练后的分类模型的步骤,包括:利用卷积神经网络提取输入分类模型的增广后的样本训练集中的每个样本的特征图;根据特征图,对增广后的样本训练集中的对应一个样本进行分类预测,得到分类结果;获取分类结果集合与样本训练集中的所有样本的标签集合的均方误差的损失函数;通过反向传播对卷积神经网络进行优化,以使得损失函数的值收敛,得到优化训练后的分类模型。Optionally, when the program is running, the device where the storage medium is located is controlled to execute a classification model trained and initialized with each augmented sample training set to obtain multiple trained classification models, including: extracting input using a convolutional neural network The feature map of each sample in the augmented sample training set of the classification model; according to the feature map, classify and predict a corresponding sample in the augmented sample training set to obtain the classification result; obtain the classification result set and the sample training set The loss function of the mean square error of the label set of all samples; the convolutional neural network is optimized by backpropagation, so that the value of the loss function converges, and the optimized training classification model is obtained.
可选地,在程序运行时控制存储介质所在设备在执行将预设的样本验证集输入每个训练后的分类模型,得到训练好的分类模型对应的分类准确度之前,还包括:从预设的样本验证集随机抽取多个验证子集;将多个验证子集分别输入每个训练后的分类模型。Optionally, when the program is running, controlling the device where the storage medium is located before executing the input of the preset sample verification set into each trained classification model to obtain the classification accuracy corresponding to the trained classification model also includes: The sample validation set of the sample validation set randomly selects multiple validation subsets; input multiple validation subsets into each trained classification model.
图4是本申请实施例提供的一种计算机设备的示意图。如图3所示,该实施例的计算机设备100包括:处理器101、存储器102以及存储在存储器102中并可在处理器101上运行的计算机程序103,处理器101执行计算机程序103时实现实施例中的图像数据的增广策略选取方法,为避免重复,此处不一一赘述。或者,该计算机程序被处理器101执行时实现实施例中图像数据的增广策略选取系统中各模型/单元的功能,为避免重复,此处不一一赘述。Fig. 4 is a schematic diagram of a computer device provided by an embodiment of the present application. As shown in FIG. 3, the computer device 100 of this embodiment includes a processor 101, a memory 102, and a computer program 103 stored in the memory 102 and running on the processor 101. The processor 101 executes the computer program 103 when the computer program 103 is executed. The method of selecting the augmentation strategy of the image data in the example is not repeated here to avoid repetition. Alternatively, when the computer program is executed by the processor 101, the function of each model/unit in the image data augmentation strategy selection system in the embodiment is realized. In order to avoid repetition, it will not be repeated here.
计算机设备100可以是桌上型计算机、笔记本、掌上电脑及云端服务器等计算设备。计算机设备可包括,但不仅限于,处理器101、存储器102。本领域技术人员可以理解, 图3仅仅是计算机设备100的示例,并不构成对计算机设备100的限定,可以包括比图示更多或更少的部件,或者组合某些部件,或者不同的部件,例如计算机设备还可以包括输入输出设备、网络接入设备、总线等。The computer device 100 may be a computing device such as a desktop computer, a notebook, a palmtop computer, and a cloud server. The computer device may include, but is not limited to, a processor 101 and a memory 102. Those skilled in the art can understand that FIG. 3 is only an example of the computer device 100 and does not constitute a limitation on the computer device 100. It may include more or less components than shown, or a combination of certain components, or different components. For example, computer equipment may also include input and output devices, network access devices, buses, and so on.
所称处理器101可以是中央处理单元(Central Processing Unit,CPU),还可以是其他通用处理器、数字信号处理器(Digital Signal Processor,DSP)、专用集成电路(Application Specific Integrated Circuit,ASIC)、现场可编程门阵列(Field-Programmable Gate Array,FPGA)或者其他可编程逻辑器件、分立门或者晶体管逻辑器件、分立硬件组件等。通用处理器可以是微处理器或者该处理器也可以是任何常规的处理器等。The so-called processor 101 may be a central processing unit (Central Processing Unit, CPU), other general-purpose processors, digital signal processors (Digital Signal Processor, DSP), application specific integrated circuits (ASIC), Field-Programmable Gate Array (FPGA) or other programmable logic devices, discrete gates or transistor logic devices, discrete hardware components, etc. The general-purpose processor may be a microprocessor or the processor may also be any conventional processor or the like.
存储器102可以是计算机设备100的内部存储单元,例如计算机设备100的硬盘或内存。存储器102也可以是计算机设备100的外部存储设备,例如计算机设备100上配备的插接式硬盘,智能存储卡(Smart Media Card,SMC),安全数字(Secure Digital,SD)卡,闪存卡(Flash Card)等。进一步地,存储器102还可以既包括计算机设备100的内部存储单元也包括外部存储设备。存储器102用于存储计算机程序以及计算机设备所需的其他程序和数据。存储器102还可以用于暂时地存储已经输出或者将要输出的数据。The memory 102 may be an internal storage unit of the computer device 100, such as a hard disk or memory of the computer device 100. The memory 102 may also be an external storage device of the computer device 100, such as a plug-in hard disk equipped on the computer device 100, a smart memory card (Smart Media Card, SMC), a Secure Digital (SD) card, and a flash memory card (Flash). Card) and so on. Further, the memory 102 may also include both an internal storage unit of the computer device 100 and an external storage device. The memory 102 is used to store computer programs and other programs and data required by the computer equipment. The memory 102 can also be used to temporarily store data that has been output or will be output.
所属领域的技术人员可以清楚地了解到,为描述的方便和简洁,上述描述的系统,装置和单元的具体工作过程,可以参考前述方法实施例中的对应过程,在此不再赘述。Those skilled in the art can clearly understand that, for the convenience and conciseness of the description, the specific working process of the above-described system, device, and unit can refer to the corresponding process in the foregoing method embodiment, which will not be repeated here.
在本申请所提供的几个实施例中,应该理解到,所揭露的系统,装置和方法,可以通过其它的方式实现。例如,以上所描述的装置实施例仅仅是示意性的,例如,所述单元的划分,仅仅为一种逻辑功能划分,实际实现时可以有另外的划分方式,例如,多个单元或组件可以结合或者可以集成到另一个系统,或一些特征可以忽略,或不执行。另一点,所显示或讨论的相互之间的耦合或直接耦合或通信连接可以是通过一些接口,装置或单元的间接耦合或通信连接,可以是电性,机械或其它的形式。In the several embodiments provided in this application, it should be understood that the disclosed system, device, and method can be implemented in other ways. For example, the device embodiments described above are only illustrative. For example, the division of the units is only a logical function division, and there may be other divisions in actual implementation, for example, multiple units or components may be combined. Or it can be integrated into another system, or some features can be ignored or not implemented. In addition, the displayed or discussed mutual coupling or direct coupling or communication connection may be indirect coupling or communication connection through some interfaces, devices or units, and may be in electrical, mechanical or other forms.
所述作为分离部件说明的单元可以是或者也可以不是物理上分开的,作为单元显示的部件可以是或者也可以不是物理单元,即可以位于一个地方,或者也可以分布到多个网络单元上。可以根据实际的需要选择其中的部分或者全部单元来实现本实施例方案的目的。The units described as separate components may or may not be physically separated, and the components displayed as units may or may not be physical units, that is, they may be located in one place, or they may be distributed on multiple network units. Some or all of the units may be selected according to actual needs to achieve the objectives of the solutions of the embodiments.
另外,在本申请各个实施例中的各功能单元可以集成在一个处理单元中,也可以是各个单元单独物理存在,也可以两个或两个以上单元集成在一个单元中。上述集成的单元既可以采用硬件的形式实现,也可以采用硬件加软件功能单元的形式实现。In addition, the functional units in the various embodiments of the present application may be integrated into one processing unit, or each unit may exist alone physically, or two or more units may be integrated into one unit. The above-mentioned integrated unit may be implemented in the form of hardware, or may be implemented in the form of hardware plus software functional units.
上述以软件功能单元的形式实现的集成的单元,可以存储在一个计算机可读取存储介质中。上述软件功能单元存储在一个存储介质中,包括若干指令用以使得一台计算机装置(可以是个人计算机,服务器,或者网络装置等)或处理器(Processor)执行本申请各个实施例所述方法的部分步骤。而前述的存储介质包括:U盘、移动硬盘、只读存储器(Read-Only Memory,ROM)、随机存取存储器(Random Access Memory,RAM)、磁碟或者光盘等各种可以存储程序代码的介质。The above-mentioned integrated unit implemented in the form of a software functional unit may be stored in a computer readable storage medium. The above-mentioned software functional unit is stored in a storage medium and includes several instructions to make a computer device (which may be a personal computer, a server, or a network device, etc.) or a processor (Processor) execute the method described in each embodiment of the present application. Part of the steps. The aforementioned storage media include: U disk, mobile hard disk, read-only memory (Read-Only Memory, ROM), random access memory (Random Access Memory, RAM), magnetic disks or optical disks and other media that can store program codes. .
以上所述,仅为本申请的具体实施方式,但本申请的保护范围并不局限于此,任何熟 悉本技术领域的技术人员在本申请揭露的技术范围内,可轻易想到变化或替换,都应涵盖在本申请的保护范围之内。因此,本申请的保护范围应以所述权利要求的保护范围为准。The above are only specific implementations of this application, but the protection scope of this application is not limited to this. Any person skilled in the art can easily think of changes or substitutions within the technical scope disclosed in this application. Should be covered within the scope of protection of this application. Therefore, the protection scope of this application should be subject to the protection scope of the claims.

Claims (20)

  1. 一种图像数据的增广策略选取方法,其中,所述方法包括:A method for selecting an augmentation strategy for image data, wherein the method includes:
    从增广策略集合中选取多个待定策略子集对预设的样本训练集进行样本增广,得到多个增广后的样本训练集,其中,每个所述待定策略子集由所述增广策略集合中至少一个增广策略组成;Select multiple undetermined strategy subsets from the augmented strategy set to perform sample augmentation on the preset sample training set to obtain multiple augmented sample training sets, wherein each of the undetermined strategy subsets is augmented by the At least one augmentation strategy in the broad strategy set;
    利用每个所述增广后的样本训练集训练初始化的分类模型,得到多个训练后的分类模型;Training an initialized classification model using each of the augmented sample training sets to obtain multiple trained classification models;
    将预设的样本验证集输入每个所述训练后的分类模型,得到训练好的所述分类模型对应的分类准确度;Input a preset sample verification set into each of the trained classification models to obtain the classification accuracy corresponding to the trained classification model;
    利用贝叶斯优化算法基于每个所述训练好的分类模型对应的分类准确度,从多个待定策略子集中确定最优策略子集。A Bayesian optimization algorithm is used to determine the optimal strategy subset from a plurality of pending strategy subsets based on the classification accuracy corresponding to each of the trained classification models.
  2. 根据权利要求1所述的方法,其中,所述利用贝叶斯优化算法基于每个所述训练好的分类模型对应的分类准确度,从多个待定策略子集中确定最优策略子集的步骤,包括:The method according to claim 1, wherein the step of using the Bayesian optimization algorithm to determine the optimal strategy subset from a plurality of pending strategy subsets based on the classification accuracy corresponding to each of the trained classification models ,include:
    基于多个样本点构建高斯过程的回归模型,其中,每个样本点包括所述训练好的分类模型的分类准确度及训练所述分类模型所采用的待定策略子集;Constructing a regression model of the Gaussian process based on multiple sample points, where each sample point includes the classification accuracy of the trained classification model and the undetermined strategy subset used to train the classification model;
    根据所述回归模型确定贝叶斯优化算法的获取函数;Determining the acquisition function of the Bayesian optimization algorithm according to the regression model;
    通过对所述获取函数的最大优化,从多个所述待定策略子集中确定最优策略子集,其中,利用所述最优策略子集增广后的样本训练集训练得到的分类模型的分类准确度最高。Through the maximum optimization of the acquisition function, an optimal strategy subset is determined from a plurality of the pending strategy subsets, wherein the classification of the classification model obtained by training the sample training set after the expansion of the optimal strategy subset is used The highest accuracy.
  3. 根据权利要求1所述的方法,其中,所述将预设的样本验证集输入每个所述训练后的分类模型,得到训练好的所述分类模型对应的分类准确度,包括:The method according to claim 1, wherein the inputting a preset sample verification set into each of the trained classification models to obtain the classification accuracy corresponding to the trained classification model comprises:
    将预设的样本验证集输入每个所述训练后的分类模型;Input a preset sample verification set into each of the trained classification models;
    获取所述分类模型输出的训练精度及验证精度;Acquiring training accuracy and verification accuracy output by the classification model;
    根据所述训练精度和所述验证精度判断所述分类模型是否拟合良好;Judging whether the classification model fits well according to the training accuracy and the verification accuracy;
    将拟合良好的所述分类模型确定为训练好的分类模型,并将所述训练好的分类模型的验证精度作为所述分类模型的分类准确度。The well-fitted classification model is determined as the trained classification model, and the verification accuracy of the trained classification model is used as the classification accuracy of the classification model.
  4. 根据权利要求1所述的方法,其中,所述利用每个所述增广后的样本训练集训练初始化的分类模型,得到多个训练后的分类模型,包括:The method according to claim 1, wherein the training an initialized classification model using each of the augmented sample training sets to obtain a plurality of trained classification models comprises:
    利用卷积神经网络提取输入分类模型的所述增广后的样本训练集中的每个样本的特征图;Extracting a feature map of each sample in the augmented sample training set of the input classification model by using a convolutional neural network;
    根据所述特征图,对所述增广后的样本训练集中的对应一个样本进行分类预测,得到分类结果;Perform classification prediction on a corresponding sample in the augmented sample training set according to the feature map to obtain a classification result;
    获取所述分类结果集合与所述样本训练集中的所有样本的标签集合的均方误差的损失函数;Acquiring a loss function of the mean square error of the classification result set and the label set of all samples in the sample training set;
    通过反向传播对所述卷积神经网络进行优化,以使得所述损失函数的值收敛,得到优化训练后的所述分类模型。The convolutional neural network is optimized by back propagation, so that the value of the loss function converges, and the optimized and trained classification model is obtained.
  5. 根据权利要求1所述的方法,其中,在所述将预设的样本验证集输入每个所述训练 后的分类模型,得到训练好的所述分类模型对应的分类准确度之前,所述方法还包括:The method according to claim 1, wherein, before the preset sample verification set is input into each of the trained classification models to obtain the classification accuracy corresponding to the trained classification model, the method Also includes:
    从所述预设的样本验证集随机抽取多个验证子集;Randomly select a plurality of verification subsets from the preset sample verification set;
    将所述多个验证子集分别输入每个所述训练后的分类模型。The multiple validation subsets are input into each of the trained classification models respectively.
  6. 根据权利要求1所述的方法,其中,所述增广策略集合包括旋转变换、翻转变换、缩放变换、平移变换、尺度变换、区域裁剪、噪声添加、分段仿射、随机掩盖、边界检测、对比度变换、颜色抖动、随机混合及复合叠加。The method according to claim 1, wherein the augmentation strategy set includes rotation transformation, flip transformation, zoom transformation, translation transformation, scale transformation, region cropping, noise addition, segment affine, random concealment, boundary detection, Contrast transformation, color dithering, random mixing and composite overlay.
  7. 一种图像数据的增广策略选取系统,其中,所述系统包括增广器、分类模型及控制器;An augmentation strategy selection system for image data, wherein the system includes an augmenter, a classification model, and a controller;
    所述增广器,用于从增广策略集合中选取多个待定策略子集对预设的样本训练集进行样本增广,得到多个增广后的样本训练集,其中,每个所述待定策略子集由所述增广策略集合中至少一个增广策略组成;The augmenter is used to select multiple undetermined strategy subsets from the augmentation strategy set to perform sample augmentation on the preset sample training set to obtain multiple augmented sample training sets, wherein each of the The undetermined strategy subset is composed of at least one augmentation strategy in the augmentation strategy set;
    所述分类模型,用于利用每个所述增广后的样本训练集训练初始化的分类模型,得到多个训练后的分类模型;并将预设的样本验证集输入每个所述训练后的分类模型,得到训练好的所述分类模型对应的分类准确度;The classification model is used to train an initialized classification model using each of the augmented sample training sets to obtain a plurality of trained classification models; and input a preset sample verification set into each of the trained samples A classification model to obtain the classification accuracy corresponding to the trained classification model;
    所述控制器,用于利用贝叶斯优化算法基于每个所述训练好的分类模型对应的分类准确度,从多个待定策略子集中确定最优策略子集。The controller is configured to use a Bayesian optimization algorithm to determine an optimal strategy subset from a plurality of pending strategy subsets based on the classification accuracy corresponding to each of the trained classification models.
  8. 根据权利要求7所述的系统,其中,所述控制器包括构建单元、第一确定单元、第二确定单元;The system according to claim 7, wherein the controller includes a construction unit, a first determination unit, and a second determination unit;
    所述构建单元,用于基于多个样本点构建高斯过程的回归模型,其中,每个样本点包括所述训练好的分类模型的分类准确度及训练所述分类模型所采用的待定策略子集;The construction unit is configured to construct a regression model of the Gaussian process based on a plurality of sample points, wherein each sample point includes the classification accuracy of the trained classification model and a subset of the pending strategy used to train the classification model ;
    所述第一确定单元,用于根据所述回归模型确定贝叶斯优化算法的获取函数;The first determining unit is configured to determine the acquisition function of the Bayesian optimization algorithm according to the regression model;
    所述第二确定单元,用于通过对所述获取函数的最大优化,从多个所述待定策略子集中确定最优策略子集,其中,利用所述最优策略子集增广后的样本训练集训练得到的分类模型的分类准确度最高。The second determining unit is configured to determine an optimal strategy subset from a plurality of the pending strategy subsets through the maximum optimization of the acquisition function, wherein the sample after the expansion of the optimal strategy subset is used The classification model trained on the training set has the highest classification accuracy.
  9. 一种计算机设备,其中,所述计算机设备包括存储器和处理器,所述存储器和所述处理器相互连接,所述存储器用于存储计算机程序,所述计算机程序被配置为由所述处理器执行,所述计算机程序配置用于执行一种图像数据的增广策略选取方法:A computer device, wherein the computer device includes a memory and a processor, the memory and the processor are connected to each other, and the memory is used to store a computer program configured to be executed by the processor , The computer program is configured to execute a method for selecting an augmentation strategy for image data:
    其中,所述方法包括:Wherein, the method includes:
    从增广策略集合中选取多个待定策略子集对预设的样本训练集进行样本增广,得到多个增广后的样本训练集,其中,每个所述待定策略子集由所述增广策略集合中至少一个增广策略组成;Select multiple undetermined strategy subsets from the augmented strategy set to perform sample augmentation on the preset sample training set to obtain multiple augmented sample training sets, wherein each of the undetermined strategy subsets is augmented by the At least one augmentation strategy in the broad strategy set;
    利用每个所述增广后的样本训练集训练初始化的分类模型,得到多个训练后的分类模型;Training an initialized classification model using each of the augmented sample training sets to obtain multiple trained classification models;
    将预设的样本验证集输入每个所述训练后的分类模型,得到训练好的所述分类模型对应的分类准确度;Input a preset sample verification set into each of the trained classification models to obtain the classification accuracy corresponding to the trained classification model;
    利用贝叶斯优化算法基于每个所述训练好的分类模型对应的分类准确度,从多个待定 策略子集中确定最优策略子集。The Bayesian optimization algorithm is used to determine the optimal strategy subset from multiple pending strategy subsets based on the classification accuracy corresponding to each of the trained classification models.
  10. 根据权利要求9所述的计算机设备,其中,所述利用贝叶斯优化算法基于每个所述训练好的分类模型对应的分类准确度,从多个待定策略子集中确定最优策略子集的步骤,包括:The computer device according to claim 9, wherein the Bayesian optimization algorithm is used to determine the optimal strategy subset from a plurality of pending strategy subsets based on the classification accuracy corresponding to each of the trained classification models The steps include:
    基于多个样本点构建高斯过程的回归模型,其中,每个样本点包括所述训练好的分类模型的分类准确度及训练所述分类模型所采用的待定策略子集;Constructing a regression model of the Gaussian process based on multiple sample points, where each sample point includes the classification accuracy of the trained classification model and the undetermined strategy subset used to train the classification model;
    根据所述回归模型确定贝叶斯优化算法的获取函数;Determining the acquisition function of the Bayesian optimization algorithm according to the regression model;
    通过对所述获取函数的最大优化,从多个所述待定策略子集中确定最优策略子集,其中,利用所述最优策略子集增广后的样本训练集训练得到的分类模型的分类准确度最高。Through the maximum optimization of the acquisition function, an optimal strategy subset is determined from a plurality of the pending strategy subsets, wherein the classification of the classification model obtained by training the sample training set after the expansion of the optimal strategy subset is used The highest accuracy.
  11. 根据权利要求9所述的计算机设备,其中,所述将预设的样本验证集输入每个所述训练后的分类模型,得到训练好的所述分类模型对应的分类准确度,包括:The computer device according to claim 9, wherein the inputting a preset sample verification set into each of the trained classification models to obtain the classification accuracy corresponding to the trained classification model comprises:
    将预设的样本验证集输入每个所述训练后的分类模型;Input a preset sample verification set into each of the trained classification models;
    获取所述分类模型输出的训练精度及验证精度;Acquiring training accuracy and verification accuracy output by the classification model;
    根据所述训练精度和所述验证精度判断所述分类模型是否拟合良好;Judging whether the classification model fits well according to the training accuracy and the verification accuracy;
    将拟合良好的所述分类模型确定为训练好的分类模型,并将所述训练好的分类模型的验证精度作为所述分类模型的分类准确度。The well-fitted classification model is determined as the trained classification model, and the verification accuracy of the trained classification model is used as the classification accuracy of the classification model.
  12. 根据权利要求9所述的计算机设备,其中,所述利用每个所述增广后的样本训练集训练初始化的分类模型,得到多个训练后的分类模型,包括:9. The computer device according to claim 9, wherein said training an initialized classification model using each of said augmented sample training sets to obtain a plurality of trained classification models comprises:
    利用卷积神经网络提取输入分类模型的所述增广后的样本训练集中的每个样本的特征图;Extracting a feature map of each sample in the augmented sample training set of the input classification model by using a convolutional neural network;
    根据所述特征图,对所述增广后的样本训练集中的对应一个样本进行分类预测,得到分类结果;Perform classification prediction on a corresponding sample in the augmented sample training set according to the feature map to obtain a classification result;
    获取所述分类结果集合与所述样本训练集中的所有样本的标签集合的均方误差的损失函数;Acquiring a loss function of the mean square error of the classification result set and the label set of all samples in the sample training set;
    通过反向传播对所述卷积神经网络进行优化,以使得所述损失函数的值收敛,得到优化训练后的所述分类模型。The convolutional neural network is optimized by back propagation, so that the value of the loss function converges, and the optimized and trained classification model is obtained.
  13. 根据权利要求9所述的计算机设备,其中,在所述将预设的样本验证集输入每个所述训练后的分类模型,得到训练好的所述分类模型对应的分类准确度之前,所述方法还包括:The computer device according to claim 9, wherein, before said inputting a preset sample verification set into each of said trained classification models to obtain the classification accuracy corresponding to said trained classification model, said Methods also include:
    从所述预设的样本验证集随机抽取多个验证子集;Randomly select a plurality of verification subsets from the preset sample verification set;
    将所述多个验证子集分别输入每个所述训练后的分类模型。The multiple validation subsets are input into each of the trained classification models respectively.
  14. 根据权利要求9所述的计算机设备,其中,所述增广策略集合包括旋转变换、翻转变换、缩放变换、平移变换、尺度变换、区域裁剪、噪声添加、分段仿射、随机掩盖、边界检测、对比度变换、颜色抖动、随机混合及复合叠加。The computer device according to claim 9, wherein the augmentation strategy set includes rotation transformation, flip transformation, zoom transformation, translation transformation, scale transformation, region cropping, noise addition, segment affine, random concealment, boundary detection , Contrast transformation, color dithering, random mixing and composite overlay.
  15. 一种计算机可读存储介质,其中,所述计算机可读存储介质存储有计算机程序,所述计算机程序被处理器执行时用于实现一种图像数据的增广策略选取方法,所述方法包 括以下步骤:A computer-readable storage medium, wherein the computer-readable storage medium stores a computer program, and when the computer program is executed by a processor, it is used to implement a method for selecting an augmentation strategy for image data. The method includes the following step:
    从增广策略集合中选取多个待定策略子集对预设的样本训练集进行样本增广,得到多个增广后的样本训练集,其中,每个所述待定策略子集由所述增广策略集合中至少一个增广策略组成;Select multiple undetermined strategy subsets from the augmented strategy set to perform sample augmentation on the preset sample training set to obtain multiple augmented sample training sets, wherein each of the undetermined strategy subsets is augmented by the At least one augmentation strategy in the broad strategy set;
    利用每个所述增广后的样本训练集训练初始化的分类模型,得到多个训练后的分类模型;Training an initialized classification model using each of the augmented sample training sets to obtain multiple trained classification models;
    将预设的样本验证集输入每个所述训练后的分类模型,得到训练好的所述分类模型对应的分类准确度;Input a preset sample verification set into each of the trained classification models to obtain the classification accuracy corresponding to the trained classification model;
    利用贝叶斯优化算法基于每个所述训练好的分类模型对应的分类准确度,从多个待定策略子集中确定最优策略子集。A Bayesian optimization algorithm is used to determine the optimal strategy subset from a plurality of pending strategy subsets based on the classification accuracy corresponding to each of the trained classification models.
  16. 根据权利要求15所述的计算机可读存储介质,其中,所述利用贝叶斯优化算法基于每个所述训练好的分类模型对应的分类准确度,从多个待定策略子集中确定最优策略子集的步骤,包括:The computer-readable storage medium according to claim 15, wherein the Bayesian optimization algorithm is used to determine the optimal strategy from a plurality of sub-sets of pending strategies based on the classification accuracy corresponding to each of the trained classification models The steps of the subset include:
    基于多个样本点构建高斯过程的回归模型,其中,每个样本点包括所述训练好的分类模型的分类准确度及训练所述分类模型所采用的待定策略子集;Constructing a regression model of the Gaussian process based on multiple sample points, where each sample point includes the classification accuracy of the trained classification model and the undetermined strategy subset used to train the classification model;
    根据所述回归模型确定贝叶斯优化算法的获取函数;Determining the acquisition function of the Bayesian optimization algorithm according to the regression model;
    通过对所述获取函数的最大优化,从多个所述待定策略子集中确定最优策略子集,其中,利用所述最优策略子集增广后的样本训练集训练得到的分类模型的分类准确度最高。Through the maximum optimization of the acquisition function, an optimal strategy subset is determined from a plurality of the pending strategy subsets, wherein the classification of the classification model obtained by training the sample training set after the expansion of the optimal strategy subset is used The highest accuracy.
  17. 根据权利要求15所述的计算机可读存储介质,其中,所述将预设的样本验证集输入每个所述训练后的分类模型,得到训练好的所述分类模型对应的分类准确度,包括:The computer-readable storage medium according to claim 15, wherein said inputting a preset sample verification set into each of said trained classification models to obtain the classification accuracy corresponding to said trained classification model comprises :
    将预设的样本验证集输入每个所述训练后的分类模型;Input a preset sample verification set into each of the trained classification models;
    获取所述分类模型输出的训练精度及验证精度;Acquiring training accuracy and verification accuracy output by the classification model;
    根据所述训练精度和所述验证精度判断所述分类模型是否拟合良好;Judging whether the classification model fits well according to the training accuracy and the verification accuracy;
    将拟合良好的所述分类模型确定为训练好的分类模型,并将所述训练好的分类模型的验证精度作为所述分类模型的分类准确度。The well-fitted classification model is determined as the trained classification model, and the verification accuracy of the trained classification model is used as the classification accuracy of the classification model.
  18. 根据权利要求15所述的计算机可读存储介质,其中,所述利用每个所述增广后的样本训练集训练初始化的分类模型,得到多个训练后的分类模型,包括:15. The computer-readable storage medium according to claim 15, wherein the training an initialized classification model using each of the augmented sample training sets to obtain a plurality of trained classification models comprises:
    利用卷积神经网络提取输入分类模型的所述增广后的样本训练集中的每个样本的特征图;Extracting a feature map of each sample in the augmented sample training set of the input classification model by using a convolutional neural network;
    根据所述特征图,对所述增广后的样本训练集中的对应一个样本进行分类预测,得到分类结果;Perform classification prediction on a corresponding sample in the augmented sample training set according to the feature map to obtain a classification result;
    获取所述分类结果集合与所述样本训练集中的所有样本的标签集合的均方误差的损失函数;Acquiring a loss function of the mean square error of the classification result set and the label set of all samples in the sample training set;
    通过反向传播对所述卷积神经网络进行优化,以使得所述损失函数的值收敛,得到优化训练后的所述分类模型。The convolutional neural network is optimized by back propagation, so that the value of the loss function converges, and the optimized and trained classification model is obtained.
  19. 根据权利要求15所述的计算机可读存储介质,其中,在所述将预设的样本验证集 输入每个所述训练后的分类模型,得到训练好的所述分类模型对应的分类准确度之前,所述方法还包括:The computer-readable storage medium according to claim 15, wherein, before inputting a preset sample verification set into each of the trained classification models to obtain the classification accuracy corresponding to the trained classification model , The method further includes:
    从所述预设的样本验证集随机抽取多个验证子集;Randomly select a plurality of verification subsets from the preset sample verification set;
    将所述多个验证子集分别输入每个所述训练后的分类模型。The multiple validation subsets are input into each of the trained classification models respectively.
  20. 根据权利要求15所述的计算机可读存储介质,其中,所述增广策略集合包括旋转变换、翻转变换、缩放变换、平移变换、尺度变换、区域裁剪、噪声添加、分段仿射、随机掩盖、边界检测、对比度变换、颜色抖动、随机混合及复合叠加。The computer-readable storage medium according to claim 15, wherein the augmentation strategy set includes rotation transformation, flip transformation, scaling transformation, translation transformation, scale transformation, region cropping, noise addition, segmentation affine, random masking , Boundary detection, contrast transformation, color dithering, random mixing and composite overlay.
PCT/CN2020/111666 2020-02-17 2020-08-27 Method and system for selecting augmentation strategy for image data WO2021164228A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN202010095784.6A CN111275129B (en) 2020-02-17 2020-02-17 Image data augmentation policy selection method and system
CN202010095784.6 2020-02-17

Publications (1)

Publication Number Publication Date
WO2021164228A1 true WO2021164228A1 (en) 2021-08-26

Family

ID=71003628

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2020/111666 WO2021164228A1 (en) 2020-02-17 2020-08-27 Method and system for selecting augmentation strategy for image data

Country Status (2)

Country Link
CN (1) CN111275129B (en)
WO (1) WO2021164228A1 (en)

Cited By (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113642667A (en) * 2021-08-30 2021-11-12 重庆紫光华山智安科技有限公司 Enhancement strategy determination method and device, electronic equipment and storage medium
CN113685972A (en) * 2021-09-07 2021-11-23 广东电网有限责任公司 Air conditioning system control strategy identification method, device, equipment and medium
CN114078218A (en) * 2021-11-24 2022-02-22 南京林业大学 Self-adaptive fusion forest smoke and fire identification data augmentation method
CN114662623A (en) * 2022-05-25 2022-06-24 山东师范大学 XGboost-based blood sample classification method and system in blood coagulation detection
CN114757104A (en) * 2022-04-28 2022-07-15 中国水利水电科学研究院 Construction method of series gate group water transfer engineering hydraulic real-time regulation model based on data driving
CN114942410A (en) * 2022-05-31 2022-08-26 哈尔滨工业大学 Interference signal identification method based on data amplification
CN115426048A (en) * 2022-07-22 2022-12-02 北京大学 Method for detecting augmented space signal, receiving device and optical communication system
CN115600121A (en) * 2022-04-26 2023-01-13 南京天洑软件有限公司(Cn) Data hierarchical classification method and device, electronic equipment and storage medium
CN115935802A (en) * 2022-11-23 2023-04-07 中国人民解放军军事科学院国防科技创新研究院 Electromagnetic scattering boundary element calculation method and device, electronic equipment and storage medium
CN115983369A (en) * 2023-02-03 2023-04-18 电子科技大学 Method for rapidly estimating uncertainty of automatic driving depth visual perception neural network
WO2023155298A1 (en) * 2022-02-21 2023-08-24 平安科技(深圳)有限公司 Data augmentation processing method and apparatus, computer device, and storage medium
WO2024125380A1 (en) * 2022-12-13 2024-06-20 广电运通集团股份有限公司 Classification recognition method, computer device, and storage medium

Families Citing this family (35)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2018176000A1 (en) 2017-03-23 2018-09-27 DeepScale, Inc. Data synthesis for autonomous control systems
US11893393B2 (en) 2017-07-24 2024-02-06 Tesla, Inc. Computational array microprocessor system with hardware arbiter managing memory requests
US10671349B2 (en) 2017-07-24 2020-06-02 Tesla, Inc. Accelerated mathematical engine
US11157441B2 (en) 2017-07-24 2021-10-26 Tesla, Inc. Computational array microprocessor system using non-consecutive data formatting
US11409692B2 (en) 2017-07-24 2022-08-09 Tesla, Inc. Vector computational unit
US11561791B2 (en) 2018-02-01 2023-01-24 Tesla, Inc. Vector computational unit receiving data elements in parallel from a last row of a computational array
US11215999B2 (en) 2018-06-20 2022-01-04 Tesla, Inc. Data pipeline and deep learning system for autonomous driving
US11361457B2 (en) 2018-07-20 2022-06-14 Tesla, Inc. Annotation cross-labeling for autonomous control systems
US11636333B2 (en) 2018-07-26 2023-04-25 Tesla, Inc. Optimizing neural network structures for embedded systems
US11562231B2 (en) 2018-09-03 2023-01-24 Tesla, Inc. Neural networks for embedded devices
WO2020077117A1 (en) 2018-10-11 2020-04-16 Tesla, Inc. Systems and methods for training machine models with augmented data
US11196678B2 (en) 2018-10-25 2021-12-07 Tesla, Inc. QOS manager for system on a chip communications
US11816585B2 (en) 2018-12-03 2023-11-14 Tesla, Inc. Machine learning models operating at different frequencies for autonomous vehicles
US11537811B2 (en) 2018-12-04 2022-12-27 Tesla, Inc. Enhanced object detection for autonomous vehicles based on field view
US11610117B2 (en) 2018-12-27 2023-03-21 Tesla, Inc. System and method for adapting a neural network model on a hardware platform
US10997461B2 (en) 2019-02-01 2021-05-04 Tesla, Inc. Generating ground truth for machine learning from time series elements
US11150664B2 (en) 2019-02-01 2021-10-19 Tesla, Inc. Predicting three-dimensional features for autonomous driving
US11567514B2 (en) 2019-02-11 2023-01-31 Tesla, Inc. Autonomous and user controlled vehicle summon to a target
US10956755B2 (en) 2019-02-19 2021-03-23 Tesla, Inc. Estimating object properties using visual image data
CN111275129B (en) * 2020-02-17 2024-08-20 平安科技(深圳)有限公司 Image data augmentation policy selection method and system
CN111797571B (en) * 2020-07-02 2024-05-28 杭州鲁尔物联科技有限公司 Landslide susceptibility evaluation method, landslide susceptibility evaluation device, landslide susceptibility evaluation equipment and storage medium
CN111815182B (en) * 2020-07-10 2024-06-14 积成电子股份有限公司 Power grid power outage overhaul plan arrangement method based on deep learning
CN113628403A (en) * 2020-07-28 2021-11-09 威海北洋光电信息技术股份公司 Optical fiber vibration sensing perimeter security intrusion behavior recognition algorithm based on multi-core support vector machine
CN111783902B (en) * 2020-07-30 2023-11-07 腾讯科技(深圳)有限公司 Data augmentation, service processing method, device, computer equipment and storage medium
CN111832666B (en) * 2020-09-15 2020-12-25 平安国际智慧城市科技股份有限公司 Medical image data amplification method, device, medium, and electronic apparatus
CN112233194B (en) * 2020-10-15 2023-06-02 平安科技(深圳)有限公司 Medical picture optimization method, device, equipment and computer readable storage medium
CN112381148B (en) * 2020-11-17 2022-06-14 华南理工大学 Semi-supervised image classification method based on random regional interpolation
CN112613543B (en) * 2020-12-15 2023-05-30 重庆紫光华山智安科技有限公司 Enhanced policy verification method, enhanced policy verification device, electronic equipment and storage medium
CN112651458B (en) * 2020-12-31 2024-04-02 深圳云天励飞技术股份有限公司 Classification model training method and device, electronic equipment and storage medium
CN113673501B (en) * 2021-08-23 2023-01-13 广东电网有限责任公司 OCR classification method, system, electronic device and storage medium
CN113869398B (en) * 2021-09-26 2024-06-21 平安科技(深圳)有限公司 Unbalanced text classification method, device, equipment and storage medium
CN114037864A (en) * 2021-10-31 2022-02-11 际络科技(上海)有限公司 Method and device for constructing image classification model, electronic equipment and storage medium
CN114627102B (en) * 2022-03-31 2024-02-13 苏州浪潮智能科技有限公司 Image anomaly detection method, device and system and readable storage medium
CN114693935A (en) * 2022-04-15 2022-07-01 湖南大学 Medical image segmentation method based on automatic data augmentation
CN116416492B (en) * 2023-03-20 2023-12-01 湖南大学 Automatic data augmentation method based on characteristic self-adaption

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20120166379A1 (en) * 2010-12-23 2012-06-28 Yahoo! Inc. Clustering cookies for identifying unique mobile devices
CN106021524A (en) * 2016-05-24 2016-10-12 成都希盟泰克科技发展有限公司 Working method for tree-augmented Navie Bayes classifier used for large data mining based on second-order dependence
CN108959395A (en) * 2018-06-04 2018-12-07 广西大学 A kind of level towards multi-source heterogeneous big data about subtracts combined cleaning method
CN111275129A (en) * 2020-02-17 2020-06-12 平安科技(深圳)有限公司 Method and system for selecting image data augmentation strategy

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR101528235B1 (en) * 2013-11-25 2015-06-12 에스케이텔레콤 주식회사 Method for path-based mobility prediction, and apparatus therefor

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20120166379A1 (en) * 2010-12-23 2012-06-28 Yahoo! Inc. Clustering cookies for identifying unique mobile devices
CN106021524A (en) * 2016-05-24 2016-10-12 成都希盟泰克科技发展有限公司 Working method for tree-augmented Navie Bayes classifier used for large data mining based on second-order dependence
CN108959395A (en) * 2018-06-04 2018-12-07 广西大学 A kind of level towards multi-source heterogeneous big data about subtracts combined cleaning method
CN111275129A (en) * 2020-02-17 2020-06-12 平安科技(深圳)有限公司 Method and system for selecting image data augmentation strategy

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
ANONYMOUS: "DeepAugment: Discover augmentation strategies tailored for your dataset[online]", BARISOZMEN, 19 May 2019 (2019-05-19), XP055838576, Retrieved from the Internet <URL:https://github.com/barisozmen/deepaugment> *

Cited By (18)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113642667A (en) * 2021-08-30 2021-11-12 重庆紫光华山智安科技有限公司 Enhancement strategy determination method and device, electronic equipment and storage medium
CN113642667B (en) * 2021-08-30 2024-02-02 重庆紫光华山智安科技有限公司 Picture enhancement strategy determination method and device, electronic equipment and storage medium
CN113685972A (en) * 2021-09-07 2021-11-23 广东电网有限责任公司 Air conditioning system control strategy identification method, device, equipment and medium
CN114078218A (en) * 2021-11-24 2022-02-22 南京林业大学 Self-adaptive fusion forest smoke and fire identification data augmentation method
CN114078218B (en) * 2021-11-24 2024-03-29 南京林业大学 Adaptive fusion forest smoke and fire identification data augmentation method
WO2023155298A1 (en) * 2022-02-21 2023-08-24 平安科技(深圳)有限公司 Data augmentation processing method and apparatus, computer device, and storage medium
CN115600121A (en) * 2022-04-26 2023-01-13 南京天洑软件有限公司(Cn) Data hierarchical classification method and device, electronic equipment and storage medium
CN115600121B (en) * 2022-04-26 2023-11-07 南京天洑软件有限公司 Data hierarchical classification method and device, electronic equipment and storage medium
CN114757104A (en) * 2022-04-28 2022-07-15 中国水利水电科学研究院 Construction method of series gate group water transfer engineering hydraulic real-time regulation model based on data driving
CN114757104B (en) * 2022-04-28 2022-11-18 中国水利水电科学研究院 Method for constructing hydraulic real-time regulation and control model of series gate group water transfer project
CN114662623A (en) * 2022-05-25 2022-06-24 山东师范大学 XGboost-based blood sample classification method and system in blood coagulation detection
CN114942410B (en) * 2022-05-31 2022-12-20 哈尔滨工业大学 Interference signal identification method based on data amplification
CN114942410A (en) * 2022-05-31 2022-08-26 哈尔滨工业大学 Interference signal identification method based on data amplification
CN115426048A (en) * 2022-07-22 2022-12-02 北京大学 Method for detecting augmented space signal, receiving device and optical communication system
CN115935802A (en) * 2022-11-23 2023-04-07 中国人民解放军军事科学院国防科技创新研究院 Electromagnetic scattering boundary element calculation method and device, electronic equipment and storage medium
CN115935802B (en) * 2022-11-23 2023-08-29 中国人民解放军军事科学院国防科技创新研究院 Electromagnetic scattering boundary element calculation method, device, electronic equipment and storage medium
WO2024125380A1 (en) * 2022-12-13 2024-06-20 广电运通集团股份有限公司 Classification recognition method, computer device, and storage medium
CN115983369A (en) * 2023-02-03 2023-04-18 电子科技大学 Method for rapidly estimating uncertainty of automatic driving depth visual perception neural network

Also Published As

Publication number Publication date
CN111275129A (en) 2020-06-12
CN111275129B (en) 2024-08-20

Similar Documents

Publication Publication Date Title
WO2021164228A1 (en) Method and system for selecting augmentation strategy for image data
CN110163080B (en) Face key point detection method and device, storage medium and electronic equipment
US10290112B2 (en) Planar region guided 3D geometry estimation from a single image
US11481869B2 (en) Cross-domain image translation
WO2019100724A1 (en) Method and device for training multi-label classification model
WO2020199468A1 (en) Image classification method and device, and computer readable storage medium
US10984272B1 (en) Defense against adversarial attacks on neural networks
WO2017148265A1 (en) Word segmentation method and apparatus
WO2017096753A1 (en) Facial key point tracking method, terminal, and nonvolatile computer readable storage medium
WO2019011249A1 (en) Method, apparatus, and device for determining pose of object in image, and storage medium
US20190164312A1 (en) Neural network-based camera calibration
CN109271930B (en) Micro-expression recognition method, device and storage medium
CN111860439A (en) Unmanned aerial vehicle inspection image defect detection method, system and equipment
US20230237771A1 (en) Self-supervised learning method and apparatus for image features, device, and storage medium
CN109413510B (en) Video abstract generation method and device, electronic equipment and computer storage medium
CN110598703B (en) OCR (optical character recognition) method and device based on deep neural network
WO2022127333A1 (en) Training method and apparatus for image segmentation model, image segmentation method and apparatus, and device
JP2014032623A (en) Image processor
WO2021043023A1 (en) Image processing method and device, classifier training method, and readable storage medium
CN114742750A (en) Abnormal cell detection method, abnormal cell detection device, terminal device and readable storage medium
CN113610016A (en) Training method, system, equipment and storage medium of video frame feature extraction model
CN111476226B (en) Text positioning method and device and model training method
CN112348008A (en) Certificate information identification method and device, terminal equipment and storage medium
Wang et al. MetaScleraSeg: an effective meta-learning framework for generalized sclera segmentation
Osuna-Coutiño et al. Structure extraction in urbanized aerial images from a single view using a CNN-based approach

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 20920023

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 20920023

Country of ref document: EP

Kind code of ref document: A1