CN110852352A - Data enhancement method for training deep neural network model for target detection - Google Patents
Data enhancement method for training deep neural network model for target detection Download PDFInfo
- Publication number
- CN110852352A CN110852352A CN201911007178.8A CN201911007178A CN110852352A CN 110852352 A CN110852352 A CN 110852352A CN 201911007178 A CN201911007178 A CN 201911007178A CN 110852352 A CN110852352 A CN 110852352A
- Authority
- CN
- China
- Prior art keywords
- target
- image
- data
- coordinates
- training
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/24—Classification techniques
- G06F18/241—Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V2201/00—Indexing scheme relating to image or video recognition or understanding
- G06V2201/07—Target detection
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- Data Mining & Analysis (AREA)
- Evolutionary Computation (AREA)
- Life Sciences & Earth Sciences (AREA)
- Artificial Intelligence (AREA)
- General Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- General Health & Medical Sciences (AREA)
- Software Systems (AREA)
- Molecular Biology (AREA)
- Computing Systems (AREA)
- Biophysics (AREA)
- Biomedical Technology (AREA)
- Mathematical Physics (AREA)
- Computational Linguistics (AREA)
- Health & Medical Sciences (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Bioinformatics & Computational Biology (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Evolutionary Biology (AREA)
- Image Analysis (AREA)
Abstract
The invention discloses a data enhancement method for training a target detection deep neural network model, which enhances data by adding three elements of three-dimensional change, lens distortion and illumination change into the traditional data enhancement method, so that a data set contains more pictures, and the enhanced new data is used for training the deep neural network to have higher precision in target detection.
Description
Technical Field
The invention belongs to the field of target detection, and particularly relates to a data enhancement method for training a target detection deep neural network model.
Technical Field
Data enhancement is an important means for improving the training effect of the deep neural network, and the essence of the data enhancement is to generate data similar to training data so as to solve the problem of insufficient training data; object detection is the detection of objects of interest from an image or a video segment, such as the detection of pedestrians, vehicles and traffic management signs in the video receiver of an unmanned vehicle; at present, two data enhancement methods for a target detection deep neural network model in the market are provided, namely a traditional enhancement method and a network generation-based method;
the conventional enhancement method is to directly carry out transformations such as rotation, scaling and noise addition on existing data in a training set to generate new data, the transformation process of the data is commonly called affine transformation, besides affine transformation, another common new data generation method is a countermeasure generation network, the image enhancement effects of the affine transformation and the countermeasure generation network are compared and found aiming at the image classification problem, the affine transformation effect is close to that of the countermeasure generation network, but the calculation amount of the affine transformation is much smaller;
however, in the prior art, the widely adopted traditional data enhancement method only adopts two-dimensional changes of images/videos to obtain new data, the obtained new data has errors compared with real data, and the target detection precision is not high by using a neural network trained by the new data.
Disclosure of Invention
In order to overcome the defects of the prior art, the invention aims to provide a data enhancement method for training a target detection deep neural network model, which can enable the generated new data to be closer to real data by adding three elements of three-dimensional change, lens distortion and illumination change in the traditional enhancement method, and has higher target detection precision by using the neural network trained by the new data.
In order to achieve the purpose, the technical scheme of the invention is realized as follows:
a data enhancement method for training a target detection deep neural network model comprises the following specific calculation steps:
the method comprises the following steps: reading images in the training set, and performing normalization processing on read image pixels;
step two: extracting all marked targets from the existing training data in the image read in the step one, namely, manually marking partial data in the training data and intercepting all marked data;
step three: randomly zooming the extracted target, namely generating a random number positioned in an interval [ a, b ] by using a computer as a zooming multiple, zooming the extracted target by utilizing a cubic spline interpolation method, and generating a zoomed target;
step four: performing three-dimensional random rotation on the zoomed target, namely generating three located intervals by using a computerRandom number of (theta)x,θyAnd thetazThe rotation angles of the target around the three axes X, Y and Z are respectively represented, and if the original coordinates of the point on the target in the three-dimensional space are (X, Y and Z) and the coordinates after rotation around the coordinate axes are (X ', Y' and Z '), then (X, Y, Z) and (X', Y ', Z') satisfy the following relations:
step five: mapping the three-dimensionally rotated target to an XY plane where the image is located, namely assigning a pixel value at the coordinate of the original target (x, y) to a pixel point of a new target round (x ', y');
step six: simulating illumination change by adopting a simplified model, and processing the mapped target to obtain new data added with the illumination change;
step seven: adding lens distortion, wherein the specific distortion process is to set the target coordinates after the fifth processing as (x, y) and the coordinates after the radial distortion as (x ', y'), and then tangentially distortion is carried out on the coordinates (x ', y') to obtain (x ', y') so as to generate a final new target;
step eight: and randomly putting the generated new target into the background of the images in the training set.
Preferably, the reading of the images in the training set and the normalization of the images in the step one specifically include:
s1, reading an image in a training set, wherein the pixel value range of the image is [0,225], the image has three channels of R, Y and B, and the image is regarded as a three-dimensional matrix of W multiplied by H multiplied by 3;
s2, carrying out normalization processing on the image pixel value, namely dividing the image pixel value by 255 to change the value range of the pixel value into [0,1 ];
wherein: w is the image width and H is the image height.
3. The method for enhancing data of target detection deep neural network model training as claimed in claim 1, wherein the specific process of transforming the three-dimensional original coordinates of the points on the target in step four from (x, y, z) around the coordinate axis to (x ', y ', z ') is as follows: and (3) rotating the point (X, Y, Z) on the target around the X axis, then rotating around the Y axis, and finally rotating around the Z axis to obtain a new coordinate point (X ', Y ', Z ').
Preferably, the specific processing procedure of adding the new data obtained by change in the step six is as follows:
s1, randomly generating four numbers mu by a computer1,μ2∈(aμ,bμ),σ1,σ2∈(aσ,bσ) Respectively as the mean and standard deviation of the two-dimensional normal distribution probability density function, wherein the two-dimensional normal distribution probability density function is as follows:
s2, randomly generating four numbers x by a computermin,xmax∈(0,w'),ymin,ymax∈(0,h');
Wherein: w 'and h' are the maximum width and the maximum height of the newly generated target matrix;
s3, dividing the interval [ x ]min,xmax]And [ ymin,ymax]Discretizing into a set { xmin,xmin+1,xmin+2,···,xmaxAnd { y }min,ymin+1,ymin+2,···,ymaxIs given by [ x ]min,xmax]Total M elements, [ y ]min,ymax]A total of N elements;
s4, sampling two-dimensional normal distribution probability density functions on the two sets to obtain an MXN two-dimensional matrix S; s m th row and n th column of f (x)min+m,ymin+n),
Wherein: f is a two-dimensional normal distribution probability density function;
s5, the newly generated target is also an image and is provided with three channels of R, G and B, the pixel value of each channel is a matrix, and the dot product of the matrix formed by the pixel values of each channel and the S is respectively calculated, namely the multiplication of the corresponding position elements of the matrix is carried out, so that the new target added with the illumination change can be obtained.
Preferably, the relationship between the target coordinates (x, y) and the coordinates (x ', y') after radial distortion in step seven satisfies:
x'=x(1+k1r2+k2r4+k3r6)
y'=y(1+k1r2+k2r4+k3r6)
wherein: r is the distance from a point to the center of the target, k1,k2And k3Is a preset parameter and can be adjusted according to the equipment for acquiring the image.
Preferably, the relationship between the coordinates (x ', y') of the radial distortion and the coordinates (x ", y") of the tangential distortion satisfies:
x”=x'+(2p1y'+p2(r2+2x'2))
y”=y'+(2p1x'+p1(r2+2y'2))
wherein: p is a radical of1、p2Is a preset parameter and can be adjusted according to the equipment for acquiring the image.
The invention has the beneficial effects that: the invention provides a data enhancement method for training a target detection deep neural network model, which is characterized in that three factors of three-dimensional change, lens distortion and illumination change are added in the traditional enhancement method to enhance data, so that a data set contains more pictures, and the enhanced new data is used for training the deep neural network to have higher precision in target detection.
Detailed Description
In order to make those skilled in the art better understand the technical solution of the present invention, the following further describes the technical solution of the present invention with reference to the embodiments.
A data enhancement method for training a target detection deep neural network model comprises the following specific calculation steps:
the method comprises the following steps: reading images in a training set, and carrying out normalization processing on extracted image pixels, wherein the specific steps are as follows;
s1, reading an image in a training set, wherein the pixel value range of the image is [0.255], the image has three channels of R, G and B, and the image is regarded as a three-dimensional matrix of W multiplied by H multiplied by 3;
wherein: w is the image width, H is the image height;
s2, carrying out normalization processing on the pixel values, namely dividing the pixel values by 255 to change the value range of the pixel values into [0,1 ].
Step two: extracting all marked targets from the existing training data of the image, namely, manually marking partial data in the training data and intercepting all marked data; the manual marking and the extraction method are the optimal extraction method of the invention, the implementation is simple, and the marking principle is as follows: if the manual marking is that the coordinates of the upper left corner of the area are (100, 200) and the coordinates of the lower right corner are (300, 400), then the rectangular areas of the two points contain the marking target; besides the method, other marking methods can be used, such as marking the target by using a matrix with only 0 and 1 values, wherein the position with 1 represents the target, and the position with 0 represents the background.
Step three: randomly zooming the extracted target, namely generating a random number positioned in an interval [ a, b ] by using a computer as a zooming multiple, zooming the extracted target by utilizing a cubic spline interpolation method, and generating a zoomed target; wherein [ a, b ] is the value range of the random number, which is adjusted according to the characteristics of the data set itself, and is not a fixed value.
Step four: performing three-dimensional random rotation on the zoomed target, namely generating three located intervals by using a computerRandom number of (theta)x,θyAnd thetazThe rotation angles of the target around the three axes of X, Y and Z are respectively represented, and the original coordinates of the point on the target in the three-dimensional space are (X, Y and Z), and the coordinates after the rotation around the coordinate axes are (X ', Y ' and Z '), then(x, y, z) and (x ', y ', z ') satisfy the following relationship:
step five: and mapping the target after three-dimensional rotation to an XY plane where the image is located, namely, assigning a pixel value at the coordinate of the original target (x, y) to a pixel point of the new target round (x ', y'), wherein for example, the pixel value at the original target (142,173) is (0.5,0.1,0.4), the pixel value at the new target (100.200) is (0.5,0.1,0.4) after the three-dimensional rotation and the mapping to the XY plane, and the point (142,173) is transformed to (100.37,200.5).
Wherein: round means rounding;
step six: simulating illumination change by using a simplified model, processing the mapped target to obtain new data added with the illumination change, and specifically comprising the following steps of:
s1, randomly generating four numbers mu by a computer1,μ2∈(aμ,bμ),σ1,σ2∈(aσ,bσ) Respectively as the mean and standard deviation of the two-dimensional normal distribution probability density function, wherein the two-dimensional normal distribution probability density function is as follows:
s2, randomly generating four numbers x by a computermin,xmax∈(0,w'),ymin,ymax∈(0,h');
Wherein: w 'and h' are the maximum width and the maximum height of the newly generated target matrix;
s3, dividing the interval [ x ]min,xmax]And [ ymin,ymax]Discretizing into a set { xmin,xmin+1,xmin+2,···,xmaxAnd { y }min,ymin+1,ymin+2,···,ymaxIs given by [ x ]min,xmax]Total M elements, [ y ]min,ymax]A total of N elements;
s4, sampling two-dimensional normal distribution probability density functions on the two sets to obtain an MXN two-dimensional matrix S; s m th row and n th column of f (x)min+m,ymin+n),
Wherein: f is a two-dimensional normal distribution probability density function;
s5, a newly generated target is also an image and is provided with three channels of R, G and B, the pixel value of each channel is a matrix, the dot product of the matrix formed by the pixel values of each channel and S is respectively calculated, namely the pixel value of the corresponding position of the matrix is multiplied, so that a new target matrix S ' w ' h ' with illumination change can be obtained, because the target is essentially an area in the image, the target is also provided with three R, G, B channels, the target can be represented into a matrix with width x height 3 no matter whether the target is a new target or an untransformed target, w ' is the width of the new target, and h ' is the height of the new target;
the S-M-N is a sample of a normal distribution function, and since the normal distribution function is continuous, only a discrete interval is available, for example, a function in the interval of (-1,1) is sampled into 20 points, and then function values at 20 points of-1, -0.9, -0.8,. 9, 0.8, 0.9, 1 are retained; the same continuous interval can sample different numbers of function values, such as-1, -0.8, -0.6, …,0.6,0.8,1, so that only 10 points are sampled;
carrying out different sampling on the continuous two-dimensional normal distribution function to generate matrixes with different sizes; the generated matrix is dot-product with each channel of the new target, i.e. the two matrices are multiplied by corresponding position elements, so that the matrix obtained by sampling the normal distribution should be the same as the maximum width and the maximum height of the new target.
Step seven: adding lens distortion:
s1, radial distortion, namely distortion distributed along the boundary direction of the lens, wherein the target coordinates after the first step to the fifth step are (x, y), and the coordinates after the radial distortion are (x ', y'), which satisfy the following relations:
x'=x(1+k1r2+k2r4+k3r6)
y'=y(1+k1r2+k2r4+k3r6)
wherein: r is the distance from a point to the center of the target, k1,k2And k3The parameters are preset parameters and can be adjusted according to equipment for acquiring images;
s2, tangential distortion, namely, setting the coordinates subjected to radial distortion as (x ', y '), and the coordinates subjected to tangential distortion as (x ', y "), wherein the coordinates satisfy the following relation:
x”=x'+(2p1y'+p2(r2+2x'2))
y”=y'+(2p1x'+p1(r2+2y'2))
wherein: p is a radical of1、p2Is a preset parameter and can be adjusted according to the equipment for acquiring the image.
Finally, the pixel value of the target coordinate (x, y) mapped in the step five is assigned to (x', y ") to generate a final new target;
step eight: and randomly putting the generated new target into the background of the images in the training set.
The first embodiment is as follows: in practice, due to the fact that the data volumes of data sets of different classes are different, a data enhancement technology (data augmentation technology) is used in training, we simply ignore the classes with the instances smaller than 100, the remaining 45 classes can be used for classification, the classes with the instances between 100 and 1000 are enhanced to 1000 by the method, and the rest is more than 1000;
then, a traffic sign detection algorithm based on a deep neural network is trained by using a newly generated data set, the training effect is evaluated by using the average accuracy, and the accuracy of the traffic sign detection algorithm is compared with the accuracy of a neural network trained by using a traditional data enhancement method, as shown in table 1:
TABLE 1 average accuracy comparison (%)
As can be seen from the table I, the average accuracy of the training data enhanced by the method is generally higher than that of the neural network trained by the traditional data enhancement method.
The foregoing shows and describes the general principles, essential features, and advantages of the invention. It will be understood by those skilled in the art that the present invention is not limited to the embodiments described above, which are described in the specification and illustrated only to illustrate the principle of the present invention, but that various changes and modifications may be made therein without departing from the spirit and scope of the present invention, which fall within the scope of the invention as claimed. The scope of the invention is defined by the appended claims and equivalents thereof.
Claims (6)
1. A data enhancement method for training a target detection deep neural network model is characterized by comprising the following specific calculation steps:
the method comprises the following steps: reading images in the training set, and performing normalization processing on read image pixels;
step two: extracting all marked targets from the existing training data in the image read in the step one, namely, manually marking partial data in the training data and intercepting all marked data;
step three: randomly zooming the extracted target, namely generating a random number positioned in an interval [ a, b ] by using a computer as a zooming multiple, zooming the extracted target by utilizing a cubic spline interpolation method, and generating a zoomed target;
step four: performing three-dimensional random rotation on the zoomed target, namely generating three located intervals by using a computerRandom number of (theta)x,θyAnd thetazRespectively representing the rotation angles of the target around the three axes of X, Y and Z, and setting the eyesThe original coordinates of the target point in the three-dimensional space are (x, y, z), and the coordinates after rotating around the coordinate axis are (x ', y', z '), then (x, y, z) and (x', y ', z') satisfy the following relations:
step five: mapping the three-dimensionally rotated target to an XY plane where the image is located, namely assigning a pixel value at the coordinate of the original target (x, y) to a pixel point of a new target round (x ', y');
step six: simulating illumination change by adopting a simplified model, and processing the mapped target to obtain new data added with the illumination change;
step seven: adding lens distortion, wherein the specific distortion process is to set the target coordinates after the fifth processing as (x, y) and the coordinates after the radial distortion as (x ', y'), and then tangentially distortion is carried out on the coordinates (x ', y') to obtain (x ', y') so as to generate a final new target;
step eight: and randomly putting the generated new target into the background of the images in the training set.
2. The method as claimed in claim 1, wherein the step one of reading the images in the training set and normalizing the images comprises the specific steps of:
s1, reading an image in a training set, wherein the pixel value range of the image is [0,225], the image has three channels of R, Y and B, and the image is regarded as a three-dimensional matrix of W multiplied by H multiplied by 3;
s2, carrying out normalization processing on the image pixel value, namely dividing the image pixel value by 255 to change the value range of the pixel value into [0,1 ];
wherein: w is the image width and H is the image height.
3. The method for enhancing data of target detection deep neural network model training as claimed in claim 1, wherein the specific process of transforming the three-dimensional original coordinates of the points on the target in step four from (x, y, z) around the coordinate axis to (x ', y ', z ') is as follows: and (3) rotating the point (X, Y, Z) on the target around the X axis, then rotating around the Y axis, and finally rotating around the Z axis to obtain a new coordinate point (X ', Y ', Z ').
4. The method for enhancing data of training the deep neural network model for target detection according to claim 1, wherein the specific process of adding new data obtained by change in the sixth step is as follows:
s1, randomly generating four numbers mu by a computer1,μ2∈(aμ,bμ),σ1,σ2∈(aσ,bσ) Respectively as the mean and standard deviation of the two-dimensional normal distribution probability density function, wherein the two-dimensional normal distribution probability density function is as follows:
s2, randomly generating four numbers x by a computermin,xmax∈(0,w'),ymin,ymax∈(0,h');
Wherein: w 'and h' are the maximum width and the maximum height of the newly generated target matrix respectively;
s3, dividing the interval [ x ]min,xmax]And [ ymin,ymax]Discretizing into a set { xmin,xmin+1,xmin+2,···,xmaxAnd { y }min,ymin+1,ymin+2,···,ymaxIs given by [ x ]min,xmax]Total M elements, [ y ]min,ymax]A total of N elements;
s4, sampling two-dimensional normal distribution probability density functions on the two sets to obtain an MXN two-dimensional matrix S; s m th row and n th column of f (x)min+m,ymin+n),
Wherein: f is a two-dimensional normal distribution probability density function;
s5, the newly generated target is also an image and is provided with three channels of R, G and B, the pixel value of each channel is a matrix, and the dot product of the matrix formed by the pixel values of each channel and the S is respectively calculated, namely the multiplication of the corresponding position elements of the matrix is carried out, so that the new target added with the illumination change can be obtained.
5. The method of claim 1, wherein the relationship between the target coordinates (x, y) and the radially distorted coordinates (x ', y') in step seven satisfies the following relationship:
x'=x(1+k1r2+k2r4+k3r6)
y'=y(1+k1r2+k2r4+k3r6)
wherein: r is the distance from a point to the center of the target, k1,k2And k3Is a preset parameter and can be adjusted according to the equipment for acquiring the image.
6. The data enhancement method for training the target detection deep neural network model of claim 5, wherein the relationship between the coordinates (x ', y') after the radial distortion and the coordinates (x ", y") of the tangential distortion satisfies:
x”=x'+(2p1y'+p2(r2+2x'2))
y”=y'+(2p1x'+p1(r2+2y'2))
wherein: p is a radical of1、p2Is a preset parameter and can be adjusted according to the equipment for acquiring the image.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201911007178.8A CN110852352B (en) | 2019-10-22 | 2019-10-22 | Data enhancement method for training deep neural network model for target detection |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201911007178.8A CN110852352B (en) | 2019-10-22 | 2019-10-22 | Data enhancement method for training deep neural network model for target detection |
Publications (2)
Publication Number | Publication Date |
---|---|
CN110852352A true CN110852352A (en) | 2020-02-28 |
CN110852352B CN110852352B (en) | 2022-07-29 |
Family
ID=69597191
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201911007178.8A Active CN110852352B (en) | 2019-10-22 | 2019-10-22 | Data enhancement method for training deep neural network model for target detection |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN110852352B (en) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN112183500A (en) * | 2020-11-27 | 2021-01-05 | 北京易真学思教育科技有限公司 | Data enhancement method, training method, computer device and storage medium |
Citations (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20110298897A1 (en) * | 2010-06-08 | 2011-12-08 | Iva Sareen | System and method for 3d virtual try-on of apparel on an avatar |
CN108230233A (en) * | 2017-05-16 | 2018-06-29 | 北京市商汤科技开发有限公司 | Data enhancing, treating method and apparatus, electronic equipment and computer storage media |
CN109190768A (en) * | 2018-08-09 | 2019-01-11 | 北京中关村科金技术有限公司 | A kind of data enhancing corpus training method in neural network |
JP2019032821A (en) * | 2017-06-26 | 2019-02-28 | コニカ ミノルタ ラボラトリー ユー.エス.エー.,インコーポレイテッド | Data augmentation techniques using style transformation with neural network |
CN109543684A (en) * | 2018-10-09 | 2019-03-29 | 广州大学 | Immediate targets tracking detection method and system based on full convolutional neural networks |
CN109801365A (en) * | 2019-01-25 | 2019-05-24 | 上海小萌科技有限公司 | A kind of three-dimensional modeling data acquisition device and its three-dimensional modeling method |
CN110175982A (en) * | 2019-04-16 | 2019-08-27 | 浙江大学城市学院 | A kind of defect inspection method based on target detection |
WO2019169895A1 (en) * | 2018-03-09 | 2019-09-12 | 华南理工大学 | Fast side-face interference resistant face detection method |
CN110348376A (en) * | 2019-07-09 | 2019-10-18 | 华南理工大学 | A kind of pedestrian's real-time detection method neural network based |
-
2019
- 2019-10-22 CN CN201911007178.8A patent/CN110852352B/en active Active
Patent Citations (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20110298897A1 (en) * | 2010-06-08 | 2011-12-08 | Iva Sareen | System and method for 3d virtual try-on of apparel on an avatar |
CN108230233A (en) * | 2017-05-16 | 2018-06-29 | 北京市商汤科技开发有限公司 | Data enhancing, treating method and apparatus, electronic equipment and computer storage media |
JP2019032821A (en) * | 2017-06-26 | 2019-02-28 | コニカ ミノルタ ラボラトリー ユー.エス.エー.,インコーポレイテッド | Data augmentation techniques using style transformation with neural network |
WO2019169895A1 (en) * | 2018-03-09 | 2019-09-12 | 华南理工大学 | Fast side-face interference resistant face detection method |
CN109190768A (en) * | 2018-08-09 | 2019-01-11 | 北京中关村科金技术有限公司 | A kind of data enhancing corpus training method in neural network |
CN109543684A (en) * | 2018-10-09 | 2019-03-29 | 广州大学 | Immediate targets tracking detection method and system based on full convolutional neural networks |
CN109801365A (en) * | 2019-01-25 | 2019-05-24 | 上海小萌科技有限公司 | A kind of three-dimensional modeling data acquisition device and its three-dimensional modeling method |
CN110175982A (en) * | 2019-04-16 | 2019-08-27 | 浙江大学城市学院 | A kind of defect inspection method based on target detection |
CN110348376A (en) * | 2019-07-09 | 2019-10-18 | 华南理工大学 | A kind of pedestrian's real-time detection method neural network based |
Non-Patent Citations (5)
Title |
---|
BARRET ZOPH 等: "Learning Data Augmentation Strategies for Object Detection", 《ARXIV:1906.11172V1》 * |
MAURICE WEILER 等: "3D Steerable CNNs: Learning Rotationally Equivariant Features in Volumetric Data", 《32ND CONFERENCE ON NEURAL INFORMATION PROCESSING SYSTEMS》 * |
YIMING YAN 等: "A Data Augmentation Strategy Based on Simulated Samples for Ship Detection in RGB Remote Sensing Images", 《INTERNATIONAL JOURNAL OF GEO-INFORMATION》 * |
卢海涛 等: "基于Faster R-CNN及数据增广的满文文档印章检测", 《大连民族大学学报》 * |
张中宝 等: "基于Faster-RCNN的遥感图像飞机检测算法", 《南京师大学报(自然科学版)》 * |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN112183500A (en) * | 2020-11-27 | 2021-01-05 | 北京易真学思教育科技有限公司 | Data enhancement method, training method, computer device and storage medium |
Also Published As
Publication number | Publication date |
---|---|
CN110852352B (en) | 2022-07-29 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN109784333B (en) | Three-dimensional target detection method and system based on point cloud weighted channel characteristics | |
CN109886896B (en) | Blue license plate segmentation and correction method | |
CN105825203B (en) | Based on point to matching and the matched ground arrow mark detection of geometry and recognition methods | |
CN110516550B (en) | FPGA-based lane line real-time detection method | |
CN111223053A (en) | Data enhancement method based on depth image | |
CN110807807B (en) | Monocular vision target positioning pattern, method, device and equipment | |
CN113592923B (en) | Batch image registration method based on depth local feature matching | |
CN106408597A (en) | Neighborhood entropy and consistency detection-based SAR (synthetic aperture radar) image registration method | |
CN113034593B (en) | 6D pose labeling method, system and storage medium | |
CN109472770B (en) | Method for quickly matching image characteristic points in printed circuit board detection | |
CN108416801B (en) | Har-SURF-RAN characteristic point matching method for stereoscopic vision three-dimensional reconstruction | |
CN109919832B (en) | Traffic image splicing method for unmanned driving | |
CN111382658B (en) | Road traffic sign detection method in natural environment based on image gray gradient consistency | |
CN110930411A (en) | Human body segmentation method and system based on depth camera | |
CN107169962A (en) | The gray level image fast partition method of Kernel fuzzy clustering is constrained based on space density | |
CN112686247A (en) | Identification card number detection method and device, readable storage medium and terminal | |
CN113627478A (en) | Target detection method, target detection device and robot | |
CN114913438B (en) | YOLOv5 garden abnormal target identification method based on anchor frame optimal clustering | |
CN110852352B (en) | Data enhancement method for training deep neural network model for target detection | |
CN112101283A (en) | Intelligent identification method and system for traffic signs | |
CN109635809B (en) | Super-pixel segmentation method for visual degradation image | |
CN111797832B (en) | Automatic generation method and system for image region of interest and image processing method | |
CN112052859A (en) | License plate accurate positioning method and device in free scene | |
CN110276260B (en) | Commodity detection method based on depth camera | |
CN116206156A (en) | Pavement crack classification and identification method under shadow interference |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |