CN112633392A - Terahertz human body security inspection image target detection model training data augmentation method - Google Patents

Terahertz human body security inspection image target detection model training data augmentation method Download PDF

Info

Publication number
CN112633392A
CN112633392A CN202011589752.8A CN202011589752A CN112633392A CN 112633392 A CN112633392 A CN 112633392A CN 202011589752 A CN202011589752 A CN 202011589752A CN 112633392 A CN112633392 A CN 112633392A
Authority
CN
China
Prior art keywords
human body
image
data
suspicious object
augmentation
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202011589752.8A
Other languages
Chinese (zh)
Other versions
CN112633392B (en
Inventor
李�诚
柳桃荣
余开
涂昊
刘泽鑫
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Brainware Terahertz Information Technology Co ltd
Original Assignee
Brainware Terahertz Information Technology Co ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Brainware Terahertz Information Technology Co ltd filed Critical Brainware Terahertz Information Technology Co ltd
Priority to CN202011589752.8A priority Critical patent/CN112633392B/en
Publication of CN112633392A publication Critical patent/CN112633392A/en
Application granted granted Critical
Publication of CN112633392B publication Critical patent/CN112633392B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • G06F18/2415Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches based on parametric or probabilistic models, e.g. based on likelihood ratio or false acceptance rate versus a false rejection rate
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/194Segmentation; Edge detection involving foreground-background segmentation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V2201/00Indexing scheme relating to image or video recognition or understanding
    • G06V2201/07Target detection
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02TCLIMATE CHANGE MITIGATION TECHNOLOGIES RELATED TO TRANSPORTATION
    • Y02T10/00Road transport of goods or passengers
    • Y02T10/10Internal combustion engine [ICE] based vehicles
    • Y02T10/40Engine management systems

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • General Physics & Mathematics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Artificial Intelligence (AREA)
  • Evolutionary Biology (AREA)
  • Evolutionary Computation (AREA)
  • General Engineering & Computer Science (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Probability & Statistics with Applications (AREA)
  • Image Analysis (AREA)

Abstract

The invention discloses a terahertz human body security inspection image target detection model training data augmentation method, which belongs to the technical field of image processing. In the training process, randomly generating an area in a human body foreground area and randomly selecting an image from the suspected object image augmentation subset to cover the image to form augmented training data. According to the invention, suspicious object image augmentation subsets are cut from a small amount of existing images with suspicious object labels, a human body foreground region is extracted by using a segmentation algorithm or a manual marking mode, a coverage region is randomly generated in a training process, the suspicious object augmentation images are randomly selected for coverage, and a data label is regenerated, so that the target detection model training convergence process can be accelerated under the condition of unbalanced samples, and the model detection accuracy is improved.

Description

Terahertz human body security inspection image target detection model training data augmentation method
Technical Field
The invention relates to the technical field of terahertz human body security inspection image processing, in particular to a terahertz human body security inspection image target detection model training data augmentation method.
Background
Target detection can classify and locate interesting targets from images, and is a common task in computer vision, and deep learning is a solution mainly adopted at present.
The terahertz human body security inspection image suspicious object target detection belongs to the field of specific application, and the task is to find and position suspicious objects hidden under human body clothes through the human body security inspection image, but problems of rare suspicious object samples, high data acquisition cost, extremely unbalanced overall sample distribution and the like exist in actual model training, so that the target detection model training effect is poor and the detection accuracy is low. The problems need to be solved urgently, and therefore a terahertz human body security inspection image target detection model training data augmentation method is provided.
Disclosure of Invention
The technical problem to be solved by the invention is as follows: the method can complete the directional data amplification according to the number ratio of suspicious object labels of various categories, and improve the detection accuracy of the target detection model.
The invention solves the technical problems through the following technical scheme, and the invention comprises the following steps:
s1: preparation phase
S11: cutting an augmented data subset from the target detection data set by using a suspicious object labeling frame, counting the number and the proportion of each type of suspicious object labels, and calculating the percentage of augmented hits;
s12: marking a human body foreground region of a training sample in a training data set by using manual marking or a semantic segmentation model, and making segmentation marking data;
s2: training phase
S21: loading single training image data, target detection labeling data and segmented human body foreground semantic label data, and setting an augmentation counter to be 1;
s22: marking by using a human body frame, calculating the coordinates of a random starting point of a covered area, adding 1 to an augmentation counter, giving up augmentation if the number of times of marking reaches the upper limit, and directly sending image data and target detection marking data to network training; s23: randomly selecting an augmentation coverage picture from the augmentation data subset according to the number ratio of the suspicious object type labels;
s24: calculating pixel proportion mu of the covered area and the human body foreground area, and intersection ratio sigma of the covered area and all suspicious object target frames, if mu and sigma meet set qualified conditions, turning to step S25, and if mu and sigma do not meet set qualified conditions, returning to step S22;
s25: and covering the original image by using the augmented subset picture, modifying the labeled data of the training picture, and sending the training picture to the network for training.
Further, in the step S11, a label (x1, y1, x2, y2, c) exists for each image, where x1 and y1 are coordinates of a point at the upper left corner of the target frame, (x2 and y2) are coordinates of a point at the lower right corner of the target frame, and c is a target category in the frame.
Further, in the step S11, the cropping mode is to extract only the image on the original image using the suspicious object labeling box [ y 1: y2, x 1: x2]In-range image data, i.e. only regions of articles of the suspect category in the image are cropped; and counting the total number of the suspicious object target frame labels and the number of the suspicious object labels of each category, and calculating the ratio of each category in the total number:
Figure BDA0002868273320000021
wherein C isiIs the number of suspect labels of category i; calculating corresponding reciprocal
Figure BDA0002868273320000022
Statistical reciprocal ratio
Figure BDA0002868273320000023
Wherein gamma isiIs the probability of random amplification of each type of suspicious objectThe probability of the expansion of the suspicious object with a small number of tags is high, and the probability of the expansion of the suspicious object with a large number of tags is low.
Further, in the step S12, it is necessary to acquire a human body segmentation foreground region for each image containing a human body frame, and generate corresponding segmentation label data, where the foreground pixel in the grayscale label image is 255 and the background pixel is 0.
Further, in step S21, the image data refers to image matrix data, the target detection annotation data refers to data of a human body frame and a possible target frame in the image, the segmented human body foreground semantic label data refers to an image label with a width and a height consistent with the image matrix, and an augmentation counter is set to 1 before each image is augmented.
Further, in the step S22, a human body box is marked (x)p1,yp1,xp2,yp2) The calculation of the random initial point coordinates of the covered area means that two random numbers r in the (0,1) interval are taken by utilizing uniform distributionx,ry(ii) a And then calculating the coordinates of the starting point:
xs=xp1+rx(xp2-xp1)
ys=yp1+ry(yp2-yp1)。
further, in the step S23, a specific process of selecting the augmented coverage picture is as follows:
s231: selecting an augmented picture from the augmented data subset according to the number ratio of the suspicious object class labels, and calculating a random hit interval under the suspicious object class i percentage;
s232: then generating a random number with the interval of (0,100) to determine the suspicious object class needing to be augmented;
s233: and selecting a specific single picture from the single-class suspicious object augmentation subset by using uniform distribution and randomness.
Further, in the step S24, let the width and height of the augmented picture be w, h, and the coordinates of the covered area frame be (x)s,ys,xs+ w,ys+ h) the area of the covered region frame pixel isAnd when S is wh, counting the total pixel sum S of the human foreground area in the framefThe proportion calculation mode is as follows:
Figure BDA0002868273320000024
further, in the step S24, the pass conditions for μ and σ are:
Figure BDA0002868273320000025
further, in the step S25, modifying the training image annotation data refers to adding the covered area coordinates and the category on the basis of the original label.
Compared with the prior art, the invention has the following advantages: according to the method for amplifying the training data of the terahertz human body security inspection image target detection model, suspicious object regions are cut out from a small amount of existing images with suspicious object labels to form suspicious object image amplification subsets, the terahertz image human body foreground regions are extracted by using a segmentation algorithm or a manual labeling mode, coverage regions are randomly generated and are randomly selected to be covered by the suspicious object amplification images in the training process, and the data labels are regenerated.
Drawings
FIG. 1 is a picture of an actual security inspection in a second embodiment of the present invention;
FIG. 2 is a schematic diagram of a data amplification process according to a second embodiment of the present invention; FIG. 3 is a flowchart illustrating a process of obtaining a data augmented subset according to a second embodiment of the present invention;
FIG. 3 is a diagram illustrating the data amplification effect according to a second embodiment of the present invention.
FIG. 4 is a diagram illustrating the data amplification effect according to a second embodiment of the present invention.
Detailed Description
The following examples are given for the detailed implementation and specific operation of the present invention, but the scope of the present invention is not limited to the following examples.
Example one
The embodiment provides a technical scheme: a terahertz human body security inspection image target detection model training data augmentation method comprises the following steps:
s1: preparation phase
S11: cutting an augmented data subset from the target detection data set by using a suspicious object labeling frame, counting the number and the proportion of each type of suspicious object labels, and calculating the percentage of augmented hits;
s12: marking a human body foreground region of a training sample in a training data set by using manual marking or a semantic segmentation model, and making segmentation marking data;
s2: training phase
S21: loading single training image data, target detection labeling data and segmented human body foreground semantic label data, and setting an augmentation counter to be 1;
s22: marking by using a human body frame, calculating the coordinates of a random starting point of a covered area, adding 1 to an augmentation counter, giving up augmentation if the number of times of marking reaches the upper limit, and directly sending image data and target detection marking data to network training; s23: randomly selecting an augmentation coverage picture from the augmentation data subset according to the number ratio of the suspicious object type labels;
s24: calculating pixel proportion mu of the covered area and the human body foreground area, and intersection ratio sigma of the covered area and all suspicious object target frames, if mu and sigma meet set qualified conditions, turning to step S25, and if mu and sigma do not meet set qualified conditions, returning to step S22;
s25: and covering the original image by using the augmented subset picture, modifying the labeled data of the training picture, and sending the training picture to the network for training.
In the step S11, there is a label (x1, y1, x2, y2, c) for each image, where x1 and y1 are coordinates of the upper left corner point of the target frame, (x2, y2) are coordinates of the lower right corner point of the target frame, and c is the target category in the frame.
In the step S11, the cropping mode is to extract only the image on the original image using the suspicious object labeling box [ y 1: y2, x 1: x2]In-range image data, i.e. only cropping picturesArea of the item in the image in the suspect category; and counting the total number of the suspicious object target frame labels and the number of the suspicious object labels of each category, and calculating the ratio of each category in the total number:
Figure BDA0002868273320000041
wherein C isiIs the number of suspect labels of category i; calculating corresponding reciprocal
Figure BDA0002868273320000042
Statistical reciprocal ratio
Figure BDA0002868273320000043
Wherein gamma isiThe probability of random amplification of each type of suspicious object is high, the probability of amplification of the suspicious object with a small number of labels is high, and the probability of amplification of the suspicious object with a large number of labels is low.
In step S12, it is necessary to acquire a human body segmentation foreground region for each image including a human body frame, and generate corresponding segmentation label data, where the foreground pixel in the grayscale label image is 255 and the background pixel is 0.
In step S21, the image data refers to image matrix data, the target detection annotation data refers to data of a human body frame and a target frame that may exist in the image, the segmented human body foreground semantic label data refers to an image label with a width and a height consistent with the image matrix, and an augmentation counter is set to 1 before each image is augmented.
In step S22, the number upper limit of the augmentation counter is set, and after the number upper limit is reached and no qualified augmented picture can be generated, the loop is exited and training is performed directly using the original image data and the target frame label.
In step S22, a human body frame is labeled (x)p1,yp1,xp2,yp2) The calculation of the random initial point coordinates of the covered area means that two random numbers r in the (0,1) interval are taken by utilizing uniform distributionx,ry(repeatable); and then calculating the coordinates of the starting point:
xs=xp1+rx(xp2-xp1)
ys=yp1+ry(yp2-yp1)。
in step S23, the specific process of selecting the augmented overlay picture is as follows:
s231: selecting an augmented picture from the augmented data subset according to the number ratio of the suspicious object class labels, and calculating a random hit interval under the suspicious object class i percentage;
s232: then generating a random number with the interval of (0,100) to determine the suspicious object class needing to be augmented;
s233: and selecting a specific single picture from the single-class suspicious object augmentation subset by using uniform distribution and randomness.
In step S24, the width and height of the enlarged picture are w, h, and the coordinates of the covered area frame are (x)s,ys,xs+w,ys+ h), the area of the covered area frame pixel is S ═ wh, the total S of the human foreground area pixels in the frame is countedfThe proportion calculation mode is as follows:
Figure BDA0002868273320000044
in step S24, it is necessary to calculate the intersection ratio of the covered area and all other suspicious object labeling boxes, and take the maximum value, if there is no suspicious object label in the current image, σ is set to 0.
In step S24, the pass conditions for μ and σ are:
Figure BDA0002868273320000051
in step S25, modifying the training image annotation data refers to adding the covered area coordinates and the category based on the original label.
Example two
As shown in fig. 1, most of image samples collected by terahertz human body security inspection equipment are normal samples which do not carry suspicious articles, only a very small number of human bodies carry suspicious articles, and specific sensitive articles such as knives are rare, so that the samples required by a target detection model are very unbalanced due to scarcity of label data of the suspicious articles, and the directly trained target detection model has low detection accuracy and cannot accurately identify the suspicious articles.
As shown in fig. 2, the data augmentation method provided by the present invention can complete random augmentation and improve the detection accuracy of the deep learning model under the condition of only a small number of specific suspicious object samples.
The specific process of the data augmentation method in this embodiment is as follows:
step S1.1: preparing training data, cutting out an augmentation subset from an original image by using a suspicious object marking frame, counting the number of each type of suspicious object, calculating the ratio of each type of suspicious object, and calculating the percentage of augmentation hit of each type of suspicious object, as shown in FIG. 3
For example, the current object detection model needs to detect suspicious objects of the following categories: rectangular articles (rectangle), pistol (gun), knife three types (knife), cutting out corresponding article area images by using the target frame marking data of the suspicious article, and renaming as rectangle _1.jpg, rectangle _2.jpg …, gun _1.jpg, gun _2.jpg …, knife _1.jpg, knife _2.jpg, ….
Assuming that 10000 security check images of human body exist in the training data set, wherein the number of rectangle frames is 400, gun is 80, and knife is 20, the ratio of the number of each category is 0.8(400/500), 0.16(80/500), 0.04(20/500), the reciprocal is 1.25(1/0.8), 6.25(1/0.16), 25(1/0.04), the percentage of the reciprocal, namely the percentage of augmented hits, is 3.9%, 19.2%, 76.9%, and the random selection of pictures from the augmented subset is completed at the later stage according to the percentage of augmented hits.
Step S1.2: preparing training data, and extracting human body foreground regions in all training samples by using an image semantic segmentation model or a manual labeling mode;
step S2.1: in the training stage, as shown in fig. 4, loading single training image data, target detection labeling data and segmented human body foreground semantic label data, and setting an augmentation counter to be 1;
step S2.2: training phase, labeling (x) with human body box (person)p1,yp1,xp2,yp2) Calculating the coordinate of a random initial point of the covered area, and adding 1 to an augmentation counter; continuously utilizing the human body frame for marking, calculating the random initial point coordinate of the covered area, adding 1 to an augmentation counter, giving up augmentation if the number of times reaches the upper limit, and directly sending the image data and the target detection marking data into a network for training;
generating the random starting point may be generated as follows:
1) two random numbers r in the interval of (0,1) are taken by utilizing uniform distributionx,ry
2) Calculating the coordinates of the starting point:
xs=xp1+rx(xp2-xp1)
ys=yp1+rx(yp2-yp1)
s2.3, training: randomly selecting an augmentation coverage picture from the augmentation data subset according to the number ratio of the suspicious object type labels;
the specific random selection method is as follows:
1) calculating random augmented hit intervals by using the augmented hit percentage obtained in the step S1.1, for example, the augmented hit percentages of rectangle, gun and knife are 3.9%, 19.2% and 76.9%, and then setting the intervals as the following corresponding intervals [0,4 ], [4,23 ], [23,100] after rounding;
2) random numbers are uniformly distributed in a [0,100] interval and fall into the interval, namely, the augmentation data are taken from the corresponding class augmentation pictures, and assuming that the random number is 21, the augmentation pictures are selected from all the augmentation subset pictures with the class gun;
3) when the sub-set pictures are amplified from a single category (such as category gun), a certain specific amplified picture can be randomly selected in a uniform distribution mode, and each picture can be ensured to be randomly accessed;
s2.4 training stage: calculating the pixel proportion of the covered area and the human body foreground area, comparing the covered area with the suspicious object target frame, if the comparison meets the requirement, turning to the step S2.5, and if the comparison does not meet the requirement, returning to the step S2.2;
the pixel intersection ratio calculation method is as follows: quiltThe coordinates of the coverage area frame are (x)n1,yn1,xn2,yn2) And the area of the covered area frame pixels is as follows: sn=(xn2-xn1)(yn2-yn1) (ii) a Counting the number S of human foreground pixels in the covered areafPixel ratio of
Figure BDA0002868273320000061
Figure BDA0002868273320000062
The calculation method of the intersection and comparison ratio between the covered area and other suspicious frames iou (intersection over union) may adopt a method common to the target detection field, and is not described herein again, it is noted that the intersection and comparison between the covered area and all suspicious frames need to be calculated, and the maximum value σ max (ious) is taken, in order to ensure that the covered area is located in the foreground area of the human body, μmay be required>0.99, while sigma may be required to avoid covering other suspects<0.1; the step S2.1 sets the augmentation counter to prevent the covered area generated randomly many times from always failing to meet the above requirement, and to avoid the process from falling into a forced exit mechanism designed for local loop.
S2.5 training stage: and covering the original image by using the augmented subset picture, modifying the labeled data of the training picture, and sending the training picture to the network for training. After the image is covered, the label needs to be adjusted as follows, and the original label is (x)11,y11,x12,y12,c1), (x21,y21,x22,y22,c2) .., the augmented subset item class label (x) should be added after augmentation11,y11,x12,y12,c1),(x21,y21,x22,y22,c2)..(x,y,x+w,y+h,cnew) After the amplification, the effect as shown in fig. 4 can be produced.
To sum up, in the method for amplifying training data of the terahertz human body security inspection image target detection model in the embodiment, the suspicious object region is cut out from a small amount of existing images with suspicious object labels to form a suspicious object image amplification subset, the terahertz image human body foreground region is extracted by using a segmentation algorithm or a manual labeling mode, the coverage region is randomly generated and the suspicious object amplification image coverage is randomly selected in the training process, and the data labels are regenerated.
Although embodiments of the present invention have been shown and described above, it is understood that the above embodiments are exemplary and should not be construed as limiting the present invention, and that variations, modifications, substitutions and alterations can be made to the above embodiments by those of ordinary skill in the art within the scope of the present invention.

Claims (10)

1. A terahertz human body security inspection image target detection model training data augmentation method is characterized by comprising the following steps:
s1: preparation phase
S11: cutting an augmented data subset from the target detection data set by using a suspicious object labeling frame, counting the number and the proportion of each type of suspicious object labels, and calculating the percentage of augmented hits;
s12: marking a human body foreground region of a training sample in a training data set by using manual marking or a semantic segmentation model, and making segmentation marking data;
s2: training phase
S21: loading single training image data, target detection labeling data and segmented human body foreground semantic label data, and setting an augmentation counter to be 1;
s22: marking by using a human body frame, calculating the coordinates of a random starting point of a covered area, adding 1 to an augmentation counter, giving up augmentation if the number of times of marking reaches the upper limit, and directly sending image data and target detection marking data to network training;
s23: randomly selecting an augmentation coverage picture from the augmentation data subset according to the number ratio of the suspicious object type labels;
s24: calculating pixel proportion mu of the covered area and the human body foreground area, and intersection ratio sigma of the covered area and all suspicious object target frames, if mu and sigma meet set qualified conditions, turning to step S25, and if mu and sigma do not meet set qualified conditions, returning to step S22;
s25: and covering the original image by using the augmented subset picture, modifying the labeled data of the training picture, and sending the training picture to the network for training.
2. The terahertz human body security inspection image target detection model training data augmentation method according to claim 1, characterized in that: in the step S11, there is a label (x1, y1, x2, y2, c) for each image, where x1 and y1 are coordinates of the upper left corner point of the target frame, (x2, y2) are coordinates of the lower right corner point of the target frame, and c is the target category in the frame.
3. The terahertz human body security inspection image target detection model training data augmentation method according to claim 2, characterized in that: the clipping mode is to extract only the image on the original image using the suspicious object labeling box [ y 1: y2, x 1: x2]In-range image data, i.e. only regions of articles of the suspect category in the image are cropped; and counting the total number of the suspicious object target frame labels and the number of the suspicious object labels of each category, and calculating the ratio of each category in the total number:
Figure FDA0002868273310000011
wherein C isiIs the number of suspect labels of category i; calculating corresponding reciprocal
Figure FDA0002868273310000012
Statistical reciprocal ratio
Figure FDA0002868273310000013
Wherein gamma isiThe probability of random amplification of each type of suspicious object is high, the probability of amplification of the suspicious object with a small number of labels is high, and the probability of amplification of the suspicious object with a large number of labels is low.
4. The terahertz human body security inspection image target detection model training data augmentation method according to claim 3, characterized in that: in step S12, it is necessary to acquire a human body segmentation foreground region for each image including a human body frame, and generate corresponding segmentation label data, where the foreground pixel in the grayscale label image is 255 and the background pixel is 0.
5. The terahertz human body security inspection image target detection model training data augmentation method according to claim 4, characterized in that: in step S21, the image data refers to image matrix data, the target detection annotation data refers to data of a human body frame and a target frame that may exist in the image, the segmented human body foreground semantic label data refers to an image label with a width and a height consistent with the image matrix, and an augmentation counter is set to 1 before each image is augmented.
6. The terahertz human body security inspection image target detection model training data augmentation method according to claim 5, characterized in that: in step S22, a human body frame is labeled (x)p1,yp1,xp2,yp2) The calculation of the random initial point coordinates of the covered area means that two random numbers r in the (0,1) interval are taken by utilizing uniform distributionx,ry(ii) a And then calculating the coordinates of the starting point:
xs=xp1+rx(xp2-xp1)
ys=yp1+ry(yp2-yp1)。
7. the method for amplifying the training data of the terahertz human body security inspection image target detection model according to claim 6, wherein in the step S23, the specific process of selecting the amplification coverage picture is as follows:
s231: selecting an augmented picture from the augmented data subset according to the number ratio of the suspicious object class labels, and calculating a random hit interval under the suspicious object class i percentage;
s232: then generating a random number with the interval of (0,100) to determine the suspicious object class needing to be augmented;
s233: and selecting a specific single picture from the single-class suspicious object augmentation subset by using uniform distribution and randomness.
8. The terahertz human body security inspection image target detection model training data augmentation method according to claim 7, characterized in that: in step S24, the width and height of the enlarged picture are w, h, and the coordinates of the covered area frame are (x)s,ys,xs+w,ys+ h), the area of the covered area frame pixel is S ═ wh, the total S of the human foreground area pixels in the frame is countedfThe proportion calculation mode is as follows:
Figure FDA0002868273310000021
9. the terahertz human body security inspection image target detection model training data augmentation method according to claim 8, characterized in that: in step S24, the pass conditions for μ and σ are:
Figure FDA0002868273310000022
10. the terahertz human body security inspection image target detection model training data augmentation method according to claim 9, characterized in that: in step S25, modifying the training image annotation data refers to adding the covered area coordinates and the category based on the original label.
CN202011589752.8A 2020-12-29 2020-12-29 Terahertz human body security inspection image target detection model training data augmentation method Active CN112633392B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202011589752.8A CN112633392B (en) 2020-12-29 2020-12-29 Terahertz human body security inspection image target detection model training data augmentation method

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202011589752.8A CN112633392B (en) 2020-12-29 2020-12-29 Terahertz human body security inspection image target detection model training data augmentation method

Publications (2)

Publication Number Publication Date
CN112633392A true CN112633392A (en) 2021-04-09
CN112633392B CN112633392B (en) 2024-08-23

Family

ID=75285915

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202011589752.8A Active CN112633392B (en) 2020-12-29 2020-12-29 Terahertz human body security inspection image target detection model training data augmentation method

Country Status (1)

Country Link
CN (1) CN112633392B (en)

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113408575A (en) * 2021-05-12 2021-09-17 桂林电子科技大学 Image data augmentation method based on discriminant area positioning
CN113807284A (en) * 2021-09-23 2021-12-17 上海亨临光电科技有限公司 Method for positioning personal object on terahertz image in human body
CN114418898A (en) * 2022-03-21 2022-04-29 南湖实验室 Data enhancement method based on target overlapping degree calculation and self-adaptive adjustment
CN115346109A (en) * 2022-08-02 2022-11-15 北京新岳纵横科技有限公司 IOU (input/output Unit) strategy based enhanced sample generation method
CN117930375A (en) * 2024-03-22 2024-04-26 国擎(山东)信息科技有限公司 Multi-dimensional detection technology fused channel type terahertz human body security inspection system

Citations (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108492343A (en) * 2018-03-28 2018-09-04 东北大学 A kind of image combining method for the training data expanding target identification
CN109948562A (en) * 2019-03-25 2019-06-28 浙江啄云智能科技有限公司 A kind of safe examination system deep learning sample generating method based on radioscopic image
EP3591582A1 (en) * 2018-07-06 2020-01-08 Tata Consultancy Services Limited Method and system for automatic object annotation using deep network
CN111144494A (en) * 2019-12-27 2020-05-12 睿魔智能科技(深圳)有限公司 Object detection model training method, object detection device, object detection equipment and object detection medium
WO2020098074A1 (en) * 2018-11-12 2020-05-22 平安科技(深圳)有限公司 Face sample picture marking method and apparatus, computer device, and storage medium
CN111260607A (en) * 2019-12-23 2020-06-09 北京无线电计量测试研究所 Automatic suspicious article detection method, terminal device, computer device and medium
CN111275080A (en) * 2020-01-14 2020-06-12 腾讯科技(深圳)有限公司 Artificial intelligence-based image classification model training method, classification method and device
CN111914723A (en) * 2020-07-27 2020-11-10 睿魔智能科技(深圳)有限公司 Data augmentation method and system for improving human body detection rate and human body detection model
CN111914725A (en) * 2020-07-27 2020-11-10 睿魔智能科技(深圳)有限公司 Data augmentation method and system and human body detection model
CN112102250A (en) * 2020-08-20 2020-12-18 西北大学 Method for establishing and detecting pathological image detection model with training data as missing label

Patent Citations (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108492343A (en) * 2018-03-28 2018-09-04 东北大学 A kind of image combining method for the training data expanding target identification
EP3591582A1 (en) * 2018-07-06 2020-01-08 Tata Consultancy Services Limited Method and system for automatic object annotation using deep network
WO2020098074A1 (en) * 2018-11-12 2020-05-22 平安科技(深圳)有限公司 Face sample picture marking method and apparatus, computer device, and storage medium
CN109948562A (en) * 2019-03-25 2019-06-28 浙江啄云智能科技有限公司 A kind of safe examination system deep learning sample generating method based on radioscopic image
CN111260607A (en) * 2019-12-23 2020-06-09 北京无线电计量测试研究所 Automatic suspicious article detection method, terminal device, computer device and medium
CN111144494A (en) * 2019-12-27 2020-05-12 睿魔智能科技(深圳)有限公司 Object detection model training method, object detection device, object detection equipment and object detection medium
CN111275080A (en) * 2020-01-14 2020-06-12 腾讯科技(深圳)有限公司 Artificial intelligence-based image classification model training method, classification method and device
CN111914723A (en) * 2020-07-27 2020-11-10 睿魔智能科技(深圳)有限公司 Data augmentation method and system for improving human body detection rate and human body detection model
CN111914725A (en) * 2020-07-27 2020-11-10 睿魔智能科技(深圳)有限公司 Data augmentation method and system and human body detection model
CN112102250A (en) * 2020-08-20 2020-12-18 西北大学 Method for establishing and detecting pathological image detection model with training data as missing label

Cited By (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113408575A (en) * 2021-05-12 2021-09-17 桂林电子科技大学 Image data augmentation method based on discriminant area positioning
CN113408575B (en) * 2021-05-12 2022-08-19 桂林电子科技大学 Image data augmentation method based on discriminant area positioning
CN113807284A (en) * 2021-09-23 2021-12-17 上海亨临光电科技有限公司 Method for positioning personal object on terahertz image in human body
CN114418898A (en) * 2022-03-21 2022-04-29 南湖实验室 Data enhancement method based on target overlapping degree calculation and self-adaptive adjustment
CN114418898B (en) * 2022-03-21 2022-07-26 南湖实验室 Data enhancement method based on target overlapping degree calculation and self-adaptive adjustment
CN115346109A (en) * 2022-08-02 2022-11-15 北京新岳纵横科技有限公司 IOU (input/output Unit) strategy based enhanced sample generation method
CN115346109B (en) * 2022-08-02 2023-07-18 北京新岳纵横科技有限公司 Enhanced sample generation method based on IOU strategy
CN117930375A (en) * 2024-03-22 2024-04-26 国擎(山东)信息科技有限公司 Multi-dimensional detection technology fused channel type terahertz human body security inspection system

Also Published As

Publication number Publication date
CN112633392B (en) 2024-08-23

Similar Documents

Publication Publication Date Title
CN112633392A (en) Terahertz human body security inspection image target detection model training data augmentation method
US11429818B2 (en) Method, system and device for multi-label object detection based on an object detection network
CN108510467B (en) SAR image target identification method based on depth deformable convolution neural network
CN103927387B (en) Image indexing system and its correlation technique and device
US10229347B2 (en) Systems and methods for identifying a target object in an image
CN105608230B (en) A kind of Business Information recommender system and method based on image retrieval
CN109829398B (en) Target detection method in video based on three-dimensional convolution network
Wang et al. Joint learning of visual attributes, object classes and visual saliency
CN104866616B (en) Monitor video Target Searching Method
Wang et al. YOLOv3‐Litchi Detection Method of Densely Distributed Litchi in Large Vision Scenes
WO2016026371A1 (en) Fast object detection method based on deformable part model (dpm)
CN109582813B (en) Retrieval method, device, equipment and storage medium for cultural relic exhibit
CN106326916B (en) Object detection method based on Analysis On Multi-scale Features estimation and high-order BING feature
Wang et al. Superpixel-based LCM detector for faint ships hidden in strong noise background SAR imagery
CN112733858B (en) Image character rapid identification method and device based on character region detection
CN111680705A (en) MB-SSD method and MB-SSD feature extraction network suitable for target detection
WO2023123924A1 (en) Target recognition method and apparatus, and electronic device and storage medium
CN111177811A (en) Automatic fire point location layout method applied to cloud platform
CN111898418A (en) Human body abnormal behavior detection method based on T-TINY-YOLO network
CN117292120B (en) Light-weight visible light insulator target detection method and system
Nebel et al. Are current monocular computer vision systems for human action recognition suitable for visual surveillance applications?
US8724890B2 (en) Vision-based object detection by part-based feature synthesis
Zhou et al. A pipeline architecture for traffic sign classification on an FPGA
CN105574880A (en) Color image segmentation method based on exponential moment pixel classification
Hao et al. Improved bags-of-words algorithm for scene recognition

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant