US20230162342A1 - Image sample generating method and system, and target detection method - Google Patents

Image sample generating method and system, and target detection method Download PDF

Info

Publication number
US20230162342A1
US20230162342A1 US17/910,346 US202017910346A US2023162342A1 US 20230162342 A1 US20230162342 A1 US 20230162342A1 US 202017910346 A US202017910346 A US 202017910346A US 2023162342 A1 US2023162342 A1 US 2023162342A1
Authority
US
United States
Prior art keywords
image
fused
security check
images
target
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
US17/910,346
Inventor
Yiqing Li
Kai Zhou
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Zhejiang Peckerai Technology Co Ltd
Zhejiang Peckerai Technology Co Ltd
Original Assignee
Zhejiang Peckerai Technology Co Ltd
Zhejiang Peckerai Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Zhejiang Peckerai Technology Co Ltd, Zhejiang Peckerai Technology Co Ltd filed Critical Zhejiang Peckerai Technology Co Ltd
Assigned to ZHEJIANG PECKERAI TECHNOLOGY CO., LTD reassignment ZHEJIANG PECKERAI TECHNOLOGY CO., LTD ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: LI, YIQING, ZHOU, Kai
Publication of US20230162342A1 publication Critical patent/US20230162342A1/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/0002Inspection of images, e.g. flaw detection
    • G06T7/0004Industrial image inspection
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/50Context or environment of the image
    • G06V20/52Surveillance or monitoring of activities, e.g. for recognising suspicious objects
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/0002Inspection of images, e.g. flaw detection
    • G06T7/0004Industrial image inspection
    • G06T7/0008Industrial image inspection checking presence/absence
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T5/00Image enhancement or restoration
    • G06T5/50Image enhancement or restoration using two or more images, e.g. averaging or subtraction
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T5/00Image enhancement or restoration
    • G06T5/70Denoising; Smoothing
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/70Determining position or orientation of objects or cameras
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/44Local feature extraction by analysis of parts of the pattern, e.g. by detecting edges, contours, loops, corners, strokes or intersections; Connectivity analysis, e.g. of connected components
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/44Local feature extraction by analysis of parts of the pattern, e.g. by detecting edges, contours, loops, corners, strokes or intersections; Connectivity analysis, e.g. of connected components
    • G06V10/443Local feature extraction by analysis of parts of the pattern, e.g. by detecting edges, contours, loops, corners, strokes or intersections; Connectivity analysis, e.g. of connected components by matching or filtering
    • G06V10/449Biologically inspired filters, e.g. difference of Gaussians [DoG] or Gabor filters
    • G06V10/451Biologically inspired filters, e.g. difference of Gaussians [DoG] or Gabor filters with interaction between the filter responses, e.g. cortical complex cells
    • G06V10/454Integrating the filters into a hierarchical structure, e.g. convolutional neural networks [CNN]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/56Extraction of image or video features relating to colour
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/74Image or video pattern matching; Proximity measures in feature spaces
    • G06V10/75Organisation of the matching processes, e.g. simultaneous or sequential comparisons of image or video features; Coarse-fine approaches, e.g. multi-scale approaches; using context analysis; Selection of dictionaries
    • G06V10/751Comparing pixel values or logical combinations thereof, or feature values having positional relevance, e.g. template matching
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/764Arrangements for image or video recognition or understanding using pattern recognition or machine learning using classification, e.g. of video objects
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/77Processing image or video features in feature spaces; using data integration or data reduction, e.g. principal component analysis [PCA] or independent component analysis [ICA] or self-organising maps [SOM]; Blind source separation
    • G06V10/774Generating sets of training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/77Processing image or video features in feature spaces; using data integration or data reduction, e.g. principal component analysis [PCA] or independent component analysis [ICA] or self-organising maps [SOM]; Blind source separation
    • G06V10/80Fusion, i.e. combining data from various sources at the sensor level, preprocessing level, feature extraction level or classification level
    • G06V10/806Fusion, i.e. combining data from various sources at the sensor level, preprocessing level, feature extraction level or classification level of extracted features
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/82Arrangements for image or video recognition or understanding using pattern recognition or machine learning using neural networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10116X-ray image
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20081Training; Learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20084Artificial neural networks [ANN]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20212Image combination
    • G06T2207/20221Image fusion; Image merging
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30108Industrial image inspection

Definitions

  • the present disclosure relates to the field of security check, for example, to an image sample generating method and system, and a target detection method.
  • the X-rays are a kind of electromagnetic radiation with a shorter wavelength than the visible light, so that the X-rays have a stronger solid and liquid penetrating ability than the visible light, and can even penetrate the steel plate of a certain thickness.
  • the internal structure of the item having different substance composition, density and thickness can absorb the X-rays to varying degrees. The larger the density and thickness, the more rays are absorbed; and the smaller the density and thickness, the less rays are absorbed.
  • a pixel value of the generated image reflects the density value of the physical item, so the ray intensity transmitted from the item can reflect internal structure information of the item.
  • the system sets the color of the security image obtained by the perspective.
  • the image color of the item belonging to the organic matter is set to be orange
  • the image color of the item belonging to the inorganic matter is set to be blue
  • the image color of the item belonging to the mixture is set to be green.
  • the specific color difference depends on the absorption degree of the item to the x-rays. The higher the absorption degree, the darker the color is; and the lower the absorption degree, the lighter the color is. Therefore, the collected X-ray image not only has shape characteristics, but also shows different colors according to the material, and the above characteristics can be used for analysis and recognition as identifying the item.
  • the radiation imaging technology is the mainstream technology widely used in security system by many countries.
  • this technology irradiates the detected item with rays (such as the X-rays), and obtains the radiographic image of the detected item by the computer processing according to the signal received by the detector, so that the security inspector can identify whether there is suspicious contraband in the image according to the shape and the ribbon of the common contraband by observing the X-ray image.
  • rays such as the X-rays
  • Such manual interpretation method has a low efficiency, a high missing check rate and a high labor cost.
  • the deep learning technology has made breakthroughs in classification, recognition, detection, segmentation, and tracking in the field of computer vision.
  • the deep convolutional neural network learns useful features from a large amount of data under the training of big data, having the advantages of high speed, high precision and low cost.
  • the deep learning is better than the conventional method, in large part, because the deep learning is based on a large amount of data, especially in the field of security check, the deep learning requires a large amount of data.
  • the mainstream approach is data enhancement, but the detection performance of the model cannot be improved simply by increasing the amount of data, difficult case samples affected by external factors such as a placement angle of the detection target, background environment and the like are also needed to restore the security check image in real scenarios.
  • the detection accuracy and recall rate of the contraband can be improved by training the detection network, which further increases the cost of collecting and identification data.
  • Sample data with identified information is mainly collected by collecting a large number of on-spot real-shot images, and then manually perform information identification on the on-spot real-shot images.
  • the invention patent No. CN201910228142.6 and the invention patent No. CN201911221349.7 provide a development method for simulating real samples in terms of difficult cases. In practice, it is found that the above existing method still has problems of complex algorithm, inflexible application for different scenarios and sample effect to be improved.
  • the present disclosure discloses an image sample generating method and system, and a target detection method, which solves problems of difficult training sample data collection and identifying for deep learning and large data amount, uses a simple algorithm to quickly provide effective training samples for detection of contraband, and can flexibly adapt to target detection tasks in different scenarios.
  • An image sample generating method is provided in the present disclosure and includes steps described below.
  • a scenario composition analysis is performed on an item to be detected in a security check place.
  • a real-shot security check image of a target scenario having a corresponding composition ratio is obtained according to the scenario composition analysis.
  • a target security check image having a label is obtained.
  • the target security check image is captured by a security check device.
  • a pixel gray value of an i-th feature layer in the real-shot security check image and a pixel gray value of an i-th feature layer in the target security check image separately are processed in the following manner:
  • a norm [i] is a pixel gray value of the i-th feature layer after the processing
  • a[i] is a pixel gray value of the i-th feature layer before the processing
  • MAX_PIXEL_VAL[i] is a theoretical maximum gray value of the i-th feature layer.
  • Images to be fused are determined.
  • the images to be fused include at least one real-shot security check image and at least one target security check image, and a number of images to be fused is marked as N, where N ⁇ 2 and N is an integer.
  • Sizes of the images to be fused are normalized.
  • the size-normalized images to be fused are fused to form a new sample, and a fusion method is as follows: for each pixel dot (i, j, k) of the new sample, in a case where each of N pixel dots in N images to be fused corresponding to the each pixel dot (i, j, k) satisfies a mean [j][k] ⁇ , a pixel value of the each pixel dot (i, j, k) is
  • a norm [i][j][k] denotes a pixel gray value of the i-th feature layer in the j-th row and k-th column of each of the size-normalized images to be fused
  • 1 ⁇ j ⁇ a maximum number of rows of each of the size-normalized images to be fused 1 ⁇ k ⁇ a maximum number of columns of each of the size-normalized images to be fused.
  • the steps in which the image to be fused is determined, the size of the image to be fused is normalized, and the size-normalized images to be fused is fused to form the new sample are performed repeatedly until a preset number of new samples are acquired as a sample composition for training.
  • An image sample generating system is further provided in the present disclosure and includes: a scenario data generating module, a target data generating module, a data preprocessing module, an image-to-be-fused preprocessing module, an image fusing module, and a sample library generating module.
  • the scenario data generating module is configured to perform a scenario composition analysis on an item to be detected in a security check place; and obtain a real-shot security check image of a target scenario having a corresponding composition ratio according to the scenario composition analysis.
  • the target data generating module is configured to obtain a target security check image having a label, where the target security check image is captured by a security check device.
  • the data preprocessing module is configured to process a pixel gray value of an i-th feature layer in the real-shot security check image and a pixel gray value of an i-th feature layer in the target security check image separately in the following manner:
  • a norm [i] is a pixel gray value of the i-th feature layer after the processing
  • a[i] is a pixel gray value of the i-th feature layer before the processing
  • MAX_PIXEL_VAL[i] is a theoretical maximum gray value of the i-th feature layer.
  • the image-to-be-fused preprocessing module is configured to determine an image to be fused, where the image to be fused includes at least one real-shot security check image and at least one target security check image, and a number of images to be fused is marked as N, where N ⁇ 2 and N is an integer; and normalize a size of the images to be fused.
  • the image fusing module is configured to fuse the size-normalized images to be fused to form a new sample, and a fusion method is as follows: for each pixel dot (i, j, k) of the new sample, in a case where each of N pixel dots in N images to be fused corresponding to the each pixel dot (i, j, k) satisfies a mean [j][k] ⁇ , a pixel value of the each pixel dot (i, j, k) is
  • a norm [i][j][k] denotes a pixel gray value of the i-th feature layer in the j-th row and k-th column of each of the size-normalized images to be fused
  • 1 ⁇ j ⁇ a maximum number of rows of each of the size-normalized images to be fused 1 ⁇ k ⁇ a maximum number of columns of each of the size-normalized images to be fused.
  • the sample library generating module is configured to perform determining the image to be fused, normalizing the size of the image to be fused, and fusing the size-normalized images to be fused to form the new sample repeatedly until a preset number of new samples are acquired as a sample composition for training.
  • a target detection method is further provided in the present disclosure and includes steps described below.
  • a security check image of an item is acquired, and the security check image is preprocessed.
  • a preset convolutional neural network extracts an image feature of the preprocessed security check image.
  • the image feature is input to a preset target detection model to obtain a target region of the security check image.
  • the preset target detection model is obtained by training the image sample obtained by the above image sample generating method.
  • a detection result of the security check image is determined according to the obtained target region of the security check image.
  • the detection result includes type information and location information of contraband.
  • FIG. 1 is a flowchart of an image sample generating method based on deep learning provided by an embodiment of the present disclosure
  • FIG. 2 is an X-ray image obtained by a sample generating method provided by an embodiment of the present disclosure.
  • FIG. 3 is an X-ray image obtained by shooting a real item provided by an embodiment of the present disclosure.
  • Contraband items that are not allowed to be manufactured, purchased, used, held, stored and transported for import or export by law, such as weapons, ammunition, explosive objects (such as explosives, detonators, fuses and the like).
  • Security check image an image acquired by a security check device.
  • the security check device or a security check machine related to the present disclosure is not limited to an X-ray security check device.
  • the security check device and/or the security check machine that can perform security check in an imaging mode are all within the protection scope of the present disclosure, such as a terahertz imaging device and the like.
  • an image sample generating method based on deep learning is provided by the present disclosure and includes steps described below.
  • S 1 a real-shot security image of a target scenario is obtained to form a scenario dataset.
  • the target scenario includes an item, such as luggage, an express parcel, a bag, a cargo and the like, that requires the security check appears in a place such as an airport, a railway station, a bus station, a government building, an embassy, a conference center, a convention and exhibition center, a hotel, a shopping mall, a large-scale event, a post office, a school, a logistics industry, an industrial detection, an express transit depot and the like.
  • a target is the contraband (such as a gun or an explosive)
  • the target scenario refers to a container in which the contraband is located, that is, a place which is configured to accommodate the contraband.
  • the target scenario does not include the target.
  • the type of the scenario is related to the place, for example, in the place such as the airport, the railway station and the like, the luggage is the main scenario, and the scenario corresponding to the express transit depot is the express parcel.
  • the scenario is different in the express transit depots having different geographical locations.
  • the scenario is generally an express parcel with clothes
  • Kunshan the majority of scenarios are an express package with electronic devices.
  • the X-ray security check device is used as an example, and the principle analysis is as follows.
  • the X-rays are a kind of electromagnetic radiation with a shorter wavelength than the visible light, so that the X-rays have a stronger solid and liquid penetrating ability than the visible light, and can even penetrate the steel plate of a certain thickness.
  • the internal structures of items with different material compositions, densities, and thicknesses can absorb the X-rays to varying degrees. The larger the density and thickness, the more rays are absorbed; and the smaller the density and thickness, the less rays are absorbed.
  • a pixel value of the generated image reflects the density value of the physical object, so the ray intensity transmitted from the object can reflect internal structure information of the item.
  • the system sets the color of the security image obtained by the perspective.
  • the image color of the item belonging to the organic matter is set to be orange
  • the image color of the item belonging to the inorganic matter is set to be blue
  • the image color of the item belonging to the mixture is set to be green.
  • the specific color difference depends on the absorption degree of the item to the x-rays. The higher the absorption degree, the darker the color; and the lower the absorption degree, the lighter the color. Therefore, the collected X-ray image not only has shape characteristics, but also shows different colors according to the material, and the above characteristics can be used for analysis and recognition as identifying the item.
  • scenario data which is essential to the sample of the present disclosure also has an emphasis according to different places.
  • a contraband detection network provided for the transit depot with the clothes as main goods
  • data of the express package with the clothes as the scenario is used as a sample or made into the sample in the method of the present disclosure. Therefore, when a real-shot security check image of the target scenario is obtained, a scenario composition analysis of an item to be detected in a security check place is performed, and a target scenario image having a corresponding ratio is selected.
  • the security check image may be acquired by using the X-ray security check device or another security check device such as the terahertz security check device.
  • the type and the model of the security check device is not limited in this embodiment as long as the device can be configured for security check and can obtain the security check image.
  • the type of the target is one or more, and the number of targets is one or more.
  • the target image is captured by the security check device, the scenario in which the target is located is not set and only the target is contained in the security check image.
  • the contraband is a general name of the targets in the embodiment of the present disclosure, and the identifying personnel identifies each target to make the target become the target having the label, and identification contents include the rectangular frame and the type of the target. The more target data, the better.
  • Images in S 1 and S 2 of this embodiment may also be performed a data enhancement and then incorporated into the scenario dataset and the target dataset separately.
  • the enhancement method includes a geometric transformation operation and/or a pixel transformation operation.
  • the geometric transformation operation includes one or more of a rotation operation, a scaling operation and a cutting operation. Synchronous transformation of identification information is acquired while the geometric transformation.
  • the pixel transformation operation includes one or more of a noise adding operation, a blur transformation, a perspective operation, a luminance operation and a contrast operation.
  • the rotation operation the image is rotated clockwise/counterclockwise by a certain angle to reduce the probability of recognition failure caused by the inclination angle of the image.
  • the scaling operation when the image sample is generated by matting, a scaling ratio is inputted, and then the scaled image is matted from the original image and then compressed into the size of the original image.
  • the cutting operation a cutting process is performed on the matted image sample, so that the probability of recognition failure caused by missing or occluded image is reduced.
  • the noise adding operation a noise matrix is generated according to a mean value and a Gaussian covariance, noises are added to the original image matrix, and then the validity of pixel values of multiple dots is judged.
  • the blur transformation is achieved by a blur function of OpenCV, that is, a blur block is added to the original image.
  • the luminance operation and the contrast operation achieve the luminance operation and the contrast operation on the image by adjusting a Red Green Blue (RGB) value of each pixel.
  • RGB Red Green Blue
  • the data in S 1 and S 2 are preprocessed, and the processing mode includes but is not limited to one or more of a processing, a denoising, a background differencing and an artifact removal of a pixel gray value.
  • pixel gray values of an i-th feature layer in the data of the S 1 and S 2 are separately processed in the following manner:
  • a norm [i] is a pixel gray value of the i-th feature layer after the processing
  • a[i] is a pixel gray value of the i-th feature layer before the processing
  • MAX_PIXEL_VAL[i] is a theoretical maximum gray value of the i-th feature layer.
  • the feature layer is a color channel.
  • a first feature layer is a Red (R) channel
  • a second feature layer is a Green (G) channel
  • a third feature layer is a Blue (B) channel.
  • R Red
  • G Green
  • B Blue
  • a corresponding relationship between a serial number of the feature layer and the color channel is not limited in this application.
  • images to be fused are determined.
  • the images to be fused include at least one real-shot security check image of the target scenario and at least one target image, and the number of images to be fused is marked as N, where N is an integer greater than or equal to 2.
  • N is an integer greater than or equal to 2.
  • the sizes of the selected images are normalized, the at least two X-ray images may be same or different, and the sizes of the at least two X-ray images may be same or different, which are both within the protection scope of the present disclosure.
  • the length and width of the normalized image are set according to the size of the minimum external rectangular frame of the image to be fused.
  • the size normalization process of each image is achieved by filling a newly added area of the image with a background color, so that the target in the original image cannot be changed.
  • the background color is related to the device collecting the X-ray image and can be adjusted according to the X-ray image.
  • the fusing method is described below.
  • a mean [j][k] is a pixel gray value in a j-th row and k-th column
  • a norm [i][j][k] is a pixel gray value of an i-th feature layer in the j-th row and k-th column.
  • S 6 S 3 , S 4 , and S 5 are iterated repeatedly until a sufficient number of samples are acquired as a sample composition for training.
  • composition of the images to be fused having pertinence can be determined according to different places, which is consistent with ideas of the step S 1 in the embodiment of the present disclosure.
  • a composition ratio of the real-shot security check image of the target scenario in the image to be fused is selected according to a scenario ratio of a daily actual situation of the airport, for example, 60% of large luggage and 30% of bags are used as the target scenario.
  • the real-shot security check image of the target scenario based on the analysis of the security check place is obtain, so that the target image having the label is obtained and the images to be fused are determined.
  • the method of obtaining the new sample by processing the images to be fused with a new fusion algorithm there is no need to capture a large number of target images in real scenarios on the spot, and there is no need to manually identify the real-shot images in the above complicate environment.
  • the algorithm is simple, which can flexibly and quickly generate the image of the new sample with the place pertinence.
  • the sample has a high realism and a high identifying accuracy, provides a large number of available sample data with identification information for the model training, and solves the sample collection problem that some contraband, such as pistols and explosives, is difficult to be obtained in the field of contraband recognition. It is found that the new sample obtained by the method in the embodiment of the present disclosure shown in FIG. 2 is almost consistent with the real-shot image containing the detection target in FIG. 3 by comparison. The new sample with the color image shows a more realistic effect, has a high realism and a high identifying accuracy, provides a large number of available sample data with identification information for the model training, and further improves the efficiency and accuracy of the target detection task executed in the intelligent security check process by using the deep learning method.
  • An image sample generating system based on deep learning includes: a scenario dataset, a target dataset, a preprocessing module, an image-to-be-fused preprocessing module, an image fusing module, and a generated sample library.
  • the scenario dataset is composed of the real-shot security check image of the target scenario according to embodiment one
  • the target dataset is composed of the target image having the label according to embodiment one.
  • the real-shot security check image and the target image are composed of X-ray images of items, and the X-ray images of items may be collected by using an X-ray security check device; and the items includes luggage, express parcels, large goods and the like.
  • data in the scenario dataset and the target dataset are preprocessed, and the processing mode includes but is not limited to one or more of a processing, a denoising, a background differencing and an artifact removal of a pixel gray value.
  • pixel gray values of i-th feature layers in the data of the scenario dataset and the target dataset are respectively processed in the following manner:
  • a norm [i] is a pixel gray value of the i-th feature layer after the processing
  • a[i] is a pixel gray value of the i-th feature layer before the processing
  • MAX_PIXEL_VAL[i] is a theoretical maximum gray value of the i-th feature layer.
  • images in the scenario dataset and the target dataset may be performed a data enhancement and the enhanced images are also components of the scenario dataset and the target dataset respectively.
  • the enhancement method includes a geometric transformation operation and/or a pixel transformation operation.
  • the geometric transformation operation includes one or more of a rotation operation, a scaling operation and a cutting operation.
  • the pixel transformation operation includes one or more of a noise adding operation, a blur transformation, a perspective operation, a luminance operation and a contrast operation.
  • the rotation operation the image is rotated clockwise/counterclockwise by a certain angle to reduce the probability of recognition failure caused by the inclination angle of the image.
  • the scaling operation when the image sample is generated by matting, a scaling ratio is inputted, and then the scaled image is matted from the original image and then compressed into the size of the original image.
  • the cutting operation a cutting process is performed on the matted image sample, so that the probability of recognition failure caused by missing or occluded image is reduced.
  • a noise matrix is generated according to a mean value and a Gaussian covariance, noises are added to the original image matrix, and then the validity of pixel values of multiple dots is judged.
  • the blur transformation is achieved by a blur function of OpenCV, that is, a blur block is added to the original image.
  • OpenCV OpenCV
  • four corner dots of the original image are transformed into four new dots according to an input perspective proportion, and then dots of the whole original image are X-rayed according to a corresponding mapping relationship between the four dots before the transformation and the four dots after the transformation.
  • the luminance operation and the contrast operation achieve the luminance operation and the contrast operation on the image by adjusting a Red Green Blue (RGB) value of each pixel.
  • RGB Red Green Blue
  • the image-to-be-fused preprocessing module is configured to select at least one image in the scenario dataset arbitrarily and at least one image in the target dataset arbitrarily and normalize sizes of the at least one image in the scenario dataset arbitrarily and the at least one image in the target dataset.
  • a size normalization module is configured to normalize sizes of N (N ⁇ 2) X-ray images taken from the original sample arbitrarily every time; the at least two X-ray images may be same or different, and the sizes of the at least two X-ray images may be same or different, which are both within the protection scope of the present disclosure.
  • the required sample quantity and quality requirements are achieved by continuously repetition of arbitrarily selection.
  • the length and width of the normalized image are set according to the size of the minimum external rectangular frame of the image to be fused.
  • the size normalization process of each image is achieved by filling a newly added area of the image with a background color, so that the target in the original image cannot be changed.
  • the background color is related to the device collecting the X-ray image and can be adjusted according to the X-ray image.
  • the image fusing module is configured to fuse the pixel dot in each position of the image obtained by the image-to-be-fused preprocessing module, and the fusing method is as follows.
  • a mean [j][k] is a pixel gray value in a j-th row and k-th column
  • a norm [i][j][k] is a pixel gray value of an i-th feature layer in the j-th row and k-th column.
  • the generated sample library includes a sample image generated by the image fusing module.
  • the number of sample images in the generated sample library is determined by execution times of the preprocessing module, the image-to-be-fused preprocessing module and the image fusing module.
  • a target detection method is also provided according to the embodiment of the present disclosure and the target detection method includes steps described below.
  • step 1 a security check image of an item is acquired and the image is preprocessed.
  • the preprocessing mode includes but is not limited to one or more of a normalizing, a denoising, a background differencing and an artifact removal of the image.
  • the image is normalized at a preset size, the case of 500*500 is used as an example in this embodiment.
  • a Gaussian smoothing algorithm is used for denoising the image.
  • a value of each dot in the Gaussian smoothed image is obtained by the weighted average of other pixel values in the image itself and the domain. For example, each pixel in the image is scanned by using a template, and a value of a pixel dot of a center of the template is replaced by a weighted average gray value of pixels in a domain determined by the template.
  • small noises on the image are removed. Although edge information in the image is weakened to a certain extent, the edge is still preserved relative to the noises.
  • step 2 an image feature of the preprocessed security check image is extracted by a preset convolutional neural network.
  • step 3 a target area of the security check image is obtained by a preset target detection model.
  • the preset target detection model is obtained by training the image sample obtained in the method of embodiment one of the present disclosure.
  • the training process of the preset target detection model mainly includes steps described below.
  • the image sample obtained in the method of embodiment one of the present disclosure is collected and a training dataset is constructed.
  • the preset deep learning network model includes a feature extraction module, a target detection network and a loss calculation module; and the preset feature extraction module and the target detection network are both convolutional neural network models.
  • the feature extraction module and the target detection network are trained by the training dataset to obtain a trained deep learning target detection model.
  • the training process includes: the image sample obtained in the method of embodiment one of the present disclosure is input into the feature extraction module for feature extraction to obtain the image feature, then the image feature is input into the target detection network model to obtain a candidate prediction of the image, the candidate prediction is input into the loss calculation module to calculate a loss function, and the preset deep learning target detection model is trained by a gradient back transmission algorithm.
  • step 4 the detection result, including information such as type information and location information of contraband, of the security check image is output.
  • the method embodiments are all expressed as a series of action combinations.
  • the embodiments of the present application are not limited by the described action sequences, and according to the present application, some steps may be performed in other sequences or concurrently.
  • the embodiments described in the present disclosure are all embodiments and that the actions and modules involved are not necessarily necessary for the present application.

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Physics & Mathematics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Multimedia (AREA)
  • Evolutionary Computation (AREA)
  • Artificial Intelligence (AREA)
  • General Health & Medical Sciences (AREA)
  • Health & Medical Sciences (AREA)
  • Software Systems (AREA)
  • Computing Systems (AREA)
  • Medical Informatics (AREA)
  • Databases & Information Systems (AREA)
  • Quality & Reliability (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Biology (AREA)
  • General Engineering & Computer Science (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Biodiversity & Conservation Biology (AREA)
  • Biomedical Technology (AREA)
  • Molecular Biology (AREA)
  • Image Analysis (AREA)

Abstract

Provided are a target detection method and an image sample generating method and system for deep learning. The image sample generating method includes performing a scenario composition analysis on an item to be detected in a security check place; obtaining a real-shot security check image of a target scenario having a corresponding composition ratio according to the scenario composition analysis; obtaining a target security check image having a label, where the target security check image is captured by a security check device; processing a pixel gray value of an i-th feature layer in the real-shot security check image and a pixel gray value of an i-th feature layer in the target security check image separately; determining images to be fused; normalizing sizes of the images to be fused; fusing the size-normalized images to be fused to form a new sample; and performing the determining the images to be fused.

Description

  • This application claims priority to Chinese Patent Application No. 202010267813.2 filed with the CNIPA on Apr. 8, 2020, the disclosure of which is incorporated herein by reference in its entirety.
  • TECHNICAL FIELD
  • The present disclosure relates to the field of security check, for example, to an image sample generating method and system, and a target detection method.
  • BACKGROUND
  • The X-rays are a kind of electromagnetic radiation with a shorter wavelength than the visible light, so that the X-rays have a stronger solid and liquid penetrating ability than the visible light, and can even penetrate the steel plate of a certain thickness. As the X-rays pass through an item, the internal structure of the item having different substance composition, density and thickness can absorb the X-rays to varying degrees. The larger the density and thickness, the more rays are absorbed; and the smaller the density and thickness, the less rays are absorbed. A pixel value of the generated image reflects the density value of the physical item, so the ray intensity transmitted from the item can reflect internal structure information of the item. Typically, to have a more intuitive understanding of the substance composition of the detected item, the system sets the color of the security image obtained by the perspective. The image color of the item belonging to the organic matter is set to be orange, the image color of the item belonging to the inorganic matter is set to be blue, and the image color of the item belonging to the mixture is set to be green. The specific color difference depends on the absorption degree of the item to the x-rays. The higher the absorption degree, the darker the color is; and the lower the absorption degree, the lighter the color is. Therefore, the collected X-ray image not only has shape characteristics, but also shows different colors according to the material, and the above characteristics can be used for analysis and recognition as identifying the item. The radiation imaging technology is the mainstream technology widely used in security system by many countries. In this technology irradiates the detected item with rays (such as the X-rays), and obtains the radiographic image of the detected item by the computer processing according to the signal received by the detector, so that the security inspector can identify whether there is suspicious contraband in the image according to the shape and the ribbon of the common contraband by observing the X-ray image. Such manual interpretation method has a low efficiency, a high missing check rate and a high labor cost.
  • With the continuous development of the artificial intelligence technology, the deep learning technology has made breakthroughs in classification, recognition, detection, segmentation, and tracking in the field of computer vision. Compared with the conventional machine vision method, the deep convolutional neural network learns useful features from a large amount of data under the training of big data, having the advantages of high speed, high precision and low cost. The deep learning is better than the conventional method, in large part, because the deep learning is based on a large amount of data, especially in the field of security check, the deep learning requires a large amount of data. How to overcome the problem that the deep learning relies on the dataset, the mainstream approach is data enhancement, but the detection performance of the model cannot be improved simply by increasing the amount of data, difficult case samples affected by external factors such as a placement angle of the detection target, background environment and the like are also needed to restore the security check image in real scenarios. The detection accuracy and recall rate of the contraband can be improved by training the detection network, which further increases the cost of collecting and identification data.
  • Sample data with identified information is mainly collected by collecting a large number of on-spot real-shot images, and then manually perform information identification on the on-spot real-shot images. On the one hand, it is difficult to obtain a large number of on-spot real-shot images; and on the other hand, problems of low identifying efficiency, high labor cost, great influence of human factors, and low accuracy still exist, which makes it difficult to generate a large number of identification data required for training the model in a short time. To solve the above problems, the invention patent No. CN201910228142.6 and the invention patent No. CN201911221349.7 provide a development method for simulating real samples in terms of difficult cases. In practice, it is found that the above existing method still has problems of complex algorithm, inflexible application for different scenarios and sample effect to be improved.
  • SUMMARY
  • The present disclosure discloses an image sample generating method and system, and a target detection method, which solves problems of difficult training sample data collection and identifying for deep learning and large data amount, uses a simple algorithm to quickly provide effective training samples for detection of contraband, and can flexibly adapt to target detection tasks in different scenarios.
  • An image sample generating method is provided in the present disclosure and includes steps described below.
  • A scenario composition analysis is performed on an item to be detected in a security check place.
  • A real-shot security check image of a target scenario having a corresponding composition ratio is obtained according to the scenario composition analysis.
  • A target security check image having a label is obtained. The target security check image is captured by a security check device.
  • A pixel gray value of an i-th feature layer in the real-shot security check image and a pixel gray value of an i-th feature layer in the target security check image separately are processed in the following manner:
  • a norm [ i ] = a [ i ] MAX_PIXEL _VAL [ i ] ,
  • where i=1, 2, 3; anorm[i] is a pixel gray value of the i-th feature layer after the processing, a[i] is a pixel gray value of the i-th feature layer before the processing, and MAX_PIXEL_VAL[i] is a theoretical maximum gray value of the i-th feature layer.
  • Images to be fused are determined. The images to be fused include at least one real-shot security check image and at least one target security check image, and a number of images to be fused is marked as N, where N≥2 and N is an integer.
  • Sizes of the images to be fused are normalized.
  • The size-normalized images to be fused are fused to form a new sample, and a fusion method is as follows: for each pixel dot (i, j, k) of the new sample, in a case where each of N pixel dots in N images to be fused corresponding to the each pixel dot (i, j, k) satisfies amean[j][k]≥δ, a pixel value of the each pixel dot (i, j, k) is
  • a new = 1 N l = 1 N a norm l [ i ] [ j ] [ k ] ;
  • and in a case where at least one pixel dot in the N images to be fused corresponding to the each pixel dot (i, j, k) does not satisfy amean[j][k]≥δ, the pixel value of the each pixel dot (i, j, k) is anewl=1 N anorm l[i][j][k], where δ is a background color threshold, 0<δ<1, l denotes an l-th picture, 1≤l≤N,
  • a mean [ j ] [ k ] = i = 1 3 a norm [ i ] [ j ] [ k ] 3
  • denotes a pixel gray value in an j-th row and k-th column of each of the size-normalized images to be fused, anorm[i][j][k] denotes a pixel gray value of the i-th feature layer in the j-th row and k-th column of each of the size-normalized images to be fused, 1≤j≤a maximum number of rows of each of the size-normalized images to be fused, and 1≤k≤a maximum number of columns of each of the size-normalized images to be fused. The steps in which the image to be fused is determined, the size of the image to be fused is normalized, and the size-normalized images to be fused is fused to form the new sample are performed repeatedly until a preset number of new samples are acquired as a sample composition for training.
  • An image sample generating system is further provided in the present disclosure and includes: a scenario data generating module, a target data generating module, a data preprocessing module, an image-to-be-fused preprocessing module, an image fusing module, and a sample library generating module.
  • The scenario data generating module is configured to perform a scenario composition analysis on an item to be detected in a security check place; and obtain a real-shot security check image of a target scenario having a corresponding composition ratio according to the scenario composition analysis.
  • The target data generating module is configured to obtain a target security check image having a label, where the target security check image is captured by a security check device.
  • The data preprocessing module is configured to process a pixel gray value of an i-th feature layer in the real-shot security check image and a pixel gray value of an i-th feature layer in the target security check image separately in the following manner:
  • a n o r m [ i ] = a [ i ] MAX_PIXEL _VAL [ i ] ,
  • where i=1, 2, 3; anorm[i] is a pixel gray value of the i-th feature layer after the processing, a[i] is a pixel gray value of the i-th feature layer before the processing, and MAX_PIXEL_VAL[i] is a theoretical maximum gray value of the i-th feature layer.
  • The image-to-be-fused preprocessing module is configured to determine an image to be fused, where the image to be fused includes at least one real-shot security check image and at least one target security check image, and a number of images to be fused is marked as N, where N≥2 and N is an integer; and normalize a size of the images to be fused.
  • The image fusing module is configured to fuse the size-normalized images to be fused to form a new sample, and a fusion method is as follows: for each pixel dot (i, j, k) of the new sample, in a case where each of N pixel dots in N images to be fused corresponding to the each pixel dot (i, j, k) satisfies amean[j][k]≥δ, a pixel value of the each pixel dot (i, j, k) is
  • a n e w = 1 N l = 1 N a n o r m l [ i ] [ j ] [ k ] ;
  • and in a case where at least one pixel dot in the N images to be fused corresponding to the each pixel dot (i, j, k) does not satisfy amean[j][k]≥δ, the pixel value of the each pixel dot (i, j, k) is anewl=1 N anorm l[i][j][k], where δ is a background color threshold, 0<δ<1, l denotes an l-th picture, 1≤l≤N,
  • a m e a n [ j ] [ k ] = i = 1 3 a n o r m [ i ] [ j ] [ k ] 3
  • denotes a pixel gray value in an j-th row and k-th column of each of the size-normalized images to be fused, anorm[i][j][k] denotes a pixel gray value of the i-th feature layer in the j-th row and k-th column of each of the size-normalized images to be fused, 1≤j≤a maximum number of rows of each of the size-normalized images to be fused, and 1≤k≤a maximum number of columns of each of the size-normalized images to be fused.
  • The sample library generating module is configured to perform determining the image to be fused, normalizing the size of the image to be fused, and fusing the size-normalized images to be fused to form the new sample repeatedly until a preset number of new samples are acquired as a sample composition for training.
  • A target detection method is further provided in the present disclosure and includes steps described below.
  • A security check image of an item is acquired, and the security check image is preprocessed.
  • A preset convolutional neural network extracts an image feature of the preprocessed security check image.
  • The image feature is input to a preset target detection model to obtain a target region of the security check image. The preset target detection model is obtained by training the image sample obtained by the above image sample generating method.
  • A detection result of the security check image is determined according to the obtained target region of the security check image. The detection result includes type information and location information of contraband.
  • BRIEF DESCRIPTION OF DRAWINGS
  • FIG. 1 is a flowchart of an image sample generating method based on deep learning provided by an embodiment of the present disclosure;
  • FIG. 2 is an X-ray image obtained by a sample generating method provided by an embodiment of the present disclosure; and
  • FIG. 3 is an X-ray image obtained by shooting a real item provided by an embodiment of the present disclosure.
  • DETAILED DESCRIPTION
  • The technical solutions in the embodiment will be described in connection with the drawing in the embodiment. The embodiment described below is part, not all, of the embodiments.
  • First, terms relating to one or more embodiments of the present disclosure are explained.
  • Contraband: items that are not allowed to be manufactured, purchased, used, held, stored and transported for import or export by law, such as weapons, ammunition, explosive objects (such as explosives, detonators, fuses and the like).
  • Security check image: an image acquired by a security check device. The security check device or a security check machine related to the present disclosure is not limited to an X-ray security check device. The security check device and/or the security check machine that can perform security check in an imaging mode are all within the protection scope of the present disclosure, such as a terahertz imaging device and the like.
  • Embodiment One
  • As shown in FIG. 1 , an image sample generating method based on deep learning is provided by the present disclosure and includes steps described below.
  • In S1: a real-shot security image of a target scenario is obtained to form a scenario dataset.
  • The target scenario includes an item, such as luggage, an express parcel, a bag, a cargo and the like, that requires the security check appears in a place such as an airport, a railway station, a bus station, a government building, an embassy, a conference center, a convention and exhibition center, a hotel, a shopping mall, a large-scale event, a post office, a school, a logistics industry, an industrial detection, an express transit depot and the like. If a target is the contraband (such as a gun or an explosive), the target scenario refers to a container in which the contraband is located, that is, a place which is configured to accommodate the contraband. In one embodiment, the target scenario does not include the target. Typically, the type of the scenario is related to the place, for example, in the place such as the airport, the railway station and the like, the luggage is the main scenario, and the scenario corresponding to the express transit depot is the express parcel. As a common phenomenon, in a case of the express transit depot, the scenario is different in the express transit depots having different geographical locations. For example, for the express transit depot located in Haining, the scenario is generally an express parcel with clothes, and for the express transit depot located in Kunshan, the majority of scenarios are an express package with electronic devices.
  • In different scenarios, the imaging effect is different. The X-ray security check device is used as an example, and the principle analysis is as follows. The X-rays are a kind of electromagnetic radiation with a shorter wavelength than the visible light, so that the X-rays have a stronger solid and liquid penetrating ability than the visible light, and can even penetrate the steel plate of a certain thickness. As the X-rays pass through items, the internal structures of items with different material compositions, densities, and thicknesses can absorb the X-rays to varying degrees. The larger the density and thickness, the more rays are absorbed; and the smaller the density and thickness, the less rays are absorbed. A pixel value of the generated image reflects the density value of the physical object, so the ray intensity transmitted from the object can reflect internal structure information of the item. Typically, to have a more intuitive understanding of the substance composition of the detected item, the system sets the color of the security image obtained by the perspective. The image color of the item belonging to the organic matter is set to be orange, the image color of the item belonging to the inorganic matter is set to be blue, and the image color of the item belonging to the mixture is set to be green. The specific color difference depends on the absorption degree of the item to the x-rays. The higher the absorption degree, the darker the color; and the lower the absorption degree, the lighter the color. Therefore, the collected X-ray image not only has shape characteristics, but also shows different colors according to the material, and the above characteristics can be used for analysis and recognition as identifying the item.
  • Based on the introduction of the above target scenario and the imaging effect, it can be known that the selection of scenario data which is essential to the sample of the present disclosure also has an emphasis according to different places. For example, for a contraband detection network provided for the transit depot with the clothes as main goods, during the detection network training, data of the express package with the clothes as the scenario is used as a sample or made into the sample in the method of the present disclosure. Therefore, when a real-shot security check image of the target scenario is obtained, a scenario composition analysis of an item to be detected in a security check place is performed, and a target scenario image having a corresponding ratio is selected.
  • The security check image may be acquired by using the X-ray security check device or another security check device such as the terahertz security check device. The type and the model of the security check device is not limited in this embodiment as long as the device can be configured for security check and can obtain the security check image.
  • In S2: a target image having a label is obtained to form a target dataset.
  • The type of the target is one or more, and the number of targets is one or more. The target image is captured by the security check device, the scenario in which the target is located is not set and only the target is contained in the security check image. As an example, in the field of security check, the contraband is a general name of the targets in the embodiment of the present disclosure, and the identifying personnel identifies each target to make the target become the target having the label, and identification contents include the rectangular frame and the type of the target. The more target data, the better.
  • Images in S1 and S2 of this embodiment may also be performed a data enhancement and then incorporated into the scenario dataset and the target dataset separately. The enhancement method includes a geometric transformation operation and/or a pixel transformation operation. The geometric transformation operation includes one or more of a rotation operation, a scaling operation and a cutting operation. Synchronous transformation of identification information is acquired while the geometric transformation. The pixel transformation operation includes one or more of a noise adding operation, a blur transformation, a perspective operation, a luminance operation and a contrast operation. In the rotation operation: the image is rotated clockwise/counterclockwise by a certain angle to reduce the probability of recognition failure caused by the inclination angle of the image. In the scaling operation: when the image sample is generated by matting, a scaling ratio is inputted, and then the scaled image is matted from the original image and then compressed into the size of the original image. In the cutting operation: a cutting process is performed on the matted image sample, so that the probability of recognition failure caused by missing or occluded image is reduced. In the noise adding operation: a noise matrix is generated according to a mean value and a Gaussian covariance, noises are added to the original image matrix, and then the validity of pixel values of multiple dots is judged. The blur transformation is achieved by a blur function of OpenCV, that is, a blur block is added to the original image. In the perspective operation, four corner dots of the original image are transformed into four new dots according to an input perspective ratio, and then dots of the whole original image are X-rayed according to a corresponding mapping relationship between the four dots before the transformation and the four dots after the transformation. The luminance operation and the contrast operation achieve the luminance operation and the contrast operation on the image by adjusting a Red Green Blue (RGB) value of each pixel.
  • As one embodiment of the present disclosure, the data in S1 and S2 are preprocessed, and the processing mode includes but is not limited to one or more of a processing, a denoising, a background differencing and an artifact removal of a pixel gray value. As one embodiment of the present disclosure, pixel gray values of an i-th feature layer in the data of the S1 and S2 are separately processed in the following manner:
  • a n o r m [ i ] = a [ i ] MAX_PIXEL _VAL [ i ] ,
  • where i=1, 2, 3; anorm[i] is a pixel gray value of the i-th feature layer after the processing, a[i] is a pixel gray value of the i-th feature layer before the processing, and MAX_PIXEL_VAL[i] is a theoretical maximum gray value of the i-th feature layer.
  • In one embodiment, the feature layer is a color channel. For example, a first feature layer is a Red (R) channel, a second feature layer is a Green (G) channel, and a third feature layer is a Blue (B) channel. A corresponding relationship between a serial number of the feature layer and the color channel is not limited in this application.
  • In S3, images to be fused are determined. The images to be fused include at least one real-shot security check image of the target scenario and at least one target image, and the number of images to be fused is marked as N, where N is an integer greater than or equal to 2. As one embodiment of the present disclosure, one image is selected from the scenario dataset arbitrarily and one image is selected from the target dataset arbitrarily to form the images to be fused, that is, N=2.
  • In S4: sizes of the images to be fused are normalized.
  • As one embodiment of the present disclosure, the sizes of the selected images are normalized, the at least two X-ray images may be same or different, and the sizes of the at least two X-ray images may be same or different, which are both within the protection scope of the present disclosure.
  • The length and width of the normalized image are set according to the size of the minimum external rectangular frame of the image to be fused. The case where there are two X-ray images are used as an example, and wnew=max(w1,w1) and hnew=max(h1,h2), the length and width of the two X-ray images are (w1,h1) and (w2,h2) respectively. The size normalization process of each image is achieved by filling a newly added area of the image with a background color, so that the target in the original image cannot be changed. The background color is related to the device collecting the X-ray image and can be adjusted according to the X-ray image.
  • In S5: the image obtained from S4 is fused to form a new sample.
  • The fusing method is described below.
  • When pixel dots corresponding to N images on pixel dots of the new sample (i, j, k) are amean[j][k]≥δ,
  • a n e w = 1 N l = 1 N a n o r m l [ i ] [ j ] [ k ] ;
  • on remaining pixels, a pixel value of the new sample is set to be anewΠl=1 N anorm l[i][j] [k], where δ is a background color threshold, 0<δ<1, l is an l-th image,
  • a m e a n [ j ] [ k ] = i = 1 3 a n o r m [ i ] [ j ] [ k ] 3
  • is a pixel gray value of each of the size-normalized images to be fused, amean[j][k] is a pixel gray value in a j-th row and k-th column, and anorm[i][j][k] is a pixel gray value of an i-th feature layer in the j-th row and k-th column.
  • In S6: S3, S4, and S5 are iterated repeatedly until a sufficient number of samples are acquired as a sample composition for training.
  • The composition of the images to be fused having pertinence can be determined according to different places, which is consistent with ideas of the step S1 in the embodiment of the present disclosure. For example, in a detection network sample for the airport, a composition ratio of the real-shot security check image of the target scenario in the image to be fused is selected according to a scenario ratio of a daily actual situation of the airport, for example, 60% of large luggage and 30% of bags are used as the target scenario.
  • In the image sample generating method based on deep learning in embodiment one, the real-shot security check image of the target scenario based on the analysis of the security check place is obtain, so that the target image having the label is obtained and the images to be fused are determined. In the method of obtaining the new sample by processing the images to be fused with a new fusion algorithm, there is no need to capture a large number of target images in real scenarios on the spot, and there is no need to manually identify the real-shot images in the above complicate environment. The algorithm is simple, which can flexibly and quickly generate the image of the new sample with the place pertinence. The sample has a high realism and a high identifying accuracy, provides a large number of available sample data with identification information for the model training, and solves the sample collection problem that some contraband, such as pistols and explosives, is difficult to be obtained in the field of contraband recognition. It is found that the new sample obtained by the method in the embodiment of the present disclosure shown in FIG. 2 is almost consistent with the real-shot image containing the detection target in FIG. 3 by comparison. The new sample with the color image shows a more realistic effect, has a high realism and a high identifying accuracy, provides a large number of available sample data with identification information for the model training, and further improves the efficiency and accuracy of the target detection task executed in the intelligent security check process by using the deep learning method.
  • Embodiment Two
  • An image sample generating system based on deep learning is provided and includes: a scenario dataset, a target dataset, a preprocessing module, an image-to-be-fused preprocessing module, an image fusing module, and a generated sample library.
  • The scenario dataset is composed of the real-shot security check image of the target scenario according to embodiment one, and the target dataset is composed of the target image having the label according to embodiment one.
  • The real-shot security check image and the target image are composed of X-ray images of items, and the X-ray images of items may be collected by using an X-ray security check device; and the items includes luggage, express parcels, large goods and the like.
  • As one embodiment of the present disclosure, data in the scenario dataset and the target dataset are preprocessed, and the processing mode includes but is not limited to one or more of a processing, a denoising, a background differencing and an artifact removal of a pixel gray value. As one embodiment of the present disclosure, pixel gray values of i-th feature layers in the data of the scenario dataset and the target dataset are respectively processed in the following manner:
  • a n o r m [ i ] = a [ i ] MAX_PIXEL _VAL [ i ] ,
  • where i=1, 2, 2; anorm[i] is a pixel gray value of the i-th feature layer after the processing, a[i] is a pixel gray value of the i-th feature layer before the processing, and MAX_PIXEL_VAL[i] is a theoretical maximum gray value of the i-th feature layer.
  • In one embodiment, images in the scenario dataset and the target dataset may be performed a data enhancement and the enhanced images are also components of the scenario dataset and the target dataset respectively. The enhancement method includes a geometric transformation operation and/or a pixel transformation operation.
  • In one exemplarily embodiment, the geometric transformation operation includes one or more of a rotation operation, a scaling operation and a cutting operation. The pixel transformation operation includes one or more of a noise adding operation, a blur transformation, a perspective operation, a luminance operation and a contrast operation. In the rotation operation: the image is rotated clockwise/counterclockwise by a certain angle to reduce the probability of recognition failure caused by the inclination angle of the image. In the scaling operation: when the image sample is generated by matting, a scaling ratio is inputted, and then the scaled image is matted from the original image and then compressed into the size of the original image. In the cutting operation: a cutting process is performed on the matted image sample, so that the probability of recognition failure caused by missing or occluded image is reduced. In the noise adding operation: a noise matrix is generated according to a mean value and a Gaussian covariance, noises are added to the original image matrix, and then the validity of pixel values of multiple dots is judged. The blur transformation is achieved by a blur function of OpenCV, that is, a blur block is added to the original image. In the perspective operation, four corner dots of the original image are transformed into four new dots according to an input perspective proportion, and then dots of the whole original image are X-rayed according to a corresponding mapping relationship between the four dots before the transformation and the four dots after the transformation. The luminance operation and the contrast operation achieve the luminance operation and the contrast operation on the image by adjusting a Red Green Blue (RGB) value of each pixel.
  • The image-to-be-fused preprocessing module is configured to select at least one image in the scenario dataset arbitrarily and at least one image in the target dataset arbitrarily and normalize sizes of the at least one image in the scenario dataset arbitrarily and the at least one image in the target dataset.
  • A size normalization module is configured to normalize sizes of N (N≥2) X-ray images taken from the original sample arbitrarily every time; the at least two X-ray images may be same or different, and the sizes of the at least two X-ray images may be same or different, which are both within the protection scope of the present disclosure. In this embodiment, the required sample quantity and quality requirements are achieved by continuously repetition of arbitrarily selection.
  • The length and width of the normalized image are set according to the size of the minimum external rectangular frame of the image to be fused. The case where there are two X-ray images are taken once as an example, and wnew=max(w1,w1) and hnew=max(h1,h2), the length and width of the two X-ray images are (w1,h1) and (w2,h2) respectively. The size normalization process of each image is achieved by filling a newly added area of the image with a background color, so that the target in the original image cannot be changed. The background color is related to the device collecting the X-ray image and can be adjusted according to the X-ray image.
  • The image fusing module is configured to fuse the pixel dot in each position of the image obtained by the image-to-be-fused preprocessing module, and the fusing method is as follows.
  • When pixel dots corresponding to N images on pixel dots of the new sample (i, j, k) are amean[j][k]≥δ,
  • a n e w = 1 N l = 1 N a n o r m l [ i ] [ j ] [ k ] ;
  • on remaining pixels, a pixel value of the new sample is set to be anewl=1 N anorm l[i] [j] [k], where δ is a background color threshold, 0<δ<1, l is an l-th image,
  • a m e a n [ j ] [ k ] = i = 1 3 a norm [ i ] [ j ] [ k ] 3
  • is a pixel gray value of each of the size-normalized images to be fused, amean[j][k] is a pixel gray value in a j-th row and k-th column, and anorm[i][j][k] is a pixel gray value of an i-th feature layer in the j-th row and k-th column.
  • The generated sample library includes a sample image generated by the image fusing module.
  • The number of sample images in the generated sample library is determined by execution times of the preprocessing module, the image-to-be-fused preprocessing module and the image fusing module.
  • Embodiment Three
  • Corresponding to an image sample generating method based on deep learning, a target detection method is also provided according to the embodiment of the present disclosure and the target detection method includes steps described below.
  • In step 1: a security check image of an item is acquired and the image is preprocessed. The preprocessing mode includes but is not limited to one or more of a normalizing, a denoising, a background differencing and an artifact removal of the image.
  • The image is normalized at a preset size, the case of 500*500 is used as an example in this embodiment.
  • A Gaussian smoothing algorithm is used for denoising the image. A value of each dot in the Gaussian smoothed image is obtained by the weighted average of other pixel values in the image itself and the domain. For example, each pixel in the image is scanned by using a template, and a value of a pixel dot of a center of the template is replaced by a weighted average gray value of pixels in a domain determined by the template. After the Gaussian smoothing, small noises on the image are removed. Although edge information in the image is weakened to a certain extent, the edge is still preserved relative to the noises. A background differencing algorithm extracts a median gray value of the whole image (500*500) as a gray value of a background, then calculates an absolute value of a difference between a gray value of each pixel dot in the image and the background: Isub=|Ifg−bg|, where bg is a median value of the whole image. It is known that foreign matter dots have a greater difference than a difference between a background dot and the gray value of the background. Therefore, the absolute value of the difference Isub is seen as a possibility of the pixel dot belonging to the foreign matter dots, and the larger the value, the more likely the corresponding pixel is the foreign matter dots.
  • In step 2: an image feature of the preprocessed security check image is extracted by a preset convolutional neural network.
  • In step 3: a target area of the security check image is obtained by a preset target detection model. The preset target detection model is obtained by training the image sample obtained in the method of embodiment one of the present disclosure.
  • The training process of the preset target detection model mainly includes steps described below.
  • 1. The image sample obtained in the method of embodiment one of the present disclosure is collected and a training dataset is constructed. 2. The preset deep learning network model includes a feature extraction module, a target detection network and a loss calculation module; and the preset feature extraction module and the target detection network are both convolutional neural network models. 3. The feature extraction module and the target detection network are trained by the training dataset to obtain a trained deep learning target detection model.
  • The training process includes: the image sample obtained in the method of embodiment one of the present disclosure is input into the feature extraction module for feature extraction to obtain the image feature, then the image feature is input into the target detection network model to obtain a candidate prediction of the image, the candidate prediction is input into the loss calculation module to calculate a loss function, and the preset deep learning target detection model is trained by a gradient back transmission algorithm.
  • In step 4: the detection result, including information such as type information and location information of contraband, of the security check image is output.
  • For the above method embodiments, for simple description, the method embodiments are all expressed as a series of action combinations. However, the embodiments of the present application are not limited by the described action sequences, and according to the present application, some steps may be performed in other sequences or concurrently. Second, the embodiments described in the present disclosure are all embodiments and that the actions and modules involved are not necessarily necessary for the present application.
  • In the embodiments described above, the description of multiple embodiments has its own emphasis. For a part not described in detail in one embodiment, reference may be made to a related description of other embodiments.
  • Although the above embodiments are all applied to security check scenarios, it can be understood from the technical solution of the present disclosure that the technical solution of the present disclosure can also be applied to a scenario where the image acquisition is carried out by using the X-ray principle in addition to the security check scenarios, for example, lesion detection and analysis in a medical image computed tomography (CT) examination scenario.

Claims (5)

1. An image sample generating method, comprising:
performing a scenario composition analysis on an item to be detected in a security check place;
obtaining a real-shot security check image of a target scenario having a corresponding composition ratio according to the scenario composition analysis;
obtaining a target security check image having a label wherein the target security check image is captured by a security check device;
processing a pixel gray value of an i-th feature layer in the real-shot security check image and a pixel gray value of an i-th feature layer in the target security check image separately in the following manner:
a n o r m [ i ] = a [ i ] MAX_PIXEL _VAL [ i ] ,
wherein i=1, 2, 3; anorm[i] is a pixel gray value of an i-th feature layer after the processing, a[i] is a pixel gray value of the i-th feature layer before the processing, and MAX_PIXEL_VAL[i] is a theoretical maximum gray value of the i-th feature layer;
determining images to be fused, wherein the images to be fused comprise at least one real-shot security check image and at least one target security check image, and a number of the images to be fused is marked as N, wherein N≥2 and N is an integer;
normalizing sizes of the images to be fused;
fusing the size-normalized images to be fused to form a new sample as follows: for each pixel point (i, j, k) of the new sample, in a case where each of N pixel points in the N images to be fused corresponding to the each pixel point (i, j, k) satisfies amean[j][k]≥δ, a pixel value of the each pixel point (i, j, k) is
a n e w = 1 N l = 1 N a n o r m l [ i ] [ j ] [ k ] ;
and in a case where at least one pixel point in the N images to be fused corresponding to the each pixel point (i, j, k) does not satisfy amean[j][k]≥δ, the pixel value of the each pixel point (i, j, k) is anewl=1 N anorm l[i][j][k], wherein δ is a background color threshold, 0<δ<1, l denotes an l-th picture, 1≤l≤N,
a m e a n [ j ] [ k ] = i = 1 3 a n o r m [ i ] [ j ] [ k ] 3
denotes a pixel gray value in a j-th row and k-th column of each of the size-normalized images to be fused, anorm[i][j][k] denotes a pixel gray value of the i-th feature layer in the j-th row and k-th column of each of the size-normalized images to be fused, 1≤j≤a maximum number of rows of each of the size-normalized images to be fused, and 1≤k≤a maximum number of columns of each of the size-normalized images to be fused; and
performing the determining the images to be fused, the normalizing the sizes of the images to be fused, and the fusing the size-normalized images to be fused to form the new sample repeatedly until a preset number of new samples are acquired as a sample composition for training.
2. The method of claim 1, wherein in the target security check image, a type of a target having the label is at least one and a number of the target is at least one.
3. The method of claim 1, after obtaining the real-shot security check image of the target scenario having the corresponding composition ratio, further comprising:
performing a data enhancement on the real-shot security check image; and
after obtaining the target security check image having the label, further comprising:
performing a data enhancement to the target security check image having the label, wherein the data enhancement comprises at least one of a geometric transformation operation or a pixel transformation operation.
4. An image sample generating system, comprising: a scenario data generating module, a target data generating module, a data preprocessing module, an image-to-be-fused preprocessing module, an image fusing module, and a sample library generating module, wherein
the scenario data generating module is configured to perform a scenario composition analysis on an item to be detected in a security check place; and obtain a real-shot security check image of a target scenario having a corresponding composition ratio according to the scenario composition analysis;
the target data generating module is configured to obtain a target security check image having an label, wherein the target security check image is captured by a security check device;
the data preprocessing module is configured to process a pixel gray value of an i-th feature layer in the real-shot security check image and a pixel gray value of an i-th feature layer in the target security check image separately in the following manner:
a n o r m [ i ] = a [ i ] MAX_PIXEL _VAL [ i ] ,
wherein i=1, 2, 3; anorm[i] is a pixel gray value of an i-th feature layer after the processing, a[i] is a pixel gray value of the i-th feature layer before the processing, and MAX_PIXEL_VAL[i] is a theoretical maximum gray value of the i-th feature layer;
the image-to-be-fused preprocessing module is configured to determine images to be fused, wherein the images to be fused comprise at least one real-shot security check image and at least one target security check image, and a number of images to be fused is marked as N, wherein N≥2 and N is an integer; and normalize a size of the images to be fused;
the image fusing module is configured to fuse the size-normalized images to be fused to form a new sample, and a fusing method is as follows: for each pixel point (i, j, k) of the new sample, in a case where each of N pixel points in N images to be fused corresponding to the each pixel point (i, j, k) satisfies amean[j][k]≥δ, a pixel value of the each pixel point (i, j, k) is
a n e w = 1 N l = 1 N a n o r m l [ i ] [ j ] [ k ] ;
and in a case where at least one pixel point in the N images to be fused corresponding to the each pixel point (i, j, k) does not satisfy amean[j][k]≥δ, the pixel value of the each pixel point (i, j, k) is anewl=1 N anorm l[i][j] [k], wherein δ denotes a background color threshold, 0<δ<1, l denotes an l-th picture, 1≤l≤N,
a m e a n [ j ] [ k ] = i = 1 3 a norm [ i ] [ j ] [ k ] 3
denotes a pixel gray value in a j-th row and k-th column of each of the size-normalized images to be fused, anorm[i][j][k] denotes a pixel gray value of the i-th feature layer in the j-th row and k-th column of each of the size-normalized images to be fused, 1≤j≤a maximum number of rows of each of the size-normalized images to be fused, and 1≤k≤a maximum number of columns of each of the size-normalized images to be fused; and
the sample library generating module is configured to perform determining the images to be fused, normalizing the sizes of the image to be fused, and fusing the size-normalized images to be fused to form the new sample repeatedly until a preset number of new samples are acquired as a sample composition for training.
5. A target detection method, comprising:
acquiring a security check image of an item, and preprocessing the security check image;
extracting, by a preset convolutional neural network, an image feature of the preprocessed security check image;
inputting the image feature to a preset target detection model to obtain a target region of the security check image, wherein the preset target detection model is obtained by training the image sample obtained according to the image sample generating method of claim 1; and
determining a detection result of the security check image according to the obtained target region of the security check image, wherein the detection result comprises type information of contraband and location information of the contraband.
US17/910,346 2020-04-08 2020-09-08 Image sample generating method and system, and target detection method Pending US20230162342A1 (en)

Applications Claiming Priority (3)

Application Number Priority Date Filing Date Title
CN202010267813.2 2020-04-08
CN202010267813.2A CN111145177B (en) 2020-04-08 2020-04-08 Image sample generation method, specific scene target detection method and system thereof
PCT/CN2020/113998 WO2021203618A1 (en) 2020-04-08 2020-09-08 Image sample generating method and system, and target detection method

Publications (1)

Publication Number Publication Date
US20230162342A1 true US20230162342A1 (en) 2023-05-25

Family

ID=70528817

Family Applications (1)

Application Number Title Priority Date Filing Date
US17/910,346 Pending US20230162342A1 (en) 2020-04-08 2020-09-08 Image sample generating method and system, and target detection method

Country Status (3)

Country Link
US (1) US20230162342A1 (en)
CN (1) CN111145177B (en)
WO (1) WO2021203618A1 (en)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN116740220A (en) * 2023-08-16 2023-09-12 海马云(天津)信息技术有限公司 Model construction method and device, and photo generation method and device
CN116994002A (en) * 2023-09-25 2023-11-03 杭州安脉盛智能技术有限公司 Image feature extraction method, device, equipment and storage medium
CN117523341A (en) * 2023-11-23 2024-02-06 中船(北京)智能装备科技有限公司 Deep learning training image sample generation method, device and equipment

Families Citing this family (16)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111145177B (en) * 2020-04-08 2020-07-31 浙江啄云智能科技有限公司 Image sample generation method, specific scene target detection method and system thereof
CN111539957B (en) * 2020-07-07 2023-04-18 浙江啄云智能科技有限公司 Image sample generation method, system and detection method for target detection
CN111709948B (en) * 2020-08-19 2021-03-02 深兰人工智能芯片研究院(江苏)有限公司 Method and device for detecting defects of container
CN112001873B (en) * 2020-08-27 2024-05-24 中广核贝谷科技有限公司 Data generation method based on container X-ray image
CN112235476A (en) * 2020-09-15 2021-01-15 南京航空航天大学 Test data generation method based on fusion variation
CN112488044A (en) * 2020-12-15 2021-03-12 中国银行股份有限公司 Picture processing method and device
CN112560698B (en) * 2020-12-18 2024-01-16 北京百度网讯科技有限公司 Image processing method, device, equipment and medium
CN115147671A (en) * 2021-03-18 2022-10-04 杭州海康威视系统技术有限公司 Object recognition model training method and device and storage medium
CN114648494B (en) * 2022-02-28 2022-12-06 扬州市苏灵农药化工有限公司 Pesticide suspending agent production control system based on factory digitization
CN114693968A (en) * 2022-03-23 2022-07-01 成都智元汇信息技术股份有限公司 Verification method and system based on intelligent image recognition box performance
CN114495017B (en) * 2022-04-14 2022-08-09 美宜佳控股有限公司 Image processing-based ground sundry detection method, device, equipment and medium
CN114821194B (en) * 2022-05-30 2023-07-25 深圳市科荣软件股份有限公司 Equipment running state identification method and device
CN115019112A (en) * 2022-08-09 2022-09-06 威海凯思信息科技有限公司 Target object detection method and device based on image and electronic equipment
CN117253144B (en) * 2023-09-07 2024-04-12 建研防火科技有限公司 Fire risk grading management and control method
CN117372275A (en) * 2023-11-02 2024-01-09 凯多智能科技(上海)有限公司 Image dataset expansion method and device and electronic equipment
CN117689980B (en) * 2024-02-04 2024-05-24 青岛海尔科技有限公司 Method for constructing environment recognition model, method, device and equipment for recognizing environment

Family Cites Families (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH10333912A (en) * 1997-05-28 1998-12-18 Oki Electric Ind Co Ltd Fuzzy rule preparation method and device therefor
CN104463196B (en) * 2014-11-11 2017-07-25 中国人民解放军理工大学 A kind of weather phenomenon recognition methods based on video
CN108932735B (en) * 2018-07-10 2021-12-28 广州众聚智能科技有限公司 Method for generating deep learning sample
US11430140B2 (en) * 2018-09-18 2022-08-30 Caide Systems, Inc. Medical image generation, localizaton, registration system
CN109948562B (en) * 2019-03-25 2021-04-30 浙江啄云智能科技有限公司 Security check system deep learning sample generation method based on X-ray image
CN109948565B (en) * 2019-03-26 2021-05-25 浙江啄云智能科技有限公司 Method for detecting contraband in postal industry without opening box
US10482603B1 (en) * 2019-06-25 2019-11-19 Artificial Intelligence, Ltd. Medical image segmentation using an integrated edge guidance module and object segmentation network
CN110910467B (en) * 2019-12-03 2023-04-18 浙江啄云智能科技有限公司 X-ray image sample generation method, system and application
CN111145177B (en) * 2020-04-08 2020-07-31 浙江啄云智能科技有限公司 Image sample generation method, specific scene target detection method and system thereof

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN116740220A (en) * 2023-08-16 2023-09-12 海马云(天津)信息技术有限公司 Model construction method and device, and photo generation method and device
CN116994002A (en) * 2023-09-25 2023-11-03 杭州安脉盛智能技术有限公司 Image feature extraction method, device, equipment and storage medium
CN117523341A (en) * 2023-11-23 2024-02-06 中船(北京)智能装备科技有限公司 Deep learning training image sample generation method, device and equipment

Also Published As

Publication number Publication date
CN111145177B (en) 2020-07-31
WO2021203618A1 (en) 2021-10-14
CN111145177A (en) 2020-05-12

Similar Documents

Publication Publication Date Title
US20230162342A1 (en) Image sample generating method and system, and target detection method
Rogers et al. Automated x-ray image analysis for cargo security: Critical review and future promise
EP3349048B1 (en) Inspection devices and methods for detecting a firearm in a luggage
Jain An evaluation of deep learning based object detection strategies for threat object detection in baggage security imagery
US10013615B2 (en) Inspection methods and devices
CN109948565B (en) Method for detecting contraband in postal industry without opening box
CN109948562B (en) Security check system deep learning sample generation method based on X-ray image
CN109902643A (en) Intelligent safety inspection method, device, system and its electronic equipment based on deep learning
Rogers et al. A deep learning framework for the automated inspection of complex dual-energy x-ray cargo imagery
US20090175411A1 (en) Methods and systems for use in security screening, with parallel processing capability
Jaccard et al. Tackling the X-ray cargo inspection challenge using machine learning
CN105389797A (en) Unmanned aerial vehicle video small-object detecting method based on super-resolution reconstruction
WO2019154383A1 (en) Tool detection method and device
Mery et al. Computer vision for x-ray testing: Imaging, systems, image databases, and algorithms
CN110910467B (en) X-ray image sample generation method, system and application
CN111539251B (en) Security check article identification method and system based on deep learning
CN105510364A (en) Nondestructive testing system for industrial part flaws based on X rays and detection method thereof
Gupta et al. Predicting detection performance on security X-ray images as a function of image quality
CN115439693A (en) Training method of target recognition network model, electronic device and program product
Chumuang et al. Analysis of X-ray for locating the weapon in the vehicle by using scale-invariant features transform
Chouai et al. CH-Net: Deep adversarial autoencoders for semantic segmentation in X-ray images of cabin baggage screening at airports
Zuo et al. A SAR-to-optical image translation method based on PIX2PIX
US11103198B2 (en) Projection of objects in CT X-ray images
CN110992324B (en) Intelligent dangerous goods detection method and system based on X-ray image
CN115081469A (en) Article category identification method, device and equipment based on X-ray security inspection equipment

Legal Events

Date Code Title Description
AS Assignment

Owner name: ZHEJIANG PECKERAI TECHNOLOGY CO., LTD, CHINA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:LI, YIQING;ZHOU, KAI;REEL/FRAME:061040/0098

Effective date: 20220824

STPP Information on status: patent application and granting procedure in general

Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION