WO2019232831A1 - Method and device for recognizing foreign object debris at airport, computer apparatus, and storage medium - Google Patents

Method and device for recognizing foreign object debris at airport, computer apparatus, and storage medium Download PDF

Info

Publication number
WO2019232831A1
WO2019232831A1 PCT/CN2018/092614 CN2018092614W WO2019232831A1 WO 2019232831 A1 WO2019232831 A1 WO 2019232831A1 CN 2018092614 W CN2018092614 W CN 2018092614W WO 2019232831 A1 WO2019232831 A1 WO 2019232831A1
Authority
WO
WIPO (PCT)
Prior art keywords
image
detection
foreign object
recognition
full
Prior art date
Application number
PCT/CN2018/092614
Other languages
French (fr)
Chinese (zh)
Inventor
叶明�
Original Assignee
平安科技(深圳)有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 平安科技(深圳)有限公司 filed Critical 平安科技(深圳)有限公司
Publication of WO2019232831A1 publication Critical patent/WO2019232831A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/35Categorising the entire scene, e.g. birthday party or wedding scene
    • G06V20/38Outdoor scenes
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • G06T5/73

Definitions

  • the present application relates to the field of image recognition, and in particular, to a method, a device, a computer device, and a storage medium for identifying foreign objects in an airport.
  • FOD Form Object Debris
  • FOD refers to some kind of foreign material that may damage aircraft or systems. It is often called runway foreign body.
  • FOD includes aircraft and engine connections (nuts, screws, washers, fuses, etc.), machine tools, flying items (nails, personal documents, pens, pencils, etc.), wild animals, leaves, stones and sand, Road surface materials, wooden blocks, plastic or polyethylene materials, paper products, ice cream in the running area, etc.
  • the existing deep learning object detection models are mainly divided into two types: two-stage detection (FastRCNN, FasterRCNN, etc.) and single-step detection (Single stage detector) models (FCN, SSD, etc.).
  • two-stage detection FastRCNN, FasterRCNN, etc.
  • single-step detection Single stage detector
  • FCN single stage detector
  • the traditional two-step detection model has a very low ratio (less than one thousandth) of the object scene, the region selection is difficult and the calculation speed is slow. It is not suitable for scenarios with certain real-time requirements.
  • the traditional single-step detection model is not sensitive enough for small objects, and for small objects, the final detection position is prone to deviation.
  • An airport foreign body identification method includes:
  • the detection result is that there is a foreign object in the detection image, obtaining a position of the foreign object in the detection image as a reference position, and extracting a feature vector of the foreign object according to the reference position as a reference feature vector;
  • a recognition result is generated according to the comparison result, and the recognition result includes confirmation as a foreign object and confirmation as a non-foreign object.
  • An airport foreign body identification device includes:
  • a detection result acquisition module configured to acquire a detection image, detect the detection image by using a foreign object detection model, and obtain a detection result
  • a reference feature vector acquisition module configured to obtain a position of the foreign object in the detection image if the detection result is that there is a foreign object in the detection image, and use it as a reference position to extract the foreign object according to the reference position; Eigenvectors as the reference eigenvectors;
  • a recognition image set composition module configured to obtain consecutive predetermined frames of images according to the detection image to form a recognition image set
  • a comparison result acquisition module configured to extract a feature vector of each recognition image in the recognition image set according to the reference position, and compare the feature vector of each recognition image and the feature vector similarity of the reference feature vector to obtain a comparison result;
  • a recognition result acquisition module is configured to generate a recognition result according to the comparison result, where the recognition result includes confirmation as a foreign object and confirmation as a non-foreign object.
  • a computer device includes a memory, a processor, and computer-readable instructions stored in the memory and executable on the processor.
  • the processor executes the computer-readable instructions, the following steps are implemented:
  • the detection result is that there is a foreign object in the detection image, obtaining a position of the foreign object in the detection image as a reference position, and extracting a feature vector of the foreign object according to the reference position as a reference feature vector;
  • a recognition result is generated according to the comparison result, and the recognition result includes confirmation as a foreign object and confirmation as a non-foreign object.
  • One or more non-volatile readable storage media storing computer-readable instructions, which when executed by one or more processors, cause the one or more processors to perform the following steps:
  • the detection result is that there is a foreign object in the detection image, obtaining a position of the foreign object in the detection image as a reference position, and extracting a feature vector of the foreign object according to the reference position as a reference feature vector;
  • a recognition result is generated according to the comparison result, and the recognition result includes confirmation as a foreign object and confirmation as a non-foreign object.
  • FIG. 1 is a schematic diagram of an application environment of an airport foreign object identification method according to an embodiment of the present application
  • FIG. 2 is an example diagram of an airport foreign object identification method in an embodiment of the present application
  • FIG. 3 is an example diagram of step S10 of an airport foreign object recognition method according to an embodiment of the present application.
  • FIG. 4 is an example diagram of step S11 of an airport foreign object recognition method in an embodiment of the present application.
  • FIG. 5 is an example diagram of step S12 of the airport foreign object recognition method in an embodiment of the present application.
  • FIG. 6 is a principle block diagram of an airport foreign body identification device according to an embodiment of the present application.
  • FIG. 7 is a schematic diagram of a computer device in an embodiment of the present application.
  • the airport foreign body identification method provided in this application can be applied in the application environment shown in FIG. 1, where a client (computer device) communicates with a server through a network.
  • the client sends a detection image to the server, and the server recognizes the detection image and generates a recognition result.
  • the client computer device
  • the server can be, but is not limited to, various personal computers, notebook computers, smart phones, tablet computers, video capture devices, and portable wearable devices.
  • the server can be implemented by an independent server or a server cluster composed of multiple servers.
  • a method for identifying foreign objects at an airport is provided.
  • the method is applied to the server in FIG. 1 as an example, and includes the following steps:
  • S10 Obtain a detection image, use a foreign object detection model to detect the detection image, and obtain a detection result.
  • the detection image is formed by dividing the video data in the surveillance video of the airport into images of a predetermined frame at a certain time interval. Preferably, the detection images are sorted in chronological order, and then the detection images are detected using a foreign object detection model to obtain detection results.
  • the detection result includes detecting the presence or absence of a foreign object in the image.
  • the foreign object detection model is a pre-trained recognition model.
  • the foreign object detection model can be a two-stage detection (FastRCNN, FasterRCNN, etc.) or a single-step detection (Single stage detection detector) model (FCN, SSD Etc.) to achieve.
  • the detection image is detected through a foreign body detection model trained in advance, and the foreign body detection model outputs a detection result.
  • the foreign object detection model is used to detect the detection image. If the foreign object detection model outputs a foreign object in the detection image, the position of the foreign object in the detection image is obtained, and a corresponding feature vector is extracted as a reference feature vector based on the position. Specifically, the detected foreign objects may be uniformly scaled to a predetermined size (such as 32 * 32), and then a feature vector may be extracted as a reference feature vector. Optionally, a color histogram and a direction gradient histogram (HOG, Histogram of Gradient) of a foreign object may be extracted to form a reference feature vector.
  • HOG Histogram of Gradient
  • S30 Obtain consecutive predetermined frames of images according to the detected images to form a recognition image set.
  • the continuous predetermined frame image refers to a predetermined predetermined frame image adjacent to the detection image in the video data where the detection image is located. For example, the next 20 frames of images corresponding to the detection image are acquired to form a recognition image set.
  • S40 Extract the feature vector of each recognition image in the recognition image set according to the reference position, and compare the feature vector of each recognition image and the feature vector similarity of the reference feature vector to obtain a comparison result.
  • the feature vector of each recognition image in the recognition image set is obtained according to the reference position, and the feature vector of each recognition image and the reference feature vector are compared to obtain the feature vector similarity.
  • algorithms such as the Ming distance, Euclidean distance, or Mahalanobis distance can be used to calculate the feature vector similarity between the feature vector of each recognition image and the reference feature vector.
  • the calculated feature vector similarity is compared with a preset similarity threshold, and a comparison result is obtained, which may specifically be similar and dissimilar. For example: when the feature vector similarity is greater than or equal to the similarity threshold, the comparison result is similar; when the feature vector similarity is less than the similarity threshold, the comparison result is dissimilar.
  • S50 Generate a recognition result according to the comparison result.
  • the recognition result includes confirmation as a foreign object and confirmation as a non-foreign object.
  • the comparison result of each recognition image in the statistical recognition image set is statistically recognized.
  • the recognition result is confirmed as a foreign object.
  • the recognition result is confirmed as a non-foreign object.
  • the determination threshold may be set by identifying the number of images in the image set. For example, the determination threshold is 60%, 80%, or 90% of the number of images in the identification image set.
  • a continuous predetermined frame image of the detection image is acquired to form a recognition image set.
  • the recognition result is obtained by identifying the similarity between the feature vector of the corresponding position of the foreign body in the image set and the feature vector of the reference feature vector, and finally generating a recognition result based on the comparison result. It can avoid the misjudgment of the recognition result caused by the changes of the surrounding environment (light, shadow, etc.) in the detection image, so as to filter out some misjudged samples during the foreign object recognition process, thereby improving the accuracy of foreign object recognition at the airport.
  • detecting a detection image by using a foreign object detection model to obtain a detection result includes the following steps:
  • S11 Preprocess the detection image to obtain an image to be identified.
  • Preprocessing the detection image refers to performing enhanced processing on the detection image to improve subsequent detection accuracy.
  • the detection image is pre-processed to improve subsequent detection accuracy.
  • an image enhancement algorithm may be used to perform global enhancement or local enhancement processing on the detected image, and then sharpen processing is performed on the enhanced detected image to obtain an image to be identified.
  • the image enhancement algorithm may be a multi-scale retina algorithm, an adaptive histogram equalization algorithm, or an optimized contrast algorithm. After performing AND processing on the detection image, an image to be identified is obtained.
  • S12 Input the image to be identified into a full-difference-pyramid feature network recognition model for recognition, and obtain a classification confidence map.
  • the full-difference-pyramid feature network recognition model refers to a full-difference (DenseNet, Densely Connected, Convolutional Networks) as the coding network and a pyramid feature (RefineNet, Multi-Path, Refinement, Networks) as the decoding network according to the coding-decoding model Neural network recognition model.
  • DenseNet Densely Connected, Convolutional Networks
  • RefineNet Multi-Path, Refinement, Networks
  • the full-difference network is a splicing of the network of different layers in the neural network model, so that the input of each layer of the network includes the output of all the layers of the previous layer, which can avoid the loss of small objects during the model upsampling process.
  • the full-difference network can improve the transmission efficiency of information and gradients in the network. Each layer can directly obtain the gradient from the loss function and directly obtain the input signal. This can train a deeper network.
  • This network structure also has regularization. Effect, the full difference network improves network performance from the perspective of feature reuse. Therefore, using a full-difference network not only reduces the phenomenon of small objects missing during the upsampling process of the model, but also improves the training speed and reduces the phenomenon of overfitting.
  • Pyramid feature network is an improved multi-path network. It extracts all the information during the downsampling process and uses a long-distance network connection to obtain a high-resolution prediction network.
  • the pyramid feature network uses the features of the fine layer, so that the semantic information of the high level can be improved.
  • a large number of RCUs (residual connection units) are used in the pyramid feature network, which makes short-range connections within the pyramid feature network, which is beneficial for training.
  • the pyramid feature network also forms a long-range connection with the full-difference network, allowing the gradient to be effectively transmitted to the entire network, increasing the impact of the underlying features on the final result, and effectively improving the positioning accuracy of objects (airport foreign objects).
  • a classification confidence map refers to an image that is labeled and displayed in different ways for different categories in the image after the image to be identified is detected.
  • different colors may be used to distinguish different categories in the image to be identified.
  • the possible objects are runways, lawns, airport equipment (non-foreign objects) and airport foreign objects. Therefore, different colors can be given to the above-mentioned different types of objects in advance.
  • the full-difference-pyramid feature network recognition model is based on different judgment results of different regions in the to-be-recognized image and combined with different colors in advance to form a classification confidence map .
  • foreign objects in the airport can also be marked with more specific objects, such as: engine connections (nuts, screws, washers, fuses, etc.), machine tools, flying items (nails, personal documents, pens, pens, Pencil, etc.) and animals. These are classified into the category of foreign objects at the airport. In this way, when the foreign objects at the airport are identified, the specific foreign object type can be further determined to facilitate the formulation of appropriate treatment measures.
  • engine connections nuts, screws, washers, fuses, etc.
  • machine tools flying items
  • flying items noils, personal documents, pens, pens, Pencil, etc.
  • S13 Obtain a detection result according to the classification confidence map, and the detection result includes detecting the presence of a foreign object in the image and the absence of a foreign object in the detection image.
  • the detection results can be obtained according to different colors on the classification confidence map.
  • the detection results include detecting the presence of foreign objects in the image and the absence of foreign objects in the image. For example, if the foreign object at the airport is set to red in the preset settings, after obtaining the classification confidence map, it is determined whether a red area exists in the classification confidence map to obtain different detection results. If there is a red area in the classification confidence image, it means that there is a foreign object in the detection image, and at this time, the detection result is that there is a foreign object in the detection image. If there is no red region in the classification confidence map, it means that there is no foreign object in the detection image, and the detection result is that there is a foreign object in the detection image.
  • the detection result may be embodied in text, voice, or signal light, or a combination of at least two of text, voice, or signal light.
  • a voice prompt may be sent and a warning light may be used as a reminder to better remind relevant personnel to process.
  • the location information of the foreign object at the airport can also be obtained.
  • the detection result also includes the location information of the foreign object at the airport.
  • an identification mark may be assigned to each of the images to be identified in advance, and used to locate the image source of the image to be identified, for example, acquired by the camera device where the identification mark is located. In this way, when a red area exists in the classification confidence map, the position of the red area in the image to be identified can be obtained, and the identification of the image to be identified can be combined to obtain the actual airport foreign object corresponding to the red area in the airport. s position.
  • This embodiment obtains a to-be-recognized image by preprocessing the detection image to improve subsequent detection accuracy.
  • a full-difference-pyramid feature network recognition model is used to recognize the image to be recognized, which ensures the recognition accuracy and positioning accuracy of small objects during the recognition process, and also improves the recognition efficiency.
  • step S11 the detection image is preprocessed to obtain an image to be identified, which specifically includes the following steps:
  • S111 Use a multi-scale retinal algorithm to perform global enhancement processing on the detected image.
  • the Multi-Scale Retinex (MSR) algorithm is an image enhancement processing algorithm, which is used to reduce the influence of various factors (such as interference noise, lack of edge details, etc.) on the original unprocessed image.
  • the multi-scale retinal algorithm is used to enhance the detection image. By removing the illumination image of the detection image, retaining the reflection image, and adjusting the gray dynamic range of the detection image, the reflection information of the reflection image corresponding to the detection image is obtained. To achieve enhanced effects.
  • the multi-scale retinal algorithm is used to perform global enhancement processing on the detection image, which specifically includes:
  • N is the number of scales
  • (x, y) is the coordinate value of the detected image pixels
  • G (x, y) is the input of the multi-scale retina algorithm, that is, the gray value of the detected image
  • R (x, y) Is the output of the multi-scale retinal algorithm, that is, the gray value of the detected image after global enhancement processing
  • w n is the weight factor of the scale
  • F n (x, y) is the n-th center wrapping function
  • ⁇ n is the scale parameter of the n-th center surround function, and the coefficient K n must satisfy:
  • the gray value G (x, y) of the detected image is obtained by an image information acquisition tool, and the value of the scale parameter ⁇ n of the input n center surround functions is determined to satisfy The value of K n , and then calculate the center surrounding functions F n (x, y) and G (x, y) according to the following formula to obtain the gray value R (x, y) of the detected image after global enhancement processing :
  • ⁇ n determines the size of the neighborhood of the center surround function, and its size determines the quality of the detected image.
  • ⁇ n is larger, the selected range of the neighborhood is larger. Detect local details of an image.
  • the number of selected scales n 3, and correspondingly set:
  • the multi-scale retinal algorithm simultaneously takes into account the three gray scales of low gray, medium gray and high gray, so as to obtain better results.
  • the multi-scale retinal algorithm can achieve good self-adaptability by combining multiple scales, highlighting the texture details of dark areas of the image, and can adjust the dynamic range of the image to achieve the purpose of image enhancement.
  • S112 The Laplace operator is used to sharpen the detection image after the global enhancement processing to obtain an image to be identified.
  • Laplacian operator is a second-order differential operator, which is suitable for improving image blur caused by diffuse reflection of light.
  • Laplace operator sharpening transformation on the image can reduce the blur of the image and improve the sharpness of the image. Therefore, by performing a sharpening process on the detection image after the global enhancement processing, the edge detail features of the detection image after the global enhancement processing are highlighted, thereby improving the contour definition of the detection image after the global enhancement processing.
  • Sharpening processing refers to the transformation of sharpening an image to enhance the target boundaries and image details in the image. After the global enhancement processing of the detected image is sharpened by the Laplacian operator, the edge details of the image are enhanced and the halo is weakened, thereby protecting the details of the detected image.
  • the Laplace operator based on second-order differential is defined as:
  • Laplace operator ⁇ 2 R is:
  • the gray value of each pixel of the detection image R (x, y) after the global enhancement process is sharpened with the Laplace operator ⁇ 2 R according to the following formula: To obtain the sharpened pixel gray value, where g (x, y) is the sharpened pixel gray value.
  • the sharpened pixel gray value is replaced with the gray value at the original (x, y) pixel to obtain the image to be identified.
  • the Laplace operator ⁇ 2 R selects a four-neighbor sharpening template matrix H:
  • a four-neighbor sharpening template matrix H is used to perform Laplace operator sharpening on the detected image after global enhancement processing.
  • the multi-scale retinal algorithm is used to perform global enhancement processing on the detection image, and the detection image after the multi-scale retinal algorithm enhancement processing is used to sharpen the Laplacian operator.
  • the edge details of the image are enhanced.
  • the halo is also weakened, thereby protecting the details of the detected image.
  • the above steps are not only simple and convenient. After processing, the edge details of the image to be identified are more prominent, and the texture features of the image to be identified are enhanced, which is beneficial to improving the accuracy of the image to be identified.
  • the airport foreign object recognition method before the steps of inputting an image to be identified into a full-difference-pyramid feature network recognition model for recognition and obtaining a classification confidence map, the airport foreign object recognition method further includes:
  • S121 Obtain a training sample set, and classify and label the training images in the training sample set.
  • the training sample set includes training images, and the training images refer to the sample images used to train the full-difference-pyramid feature network recognition model.
  • the training image may be obtained by setting a video capture device or an image capture device at different locations in the airport to collect corresponding data, and the video capture device or the image capture device collects the corresponding data and sends it to the server. If the server obtains video data, the video data may be framed at a predetermined frame rate to obtain a training image.
  • Classifying and labeling training images refers to classifying different objects in the training images. For example, in the training image, the objects that may appear are runways, lawns, airport equipment (non-foreign objects), and airport foreign objects. By assigning different labeling information to different objects in the training image, the classification labeling of the training image is completed.
  • S122 Use the training images in the training sample set to classify and label the training network to obtain the target output vector.
  • the training network is trained using the training images in the training sample set to label the full difference network.
  • the training image input is set to x 0.
  • the full difference network consists of L-layer structures, and each layer is completely different.
  • the networks all contain a non-linear transformation H l ( ⁇ ).
  • the non-linear transformation may include ReLU (Recified Linear Units) and Pooling, or BN (Batch Normalization), ReLU, and convolutional layers, or BN, ReLU, and Pooling.
  • BN is to adjust the distribution of the input value of any neuron in each layer of the neural network to a normal normal distribution with a mean value of 0 and a variance of 1 by means of normalization, so that the activation input value falls in a non-linear function that is sensitive to the input Region, make the gradient larger, avoid the problem of gradient disappearance, and greatly speed up the training speed.
  • ReLU is a piecewise linear function and a one-sided suppression function. It can output all the negative values of the input to 0, while the positive values of the input remain unchanged. ReLU can realize the sparse model to better mine related features and fit the training data.
  • x l H l ([x 0 , x 1 , ..., x l-1 ]);
  • the output of the corresponding layer in the full difference network constitutes the target output vector for subsequent training of the pyramid feature network using the target output vector.
  • S123 Use the target output vector to train the pyramid feature network to obtain a full-difference-pyramid feature network recognition model.
  • the output of each layer in the target output vector in the full-difference network is connected to the RCU unit of the pyramid feature network, respectively. That is, there are RCU units in the pyramid feature network that have the same number of layers as the target output vector in the full-difference network.
  • the RCU unit refers to the unit structure extracted from the full-difference network, and specifically includes ReLU, convolution and summation.
  • the target output vectors of each layer obtained in the full-difference network are respectively subjected to ReLU, convolution and summing operations.
  • the output of each layer of the RCU unit is processed using Multi-resolution fusion to obtain different output feature maps.
  • the output feature maps of each layer of the RCU unit are adaptively processed by a convolution layer. Then upsampling to the maximum resolution of this layer. Chained residual pooling samples the output feature maps of different resolutions of the input to the same size as the maximum output feature map and then superimposes them. Finally, the superimposed output feature map is convolved by an RCU to obtain a fine feature map.
  • the function of the pyramid feature network is to fuse feature maps with different resolutions. First divide the pre-trained full-difference network into several full-difference blocks according to the resolution of the feature map, and then fuse these blocks to the right as several paths to fuse through the pyramid feature network, and finally obtain a fine feature map (subsequent connection softmax layer, and then output through bilinear interpolation).
  • the target feature vector of the full-difference network is used to train the pyramid feature network to form a preliminary training network.
  • the verification samples are then used to verify and adjust the pyramid feature network until a preset classification accuracy rate is obtained, and the training ends.
  • the preset classification accuracy can be set according to the needs of the actual recognition model.
  • a full-difference-pyramid feature network recognition model is obtained by training with a training sample set after classification and labeling, which ensures the recognition accuracy and speed of the full-difference-pyramid feature network recognition model.
  • training a full-difference network specifically includes:
  • the convolution layer is used to extract the features of the input image, and the initial convolution layer extracts the features of the training image.
  • the initial convolution layer uses a 7 * 7 convolution kernel.
  • the maximum pooling layer in the full-difference network is used for downsampling. During the sampling process, if the new sampling rate is less than the original sampling rate, it is downsampling. Max-pooling refers to the maximum value of all neurons in the area taken by the sampling function.
  • the input image through the initial convolution layer is subjected to maximum pooling processing, feature compression, main features are extracted, and network computation complexity is simplified.
  • Each full-difference network module includes a full-difference convolution layer and a full-difference activation layer.
  • the activation function in the full-difference activation layer adopts a linear activation function.
  • each fully differential network module is the combination of the outputs of all the previous modules, that is:
  • each H l ( ⁇ ) is a combination of two operations of the convolutional layer and the activation layer: Conv-> ReLU.
  • the size of the convolution kernel in the full-difference convolution layer is 3 * 3.
  • the number of features output by each H l ( ⁇ ) is the feature growth rate.
  • the feature growth rate is set to 16
  • the number of output features output by the three-layer fully-differential network module is 48.
  • the linear activation function formula is:
  • the time of the training process can be quickly converged.
  • a transmission layer is set between the fully differential network modules, and each transmission layer includes a normalization layer, a transmission activation layer, and an average pooling layer.
  • the output characteristics of each full-difference network module are increasing.
  • the output characteristic of the three-layer full-difference network module output is 48.
  • the transmission parameter is 0.6, that is, the input of the transmission layer is reduced to the original 0.6.
  • the training speed and accuracy of the full-difference network are guaranteed.
  • the loss function is implemented using a Focal Loss function:
  • p t is the prediction value of the full difference-pyramid feature network recognition model for the training image
  • y is the labeled value of the training image
  • is the adjustment parameter.
  • Loss function refers to a function that maps an event (an element in a sample space) to a real number that expresses the economic or opportunity cost associated with its event.
  • a loss function is used to measure the prediction of the full-difference-pyramid feature network recognition model. The smaller the loss function, the better the predictive ability of the recognition model.
  • the number of sample images of each classification in the training images of the training sample set may be uneven, especially the training images containing foreign objects in the airport may be less. The prediction function was selected for the loss function.
  • the Focal Loss function is used to implement the loss function.
  • the Focal loss function adds an adjustment factor (1-p t ) ⁇ , where the value of the adjustment parameter ⁇ is between [0,5].
  • P t is small.
  • the adjustment factor (1-p t ) ⁇ is close to 1, and the loss will not have a great impact.
  • the value of P t is large, it will approach 1.
  • the value of the adjustment factor approaches 0, so the loss value for the correctly classified samples is reduced.
  • the FocalLoss function is used in the process of training the full difference-pyramid feature network recognition model, which can reduce the impact of uneven sample classification on training the full difference-pyramid feature network recognition model, and has also improved the follow-up. The effect of detection accuracy.
  • an airport foreign body identification device corresponds to the airport foreign body identification method in the above embodiment in a one-to-one correspondence.
  • the airport foreign object recognition device includes a detection result acquisition module 10, a reference feature vector acquisition module 20, a recognition image set composition module 30, a comparison result acquisition module 40, and a recognition result acquisition module 50.
  • the detailed description of each function module is as follows:
  • the detection result acquiring module 10 is configured to acquire a detection image, detect the detection image by using a foreign object detection model, and obtain a detection result.
  • a reference feature vector acquisition module 20 is configured to obtain a position of the foreign object in the detection image if the detection result is that there is a foreign object in the detection image, and use the reference position to extract a feature vector of the foreign object according to the reference position as a reference feature vector.
  • the recognition image set composition module 30 is configured to obtain consecutive predetermined frames of images according to the detection image to form a recognition image set.
  • the comparison result acquisition module 40 is configured to extract a feature vector of each recognition image in the recognition image set according to the reference position, and compare the feature vector of each recognition image and the feature vector similarity of the reference feature vector to obtain a comparison result.
  • the recognition result acquisition module 50 is configured to generate a recognition result according to the comparison result.
  • the recognition result includes confirmation as a foreign object and confirmation as a non-foreign object.
  • the detection result acquisition module 10 includes an image-to-be-identified image acquisition unit 11, a classification confidence map acquisition unit 12, and a detection result acquisition unit 13.
  • the to-be-recognized image obtaining unit 11 is configured to preprocess the detection image to obtain the to-be-recognized image.
  • a classification confidence map acquisition unit 12 is configured to input an image to be identified into a full-difference-pyramid feature network recognition model for recognition, and obtain a classification confidence map.
  • the detection result obtaining unit 13 is configured to obtain a detection result according to the classification confidence map, and the detection result includes detecting the presence of a foreign object in the image and the absence of a foreign object in the detection image.
  • the image acquisition unit 11 includes a global enhancement processing sub-unit 111 and a sharpening processing sub-unit 112.
  • the global enhancement processing sub-unit 111 is configured to perform global enhancement processing on the original image by using a multi-scale retinal algorithm.
  • a sharpening processing unit 112 is configured to use a Laplacian to sharpen the original image after global enhancement processing to obtain an image to be identified.
  • the airport foreign body identification device further includes a training sample set acquisition module 121, a target output vector acquisition module 122, and a recognition model acquisition module 123.
  • the training sample set obtaining module 121 is configured to obtain a training sample set and classify and label the training images in the training sample set.
  • a target output vector obtaining module 122 is configured to train a full difference network using training images in which training samples are classified and labeled to obtain a target output vector.
  • a recognition model acquisition module 123 is used to train a pyramid feature network using a target output vector to obtain a full-difference-pyramid feature network recognition model.
  • Each module in the above-mentioned airport foreign body identification device may be implemented in whole or in part by software, hardware, and a combination thereof.
  • the above-mentioned modules may be embedded in the hardware in or independent of the processor in the computer device, or may be stored in the memory of the computer device in the form of software, so that the processor can call and execute the operations corresponding to the above modules.
  • a computer device is provided.
  • the computer device may be a server, and its internal structure diagram may be as shown in FIG. 7.
  • the computer device includes a processor, a memory, a network interface, and a database connected through a system bus.
  • the processor of the computer device is used to provide computing and control capabilities.
  • the memory of the computer device includes a non-volatile storage medium and an internal memory.
  • the non-volatile storage medium stores an operating system, computer-readable instructions, and a database.
  • the internal memory provides an environment for the operation of the operating system and computer-readable instructions in a non-volatile storage medium.
  • the database of the computer equipment is used to store detection images and foreign object detection model data.
  • the network interface of the computer device is used to communicate with an external terminal through a network connection.
  • the computer-readable instructions are executed by a processor to implement an airport foreign object identification method.
  • a computer device including a memory, a processor, and computer-readable instructions stored on the memory and executable on the processor.
  • the processor executes the computer-readable instructions, the following steps are implemented:
  • the detection result is that there is a foreign object in the detection image
  • the position of the foreign object in the detection image is obtained as a reference position
  • a feature vector of the foreign object is extracted according to the reference position as a reference feature vector.
  • Consecutive predetermined frame images are acquired according to the detected images to form a recognition image set.
  • the feature vector of each recognition image in the recognition image set is extracted according to the reference position, and the feature vector of each recognition image and the feature vector similarity of the reference feature vector are compared to obtain a comparison result.
  • a recognition result is generated based on the comparison result, and the recognition result includes confirmation as a foreign object and confirmation as a non-foreign object.
  • one or more non-volatile readable storage media storing computer-readable instructions are provided, and when the computer-readable instructions are executed by one or more processors, the one or more Each processor performs the following steps:
  • the detection result is that there is a foreign object in the detection image
  • the position of the foreign object in the detection image is obtained as a reference position
  • a feature vector of the foreign object is extracted according to the reference position as a reference feature vector.
  • Consecutive predetermined frame images are acquired according to the detected images to form a recognition image set.
  • the feature vector of each recognition image in the recognition image set is extracted according to the reference position, and the feature vector of each recognition image and the feature vector similarity of the reference feature vector are compared to obtain a comparison result.
  • a recognition result is generated based on the comparison result, and the recognition result includes confirmation as a foreign object and confirmation as a non-foreign object.
  • Non-volatile memory may include read-only memory (ROM), programmable ROM (PROM), electrically programmable ROM (EPROM), electrically erasable programmable ROM (EEPROM), or flash memory.
  • Volatile memory can include random access memory (RAM) or external cache memory.
  • RAM is available in various forms, such as static RAM (SRAM), dynamic RAM (DRAM), synchronous DRAM (SDRAM), dual data rate SDRAM (DDRSDRAM), enhanced SDRAM (ESDRAM), synchronous chain (Synchlink) DRAM (SLDRAM), memory bus (Rambus) direct RAM (RDRAM), direct memory bus dynamic RAM (DRDRAM), and memory bus dynamic RAM (RDRAM).

Abstract

A method and device for recognizing foreign object debris at airport, a computer apparatus, and a storage medium. The method comprises: if a foreign object debris detection model detects foreign object debris in an image undergoing detection, acquiring a preset number of images adjacent to said image to form a set of recognition images; acquiring a comparison result according to the level of similarity between feature vectors of the foreign object debris at corresponding positions in the set of recognition images and a reference feature vector; and generating a recognition result according to the comparison result. The invention can prevent an image undergoing detection from being affected by changes in the surrounding environment, such as changes in light and shadow, so as to avoid incorrect determination of a recognition result, such that certain incorrectly determined samples can be filtered out in a process of foreign object debris recognition, thereby enhancing precision of foreign object debris recognition.

Description

机场异物识别方法、装置、计算机设备及存储介质Airport foreign object identification method, device, computer equipment and storage medium
本申请以2018年06月06日提交的申请号为201810574129.1,名称为“机场异物识别方法、装置、计算机设备及存储介质”的中国发明专利申请为基础,并要求其优先权。This application is based on a Chinese invention patent application filed on June 06, 2018 with application number 201810574129.1 and entitled "Airport Foreign Object Identification Method, Device, Computer Equipment, and Storage Medium", and claims its priority.
技术领域Technical field
本申请涉及图像识别领域,尤其涉及一种机场异物识别方法、装置、计算机设备及存储介质。The present application relates to the field of image recognition, and in particular, to a method, a device, a computer device, and a storage medium for identifying foreign objects in an airport.
背景技术Background technique
在机场跑道中经常会出现各种异常物体,被称为FOD(Foreign Object Debris),FOD泛指可能损伤航空器或系统的某种外来的物质,常称为跑道异物。FOD的种类相当多,如飞机和发动机连接件(螺帽、螺钉、垫圈、保险丝等)、机械工具、飞行物品(钉子、私人证件、钢笔、铅笔等)、野生动物、树叶、石头和沙子、道面材料、木块、塑料或聚乙烯材料、纸制品、运行区的冰碴儿等等。实验和案例都表明,机场道面上的外来物可以很容易被吸入发动机,导致发动机失效。碎片也会堆积在机械装置中,影响起落架、襟翼等设备的正常运行。Various anomalous objects often appear in airport runways. They are called FOD (Foreign Object Debris). FOD refers to some kind of foreign material that may damage aircraft or systems. It is often called runway foreign body. There are many types of FOD, such as aircraft and engine connections (nuts, screws, washers, fuses, etc.), machine tools, flying items (nails, personal documents, pens, pencils, etc.), wild animals, leaves, stones and sand, Road surface materials, wooden blocks, plastic or polyethylene materials, paper products, ice cream in the running area, etc. Experiments and cases have shown that foreign objects on the airport pavement can be easily drawn into the engine, causing the engine to fail. Debris can also accumulate in mechanical devices, affecting the normal operation of equipment such as landing gear and flaps.
而由于人工智能的发展,开始尝试用深度学习物体检测模型来实现对机场异物的检测。然而,现有深度学习物体检测模型主要分为两步检测(Two stage detecotr)模型(FastRCNN,FasterRCNN等)和单步检测(Single stage detector)模型(FCN,SSD等)两类。传统的两步检测模型对于物体场景占比率极低(不足千分之一)的情况下,区域选取困难且运算速度慢。不适合有一定实时性要求的场景。而传统的单步检测模型对于微小物体不够敏感,且对于微小物体而言,最终检测位置容易产生偏差。However, due to the development of artificial intelligence, attempts have been made to use deep learning object detection models to detect foreign objects in airports. However, the existing deep learning object detection models are mainly divided into two types: two-stage detection (FastRCNN, FasterRCNN, etc.) and single-step detection (Single stage detector) models (FCN, SSD, etc.). In the case where the traditional two-step detection model has a very low ratio (less than one thousandth) of the object scene, the region selection is difficult and the calculation speed is slow. It is not suitable for scenarios with certain real-time requirements. The traditional single-step detection model is not sensitive enough for small objects, and for small objects, the final detection position is prone to deviation.
发明内容Summary of the Invention
基于此,有必要针对上述技术问题,提供一种可以提高机场异物识别精度的机场异物识别方法、装置、计算机设备及存储介质。Based on this, it is necessary to provide an airport foreign object recognition method, device, computer equipment, and storage medium that can improve the accuracy of airport foreign object recognition in view of the above technical problems.
一种机场异物识别方法,包括:An airport foreign body identification method includes:
获取检测图像,采用异物检测模型对所述检测图像进行检测,获取检测结果;Acquiring a detection image, detecting the detection image by using a foreign object detection model, and acquiring a detection result;
若所述检测结果为所述检测图像中存在异物,则获取所述异物在所述检测图像中的位置,作为基准位置,根据所述基准位置提取所述异物的特征向量,作为基准特征向量;If the detection result is that there is a foreign object in the detection image, obtaining a position of the foreign object in the detection image as a reference position, and extracting a feature vector of the foreign object according to the reference position as a reference feature vector;
根据所述检测图像获取连续预定帧图像,组成识别图像集;Acquiring successive predetermined frames of images according to the detected images to form a recognition image set;
根据所述基准位置提取所述识别图像集中每一识别图像的特征向量,并比对每一识别图像的特征向量和所述基准特征向量的特征向量相似度,获取比较结果;Extracting a feature vector of each recognition image in the recognition image set according to the reference position, and comparing the feature vector of each recognition image and the feature vector similarity of the reference feature vector to obtain a comparison result;
根据所述比较结果生成识别结果,所述识别结果包括确认为异物和确认为非异物。A recognition result is generated according to the comparison result, and the recognition result includes confirmation as a foreign object and confirmation as a non-foreign object.
一种机场异物识别装置,包括:An airport foreign body identification device includes:
检测结果获取模块,用于获取检测图像,采用异物检测模型对所述检测图像进行检测,获取检测结果;A detection result acquisition module, configured to acquire a detection image, detect the detection image by using a foreign object detection model, and obtain a detection result;
基准特征向量获取模块,用于若所述检测结果为所述检测图像中存在异物,则获取所述异物在所述检测图像中的位置,作为基准位置,根据所述基准位置提取所述异物的特征向量,作为基准特征向量;A reference feature vector acquisition module, configured to obtain a position of the foreign object in the detection image if the detection result is that there is a foreign object in the detection image, and use it as a reference position to extract the foreign object according to the reference position; Eigenvectors as the reference eigenvectors;
识别图像集组成模块,用于根据所述检测图像获取连续预定帧图像,组成识别图像集;A recognition image set composition module, configured to obtain consecutive predetermined frames of images according to the detection image to form a recognition image set;
比较结果获取模块,用于根据所述基准位置提取所述识别图像集中每一识别图像的特征向量,并比对每一识别图像的特征向量和所述基准特征向量的特征向量相似度,获取比 较结果;A comparison result acquisition module, configured to extract a feature vector of each recognition image in the recognition image set according to the reference position, and compare the feature vector of each recognition image and the feature vector similarity of the reference feature vector to obtain a comparison result;
识别结果获取模块,用于根据所述比较结果生成识别结果,所述识别结果包括确认为异物和确认为非异物。A recognition result acquisition module is configured to generate a recognition result according to the comparison result, where the recognition result includes confirmation as a foreign object and confirmation as a non-foreign object.
一种计算机设备,包括存储器、处理器以及存储在所述存储器中并可在所述处理器上运行的计算机可读指令,所述处理器执行所述计算机可读指令时实现如下步骤:A computer device includes a memory, a processor, and computer-readable instructions stored in the memory and executable on the processor. When the processor executes the computer-readable instructions, the following steps are implemented:
获取检测图像,采用异物检测模型对所述检测图像进行检测,获取检测结果;Acquiring a detection image, detecting the detection image by using a foreign object detection model, and acquiring a detection result;
若所述检测结果为所述检测图像中存在异物,则获取所述异物在所述检测图像中的位置,作为基准位置,根据所述基准位置提取所述异物的特征向量,作为基准特征向量;If the detection result is that there is a foreign object in the detection image, obtaining a position of the foreign object in the detection image as a reference position, and extracting a feature vector of the foreign object according to the reference position as a reference feature vector;
根据所述检测图像获取连续预定帧图像,组成识别图像集;Acquiring successive predetermined frames of images according to the detected images to form a recognition image set;
根据所述基准位置提取所述识别图像集中每一识别图像的特征向量,并比对每一识别图像的特征向量和所述基准特征向量的特征向量相似度,获取比较结果;Extracting a feature vector of each recognition image in the recognition image set according to the reference position, and comparing the feature vector of each recognition image and the feature vector similarity of the reference feature vector to obtain a comparison result;
根据所述比较结果生成识别结果,所述识别结果包括确认为异物和确认为非异物。A recognition result is generated according to the comparison result, and the recognition result includes confirmation as a foreign object and confirmation as a non-foreign object.
一个或多个存储有计算机可读指令的非易失性可读存储介质,所述计算机可读指令被一个或多个处理器执行时,使得所述一个或多个处理器执行如下步骤:One or more non-volatile readable storage media storing computer-readable instructions, which when executed by one or more processors, cause the one or more processors to perform the following steps:
获取检测图像,采用异物检测模型对所述检测图像进行检测,获取检测结果;Acquiring a detection image, detecting the detection image by using a foreign object detection model, and acquiring a detection result;
若所述检测结果为所述检测图像中存在异物,则获取所述异物在所述检测图像中的位置,作为基准位置,根据所述基准位置提取所述异物的特征向量,作为基准特征向量;If the detection result is that there is a foreign object in the detection image, obtaining a position of the foreign object in the detection image as a reference position, and extracting a feature vector of the foreign object according to the reference position as a reference feature vector;
根据所述检测图像获取连续预定帧图像,组成识别图像集;Acquiring successive predetermined frames of images according to the detected images to form a recognition image set;
根据所述基准位置提取所述识别图像集中每一识别图像的特征向量,并比对每一识别图像的特征向量和所述基准特征向量的特征向量相似度,获取比较结果;Extracting a feature vector of each recognition image in the recognition image set according to the reference position, and comparing the feature vector of each recognition image and the feature vector similarity of the reference feature vector to obtain a comparison result;
根据所述比较结果生成识别结果,所述识别结果包括确认为异物和确认为非异物。A recognition result is generated according to the comparison result, and the recognition result includes confirmation as a foreign object and confirmation as a non-foreign object.
本申请的一个或多个实施例的细节在下面的附图和描述中提出,本申请的其他特征和优点将从说明书、附图以及权利要求变得明显。Details of one or more embodiments of the present application are set forth in the accompanying drawings and description below, and other features and advantages of the present application will become apparent from the description, the drawings, and the claims.
附图说明BRIEF DESCRIPTION OF THE DRAWINGS
为了更清楚地说明本申请实施例的技术方案,下面将对本申请实施例的描述中所需要使用的附图作简单地介绍,显而易见地,下面描述中的附图仅仅是本申请的一些实施例,对于本领域普通技术人员来讲,在不付出创造性劳动性的前提下,还可以根据这些附图获得其他的附图。In order to explain the technical solutions of the embodiments of the present application more clearly, the drawings used in the description of the embodiments of the application will be briefly introduced below. Obviously, the drawings in the following description are just some embodiments of the application. For those of ordinary skill in the art, other drawings can be obtained based on these drawings without paying creative labor.
图1是本申请一实施例中机场异物识别方法的一应用环境示意图;FIG. 1 is a schematic diagram of an application environment of an airport foreign object identification method according to an embodiment of the present application; FIG.
图2是本申请一实施例中机场异物识别方法的一示例图;FIG. 2 is an example diagram of an airport foreign object identification method in an embodiment of the present application; FIG.
图3是本申请一实施例中机场异物识别方法的步骤S10的一示例图;FIG. 3 is an example diagram of step S10 of an airport foreign object recognition method according to an embodiment of the present application;
图4是本申请一实施例中机场异物识别方法的步骤S11的一示例图;FIG. 4 is an example diagram of step S11 of an airport foreign object recognition method in an embodiment of the present application; FIG.
图5是本申请一实施例中机场异物识别方法的步骤S12的一示例图;FIG. 5 is an example diagram of step S12 of the airport foreign object recognition method in an embodiment of the present application; FIG.
图6是本申请一实施例中机场异物识别装置的一原理框图;FIG. 6 is a principle block diagram of an airport foreign body identification device according to an embodiment of the present application; FIG.
图7是本申请一实施例中计算机设备的一示意图。FIG. 7 is a schematic diagram of a computer device in an embodiment of the present application.
具体实施方式Detailed ways
下面将结合本申请实施例中的附图,对本申请实施例中的技术方案进行清楚、完整地描述,显然,所描述的实施例是本申请一部分实施例,而不是全部的实施例。基于本申请中的实施例,本领域普通技术人员在没有作出创造性劳动前提下所获得的所有其他实施例,都属于本申请保护的范围。In the following, the technical solutions in the embodiments of the present application will be clearly and completely described with reference to the drawings in the embodiments of the present application. Obviously, the described embodiments are part of the embodiments of the present application, but not all of the embodiments. Based on the embodiments in this application, all other embodiments obtained by a person of ordinary skill in the art without creative efforts shall fall within the protection scope of this application.
本申请提供的机场异物识别方法,可应用在如图1的应用环境中,其中,客户端(计算机设备)通过网络与服务端进行通信。客户端发送检测图像到服务端中,服务端对检测图像进行识别,生成识别结果。其中,客户端(计算机设备)可以但不限于是各种个人计 算机、笔记本电脑、智能手机、平板电脑、视频采集设备和便携式可穿戴设备。服务端可以用独立的服务器或者是多个服务器组成的服务器集群来实现。The airport foreign body identification method provided in this application can be applied in the application environment shown in FIG. 1, where a client (computer device) communicates with a server through a network. The client sends a detection image to the server, and the server recognizes the detection image and generates a recognition result. Among them, the client (computer device) can be, but is not limited to, various personal computers, notebook computers, smart phones, tablet computers, video capture devices, and portable wearable devices. The server can be implemented by an independent server or a server cluster composed of multiple servers.
在一实施例中,如图2所示,提供一种机场异物识别方法,以该方法应用在图1中的服务端为例进行说明,包括如下步骤:In one embodiment, as shown in FIG. 2, a method for identifying foreign objects at an airport is provided. The method is applied to the server in FIG. 1 as an example, and includes the following steps:
S10:获取检测图像,采用异物检测模型对检测图像进行检测,获取检测结果。S10: Obtain a detection image, use a foreign object detection model to detect the detection image, and obtain a detection result.
其中,检测图像是指将机场的监控视频中的视频数据按照一定的时间间隔分割成预定帧的图像而形成的。优选地,将检测图像按照时间先后顺序进行排序,然后采用异物检测模型对检测图像进行检测,获取检测结果。可选地,检测结果包括检测图像中存在异物和不存在异物两种情况。异物检测模型是一个预先训练好的识别模型,可选地,异物检测模型可以采用两步检测(Two stage detecotr)模型(FastRCNN,FasterRCNN等)或单步检测(Single stage detector)模型(FCN,SSD等)实现。通过预先训练好的异物检测模型对检测图像进行检测,异物检测模型输出一个检测结果。The detection image is formed by dividing the video data in the surveillance video of the airport into images of a predetermined frame at a certain time interval. Preferably, the detection images are sorted in chronological order, and then the detection images are detected using a foreign object detection model to obtain detection results. Optionally, the detection result includes detecting the presence or absence of a foreign object in the image. The foreign object detection model is a pre-trained recognition model. Optionally, the foreign object detection model can be a two-stage detection (FastRCNN, FasterRCNN, etc.) or a single-step detection (Single stage detection detector) model (FCN, SSD Etc.) to achieve. The detection image is detected through a foreign body detection model trained in advance, and the foreign body detection model outputs a detection result.
S20:若检测结果为检测图像中存在异物,则获取异物在检测图像中的位置,作为基准位置,根据基准位置提取异物的特征向量,作为基准特征向量。S20: If the detection result is that there is a foreign object in the detection image, obtain the position of the foreign object in the detection image as a reference position, and extract a feature vector of the foreign object according to the reference position as a reference feature vector.
采用异物检测模型对检测图像进行检测,若异物检测模型输出的结果是该检测图像中存在异物,则获取异物在检测图像中的位置,并基于该位置提取对应的特征向量作为基准特征向量。具体地,可以将检测到的异物先统一缩放到预定大小(比如32*32),然后提取特征向量,作为基准特征向量。可选地,可以提取异物的颜色直方图和方向梯度直方图(HOG,Histogram of Gradient)组成基准特征向量。The foreign object detection model is used to detect the detection image. If the foreign object detection model outputs a foreign object in the detection image, the position of the foreign object in the detection image is obtained, and a corresponding feature vector is extracted as a reference feature vector based on the position. Specifically, the detected foreign objects may be uniformly scaled to a predetermined size (such as 32 * 32), and then a feature vector may be extracted as a reference feature vector. Optionally, a color histogram and a direction gradient histogram (HOG, Histogram of Gradient) of a foreign object may be extracted to form a reference feature vector.
S30:根据检测图像获取连续预定帧图像,组成识别图像集。S30: Obtain consecutive predetermined frames of images according to the detected images to form a recognition image set.
基于该检测图像获取对应的连续预定帧图像,组成识别图像集。连续预定帧图像是指在检测图像所在的视频数据中和检测图像相邻且连续的预定帧图像。例如,获取该检测图像对应的下20帧图像,组成识别图像集。Based on the detected image, corresponding consecutive predetermined frames of images are acquired to form a recognition image set. The continuous predetermined frame image refers to a predetermined predetermined frame image adjacent to the detection image in the video data where the detection image is located. For example, the next 20 frames of images corresponding to the detection image are acquired to form a recognition image set.
S40:根据基准位置提取识别图像集中每一识别图像的特征向量,并比对每一识别图像的特征向量和基准特征向量的特征向量相似度,获取比较结果。S40: Extract the feature vector of each recognition image in the recognition image set according to the reference position, and compare the feature vector of each recognition image and the feature vector similarity of the reference feature vector to obtain a comparison result.
根据基准位置获取到识别图像集中每一识别图像的特征向量,并将每一识别图像的特征向量和基准特征向量进行比较,获取特征向量相似度。具体地,可以采用明氏距离、欧氏距离或马氏距离等算法来计算每一识别图像的特征向量和基准特征向量的特征向量相似度。其中,将计算得到的特征向量相似度和预设的相似度阈值进行比较,并得出一比较结果,具体可以为:相似和不相似。例如:当特征向量相似度大于或等于相似度阈值时,比较结果为相似;当特征向量相似度小于相似度阈值时,比较结果为不相似。The feature vector of each recognition image in the recognition image set is obtained according to the reference position, and the feature vector of each recognition image and the reference feature vector are compared to obtain the feature vector similarity. Specifically, algorithms such as the Ming distance, Euclidean distance, or Mahalanobis distance can be used to calculate the feature vector similarity between the feature vector of each recognition image and the reference feature vector. The calculated feature vector similarity is compared with a preset similarity threshold, and a comparison result is obtained, which may specifically be similar and dissimilar. For example: when the feature vector similarity is greater than or equal to the similarity threshold, the comparison result is similar; when the feature vector similarity is less than the similarity threshold, the comparison result is dissimilar.
S50:根据比较结果生成识别结果,识别结果包括确认为异物和确认为非异物。S50: Generate a recognition result according to the comparison result. The recognition result includes confirmation as a foreign object and confirmation as a non-foreign object.
统计识别图像集中每一识别图像的比较结果,当比较结果中相似的数量大于或等于判定阈值时,识别结果为确认为异物。当比较结果中相似的数量小于判定阈值时,识别结果为确认为非异物。该判定阈值可以通过识别图像集中的图像的数量来设定,例如,判定阈值为识别图像集中的图像的数量的60%、80%或者90%。The comparison result of each recognition image in the statistical recognition image set is statistically recognized. When the number of similarities in the comparison result is greater than or equal to the determination threshold, the recognition result is confirmed as a foreign object. When the number of similarities in the comparison result is less than the determination threshold, the recognition result is confirmed as a non-foreign object. The determination threshold may be set by identifying the number of images in the image set. For example, the determination threshold is 60%, 80%, or 90% of the number of images in the identification image set.
在检测图像中时,有可能是因为周围环境变化(光照,阴影等)的影响而在检测图像中形成了阴影。在采用异物检测模型对检测图像进行检测时,就有可能将该阴影认定成异物。在异物检测模型的检测结果为存在异物时,通过进一步比对检测图像连续预定帧图像中对应位置处的特征向量相似度,来排除检测图像中存在的阴影对识别结果的影响。When detecting an image, it is possible that shadows are formed in the detection image due to the influence of changes in the surrounding environment (lighting, shadows, etc.). When using a foreign object detection model to detect a detection image, it is possible to identify the shadow as a foreign object. When the detection result of the foreign object detection model is the presence of foreign objects, the influence of the presence of shadows in the detection image on the recognition results is excluded by further comparing the similarity of the feature vectors at corresponding positions in the consecutive predetermined frames of the detection image.
在本实施例中,在通过异物检测模型检测出检测图像中存在异物时,通过获取该检测图像的连续预定帧图像,组成识别图像集。通过识别图像集中异物对应位置的特征向量和基准特征向量的特征向量相似度,来获取比较结果,最后根据比较结果生成识别结果。可以避免检测图像受周围环境变化(光照,阴影等)影响而对识别结果造成的误判,以在异物识别过程中筛除一部分错判样本,从而提高了机场异物的识别精度。In this embodiment, when a foreign object is detected in the detection image through the foreign object detection model, a continuous predetermined frame image of the detection image is acquired to form a recognition image set. The recognition result is obtained by identifying the similarity between the feature vector of the corresponding position of the foreign body in the image set and the feature vector of the reference feature vector, and finally generating a recognition result based on the comparison result. It can avoid the misjudgment of the recognition result caused by the changes of the surrounding environment (light, shadow, etc.) in the detection image, so as to filter out some misjudged samples during the foreign object recognition process, thereby improving the accuracy of foreign object recognition at the airport.
在一实施例中,如图3所示,采用异物检测模型对检测图像进行检测,获取检测结果,包括如下步骤:In an embodiment, as shown in FIG. 3, detecting a detection image by using a foreign object detection model to obtain a detection result includes the following steps:
S11:对检测图像进行预处理,得到待识别图像。S11: Preprocess the detection image to obtain an image to be identified.
对检测图像进行预处理是指对检测图像进行增强处理,以提高后续的检测精度。在获取检测图像时,影响检测图像的因素有很多,例如:光照度不均匀、采集设备的限制和采集环境的不同等都会导致检测图像的清晰度不够,导致后续的识别精度的降低。因此,在该步骤中通过对检测图像进行预处理,以提高后续的检测精度。可选地,可以对检测图像采用图像增强算法进行全局增强或者局部增强处理,再对增强后的检测图像进行锐化处理,得到待识别图像。优选地,图像增强算法可以为多尺度视网膜算法、自适应直方图均衡化算法或者优化对比度算法等。通过对检测图像进行与处理之后,就得到待识别图像。Preprocessing the detection image refers to performing enhanced processing on the detection image to improve subsequent detection accuracy. When acquiring a detection image, there are many factors that affect the detection image, such as: uneven illumination, restrictions on acquisition equipment, and different acquisition environments will cause the clarity of the detection image to be insufficient, resulting in a reduction in subsequent recognition accuracy. Therefore, in this step, the detection image is pre-processed to improve subsequent detection accuracy. Optionally, an image enhancement algorithm may be used to perform global enhancement or local enhancement processing on the detected image, and then sharpen processing is performed on the enhanced detected image to obtain an image to be identified. Preferably, the image enhancement algorithm may be a multi-scale retina algorithm, an adaptive histogram equalization algorithm, or an optimized contrast algorithm. After performing AND processing on the detection image, an image to be identified is obtained.
S12:将待识别图像输入到全差-金字塔特征网络识别模型中进行识别,获取分类置信图。S12: Input the image to be identified into a full-difference-pyramid feature network recognition model for recognition, and obtain a classification confidence map.
其中,全差-金字塔特征网络识别模型是指根据编码-解码模型,采用全差(DenseNet,Densely Connected Convolutional Networks)作为编码网络,采用金字塔特征(RefineNet,Multi-Path Refinement Networks)作为解码网络而构成的神经网络识别模型。Among them, the full-difference-pyramid feature network recognition model refers to a full-difference (DenseNet, Densely Connected, Convolutional Networks) as the coding network and a pyramid feature (RefineNet, Multi-Path, Refinement, Networks) as the decoding network according to the coding-decoding model Neural network recognition model.
具体地,全差网络是在神经网络模型通过将不同层的网络做一个拼接,使得每一层网络的输入包括前面所有层网络的输出,这样可以避免模型上采样过程中微小物体丢失。全差网络可以提升信息和梯度在网络中的传输效率,每层都能直接从损失函数拿到梯度,并且直接得到输入信号,这样就能训练更深的网络,这种网络结构还有正则化的效果,全差网络从特征重用的角度来提升网络性能。因此,采用全差网络不仅降低了模型上采样过程中微小物体丢失的现象,同时也提高了训练速度并减小了过拟合现象。Specifically, the full-difference network is a splicing of the network of different layers in the neural network model, so that the input of each layer of the network includes the output of all the layers of the previous layer, which can avoid the loss of small objects during the model upsampling process. The full-difference network can improve the transmission efficiency of information and gradients in the network. Each layer can directly obtain the gradient from the loss function and directly obtain the input signal. This can train a deeper network. This network structure also has regularization. Effect, the full difference network improves network performance from the perspective of feature reuse. Therefore, using a full-difference network not only reduces the phenomenon of small objects missing during the upsampling process of the model, but also improves the training speed and reduces the phenomenon of overfitting.
金字塔特征网络是一个多路径的改进网络,其提取下采样过程中所有信息,使用长距离网络连接获得高分辨率的预测网络。金字塔特征网络用精细层的特征,使得高层的语义信息可以得到改善。金字塔特征网络中使用了较多的RCU(残差连接单元,residual connection units),使得金字塔特征网络内部形成了short-range的连接,对训练有益。此外,金字塔特征网络还与全差网络形成了long-range的连接,让梯度能够有效传送到整个网络中,增加了底层特征对最终结果的影响,有效提高了物体(机场异物)的定位精度。Pyramid feature network is an improved multi-path network. It extracts all the information during the downsampling process and uses a long-distance network connection to obtain a high-resolution prediction network. The pyramid feature network uses the features of the fine layer, so that the semantic information of the high level can be improved. A large number of RCUs (residual connection units) are used in the pyramid feature network, which makes short-range connections within the pyramid feature network, which is beneficial for training. In addition, the pyramid feature network also forms a long-range connection with the full-difference network, allowing the gradient to be effectively transmitted to the entire network, increasing the impact of the underlying features on the final result, and effectively improving the positioning accuracy of objects (airport foreign objects).
分类置信图是指对待识别图像进行检测之后对图像中不同类别采用不同方式标注出来而呈现的图像。可选地,可以采用不同的颜色对待识别图像中不同类别进行区分。例如:在检测图像中,有可能出现的物体为跑道、草坪、机场设备(非异物)和机场异物等。因此,可以为上述不同类别的物体提前赋予不同的颜色。在将待识别图像输入到全差-金字塔特征网络识别模型中进行识别之后,全差-金字塔特征网络识别模型根据待识别图像中不同区域的不同判断结果再结合提前赋予不同的颜色形成分类置信图。A classification confidence map refers to an image that is labeled and displayed in different ways for different categories in the image after the image to be identified is detected. Optionally, different colors may be used to distinguish different categories in the image to be identified. For example: in the detection image, the possible objects are runways, lawns, airport equipment (non-foreign objects) and airport foreign objects. Therefore, different colors can be given to the above-mentioned different types of objects in advance. After inputting the image to be identified into the full-difference-pyramid feature network recognition model for recognition, the full-difference-pyramid feature network recognition model is based on different judgment results of different regions in the to-be-recognized image and combined with different colors in advance to form a classification confidence map .
在一个具体实施方式中,也可以将机场异物用更加具体的物体进行标注,例如:发动机连接件(螺帽、螺钉、垫圈、保险丝等)、机械工具、飞行物品(钉子、私人证件、钢笔、铅笔等)和动物等。并将这些都归属到机场异物的类别,如此,可以在识别出机场异物时进一步确定出具体的异物类型,方便制定合适的处理措施。In a specific embodiment, foreign objects in the airport can also be marked with more specific objects, such as: engine connections (nuts, screws, washers, fuses, etc.), machine tools, flying items (nails, personal documents, pens, pens, Pencil, etc.) and animals. These are classified into the category of foreign objects at the airport. In this way, when the foreign objects at the airport are identified, the specific foreign object type can be further determined to facilitate the formulation of appropriate treatment measures.
S13:根据分类置信图获取检测结果,检测结果包括检测图像中存在异物和检测图像中不存在异物。S13: Obtain a detection result according to the classification confidence map, and the detection result includes detecting the presence of a foreign object in the image and the absence of a foreign object in the detection image.
在获取分类置信图之后,可以根据分类置信图上的不同颜色来获取检测结果,检测结果包括检测图像中存在异物和检测图像中不存在异物。例如,若在预先的设定中,将机场异物设定为红色,则在获取到分类置信图之后,判断分类置信图中是否存在红色区域从而得到不同的检测结果。若分类置信图中存在红色区域,则说明检测图像中存在异物,此时检测结果为检测图像中存在异物。若分类置信图中不存在红色区域,则说明检测图像中不存在异物,此时检测结果为检测图像中存在异物。可选地,检测结果可以通过文字、语音 或者信号灯等方式体现,也可以是文字、语音或者信号灯至少两项的结合。例如,当检测结果为检测图像中存在异物时,可以发送语音提示,并采用警示灯的方式进行提示,以更好地提醒相关人员进行处理。After obtaining the classification confidence map, the detection results can be obtained according to different colors on the classification confidence map. The detection results include detecting the presence of foreign objects in the image and the absence of foreign objects in the image. For example, if the foreign object at the airport is set to red in the preset settings, after obtaining the classification confidence map, it is determined whether a red area exists in the classification confidence map to obtain different detection results. If there is a red area in the classification confidence image, it means that there is a foreign object in the detection image, and at this time, the detection result is that there is a foreign object in the detection image. If there is no red region in the classification confidence map, it means that there is no foreign object in the detection image, and the detection result is that there is a foreign object in the detection image. Optionally, the detection result may be embodied in text, voice, or signal light, or a combination of at least two of text, voice, or signal light. For example, when the detection result is that there is a foreign object in the detected image, a voice prompt may be sent and a warning light may be used as a reminder to better remind relevant personnel to process.
在一个实施方式中,当分类置信图中存在红色区域时,还可以获取该机场异物的位置信息,此时检测结果还包括机场异物的位置信息。具体地,可以预先为每一待识别图像赋予一识别标识,用于定位该待识别图像的图像来源,例如通过该识别标识定位到是哪个位置的摄像装置采集得到的。如此,当分类置信图中存在红色区域时,可以通过获取该红色区域在待识别图像中的位置,再结合该待识别图像的识别标识,从而得出该红色区域对应的机场异物在机场中实际的位置。In one embodiment, when there is a red area in the classification confidence map, the location information of the foreign object at the airport can also be obtained. At this time, the detection result also includes the location information of the foreign object at the airport. Specifically, an identification mark may be assigned to each of the images to be identified in advance, and used to locate the image source of the image to be identified, for example, acquired by the camera device where the identification mark is located. In this way, when a red area exists in the classification confidence map, the position of the red area in the image to be identified can be obtained, and the identification of the image to be identified can be combined to obtain the actual airport foreign object corresponding to the red area in the airport. s position.
本实施例通过对检测图像进行预处理,得到待识别图像,以提高后续的检测精度。并采用全差-金字塔特征网络识别模型对待识别图像进行识别,保证了在识别过程中对微小物体的识别精度和定位精度,也提高了识别效率。This embodiment obtains a to-be-recognized image by preprocessing the detection image to improve subsequent detection accuracy. A full-difference-pyramid feature network recognition model is used to recognize the image to be recognized, which ensures the recognition accuracy and positioning accuracy of small objects during the recognition process, and also improves the recognition efficiency.
在一实施例中,如图4所示,步骤S11中,即对检测图像进行预处理,得到待识别图像,具体包括如下步骤:In an embodiment, as shown in FIG. 4, in step S11, the detection image is preprocessed to obtain an image to be identified, which specifically includes the following steps:
S111:采用多尺度视网膜算法对检测图像进行全局增强处理。S111: Use a multi-scale retinal algorithm to perform global enhancement processing on the detected image.
其中,多尺度视网膜(Multi-Scale Retinex,MSR)算法是一种图像增强处理的算法,用于减弱未经处理的原图像的各种因素(如干扰噪声、边缘细节缺失等)的影响。采用多尺度视网膜算法对检测图像进行增强处理,通过将检测图像的照度图像去掉,保留反射图像,并对检测图像的灰度动态范围进行调整,得到检测图像对应的反射图像的反射信息,据此来达到增强效果。Among them, the Multi-Scale Retinex (MSR) algorithm is an image enhancement processing algorithm, which is used to reduce the influence of various factors (such as interference noise, lack of edge details, etc.) on the original unprocessed image. The multi-scale retinal algorithm is used to enhance the detection image. By removing the illumination image of the detection image, retaining the reflection image, and adjusting the gray dynamic range of the detection image, the reflection information of the reflection image corresponding to the detection image is obtained. To achieve enhanced effects.
优选地,采用多尺度视网膜算法对检测图像进行全局增强处理,具体包括:Preferably, the multi-scale retinal algorithm is used to perform global enhancement processing on the detection image, which specifically includes:
采用如下公式对检测图像进行全局增强处理:Use the following formula to globally enhance the detection image:
Figure PCTCN2018092614-appb-000001
Figure PCTCN2018092614-appb-000001
其中,N为尺度的个数,(x,y)为检测图像像素的坐标值,G(x,y)为多尺度视网膜算法的输入,即检测图像的灰度值,R(x,y)为多尺度视网膜算法的输出,即全局增强处理后的检测图像的灰度值,w n为尺度的权重因子,其约束条件为
Figure PCTCN2018092614-appb-000002
F n(x,y)为第n个中心环绕函数,表达式为:
Among them, N is the number of scales, (x, y) is the coordinate value of the detected image pixels, and G (x, y) is the input of the multi-scale retina algorithm, that is, the gray value of the detected image, R (x, y) Is the output of the multi-scale retinal algorithm, that is, the gray value of the detected image after global enhancement processing, w n is the weight factor of the scale, and its constraint is
Figure PCTCN2018092614-appb-000002
F n (x, y) is the n-th center wrapping function, and the expression is:
Figure PCTCN2018092614-appb-000003
Figure PCTCN2018092614-appb-000003
式中,σ n为第n个中心环绕函数的尺度参数,系数K n须满足: In the formula, σ n is the scale parameter of the n-th center surround function, and the coefficient K n must satisfy:
Figure PCTCN2018092614-appb-000004
Figure PCTCN2018092614-appb-000004
具体地,通过图像信息获取工具获取检测图像的灰度值G(x,y),根据输入的n个中心环绕函数的尺度参数σ n的值,确定满足
Figure PCTCN2018092614-appb-000005
的K n的值,然后将中心环绕函数F n(x,y)和G(x,y)依据如下公式进行计算后,得到全局增强处理后的检测图像的灰度值R(x,y):
Specifically, the gray value G (x, y) of the detected image is obtained by an image information acquisition tool, and the value of the scale parameter σ n of the input n center surround functions is determined to satisfy
Figure PCTCN2018092614-appb-000005
The value of K n , and then calculate the center surrounding functions F n (x, y) and G (x, y) according to the following formula to obtain the gray value R (x, y) of the detected image after global enhancement processing :
Figure PCTCN2018092614-appb-000006
Figure PCTCN2018092614-appb-000006
其中,σ n决定中心环绕函数邻域大小,其大小决定了检测图像的质量,σ n取较大时, 选取的邻域范围就较大,检测图像像素受到其周围像素的影响越小,突出检测图像的局部细节。 Among them, σ n determines the size of the neighborhood of the center surround function, and its size determines the quality of the detected image. When σ n is larger, the selected range of the neighborhood is larger. Detect local details of an image.
在一具体实施方式中,选取尺度的个数n=3,相应地设置:In a specific embodiment, the number of selected scales n = 3, and correspondingly set:
σ 1=30,σ 2=110,σ 3=200; σ 1 = 30, σ 2 = 110, σ 3 = 200;
其中,σ 1、σ 2和σ 3分别对应检测图像的灰度值区间[0,255]的低灰度、中灰度和高灰度,并且设置w 1=w 2=w 3=1/3。根据上述参数的设置,多尺度视网膜算法同时兼顾了低灰度、中灰度和高灰度这3个灰度尺度,从而获得较好的效果。多尺度视网膜算法通过结合多个尺度可以实现很好的自适应性,突出了图像暗区纹理细节,并可以实现图像动态范围的调整进而达到图像增强的目的。 Among them, σ 1 , σ 2, and σ 3 correspond to low gray, middle gray, and high gray of the gray value interval [0,255] of the detection image, respectively, and set w 1 = w 2 = w 3 = 1/3. According to the setting of the above parameters, the multi-scale retinal algorithm simultaneously takes into account the three gray scales of low gray, medium gray and high gray, so as to obtain better results. The multi-scale retinal algorithm can achieve good self-adaptability by combining multiple scales, highlighting the texture details of dark areas of the image, and can adjust the dynamic range of the image to achieve the purpose of image enhancement.
S112:采用拉普拉斯算子对全局增强处理后的检测图像进行锐化处理,得到待识别图像。S112: The Laplace operator is used to sharpen the detection image after the global enhancement processing to obtain an image to be identified.
拉普拉斯算子(Laplacian operator)是一种二阶微分算子,适用于改善因为光线的漫反射造成的图像模糊。对图像进行拉普拉斯算子锐化变换可以减少图像的模糊,提高图像的清晰度。因此,通过对全局增强处理后的检测图像进行锐化处理,突出全局增强处理后的检测图像的边缘细节特征,提高了全局增强处理后的检测图像的轮廓清晰度。锐化处理是指对图像进行锐化的变换,用于加强图像中的目标边界和图像细节。全局增强处理后的检测图像经过拉普拉斯算子锐化处理后,图像边缘细节特征被加强的同时也削弱了光晕,从而保护了检测图像的细节。Laplacian operator is a second-order differential operator, which is suitable for improving image blur caused by diffuse reflection of light. Laplace operator sharpening transformation on the image can reduce the blur of the image and improve the sharpness of the image. Therefore, by performing a sharpening process on the detection image after the global enhancement processing, the edge detail features of the detection image after the global enhancement processing are highlighted, thereby improving the contour definition of the detection image after the global enhancement processing. Sharpening processing refers to the transformation of sharpening an image to enhance the target boundaries and image details in the image. After the global enhancement processing of the detected image is sharpened by the Laplacian operator, the edge details of the image are enhanced and the halo is weakened, thereby protecting the details of the detected image.
基于二阶微分的拉普拉斯算子定义为:The Laplace operator based on second-order differential is defined as:
Figure PCTCN2018092614-appb-000007
Figure PCTCN2018092614-appb-000007
对于全局增强处理后的检测图像R(x,y),其二阶导数为:For the detection image R (x, y) after global enhancement processing, its second derivative is:
Figure PCTCN2018092614-appb-000008
Figure PCTCN2018092614-appb-000008
因此,拉普拉斯算子▽ 2R为: Therefore, Laplace operator ▽ 2 R is:
2R=R(x+1,y)+R(x-1,y)+R(x,y+1)+R(x,y-1)-4R(x,y); 2 R = R (x + 1, y) + R (x-1, y) + R (x, y + 1) + R (x, y-1) -4R (x, y);
得到拉普拉斯算子▽ 2R之后,用拉普拉斯算子▽ 2R对全局增强处理后的检测图像R(x,y)的每一像素灰度值都根据下述公式进行锐化,得到锐化后的像素灰度值,式中,g(x,y)为锐化后的像素灰度值。 After the Laplace operator ▽ 2 R is obtained, the gray value of each pixel of the detection image R (x, y) after the global enhancement process is sharpened with the Laplace operator ▽ 2 R according to the following formula: To obtain the sharpened pixel gray value, where g (x, y) is the sharpened pixel gray value.
Figure PCTCN2018092614-appb-000009
Figure PCTCN2018092614-appb-000009
将锐化后的像素灰度值替换原(x,y)像素处的灰度值得到待识别图像。The sharpened pixel gray value is replaced with the gray value at the original (x, y) pixel to obtain the image to be identified.
在一个具体实施方式中,拉普拉斯算子▽ 2R选用四邻域锐化模板矩阵H: In a specific embodiment, the Laplace operator ▽ 2 R selects a four-neighbor sharpening template matrix H:
Figure PCTCN2018092614-appb-000010
Figure PCTCN2018092614-appb-000010
采用四邻域锐化模板矩阵H对全局增强处理后的检测图像进行拉普拉斯算子锐化。A four-neighbor sharpening template matrix H is used to perform Laplace operator sharpening on the detected image after global enhancement processing.
在本实施例中,采用多尺度视网膜算法对检测图像进行全局增强处理,将经过多尺度视网膜算法增强处理之后的检测图像采用拉普拉斯算子进行锐化,图像边缘细节特征被加强的同时也削弱了光晕,从而保护了检测图像的细节。此外,上述步骤不仅简单方便,处理后得到待识别图像边缘细节特征更加突出明了,增强了待识别图像的纹理特征,有利于提高待识别图像识别的准确率。In this embodiment, the multi-scale retinal algorithm is used to perform global enhancement processing on the detection image, and the detection image after the multi-scale retinal algorithm enhancement processing is used to sharpen the Laplacian operator. At the same time, the edge details of the image are enhanced. The halo is also weakened, thereby protecting the details of the detected image. In addition, the above steps are not only simple and convenient. After processing, the edge details of the image to be identified are more prominent, and the texture features of the image to be identified are enhanced, which is beneficial to improving the accuracy of the image to be identified.
在一实施例中,如图5所示,在将待识别图像输入到全差-金字塔特征网络识别模型中进行识别,获取分类置信图的步骤之前,该机场异物识别方法还包括:In an embodiment, as shown in FIG. 5, before the steps of inputting an image to be identified into a full-difference-pyramid feature network recognition model for recognition and obtaining a classification confidence map, the airport foreign object recognition method further includes:
S121:获取训练样本集,对训练样本集中的训练图像进行分类标注。S121: Obtain a training sample set, and classify and label the training images in the training sample set.
其中,训练样本集中包括了训练图像,训练图像是指用于训练全差-金字塔特征网络识别模型的样本图像。可选地,该训练图像可以通过在机场不同位置设置视频采集设备或者图像采集设备来采集对应的数据来获得,视频采集设备或者图像采集设备采集到对应的数据之后发送到服务端。若服务端获取到的是视频数据,可以对视频数据按照预定的帧率进行分帧处理以得到训练图像。对训练图像进行分类标注是指对训练图像中的不同物体进行分类。例如,在训练图像中,有可能出现的物体为跑道、草坪、机场设备(非异物)和机场异物等。通过对训练图像中的不同物体赋予不同的标注信息从而完成对训练图像的分类标注。The training sample set includes training images, and the training images refer to the sample images used to train the full-difference-pyramid feature network recognition model. Optionally, the training image may be obtained by setting a video capture device or an image capture device at different locations in the airport to collect corresponding data, and the video capture device or the image capture device collects the corresponding data and sends it to the server. If the server obtains video data, the video data may be framed at a predetermined frame rate to obtain a training image. Classifying and labeling training images refers to classifying different objects in the training images. For example, in the training image, the objects that may appear are runways, lawns, airport equipment (non-foreign objects), and airport foreign objects. By assigning different labeling information to different objects in the training image, the classification labeling of the training image is completed.
S122:采用训练样本集中分类标注的训练图像训练全差网络,得到目标输出向量。S122: Use the training images in the training sample set to classify and label the training network to obtain the target output vector.
在该步骤中,采用训练样本集中分类标注的训练图像来训练全差网络,而在全差网络中,设定训练图像输入为x 0,全差网络由L层结构组成,每一层全差网络都包含一个非线性变换H l(·)。可选地,非线性变换可以包括ReLU(激活函数,Rectified Linear Units)和Pooling(池化),或者BN(规范化层,Batch Normalization)、ReLU和卷积层,或者BN、ReLU和Pooling。其中,BN就是通过规范化手段,把每层神经网络任意神经元的输入值的分布调整到均值为0和方差为1的标准正态分布,使得激活输入值落在非线性函数对输入比较敏感的区域,让梯度变大,避免梯度消失问题产生,能大大加快训练速度。ReLU是一个分段线性函数,也是一种单侧抑制函数,可以将输入的所有的负值都输出为0,而输入的正值则保持不变。通过ReLU可以实现稀疏后的模型能够更好地挖掘相关特征,拟合训练数据。 In this step, the training network is trained using the training images in the training sample set to label the full difference network. In the full difference network, the training image input is set to x 0. The full difference network consists of L-layer structures, and each layer is completely different. The networks all contain a non-linear transformation H l (·). Optionally, the non-linear transformation may include ReLU (Recified Linear Units) and Pooling, or BN (Batch Normalization), ReLU, and convolutional layers, or BN, ReLU, and Pooling. Among them, BN is to adjust the distribution of the input value of any neuron in each layer of the neural network to a normal normal distribution with a mean value of 0 and a variance of 1 by means of normalization, so that the activation input value falls in a non-linear function that is sensitive to the input Region, make the gradient larger, avoid the problem of gradient disappearance, and greatly speed up the training speed. ReLU is a piecewise linear function and a one-sided suppression function. It can output all the negative values of the input to 0, while the positive values of the input remain unchanged. ReLU can realize the sparse model to better mine related features and fit the training data.
在本实施例中,设全差网络中第l层的输出为x l,全差网络中每一层都和前面所有的层直接连接,即: In this embodiment, suppose that the output of the first layer in the full-difference network is x l , and each layer in the full-difference network is directly connected to all the previous layers, that is:
x l=H l([x 0,x 1,...,x l-1]); x l = H l ([x 0 , x 1 , ..., x l-1 ]);
而全差网络中对应层的输出就构成了目标输出向量,以供后续采用该目标输出向量训练金字塔特征网络。The output of the corresponding layer in the full difference network constitutes the target output vector for subsequent training of the pyramid feature network using the target output vector.
S123:采用目标输出向量训练金字塔特征网络,得到全差-金字塔特征网络识别模型。S123: Use the target output vector to train the pyramid feature network to obtain a full-difference-pyramid feature network recognition model.
在金字塔特征网络中,全差网络中的目标输出向量中的各层输出会分别和金字塔特征网络的RCU单元连接。即金字塔特征网络中存在和全差网络中的目标输出向量的层数相同的RCU单元。In the pyramid feature network, the output of each layer in the target output vector in the full-difference network is connected to the RCU unit of the pyramid feature network, respectively. That is, there are RCU units in the pyramid feature network that have the same number of layers as the target output vector in the full-difference network.
RCU单元是指从全差网络中提取出来的单元结构,具体包括了ReLU、卷积和求和等部分。通过将全差网络中获取的各层目标输出向量分别经过ReLU、卷积和求和操作。RCU单元的各层输出都采用Multi-resolution fusion(多分辨率融合)进行处理,从而得到不同 的输出特征图:先对RCU单元的各层输出特征图都用一个卷积层进行自适应处理,再进行上采样到该层的最大分辨率。Chained residual pooling(链式残差池化)将输入的不同分辨率的输出特征图上采样到和最大输出特征图相等的尺寸然后进行叠加。最后将叠加后的输出特征图再经过一个RCU进行卷积,即得到一个精细特征图。The RCU unit refers to the unit structure extracted from the full-difference network, and specifically includes ReLU, convolution and summation. The target output vectors of each layer obtained in the full-difference network are respectively subjected to ReLU, convolution and summing operations. The output of each layer of the RCU unit is processed using Multi-resolution fusion to obtain different output feature maps. First, the output feature maps of each layer of the RCU unit are adaptively processed by a convolution layer. Then upsampling to the maximum resolution of this layer. Chained residual pooling samples the output feature maps of different resolutions of the input to the same size as the maximum output feature map and then superimposes them. Finally, the superimposed output feature map is convolved by an RCU to obtain a fine feature map.
金字塔特征网络的作用就是把不同分辨率的特征图进行融合。先把预先训练的全差网络按特征图的分辨率分成若干个全差blocks,然后向右把这若干个blocks分别作为若干个path通过金字塔特征网络进行融合,最后得到一个精细特征图(后续连接softmax层,再通过双线性插值输出)。The function of the pyramid feature network is to fuse feature maps with different resolutions. First divide the pre-trained full-difference network into several full-difference blocks according to the resolution of the feature map, and then fuse these blocks to the right as several paths to fuse through the pyramid feature network, and finally obtain a fine feature map (subsequent connection softmax layer, and then output through bilinear interpolation).
在金字塔特征网络中,通过全差网络中的目标输出向量训练金字塔特征网络形成一个初步训练网络,再采用验证样本对金字塔特征网络进行验证和调整,直至得到预设的分类准确率,则训练结束。其中,预设的分类准确率可以根据实际的识别模型的需要而设置。In the pyramid feature network, the target feature vector of the full-difference network is used to train the pyramid feature network to form a preliminary training network. The verification samples are then used to verify and adjust the pyramid feature network until a preset classification accuracy rate is obtained, and the training ends. . The preset classification accuracy can be set according to the needs of the actual recognition model.
在这个实施例中,通过采用分类标注后的训练样本集训练得到全差-金字塔特征网络识别模型,保证了该全差-金字塔特征网络识别模型的识别精度和速度。In this embodiment, a full-difference-pyramid feature network recognition model is obtained by training with a training sample set after classification and labeling, which ensures the recognition accuracy and speed of the full-difference-pyramid feature network recognition model.
在一实施例中,训练全差网络,具体包括:In an embodiment, training a full-difference network specifically includes:
设置全差网络的初始卷积层,并采用全差网络中的最大池化层进行下采样。Set the initial convolution layer of the full-difference network, and use the largest pooling layer in the full-difference network for downsampling.
卷积层用于对输入图像的特征提取,而初始卷积层提取的就是训练图像的特征,可选地,初始卷积层采用7*7的卷积核。采用全差网络中的最大池化层进行下采样,在进行采样的过程中,如果新采样率小于原采样率则为下采样。最大池化(max-pooling)是指采样函数取区域内所有神经元的最大值。通过经过初始卷积层的输入图像进行最大池化处理,进行特征压缩,提取主要特征,并简化网络计算复杂度。The convolution layer is used to extract the features of the input image, and the initial convolution layer extracts the features of the training image. Optionally, the initial convolution layer uses a 7 * 7 convolution kernel. The maximum pooling layer in the full-difference network is used for downsampling. During the sampling process, if the new sampling rate is less than the original sampling rate, it is downsampling. Max-pooling refers to the maximum value of all neurons in the area taken by the sampling function. The input image through the initial convolution layer is subjected to maximum pooling processing, feature compression, main features are extracted, and network computation complexity is simplified.
设置三层全差网络模块,每一全差网络模块包括一个全差卷积层和一个全差激活层,全差激活层中的激活函数采用线性激活函数。A three-layer full-difference network module is provided. Each full-difference network module includes a full-difference convolution layer and a full-difference activation layer. The activation function in the full-difference activation layer adopts a linear activation function.
在这三层全差网络模块中,每一个全差网络模块的输出都为前面所有模块输出的结合,即:In these three layers of fully differential network modules, the output of each fully differential network module is the combination of the outputs of all the previous modules, that is:
x l=H l([x 0,x 1,...,x l-1]),l=1,2,3; x l = H l ([x 0 , x 1 , ..., x l-1 ]), l = 1, 2, 3;
其中,每一个H l(·)都为卷积层和激活层两层操作的组合:Conv->ReLU。可选地,全差卷积层中的卷积核大小为3*3。每一H l(·)输出的特征数量即为特征增长率,可选地,设置特征增长率为16,则三层全差网络模块输出的输出特征数量为48。而线性激活函数公式为: Among them, each H l (·) is a combination of two operations of the convolutional layer and the activation layer: Conv-> ReLU. Optionally, the size of the convolution kernel in the full-difference convolution layer is 3 * 3. The number of features output by each H l (·) is the feature growth rate. Optionally, if the feature growth rate is set to 16, then the number of output features output by the three-layer fully-differential network module is 48. The linear activation function formula is:
Figure PCTCN2018092614-appb-000011
Figure PCTCN2018092614-appb-000011
通过线性激活函数的转换,可以使得训练过程的时间快速地收敛。By transforming the linear activation function, the time of the training process can be quickly converged.
在全差网络模块之间设置传输层,每一传输层包括规范化层、传输激活层和平均池化层。A transmission layer is set between the fully differential network modules, and each transmission layer includes a normalization layer, a transmission activation layer, and an average pooling layer.
在全差网络模块中,每个全差网络模块的输出特征都是在增加的,在上述设置中,若特征增长率为16,则三层全差网络模块输出的输出特征为48。如此,计算量是逐步增大的,因此引入了传输层,设置传输参数来表示将传输层的输入缩小到原来的多少倍。示例地,传输参数为0.6,即将传输层的输入缩小到原来的0.6。In the full-difference network module, the output characteristics of each full-difference network module are increasing. In the above setting, if the feature growth rate is 16, the output characteristic of the three-layer full-difference network module output is 48. In this way, the amount of calculation is gradually increased, so the transmission layer is introduced, and the transmission parameters are set to indicate how many times the input of the transmission layer is reduced. For example, the transmission parameter is 0.6, that is, the input of the transmission layer is reduced to the original 0.6.
在本实施例中,通过对全差网络中网络结构和各参数的设置,保证了全差网络的训练速度和精度。In this embodiment, by setting the network structure and various parameters in the full-difference network, the training speed and accuracy of the full-difference network are guaranteed.
在一实施例中,在训练全差-金字塔特征网络识别模型的过程中,损失函数采用Focal Loss函数实现:In an embodiment, in the process of training the full-difference-pyramid feature network recognition model, the loss function is implemented using a Focal Loss function:
FL(p t)=-(1-p t) γlog(p t); FL (p t ) =-(1-p t ) γ log (p t );
其中,
Figure PCTCN2018092614-appb-000012
p t是全差-金字塔特征网络识别模型对训练图像的预测值,p为模型对于训练图像y=1的估计概率,p∈[0,1],y为训练图像的标注值,γ为调节参数。
among them,
Figure PCTCN2018092614-appb-000012
p t is the prediction value of the full difference-pyramid feature network recognition model for the training image, p is the estimated probability of the model for the training image y = 1, p ∈ [0,1], y is the labeled value of the training image, and γ is the adjustment parameter.
损失函数指一种将一个事件(在一个样本空间中的一个元素)映射到一个表达与其事件相关的经济成本或机会成本的实数上的一种函数。在本实施例中,在训练全差-金字塔特征网络识别模型时,采用损失函数衡量这个全差-金字塔特征网络识别模型预测的好坏,损失函数越小,该识别模型的预测能力越好。在本申请实施例中,训练样本集的训练图像中各个分类的样本图像数量可能不均衡,特别是包含机场异物的训练图像可能较少,为了更好地提升全差-金字塔特征网络识别模型的预测能力而对损失函数进行了选择。Loss function refers to a function that maps an event (an element in a sample space) to a real number that expresses the economic or opportunity cost associated with its event. In this embodiment, when training the full-difference-pyramid feature network recognition model, a loss function is used to measure the prediction of the full-difference-pyramid feature network recognition model. The smaller the loss function, the better the predictive ability of the recognition model. In the embodiment of the present application, the number of sample images of each classification in the training images of the training sample set may be uneven, especially the training images containing foreign objects in the airport may be less. The prediction function was selected for the loss function.
因此损失函数采用了Focal Loss函数实现,Focal loss函数增加了一个调节因子(1-p t) γ,其中,调节参数γ的取值在[0,5]之间。y为训练图像的标注值,示例地,对于训练图像中的异物的标注,如果是异物则标注为y=1,非异物则标注为y=-1。当一个训练图像被误分类时,P t很小,此时调节因子(1-p t) γ接近1,该损失不会有很大影响;当P t值很大,趋近于1的时候,调节因子的值趋近于0,因此对于正确分类的样本的loss值被缩小了。 Therefore, the Focal Loss function is used to implement the loss function. The Focal loss function adds an adjustment factor (1-p t ) γ , where the value of the adjustment parameter γ is between [0,5]. y is the labeling value of the training image. For example, for foreign object labels in the training image, if it is a foreign object, it is labeled as y = 1, and for non-foreign objects, it is labeled as y = -1. When a training image is misclassified, P t is small. At this time, the adjustment factor (1-p t ) γ is close to 1, and the loss will not have a great impact. When the value of P t is large, it will approach 1. The value of the adjustment factor approaches 0, so the loss value for the correctly classified samples is reduced.
在本实施例中,在训练全差-金字塔特征网络识别模型的过程中采用Focal Loss函数,可以减小了分类样本不均对训练全差-金字塔特征网络识别模型的影响,也达到了提高后续检测精度的效果。In this embodiment, the FocalLoss function is used in the process of training the full difference-pyramid feature network recognition model, which can reduce the impact of uneven sample classification on training the full difference-pyramid feature network recognition model, and has also improved the follow-up. The effect of detection accuracy.
应理解,上述实施例中各步骤的序号的大小并不意味着执行顺序的先后,各过程的执行顺序应以其功能和内在逻辑确定,而不应对本申请实施例的实施过程构成任何限定。It should be understood that the size of the sequence numbers of the steps in the above embodiments does not mean the order of execution. The execution order of each process should be determined by its function and internal logic, and should not constitute any limitation on the implementation process of the embodiments of this application.
在一实施例中,提供一种机场异物识别装置,该机场异物识别装置与上述实施例中机场异物识别方法一一对应。如图6所示,该机场异物识别装置包括检测结果获取模块10、基准特征向量获取模块20、识别图像集组成模块30、比较结果获取模块40和识别结果获取模块50。各功能模块详细说明如下:In one embodiment, an airport foreign body identification device is provided. The airport foreign body identification device corresponds to the airport foreign body identification method in the above embodiment in a one-to-one correspondence. As shown in FIG. 6, the airport foreign object recognition device includes a detection result acquisition module 10, a reference feature vector acquisition module 20, a recognition image set composition module 30, a comparison result acquisition module 40, and a recognition result acquisition module 50. The detailed description of each function module is as follows:
检测结果获取模块10,用于获取检测图像,采用异物检测模型对检测图像进行检测,获取检测结果。The detection result acquiring module 10 is configured to acquire a detection image, detect the detection image by using a foreign object detection model, and obtain a detection result.
基准特征向量获取模块20,用于若检测结果为检测图像中存在异物,则获取异物在检测图像中的位置,作为基准位置,根据基准位置提取异物的特征向量,作为基准特征向量。A reference feature vector acquisition module 20 is configured to obtain a position of the foreign object in the detection image if the detection result is that there is a foreign object in the detection image, and use the reference position to extract a feature vector of the foreign object according to the reference position as a reference feature vector.
识别图像集组成模块30,用于根据检测图像获取连续预定帧图像,组成识别图像集。The recognition image set composition module 30 is configured to obtain consecutive predetermined frames of images according to the detection image to form a recognition image set.
比较结果获取模块40,用于根据基准位置提取识别图像集中每一识别图像的特征向量,并比对每一识别图像的特征向量和基准特征向量的特征向量相似度,获取比较结果。The comparison result acquisition module 40 is configured to extract a feature vector of each recognition image in the recognition image set according to the reference position, and compare the feature vector of each recognition image and the feature vector similarity of the reference feature vector to obtain a comparison result.
识别结果获取模块50,用于根据比较结果生成识别结果,识别结果包括确认为异物和确认为非异物。The recognition result acquisition module 50 is configured to generate a recognition result according to the comparison result. The recognition result includes confirmation as a foreign object and confirmation as a non-foreign object.
优选地,检测结果获取模块10包括待识别图像获取单元11、分类置信图获取单元12和检测结果获取单元13。Preferably, the detection result acquisition module 10 includes an image-to-be-identified image acquisition unit 11, a classification confidence map acquisition unit 12, and a detection result acquisition unit 13.
待识别图像获取单元11,用于对检测图像进行预处理,得到待识别图像。The to-be-recognized image obtaining unit 11 is configured to preprocess the detection image to obtain the to-be-recognized image.
分类置信图获取单元12,用于将待识别图像输入到全差-金字塔特征网络识别模型中进行识别,获取分类置信图。A classification confidence map acquisition unit 12 is configured to input an image to be identified into a full-difference-pyramid feature network recognition model for recognition, and obtain a classification confidence map.
检测结果获取单元13,用于根据分类置信图获取检测结果,检测结果包括检测图像中存在异物和检测图像中不存在异物。The detection result obtaining unit 13 is configured to obtain a detection result according to the classification confidence map, and the detection result includes detecting the presence of a foreign object in the image and the absence of a foreign object in the detection image.
优选地,待识别图像获取单元11包括全局增强处理子单元111和锐化处理子单元112。Preferably, the image acquisition unit 11 includes a global enhancement processing sub-unit 111 and a sharpening processing sub-unit 112.
全局增强处理子单元111,用于采用多尺度视网膜算法对原始图像进行全局增强处理。The global enhancement processing sub-unit 111 is configured to perform global enhancement processing on the original image by using a multi-scale retinal algorithm.
锐化处理单元子112,用于采用拉普拉斯算子对全局增强处理后的原始图像进行锐化处理,得到待识别图像。A sharpening processing unit 112 is configured to use a Laplacian to sharpen the original image after global enhancement processing to obtain an image to be identified.
优选地,该机场异物识别装置还包括训练样本集获取模块121、目标输出向量获取模块122和识别模型获取模块123。Preferably, the airport foreign body identification device further includes a training sample set acquisition module 121, a target output vector acquisition module 122, and a recognition model acquisition module 123.
训练样本集获取模块121,用于获取训练样本集,对训练样本集中的训练图像进行分类标注。The training sample set obtaining module 121 is configured to obtain a training sample set and classify and label the training images in the training sample set.
目标输出向量获取模块122,用于采用训练样本集中分类标注的训练图像训练全差网络,得到目标输出向量。A target output vector obtaining module 122 is configured to train a full difference network using training images in which training samples are classified and labeled to obtain a target output vector.
识别模型获取模块123,用于采用目标输出向量训练金字塔特征网络,得到全差-金字塔特征网络识别模型。A recognition model acquisition module 123 is used to train a pyramid feature network using a target output vector to obtain a full-difference-pyramid feature network recognition model.
关于机场异物识别装置的具体限定可以参见上文中对于机场异物识别方法的限定,在此不再赘述。上述机场异物识别装置中的各个模块可全部或部分通过软件、硬件及其组合来实现。上述各模块可以硬件形式内嵌于或独立于计算机设备中的处理器中,也可以以软件形式存储于计算机设备中的存储器中,以便于处理器调用执行以上各个模块对应的操作。For specific limitations on the airport foreign body identification device, refer to the foregoing limitation on the airport foreign body identification method, which will not be repeated here. Each module in the above-mentioned airport foreign body identification device may be implemented in whole or in part by software, hardware, and a combination thereof. The above-mentioned modules may be embedded in the hardware in or independent of the processor in the computer device, or may be stored in the memory of the computer device in the form of software, so that the processor can call and execute the operations corresponding to the above modules.
在一个实施例中,提供了一种计算机设备,该计算机设备可以是服务端,其内部结构图可以如图7所示。该计算机设备包括通过系统总线连接的处理器、存储器、网络接口和数据库。其中,该计算机设备的处理器用于提供计算和控制能力。该计算机设备的存储器包括非易失性存储介质、内存储器。该非易失性存储介质存储有操作系统、计算机可读指令和数据库。该内存储器为非易失性存储介质中的操作系统和计算机可读指令的运行提供环境。该计算机设备的数据库用于存储检测图像和异物检测模型数据。该计算机设备的网络接口用于与外部的终端通过网络连接通信。该计算机可读指令被处理器执行时以实现一种机场异物识别方法。In one embodiment, a computer device is provided. The computer device may be a server, and its internal structure diagram may be as shown in FIG. 7. The computer device includes a processor, a memory, a network interface, and a database connected through a system bus. The processor of the computer device is used to provide computing and control capabilities. The memory of the computer device includes a non-volatile storage medium and an internal memory. The non-volatile storage medium stores an operating system, computer-readable instructions, and a database. The internal memory provides an environment for the operation of the operating system and computer-readable instructions in a non-volatile storage medium. The database of the computer equipment is used to store detection images and foreign object detection model data. The network interface of the computer device is used to communicate with an external terminal through a network connection. The computer-readable instructions are executed by a processor to implement an airport foreign object identification method.
在一个实施例中,提供了一种计算机设备,包括存储器、处理器及存储在存储器上并可在处理器上运行的计算机可读指令,处理器执行计算机可读指令时实现以下步骤:In one embodiment, a computer device is provided, including a memory, a processor, and computer-readable instructions stored on the memory and executable on the processor. When the processor executes the computer-readable instructions, the following steps are implemented:
获取检测图像,采用异物检测模型对检测图像进行检测,获取检测结果。Obtain the detection image, use the foreign object detection model to detect the detection image, and obtain the detection result.
若所述检测结果为检测图像中存在异物,则获取异物在检测图像中的位置,作为基准位置,根据基准位置提取异物的特征向量,作为基准特征向量。If the detection result is that there is a foreign object in the detection image, the position of the foreign object in the detection image is obtained as a reference position, and a feature vector of the foreign object is extracted according to the reference position as a reference feature vector.
根据检测图像获取连续预定帧图像,组成识别图像集。Consecutive predetermined frame images are acquired according to the detected images to form a recognition image set.
根据基准位置提取识别图像集中每一识别图像的特征向量,并比对每一识别图像的特征向量和基准特征向量的特征向量相似度,获取比较结果。The feature vector of each recognition image in the recognition image set is extracted according to the reference position, and the feature vector of each recognition image and the feature vector similarity of the reference feature vector are compared to obtain a comparison result.
根据比较结果生成识别结果,识别结果包括确认为异物和确认为非异物。A recognition result is generated based on the comparison result, and the recognition result includes confirmation as a foreign object and confirmation as a non-foreign object.
在一个实施例中,提供了一个或多个存储有计算机可读指令的非易失性可读存储介质,所述计算机可读指令被一个或多个处理器执行时,使得所述一个或多个处理器执行如下步骤:In one embodiment, one or more non-volatile readable storage media storing computer-readable instructions are provided, and when the computer-readable instructions are executed by one or more processors, the one or more Each processor performs the following steps:
获取检测图像,采用异物检测模型对检测图像进行检测,获取检测结果。Obtain the detection image, use the foreign object detection model to detect the detection image, and obtain the detection result.
若所述检测结果为检测图像中存在异物,则获取异物在检测图像中的位置,作为基准位置,根据基准位置提取异物的特征向量,作为基准特征向量。If the detection result is that there is a foreign object in the detection image, the position of the foreign object in the detection image is obtained as a reference position, and a feature vector of the foreign object is extracted according to the reference position as a reference feature vector.
根据检测图像获取连续预定帧图像,组成识别图像集。Consecutive predetermined frame images are acquired according to the detected images to form a recognition image set.
根据基准位置提取识别图像集中每一识别图像的特征向量,并比对每一识别图像的特征向量和基准特征向量的特征向量相似度,获取比较结果。The feature vector of each recognition image in the recognition image set is extracted according to the reference position, and the feature vector of each recognition image and the feature vector similarity of the reference feature vector are compared to obtain a comparison result.
根据比较结果生成识别结果,识别结果包括确认为异物和确认为非异物。A recognition result is generated based on the comparison result, and the recognition result includes confirmation as a foreign object and confirmation as a non-foreign object.
本领域普通技术人员可以理解实现上述实施例方法中的全部或部分流程,是可以通过 计算机可读指令来指令相关的硬件来完成,所述的计算机可读指令可存储于一非易失性计算机可读取存储介质中,该计算机可读指令在执行时,可包括如上述各方法的实施例的流程。其中,本申请所提供的各实施例中所使用的对存储器、存储、数据库或其它介质的任何引用,均可包括非易失性和/或易失性存储器。非易失性存储器可包括只读存储器(ROM)、可编程ROM(PROM)、电可编程ROM(EPROM)、电可擦除可编程ROM(EEPROM)或闪存。易失性存储器可包括随机存取存储器(RAM)或者外部高速缓冲存储器。作为说明而非局限,RAM以多种形式可得,诸如静态RAM(SRAM)、动态RAM(DRAM)、同步DRAM(SDRAM)、双数据率SDRAM(DDRSDRAM)、增强型SDRAM(ESDRAM)、同步链路(Synchlink)DRAM(SLDRAM)、存储器总线(Rambus)直接RAM(RDRAM)、直接存储器总线动态RAM(DRDRAM)、以及存储器总线动态RAM(RDRAM)等。A person of ordinary skill in the art can understand that all or part of the processes in the methods of the foregoing embodiments can be implemented by using computer-readable instructions to instruct related hardware. The computer-readable instructions can be stored in a non-volatile computer. In the readable storage medium, the computer-readable instructions, when executed, may include the processes of the embodiments of the methods described above. Wherein, any reference to the storage, storage, database, or other media used in the embodiments provided in this application may include non-volatile and / or volatile storage. Non-volatile memory may include read-only memory (ROM), programmable ROM (PROM), electrically programmable ROM (EPROM), electrically erasable programmable ROM (EEPROM), or flash memory. Volatile memory can include random access memory (RAM) or external cache memory. By way of illustration and not limitation, RAM is available in various forms, such as static RAM (SRAM), dynamic RAM (DRAM), synchronous DRAM (SDRAM), dual data rate SDRAM (DDRSDRAM), enhanced SDRAM (ESDRAM), synchronous chain (Synchlink) DRAM (SLDRAM), memory bus (Rambus) direct RAM (RDRAM), direct memory bus dynamic RAM (DRDRAM), and memory bus dynamic RAM (RDRAM).
所属领域的技术人员可以清楚地了解到,为了描述的方便和简洁,仅以上述各功能单元、模块的划分进行举例说明,实际应用中,可以根据需要而将上述功能分配由不同的功能单元、模块完成,即将所述装置的内部结构划分成不同的功能单元或模块,以完成以上描述的全部或者部分功能。Those skilled in the art can clearly understand that, for the convenience and brevity of the description, only the above-mentioned division of functional units and modules is used as an example. In practical applications, the above functions can be assigned by different functional units, Module completion, that is, dividing the internal structure of the device into different functional units or modules to complete all or part of the functions described above.
以上所述实施例仅用以说明本申请的技术方案,而非对其限制;尽管参照前述实施例对本申请进行了详细的说明,本领域的普通技术人员应当理解:其依然可以对前述各实施例所记载的技术方案进行修改,或者对其中部分技术特征进行等同替换;而这些修改或者替换,并不使相应技术方案的本质脱离本申请各实施例技术方案的精神和范围,均应包含在本申请的保护范围之内。The above-mentioned embodiments are only used to describe the technical solution of the present application, but not limited thereto. Although the present application has been described in detail with reference to the foregoing embodiments, those skilled in the art should understand that they can still implement the foregoing implementations. The technical solutions described in the examples are modified, or some of the technical features are equivalently replaced; and these modifications or replacements do not deviate the essence of the corresponding technical solutions from the spirit and scope of the technical solutions of the embodiments of the application, and should be included in Within the scope of this application.

Claims (20)

  1. 一种机场异物识别方法,其特征在于,包括:An airport foreign body identification method, comprising:
    获取检测图像,采用异物检测模型对所述检测图像进行检测,获取检测结果;Acquiring a detection image, detecting the detection image by using a foreign object detection model, and acquiring a detection result;
    若所述检测结果为所述检测图像中存在异物,则获取所述异物在所述检测图像中的位置,作为基准位置,根据所述基准位置提取所述异物的特征向量,作为基准特征向量;If the detection result is that there is a foreign object in the detection image, obtaining a position of the foreign object in the detection image as a reference position, and extracting a feature vector of the foreign object according to the reference position as a reference feature vector;
    根据所述检测图像获取连续预定帧图像,组成识别图像集;Acquiring successive predetermined frames of images according to the detected images to form a recognition image set;
    根据所述基准位置提取所述识别图像集中每一识别图像的特征向量,并比对每一识别图像的特征向量和所述基准特征向量的特征向量相似度,获取比较结果;Extracting a feature vector of each recognition image in the recognition image set according to the reference position, and comparing the feature vector of each recognition image and the feature vector similarity of the reference feature vector to obtain a comparison result;
    根据所述比较结果生成识别结果,所述识别结果包括确认为异物和确认为非异物。A recognition result is generated according to the comparison result, and the recognition result includes confirmation as a foreign object and confirmation as a non-foreign object.
  2. 如权利要求1所述的机场异物识别方法,其特征在于,所述采用异物检测模型对所述检测图像进行检测,获取检测结果,具体包括:The foreign object identification method of an airport according to claim 1, wherein the detecting the detected image by using a foreign object detection model to obtain a detection result specifically includes:
    对所述检测图像进行预处理,得到待识别图像;Preprocessing the detection image to obtain an image to be identified;
    将所述待识别图像输入到全差-金字塔特征网络识别模型中进行识别,获取分类置信图;Inputting the to-be-recognized image into a full-difference-pyramid feature network recognition model for recognition, and obtaining a classification confidence map;
    根据所述分类置信图获取检测结果,所述检测结果包括检测图像中存在异物和检测图像中不存在异物。A detection result is obtained according to the classification confidence map, and the detection result includes detecting the presence of a foreign object in the image and the absence of a foreign object in the detection image.
  3. 如权利要求2所述的机场异物识别方法,其特征在于,所述对所述检测图像进行预处理,得到待识别图像,具体包括:The method for identifying foreign objects in an airport according to claim 2, wherein the preprocessing the detection image to obtain an image to be identified specifically includes:
    采用多尺度视网膜算法对所述检测图像进行全局增强处理;Using a multi-scale retinal algorithm to globally enhance the detection image;
    采用拉普拉斯算子对全局增强处理后的所述检测图像进行锐化处理,得到待识别图像。A Laplace operator is used to sharpen the detection image after the global enhancement processing to obtain an image to be identified.
  4. 如权利要求2所述的机场异物识别方法,其特征在于,在所述将所述待识别图像输入到全差-金字塔特征网络识别模型中进行识别,获取分类置信图的步骤之前,所述机场异物识别方法还包括:The foreign object identification method for an airport according to claim 2, wherein before the step of inputting the image to be identified into a full-difference-pyramid feature network identification model to obtain a classification confidence map, the airport Foreign object identification methods also include:
    获取训练样本集,对所述训练样本集中的训练图像进行分类标注;Acquiring a training sample set, and classifying the training images in the training sample set;
    采用所述训练样本集中所述分类标注的训练图像训练全差网络,得到目标输出向量;Training a full-difference network by using the classified labeled training images in the training sample set to obtain a target output vector;
    采用所述目标输出向量训练金字塔特征网络,得到所述全差-金字塔特征网络识别模型。The target output vector is used to train a pyramid feature network to obtain the full-difference-pyramid feature network recognition model.
  5. 如权利要求4所述的机场异物识别方法,其特征在于,所述训练全差网络,具体包括:The method for identifying foreign objects in an airport according to claim 4, wherein the training of the full difference network specifically comprises:
    设置所述全差网络的初始卷积层,并采用所述全差网络中的最大池化层进行下采样;Setting an initial convolution layer of the full-difference network, and using the largest pooling layer in the full-difference network for downsampling;
    设置三层全差网络模块,每一所述全差网络模块包括一个全差卷积层和一个全差激活层,所述全差激活层中的激活函数采用线性激活函数;Setting up three layers of full-difference network modules, each of which includes a full-difference convolution layer and a full-difference activation layer, and the activation function in the full-difference activation layer adopts a linear activation function;
    在所述全差网络模块之间设置传输层,每一所述传输层包括规范化层、传输激活层和平均池化层。A transmission layer is provided between the fully differential network modules, and each of the transmission layers includes a normalization layer, a transmission activation layer, and an average pooling layer.
  6. 如权利要求4所述的机场异物识别方法,其特征在于,在训练全差-金字塔特征网络识别模型的过程中,损失函数采用Focal Loss函数实现:The method for identifying foreign objects in an airport according to claim 4, characterized in that, in the process of training the full difference-pyramid feature network recognition model, the loss function is implemented using a FocalLoss function:
    FL(p t)=-(1-p t) γlog(p t); FL (p t ) =-(1-p t ) γ log (p t );
    其中,
    Figure PCTCN2018092614-appb-100001
    p t是所述全差-金字塔特征网络识别模型对所述训练图像的预测值,p为模型对于所述训练图像y=1的估计概率,p∈[0,1],y为所述训练图像的标注值,γ为调节参数。
    among them,
    Figure PCTCN2018092614-appb-100001
    p t is the predicted value of the full difference-pyramid feature network recognition model for the training image, p is the estimated probability of the model for the training image y = 1, p ∈ [0,1], and y is the training The label value of the image, γ is the adjustment parameter.
  7. 一种机场异物识别装置,其特征在于,包括:An airport foreign body identification device, comprising:
    检测结果获取模块,用于获取检测图像,采用异物检测模型对所述检测图像进行检测,获取检测结果;A detection result acquisition module, configured to acquire a detection image, detect the detection image by using a foreign object detection model, and obtain a detection result;
    基准特征向量获取模块,用于若所述检测结果为所述检测图像中存在异物,则获取所述异物在所述检测图像中的位置,作为基准位置,根据所述基准位置提取所述异物的特征向量,作为基准特征向量;A reference feature vector acquisition module, configured to obtain a position of the foreign object in the detection image if the detection result is that there is a foreign object in the detection image, and use it as a reference position to extract the foreign object according to the reference position; Eigenvectors as the reference eigenvectors;
    识别图像集组成模块,用于根据所述检测图像获取连续预定帧图像,组成识别图像集;A recognition image set composition module, configured to obtain consecutive predetermined frames of images according to the detection image to form a recognition image set;
    比较结果获取模块,用于根据所述基准位置提取所述识别图像集中每一识别图像的特征向量,并比对每一识别图像的特征向量和所述基准特征向量的特征向量相似度,获取比较结果;A comparison result acquisition module, configured to extract a feature vector of each recognition image in the recognition image set according to the reference position, and compare the feature vector of each recognition image and the feature vector similarity of the reference feature vector to obtain a comparison result;
    识别结果获取模块,用于根据所述比较结果生成识别结果,所述识别结果包括确认为异物和确认为非异物。A recognition result acquisition module is configured to generate a recognition result according to the comparison result, where the recognition result includes confirmation as a foreign object and confirmation as a non-foreign object.
  8. 如权利要求7所述的机场异物识别装置,其特征在于,检测结果获取模块包括:The foreign body identification device for an airport according to claim 7, wherein the detection result acquisition module comprises:
    待识别图像获取单元,用于对所述检测图像进行预处理,得到待识别图像;An image-to-be-identified obtaining unit, configured to pre-process the detection image to obtain an image to be identified;
    分类置信图获取单元,用于将所述待识别图像输入到全差-金字塔特征网络识别模型中进行识别,获取分类置信图;A classification confidence map acquisition unit, configured to input the image to be identified into a full-difference-pyramid feature network recognition model for recognition, and obtain a classification confidence map;
    检测结果获取单元,用于根据所述分类置信图获取检测结果,所述检测结果包括检测图像中存在异物和检测图像中不存在异物。A detection result obtaining unit is configured to obtain a detection result according to the classification confidence map, where the detection result includes a detection of the presence of a foreign object in the image and a detection of the absence of a foreign object in the image.
  9. 一种计算机设备,包括存储器、处理器以及存储在所述存储器中并可在所述处理器上运行的计算机可读指令,其特征在于,所述处理器执行所述计算机可读指令时实现如下步骤:A computer device includes a memory, a processor, and computer-readable instructions stored in the memory and executable on the processor, and is characterized in that the processor implements the computer-readable instructions as follows step:
    获取检测图像,采用异物检测模型对所述检测图像进行检测,获取检测结果;Acquiring a detection image, detecting the detection image by using a foreign object detection model, and acquiring a detection result;
    若所述检测结果为所述检测图像中存在异物,则获取所述异物在所述检测图像中的位置,作为基准位置,根据所述基准位置提取所述异物的特征向量,作为基准特征向量;If the detection result is that there is a foreign object in the detection image, obtaining a position of the foreign object in the detection image as a reference position, and extracting a feature vector of the foreign object according to the reference position as a reference feature vector;
    根据所述检测图像获取连续预定帧图像,组成识别图像集;Acquiring successive predetermined frames of images according to the detected images to form a recognition image set;
    根据所述基准位置提取所述识别图像集中每一识别图像的特征向量,并比对每一识别图像的特征向量和所述基准特征向量的特征向量相似度,获取比较结果;Extracting a feature vector of each recognition image in the recognition image set according to the reference position, and comparing the feature vector of each recognition image and the feature vector similarity of the reference feature vector to obtain a comparison result;
    根据所述比较结果生成识别结果,所述识别结果包括确认为异物和确认为非异物。A recognition result is generated according to the comparison result, and the recognition result includes confirmation as a foreign object and confirmation as a non-foreign object.
  10. 如权利要求9所述的计算机设备,其特征在于,所述采用异物检测模型对所述检测图像进行检测,获取检测结果,具体包括:The computer device according to claim 9, wherein the detecting the detected image by using a foreign object detection model to obtain a detection result specifically comprises:
    对所述检测图像进行预处理,得到待识别图像;Preprocessing the detection image to obtain an image to be identified;
    将所述待识别图像输入到全差-金字塔特征网络识别模型中进行识别,获取分类置信图;Inputting the to-be-recognized image into a full-difference-pyramid feature network recognition model for recognition, and obtaining a classification confidence map;
    根据所述分类置信图获取检测结果,所述检测结果包括检测图像中存在异物和检测图像中不存在异物。A detection result is obtained according to the classification confidence map, and the detection result includes detecting the presence of a foreign object in the image and the absence of a foreign object in the detection image.
  11. 如权利要求10所述的计算机设备,其特征在于,所述对所述检测图像进行预处理,得到待识别图像,具体包括:The computer device according to claim 10, wherein the preprocessing the detected image to obtain an image to be identified specifically comprises:
    采用多尺度视网膜算法对所述检测图像进行全局增强处理;Using a multi-scale retinal algorithm to globally enhance the detection image;
    采用拉普拉斯算子对全局增强处理后的所述检测图像进行锐化处理,得到待识别图像。A Laplace operator is used to sharpen the detection image after the global enhancement processing to obtain an image to be identified.
  12. 如权利要求10所述的计算机设备,其特征在于,在所述将所述待识别图像输入到全差-金字塔特征网络识别模型中进行识别,获取分类置信图的步骤之前,还包括:The computer device according to claim 10, wherein before the step of inputting the to-be-recognized image into a full-difference-pyramid feature network recognition model for recognition and obtaining a classification confidence map, further comprising:
    获取训练样本集,对所述训练样本集中的训练图像进行分类标注;Acquiring a training sample set, and classifying the training images in the training sample set;
    采用所述训练样本集中所述分类标注的训练图像训练全差网络,得到目标输出向量;Training a full-difference network by using the classified labeled training images in the training sample set to obtain a target output vector;
    采用所述目标输出向量训练金字塔特征网络,得到所述全差-金字塔特征网络识别模 型。The target output vector is used to train a pyramid feature network to obtain the full-difference-pyramid feature network recognition model.
  13. 如权利要求12所述的计算机设备,其特征在于,所述训练全差网络,具体包括:The computer device according to claim 12, wherein the training of the full difference network specifically comprises:
    设置所述全差网络的初始卷积层,并采用所述全差网络中的最大池化层进行下采样;Setting an initial convolution layer of the full-difference network, and using the largest pooling layer in the full-difference network for downsampling;
    设置三层全差网络模块,每一所述全差网络模块包括一个全差卷积层和一个全差激活层,所述全差激活层中的激活函数采用线性激活函数;Setting up three layers of full-difference network modules, each of which includes a full-difference convolution layer and a full-difference activation layer, and the activation function in the full-difference activation layer adopts a linear activation function;
    在所述全差网络模块之间设置传输层,每一所述传输层包括规范化层、传输激活层和平均池化层。A transmission layer is provided between the fully differential network modules, and each of the transmission layers includes a normalization layer, a transmission activation layer, and an average pooling layer.
  14. 如权利要求12所述的计算机设备,其特征在于,在训练全差-金字塔特征网络识别模型的过程中,损失函数采用Focal Loss函数实现:The computer device according to claim 12, characterized in that, in the process of training the full difference-pyramid feature network recognition model, the loss function is implemented using a FocalLoss function:
    FL(p t)=-(1-p t) γlog(p t); FL (p t ) =-(1-p t ) γ log (p t );
    其中,
    Figure PCTCN2018092614-appb-100002
    p t是所述全差-金字塔特征网络识别模型对所述训练图像的预测值,p为模型对于所述训练图像y=1的估计概率,p∈[0,1],y为所述训练图像的标注值,γ为调节参数。
    among them,
    Figure PCTCN2018092614-appb-100002
    p t is the predicted value of the full difference-pyramid feature network recognition model for the training image, p is the estimated probability of the model for the training image y = 1, p ∈ [0,1], and y is the training The label value of the image, γ is the adjustment parameter.
  15. 一个或多个存储有计算机可读指令的非易失性可读存储介质,所述计算机可读指令被一个或多个处理器执行时,使得所述一个或多个处理器执行如下步骤:One or more non-volatile readable storage media storing computer-readable instructions, which when executed by one or more processors, cause the one or more processors to perform the following steps:
    获取检测图像,采用异物检测模型对所述检测图像进行检测,获取检测结果;Acquiring a detection image, detecting the detection image by using a foreign object detection model, and acquiring a detection result;
    若所述检测结果为所述检测图像中存在异物,则获取所述异物在所述检测图像中的位置,作为基准位置,根据所述基准位置提取所述异物的特征向量,作为基准特征向量;If the detection result is that there is a foreign object in the detection image, obtaining a position of the foreign object in the detection image as a reference position, and extracting a feature vector of the foreign object according to the reference position as a reference feature vector;
    根据所述检测图像获取连续预定帧图像,组成识别图像集;Acquiring successive predetermined frames of images according to the detected images to form a recognition image set;
    根据所述基准位置提取所述识别图像集中每一识别图像的特征向量,并比对每一识别图像的特征向量和所述基准特征向量的特征向量相似度,获取比较结果;Extracting a feature vector of each recognition image in the recognition image set according to the reference position, and comparing the feature vector of each recognition image and the feature vector similarity of the reference feature vector to obtain a comparison result;
    根据所述比较结果生成识别结果,所述识别结果包括确认为异物和确认为非异物。A recognition result is generated according to the comparison result, and the recognition result includes confirmation as a foreign object and confirmation as a non-foreign object.
  16. 如权利要求15所述的非易失性可读存储介质,其特征在于,所述采用异物检测模型对所述检测图像进行检测,获取检测结果,具体包括:The non-volatile readable storage medium according to claim 15, wherein the detecting the detection image by using a foreign object detection model to obtain a detection result specifically comprises:
    对所述检测图像进行预处理,得到待识别图像;Preprocessing the detection image to obtain an image to be identified;
    将所述待识别图像输入到全差-金字塔特征网络识别模型中进行识别,获取分类置信图;Inputting the to-be-recognized image into a full-difference-pyramid feature network recognition model for recognition, and obtaining a classification confidence map;
    根据所述分类置信图获取检测结果,所述检测结果包括检测图像中存在异物和检测图像中不存在异物。A detection result is obtained according to the classification confidence map, and the detection result includes detecting the presence of a foreign object in the image and the absence of a foreign object in the detection image.
  17. 如权利要求16所述的非易失性可读存储介质,其特征在于,所述对所述检测图像进行预处理,得到待识别图像,具体包括:The non-volatile readable storage medium according to claim 16, wherein the preprocessing the detection image to obtain an image to be identified specifically comprises:
    采用多尺度视网膜算法对所述检测图像进行全局增强处理;Using a multi-scale retinal algorithm to globally enhance the detection image;
    采用拉普拉斯算子对全局增强处理后的所述检测图像进行锐化处理,得到待识别图像。A Laplace operator is used to sharpen the detection image after the global enhancement processing to obtain an image to be identified.
  18. 如权利要求16所述的非易失性可读存储介质,其特征在于,在所述将所述待识别图像输入到全差-金字塔特征网络识别模型中进行识别,获取分类置信图的步骤之前,还包括:The non-volatile readable storage medium according to claim 16, characterized in that before the step of inputting the image to be identified into a full-difference-pyramid feature network identification model to obtain a classification confidence map ,Also includes:
    获取训练样本集,对所述训练样本集中的训练图像进行分类标注;Acquiring a training sample set, and classifying the training images in the training sample set;
    采用所述训练样本集中所述分类标注的训练图像训练全差网络,得到目标输出向量;Training a full-difference network by using the classified labeled training images in the training sample set to obtain a target output vector;
    采用所述目标输出向量训练金字塔特征网络,得到所述全差-金字塔特征网络识别模型。The target output vector is used to train a pyramid feature network to obtain the full-difference-pyramid feature network recognition model.
  19. 如权利要求18所述的非易失性可读存储介质,其特征在于,所述训练全差网络,具体包括:The non-volatile readable storage medium according to claim 18, wherein the training of the full difference network specifically comprises:
    设置所述全差网络的初始卷积层,并采用所述全差网络中的最大池化层进行下采样;Setting an initial convolution layer of the full-difference network, and using the largest pooling layer in the full-difference network for downsampling;
    设置三层全差网络模块,每一所述全差网络模块包括一个全差卷积层和一个全差激活层,所述全差激活层中的激活函数采用线性激活函数;Setting up three layers of full-difference network modules, each of which includes a full-difference convolution layer and a full-difference activation layer, and the activation function in the full-difference activation layer adopts a linear activation function;
    在所述全差网络模块之间设置传输层,每一所述传输层包括规范化层、传输激活层和平均池化层。A transmission layer is provided between the fully differential network modules, and each of the transmission layers includes a normalization layer, a transmission activation layer, and an average pooling layer.
  20. 如权利要求18所述的非易失性可读存储介质,其特征在于,在训练全差-金字塔特征网络识别模型的过程中,损失函数采用Focal Loss函数实现:The non-volatile readable storage medium according to claim 18, wherein in the process of training the full-difference-pyramid feature network recognition model, the loss function is implemented using a FocalLoss function:
    FL(p t)=-(1-p t) γlog(p t); FL (p t ) =-(1-p t ) γ log (p t );
    其中,
    Figure PCTCN2018092614-appb-100003
    p t是所述全差-金字塔特征网络识别模型对所述训练图像的预测值,p为模型对于所述训练图像y=1的估计概率,p∈[0,1],y为所述训练图像的标注值,γ为调节参数。
    among them,
    Figure PCTCN2018092614-appb-100003
    p t is the predicted value of the full difference-pyramid feature network recognition model for the training image, p is the estimated probability of the model for the training image y = 1, p ∈ [0,1], and y is the training The label value of the image, γ is the adjustment parameter.
PCT/CN2018/092614 2018-06-06 2018-06-25 Method and device for recognizing foreign object debris at airport, computer apparatus, and storage medium WO2019232831A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN201810574129.1 2018-06-06
CN201810574129.1A CN108764202B (en) 2018-06-06 2018-06-06 Airport foreign matter identification method and device, computer equipment and storage medium

Publications (1)

Publication Number Publication Date
WO2019232831A1 true WO2019232831A1 (en) 2019-12-12

Family

ID=63999142

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2018/092614 WO2019232831A1 (en) 2018-06-06 2018-06-25 Method and device for recognizing foreign object debris at airport, computer apparatus, and storage medium

Country Status (2)

Country Link
CN (1) CN108764202B (en)
WO (1) WO2019232831A1 (en)

Cited By (29)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111178200A (en) * 2019-12-20 2020-05-19 海南车智易通信息技术有限公司 Identification method of instrument panel indicator lamp and computing equipment
CN111259763A (en) * 2020-01-13 2020-06-09 华雁智能科技(集团)股份有限公司 Target detection method and device, electronic equipment and readable storage medium
CN111310645A (en) * 2020-02-12 2020-06-19 上海东普信息科技有限公司 Overflow bin early warning method, device, equipment and storage medium for cargo accumulation amount
CN111340796A (en) * 2020-03-10 2020-06-26 创新奇智(成都)科技有限公司 Defect detection method and device, electronic equipment and storage medium
CN111340137A (en) * 2020-03-26 2020-06-26 上海眼控科技股份有限公司 Image recognition method, device and storage medium
CN111462059A (en) * 2020-03-24 2020-07-28 湖南大学 Parallel processing method and device for intelligent target detection of fetal ultrasound image
CN111897025A (en) * 2020-08-06 2020-11-06 航泰众联(北京)科技有限公司 Airport pavement foreign matter detection equipment and system based on 3D/2D integration detection
CN112308073A (en) * 2020-11-06 2021-02-02 中冶赛迪重庆信息技术有限公司 Method, system, equipment and medium for identifying loading and unloading transshipment state of scrap steel train
CN112766151A (en) * 2021-01-19 2021-05-07 北京深睿博联科技有限责任公司 Binocular target detection method and system for blind guiding glasses
CN112949474A (en) * 2021-02-26 2021-06-11 山东鹰格信息工程有限公司 Airport FOD monitoring method, equipment, storage medium and device
CN113096080A (en) * 2021-03-30 2021-07-09 四川大学华西第二医院 Image analysis method and system
CN113111703A (en) * 2021-03-02 2021-07-13 郑州大学 Airport pavement disease foreign matter detection method based on fusion of multiple convolutional neural networks
CN113111810A (en) * 2021-04-20 2021-07-13 北京嘀嘀无限科技发展有限公司 Target identification method and system
CN113159089A (en) * 2021-01-18 2021-07-23 安徽建筑大学 Pavement damage identification method, system, computer equipment and storage medium
CN113177459A (en) * 2021-04-25 2021-07-27 云赛智联股份有限公司 Intelligent video analysis method and system for intelligent airport service
US20210272236A1 (en) * 2019-02-28 2021-09-02 Tencent Technology (Shenzhen) Company Limited Image enhancement method and apparatus, and storage medium
CN113344900A (en) * 2021-06-25 2021-09-03 北京市商汤科技开发有限公司 Airport runway intrusion detection method, airport runway intrusion detection device, storage medium and electronic equipment
CN113449554A (en) * 2020-03-25 2021-09-28 北京灵汐科技有限公司 Target detection and identification method and system
CN113658135A (en) * 2021-08-17 2021-11-16 中国矿业大学 Fuzzy PID (proportion integration differentiation) -based self-adaptive dimming belt foreign matter detection method and system
CN113780074A (en) * 2021-08-04 2021-12-10 五邑大学 Method and device for detecting quality of wrapping paper and storage medium
CN113807227A (en) * 2021-09-11 2021-12-17 浙江浙能嘉华发电有限公司 Safety monitoring method, device and equipment based on image recognition and storage medium
CN114155746A (en) * 2021-12-01 2022-03-08 南京莱斯电子设备有限公司 FOD alarm accuracy rate and FOD alarm false alarm rate calculation method
CN114821194A (en) * 2022-05-30 2022-07-29 深圳市科荣软件股份有限公司 Equipment running state identification method and device
CN115147770A (en) * 2022-08-30 2022-10-04 山东千颐科技有限公司 Belt foreign matter vision recognition system based on image processing
CN116269378A (en) * 2023-01-09 2023-06-23 西安电子科技大学 Psychological health state detection device based on skin nicotinic acid response video analysis
CN116704446A (en) * 2023-08-04 2023-09-05 武汉工程大学 Real-time detection method and system for foreign matters on airport runway pavement
CN117523318A (en) * 2023-12-26 2024-02-06 宁波微科光电股份有限公司 Anti-light interference subway shielding door foreign matter detection method, device and medium
CN111539443B (en) * 2020-01-22 2024-02-09 北京小米松果电子有限公司 Image recognition model training method and device and storage medium
CN117726882A (en) * 2024-02-07 2024-03-19 杭州宇泛智能科技有限公司 Tower crane object identification method, system and electronic equipment

Families Citing this family (15)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109218233B (en) * 2018-11-14 2021-07-20 国家电网有限公司 OFDM channel estimation method based on depth feature fusion network
CN109583502B (en) * 2018-11-30 2022-11-18 天津师范大学 Pedestrian re-identification method based on anti-erasure attention mechanism
CN109766884A (en) * 2018-12-26 2019-05-17 哈尔滨工程大学 A kind of airfield runway foreign matter detecting method based on Faster-RCNN
CN109765557B (en) * 2018-12-30 2021-05-04 上海微波技术研究所(中国电子科技集团公司第五十研究所) FOD target self-adaptive rapid classification identification method, system and medium based on distribution characteristics
CN109829501B (en) * 2019-02-01 2021-02-19 北京市商汤科技开发有限公司 Image processing method and device, electronic equipment and storage medium
CN111862105A (en) * 2019-04-29 2020-10-30 北京字节跳动网络技术有限公司 Image area processing method and device and electronic equipment
CN110268420B (en) * 2019-05-09 2023-07-28 京东方科技集团股份有限公司 Computer-implemented method of detecting a foreign object on a background object in an image, device for detecting a foreign object on a background object in an image, and computer program product
CN111553277B (en) * 2020-04-28 2022-04-26 电子科技大学 Chinese signature identification method and terminal introducing consistency constraint
CN111680609B (en) * 2020-06-03 2023-02-07 合肥中科类脑智能技术有限公司 Foreign matter identification system and method based on image registration and target detection
CN112037197A (en) * 2020-08-31 2020-12-04 中冶赛迪重庆信息技术有限公司 Hot-rolled bar cold-shearing material accumulation detection method, system and medium
CN112580509B (en) * 2020-12-18 2022-04-15 中国民用航空总局第二研究所 Logical reasoning type road surface detection method and system
CN112686172B (en) * 2020-12-31 2023-06-13 上海微波技术研究所(中国电子科技集团公司第五十研究所) Airport runway foreign matter detection method, device and storage medium
CN113177922A (en) * 2021-05-06 2021-07-27 中冶赛迪重庆信息技术有限公司 Raw material foreign matter identification method, system, medium and electronic terminal
CN113393401B (en) * 2021-06-24 2023-09-05 上海科技大学 Object detection hardware accelerator, system, method, apparatus and medium
WO2023118937A1 (en) * 2021-12-20 2023-06-29 Sensetime International Pte. Ltd. Object recognition method, apparatus, device and storage medium

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105160362A (en) * 2015-10-22 2015-12-16 中国民用航空总局第二研究所 Runway FOD (Foreign Object Debris) image detection method and device
CN105740910A (en) * 2016-02-02 2016-07-06 北京格灵深瞳信息技术有限公司 Vehicle object detection method and device
EP3151164A2 (en) * 2016-12-26 2017-04-05 Argosai Teknoloji Anonim Sirketi A method for foreign object debris detection
CN107481233A (en) * 2017-08-22 2017-12-15 广州辰创科技发展有限公司 A kind of image-recognizing method being applied in FOD foreign bodies detection radars
CN107766864A (en) * 2016-08-23 2018-03-06 阿里巴巴集团控股有限公司 Extract method and apparatus, the method and apparatus of object identification of feature

Family Cites Families (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2017107188A1 (en) * 2015-12-25 2017-06-29 中国科学院深圳先进技术研究院 Method and apparatus for rapidly recognizing video classification
CN107423690B (en) * 2017-06-26 2020-11-13 广东工业大学 Face recognition method and device
CN107545263B (en) * 2017-08-02 2020-12-15 清华大学 Object detection method and device
CN107562900B (en) * 2017-09-07 2018-10-23 广州辰创科技发展有限公司 Method and system for analyzing airfield runway foreign matter based on big data mode

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105160362A (en) * 2015-10-22 2015-12-16 中国民用航空总局第二研究所 Runway FOD (Foreign Object Debris) image detection method and device
CN105740910A (en) * 2016-02-02 2016-07-06 北京格灵深瞳信息技术有限公司 Vehicle object detection method and device
CN107766864A (en) * 2016-08-23 2018-03-06 阿里巴巴集团控股有限公司 Extract method and apparatus, the method and apparatus of object identification of feature
EP3151164A2 (en) * 2016-12-26 2017-04-05 Argosai Teknoloji Anonim Sirketi A method for foreign object debris detection
CN107481233A (en) * 2017-08-22 2017-12-15 广州辰创科技发展有限公司 A kind of image-recognizing method being applied in FOD foreign bodies detection radars

Cited By (49)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11790497B2 (en) * 2019-02-28 2023-10-17 Tencent Technology (Shenzhen) Company Limited Image enhancement method and apparatus, and storage medium
US20210272236A1 (en) * 2019-02-28 2021-09-02 Tencent Technology (Shenzhen) Company Limited Image enhancement method and apparatus, and storage medium
CN111178200A (en) * 2019-12-20 2020-05-19 海南车智易通信息技术有限公司 Identification method of instrument panel indicator lamp and computing equipment
CN111178200B (en) * 2019-12-20 2023-06-02 海南车智易通信息技术有限公司 Method for identifying instrument panel indicator lamp and computing equipment
CN111259763B (en) * 2020-01-13 2024-02-02 华雁智能科技(集团)股份有限公司 Target detection method, target detection device, electronic equipment and readable storage medium
CN111259763A (en) * 2020-01-13 2020-06-09 华雁智能科技(集团)股份有限公司 Target detection method and device, electronic equipment and readable storage medium
CN111539443B (en) * 2020-01-22 2024-02-09 北京小米松果电子有限公司 Image recognition model training method and device and storage medium
CN111310645B (en) * 2020-02-12 2023-06-13 上海东普信息科技有限公司 Method, device, equipment and storage medium for warning overflow bin of goods accumulation
CN111310645A (en) * 2020-02-12 2020-06-19 上海东普信息科技有限公司 Overflow bin early warning method, device, equipment and storage medium for cargo accumulation amount
CN111340796A (en) * 2020-03-10 2020-06-26 创新奇智(成都)科技有限公司 Defect detection method and device, electronic equipment and storage medium
CN111340796B (en) * 2020-03-10 2023-07-21 创新奇智(成都)科技有限公司 Defect detection method and device, electronic equipment and storage medium
CN111462059A (en) * 2020-03-24 2020-07-28 湖南大学 Parallel processing method and device for intelligent target detection of fetal ultrasound image
CN111462059B (en) * 2020-03-24 2023-09-29 湖南大学 Parallel processing method and device for intelligent target detection of fetal ultrasonic image
CN113449554A (en) * 2020-03-25 2021-09-28 北京灵汐科技有限公司 Target detection and identification method and system
CN113449554B (en) * 2020-03-25 2024-03-08 北京灵汐科技有限公司 Target detection and identification method and system
CN111340137A (en) * 2020-03-26 2020-06-26 上海眼控科技股份有限公司 Image recognition method, device and storage medium
CN111897025A (en) * 2020-08-06 2020-11-06 航泰众联(北京)科技有限公司 Airport pavement foreign matter detection equipment and system based on 3D/2D integration detection
CN112308073B (en) * 2020-11-06 2023-08-25 中冶赛迪信息技术(重庆)有限公司 Method, system, equipment and medium for identifying loading and unloading and transferring states of scrap steel train
CN112308073A (en) * 2020-11-06 2021-02-02 中冶赛迪重庆信息技术有限公司 Method, system, equipment and medium for identifying loading and unloading transshipment state of scrap steel train
CN113159089A (en) * 2021-01-18 2021-07-23 安徽建筑大学 Pavement damage identification method, system, computer equipment and storage medium
CN112766151A (en) * 2021-01-19 2021-05-07 北京深睿博联科技有限责任公司 Binocular target detection method and system for blind guiding glasses
CN112949474A (en) * 2021-02-26 2021-06-11 山东鹰格信息工程有限公司 Airport FOD monitoring method, equipment, storage medium and device
CN113111703B (en) * 2021-03-02 2023-07-28 郑州大学 Airport pavement disease foreign matter detection method based on fusion of multiple convolutional neural networks
CN113111703A (en) * 2021-03-02 2021-07-13 郑州大学 Airport pavement disease foreign matter detection method based on fusion of multiple convolutional neural networks
CN113096080B (en) * 2021-03-30 2024-01-16 四川大学华西第二医院 Image analysis method and system
CN113096080A (en) * 2021-03-30 2021-07-09 四川大学华西第二医院 Image analysis method and system
CN113111810A (en) * 2021-04-20 2021-07-13 北京嘀嘀无限科技发展有限公司 Target identification method and system
CN113111810B (en) * 2021-04-20 2023-12-08 北京嘀嘀无限科技发展有限公司 Target identification method and system
CN113177459A (en) * 2021-04-25 2021-07-27 云赛智联股份有限公司 Intelligent video analysis method and system for intelligent airport service
CN113344900B (en) * 2021-06-25 2023-04-18 北京市商汤科技开发有限公司 Airport runway intrusion detection method, airport runway intrusion detection device, storage medium and electronic device
CN113344900A (en) * 2021-06-25 2021-09-03 北京市商汤科技开发有限公司 Airport runway intrusion detection method, airport runway intrusion detection device, storage medium and electronic equipment
CN113780074A (en) * 2021-08-04 2021-12-10 五邑大学 Method and device for detecting quality of wrapping paper and storage medium
CN113658135B (en) * 2021-08-17 2024-02-02 中国矿业大学 Fuzzy PID-based self-adaptive dimming belt foreign matter detection method and system
CN113658135A (en) * 2021-08-17 2021-11-16 中国矿业大学 Fuzzy PID (proportion integration differentiation) -based self-adaptive dimming belt foreign matter detection method and system
CN113807227A (en) * 2021-09-11 2021-12-17 浙江浙能嘉华发电有限公司 Safety monitoring method, device and equipment based on image recognition and storage medium
CN113807227B (en) * 2021-09-11 2023-07-25 浙江浙能嘉华发电有限公司 Safety monitoring method, device, equipment and storage medium based on image recognition
CN114155746A (en) * 2021-12-01 2022-03-08 南京莱斯电子设备有限公司 FOD alarm accuracy rate and FOD alarm false alarm rate calculation method
CN114155746B (en) * 2021-12-01 2022-09-13 南京莱斯电子设备有限公司 FOD alarm accuracy rate and FOD alarm false alarm rate calculation method
CN114821194A (en) * 2022-05-30 2022-07-29 深圳市科荣软件股份有限公司 Equipment running state identification method and device
CN114821194B (en) * 2022-05-30 2023-07-25 深圳市科荣软件股份有限公司 Equipment running state identification method and device
CN115147770B (en) * 2022-08-30 2022-12-02 山东千颐科技有限公司 Belt foreign matter vision recognition system based on image processing
CN115147770A (en) * 2022-08-30 2022-10-04 山东千颐科技有限公司 Belt foreign matter vision recognition system based on image processing
CN116269378A (en) * 2023-01-09 2023-06-23 西安电子科技大学 Psychological health state detection device based on skin nicotinic acid response video analysis
CN116269378B (en) * 2023-01-09 2023-11-17 西安电子科技大学 Psychological health state detection device based on skin nicotinic acid response video analysis
CN116704446B (en) * 2023-08-04 2023-10-24 武汉工程大学 Real-time detection method and system for foreign matters on airport runway pavement
CN116704446A (en) * 2023-08-04 2023-09-05 武汉工程大学 Real-time detection method and system for foreign matters on airport runway pavement
CN117523318A (en) * 2023-12-26 2024-02-06 宁波微科光电股份有限公司 Anti-light interference subway shielding door foreign matter detection method, device and medium
CN117523318B (en) * 2023-12-26 2024-04-16 宁波微科光电股份有限公司 Anti-light interference subway shielding door foreign matter detection method, device and medium
CN117726882A (en) * 2024-02-07 2024-03-19 杭州宇泛智能科技有限公司 Tower crane object identification method, system and electronic equipment

Also Published As

Publication number Publication date
CN108764202B (en) 2023-04-18
CN108764202A (en) 2018-11-06

Similar Documents

Publication Publication Date Title
WO2019232831A1 (en) Method and device for recognizing foreign object debris at airport, computer apparatus, and storage medium
WO2019232830A1 (en) Method and device for detecting foreign object debris at airport, computer apparatus, and storage medium
US11430259B2 (en) Object detection based on joint feature extraction
CN108304873B (en) Target detection method and system based on high-resolution optical satellite remote sensing image
CA3017646C (en) Label and field identification without optical character recognition (ocr)
Alvarez et al. Road scene segmentation from a single image
Khan et al. An efficient contour based fine-grained algorithm for multi category object detection
Neto et al. Brazilian vehicle identification using a new embedded plate recognition system
WO2019154383A1 (en) Tool detection method and device
US9224207B2 (en) Segmentation co-clustering
Do et al. Automatic license plate recognition using mobile device
CN110826408B (en) Face recognition method by regional feature extraction
Harini et al. Sign language translation
KR20210020065A (en) Systems and methods for finding and classifying patterns in images with vision systems
CN115761834A (en) Multi-task mixed model for face recognition and face recognition method
CN112784712B (en) Missing child early warning implementation method and device based on real-time monitoring
CN112686248B (en) Certificate increase and decrease type detection method and device, readable storage medium and terminal
CN116934762B (en) System and method for detecting surface defects of lithium battery pole piece
Singh Gaussian elliptical fitting based skin color modeling for human detection
Jin et al. Road curvature estimation using a new lane detection method
CN111488811A (en) Face recognition method and device, terminal equipment and computer readable medium
CN108460772B (en) Advertisement harassment fax image detection system and method based on convolutional neural network
US10366278B2 (en) Curvature-based face detector
Lin et al. Face detection algorithm based on multi-orientation gabor filters and feature fusion
CN110738225A (en) Image recognition method and device

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 18922071

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

32PN Ep: public notification in the ep bulletin as address of the adressee cannot be established

Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205A DATED 12/03/2021)

122 Ep: pct application non-entry in european phase

Ref document number: 18922071

Country of ref document: EP

Kind code of ref document: A1