WO2020114097A1 - 一种边界框确定方法、装置、电子设备及存储介质 - Google Patents

一种边界框确定方法、装置、电子设备及存储介质 Download PDF

Info

Publication number
WO2020114097A1
WO2020114097A1 PCT/CN2019/111645 CN2019111645W WO2020114097A1 WO 2020114097 A1 WO2020114097 A1 WO 2020114097A1 CN 2019111645 W CN2019111645 W CN 2019111645W WO 2020114097 A1 WO2020114097 A1 WO 2020114097A1
Authority
WO
WIPO (PCT)
Prior art keywords
bounding box
target object
image
feature
neural
Prior art date
Application number
PCT/CN2019/111645
Other languages
English (en)
French (fr)
Inventor
刘裕峰
胡玉坤
郑文
Original Assignee
北京达佳互联信息技术有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 北京达佳互联信息技术有限公司 filed Critical 北京达佳互联信息技术有限公司
Publication of WO2020114097A1 publication Critical patent/WO2020114097A1/zh

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/20Image preprocessing
    • G06V10/24Aligning, centring, orientation detection or correction of the image
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features

Definitions

  • the present application relates to the field of image processing technology, and in particular, to a method, device, electronic device, and storage medium for determining a bounding box.
  • the bounding box is the boundary that defines the edge of the target object during object detection.
  • the boundary of the target object is mainly divided by the first position parameter value corresponding to the bounding box.
  • the convolutional neural sub-network taking the image to be detected as input, after extracting image features through the convolutional neural sub-network, and obtaining the image features of the image to be detected, the image features of the image to be detected are passed through the second nerve
  • the subnet outputs the bounding box of the target object in the image to be detected.
  • the boundary box obtained based on the second neural sub-network has a problem of inaccurate recognition.
  • the present application provides a method and device for determining a bounding box.
  • a method for determining a bounding box including:
  • the image to be detected into the first convolutional neural sub-network to obtain the first feature of the target object in the image to be detected;
  • the first feature includes: the boundary image feature of the target object in the image to be detected;
  • the target bounding box of the target object in the image to be detected is determined.
  • a device for determining a bounding box including:
  • a first feature acquisition module configured to input the image to be detected into the first convolutional neural sub-network to acquire the first feature of the target object in the image to be detected; the first feature includes: Describe the boundary image features of the target object;
  • the first bounding box determining module is configured to input the first features into n second neural sub-networks respectively to obtain n first bounding boxes of the target object; the n is an integer greater than or equal to 2; Each first bounding box includes a first position parameter value;
  • the target bounding box determining module is configured to determine the target bounding box of the target object in the image to be detected based on the first position parameter values of the n first bounding boxes.
  • an electronic device including: a processor; a memory for storing processor executable instructions; wherein the processor is configured to: execute the above-mentioned bounding box determination method.
  • a non-transitory computer-readable storage medium which when a command in the storage medium is executed by a processor of a mobile terminal, enables the mobile terminal to perform the above-mentioned bounding box determination method .
  • a computer program product which, when instructions in the computer program product are executed by a processor of a mobile terminal, enables the mobile terminal to execute the above-mentioned bounding box determination method.
  • the image to be detected is input into the first convolutional neural network to obtain the first feature of the target object in the image to be detected;
  • the first feature includes: the boundary image feature of the target object in the image to be detected ; Enter the first feature into n second neural network respectively to get n first bounding boxes of the target object;
  • n is an integer greater than or equal to 2; each first bounding box includes the first position parameter value; based on n
  • Each first position parameter value of the first bounding box determines the target bounding box of the target object in the image to be detected.
  • the target bounding box of the target object is obtained based on n second neural network at the same time, the determined target bounding box can be located in a relatively middle region of each first position parameter value of each first bounding box, which can avoid a single second neuron
  • the problem of inaccurate division of the target object caused by the instability of the network can improve the robustness of determining the bounding box of the target object, can effectively avoid the problem of overfitting, and thus can improve the accuracy of the bounding box of the target object .
  • Fig. 1 is a flow chart of a method for determining a bounding box according to an exemplary embodiment.
  • Fig. 2 is a flowchart of another method for determining a bounding box according to an exemplary embodiment.
  • Fig. 3 is a block diagram of a device for determining a bounding box according to an exemplary embodiment.
  • Fig. 4 is a block diagram of another apparatus for determining a bounding box according to an exemplary embodiment.
  • Fig. 5 is a block diagram of an electronic device according to an exemplary embodiment.
  • Fig. 6 is a block diagram of another electronic device according to an exemplary embodiment.
  • Fig. 1 is a flowchart of a method for determining a bounding box according to an exemplary embodiment. As shown in Fig. 1, the method for determining a bounding box is used in a terminal, and may include the following steps.
  • step S11 the image to be detected is input into the first convolutional neural sub-network to obtain the first feature of the target object in the image to be detected.
  • the first feature may include: a boundary image feature of the target object in the image to be detected.
  • the image to be detected may be in RGB format, YUV format, or the like. In the embodiments of the present application, this is not specifically limited.
  • the above-mentioned image to be detected may be a captured image, or an image downloaded based on a network, or may be one or more frames of images in a video, etc. In the embodiment of the present application, this is not specifically limited. .
  • the target object in the image to be detected may be any subject in the image to be detected, for example, if the image to be detected includes: puppy, kitten, flower, etc., the image to be detected
  • the target object in the image may be a puppy in the image to be detected, or the target object in the image to be detected may be a kitten in the image to be detected, or the target object in the image to be detected may be the image to be detected Flowers in etc.
  • the first convolutional neural network may be a convolutional neural network extracted for the first feature of the target object.
  • the first convolutional neural sub-network may be a convolutional neural sub-network extracted for the first feature of the puppy. In the embodiments of the present application, this is not specifically limited.
  • the first convolutional neural network may include: an input layer, a convolutional layer, a pooling layer, etc. In the embodiment of the present application, this is not specifically limited.
  • the first convolutional neural network may also include: a dense convolutional network.
  • the input of each first network layer in the dense convolutional network includes the inputs of all second network layers in the dense convolutional network, where the first network layer may be any network layer in the dense convolutional network, the The second network layer may be each network layer before the first network layer in the dense convolutional network.
  • the input of the fourth layer may include: in the dense convolutional network, the first layer input, the second layer input, the third The input of the layer, and the output of the third layer.
  • the dense convolutional network can reduce the phenomenon of gradient disappearance, and at the same time make the first feature transfer of the target object in the image to be detected more effective, which is convenient for network training.
  • the problem of overfitting can be suppressed to a certain extent, which further makes the first feature of the target object in the extracted image to be detected more accurate based on the first convolutional neural network.
  • the first feature of the target object in the image to be detected may include: a boundary image feature of the target object in the image to be detected.
  • the boundary image features of the target object in the image to be detected may be: contour image features, shape image features, spatial position image features of the target object in the image to be detected, and so on. In the embodiments of the present application, this is not specifically limited.
  • the image to be detected is input into the first convolutional neural network to obtain the first feature of the target object in the image to be detected, and the first feature may include: the target object's feature in the image to be detected Boundary image features.
  • step S12 the first feature is input into n second neural sub-networks respectively to obtain n first bounding boxes of the target object.
  • n is an integer greater than or equal to 2
  • each first bounding box includes a first position parameter value.
  • n is an integer greater than or equal to 2, which can be specifically set according to actual needs.
  • n may be 2
  • n may be 8
  • n may be 20, etc. In the embodiment of the present application, this is not specifically limited.
  • each of the above-mentioned second neural sub-networks may output the boundary of the target object based on the first feature of the target object in the image to be detected, and then each first bounding box of the target object may be obtained.
  • the second neural sub-network may include: at least one of a convolution layer, a pooling layer, a fully connected layer, etc., which is not specifically limited in the embodiment of the present application.
  • the first bounding box of the target object may be a bounding box characterizing the outline of the target object in the image to be detected, the bounding box may be a rectangle, and the first bounding box may include the first position parameter value ,
  • the first position parameter value of the first bounding box may be a position parameter capable of defining the contour of the target object in the image to be detected, for example, the first position parameter of the first bounding box may be: The center coordinates, and the length and width of the bounding box, or the first position parameter of the first bounding box may be: the coordinates of the vertices on at least one diagonal of the first bounding box, for example, the first A position parameter may be: the vertex coordinates of the upper left corner of the bounding box and the vertex coordinates of the lower right corner. In the embodiments of the present application, this is not specifically limited.
  • n second neural subnetworks may be n identical second neural subnetworks, or n second neural subnetworks may be n different second neural subnetworks, or, n
  • the second neural network may include m groups of third neural networks, where m is an integer greater than 0 and less than n, and each group of third neural networks may include: y identical fourth neural networks, and y is greater than or equal to 2 an integer less than or equal to n, and x second neural network different from the third neural network or the fourth neural network and so on. In the embodiments of the present application, this is not specifically limited.
  • the eight second neural sub-networks may include: eight identical second neural sub-networks, for example, A1, A2, A3, A4, A5, A6, A7, A8, or, the The eight second neural sub-networks may include: eight different second neural sub-networks, such as A1, B1, C1, D1, E1, F1, G1, H1, or, the eight second neurons
  • the network may include: 3 groups of third neural sub-networks, then m may be 3, where the first group of third neural sub-networks includes 3 identical fourth neural sub-networks, such as: A1, A2, A3, then y1 Can be 3, the second group of third neural subnetworks can include 2 identical fourth neural subnetworks, such as B1, B2, then y2 can be 2, and the third group of third neural subnetworks can include 2 identical
  • the fourth neural network such as C1 and C2, y3 can be 2 and a second neural network different from the third neural network or the fourth neural network, for example, D1, then x can be 1.
  • the above first features may be input into the above-mentioned n second neural network to determine the boundary of the target object to obtain n first boundary boxes of the target object, each of the first boundary boxes includes The first position parameter value.
  • the first feature can be input into the A1, B1, C1, D1, E1, F1, G1, H1, 8 second neural network, refer to the following table 1, table 1 shows the above A model structure of 8 second neural network.
  • Mobile_unit may be a fifth neural network composed of e first convolutional layers, f first pooling layers, and p first fully connected layers, and e, f, p, s, and w may be greater than An integer of 0, s*w can be the size of the above first feature, input can be the number of the above first feature channels, and the size and channel number of the first feature can be determined according to the actual situation of the first feature, for example, s*w can Is: 16*16, the number of channels can be 1, or the number of channels can be 3, or the number of channels can be 4, etc., AvgPool can be the average pooling layer, conv_layer can be the convolutional layer, MaxPool can be the maximum pooling layer . In Table 1, conv_layer1, conv_layer9, conv_layer10, and conv_layer12 may be the same or different, which is not specifically limited in the embodiment of the present application.
  • each second neural sub-network the number of each layer can be set according to actual conditions. In the embodiments of the present application, this is not specifically limited.
  • the first feature may be input into the A1 second neural network to obtain the first first bounding box
  • the first feature may be input into the B1 second neural network to obtain the second First bounding box, input the first feature into the C1 second neural network, and then obtain the third first bounding box, input the first feature into the D1 second neural network, and obtain the fourth A bounding box, input the first feature into the E1 second neural network to obtain a fifth first bounding box, and input the first feature into the F1 second neural network to obtain a sixth first boundary Box, input the first feature into the G1 second neural network to obtain a seventh first bounding box, and input the first feature into the H1 second neural network to obtain an eighth first bounding box.
  • each first bounding box includes a first position parameter value.
  • step S13 based on the respective first position parameter values of the n first bounding boxes, the target bounding box of the target object in the image to be detected is determined.
  • the target bounding box of the target object in the image to be detected is determined based on the first position parameter values of the n first bounding boxes.
  • the first position parameter values of the n first bounding boxes may be averaged, etc., to further determine the target bounding box of the target object in the image to be detected.
  • the mean value calculation may be geometric mean value calculation, arithmetic mean value calculation, weighted mean value calculation, root mean square mean value calculation, harmonic mean value calculation, etc., which are not specifically limited in the embodiments of the present application.
  • the above first feature is input into the above A1, B1, C1, D1, E1, F1, G1, H1, 8 second neural sub-networks to obtain 8 first bounding boxes, each of the first boundary
  • the frame includes first position parameter values, and based on the first position parameter values of the eight first bounding boxes, the target bounding box of the target object in the image to be detected is determined.
  • the determined target bounding box may be located in a relatively middle region of each first position parameter value of each first bounding box, which can avoid the problem of inaccurate division of the target object caused by the instability of the single second neural sub-network. Improve the accuracy of the target object's bounding box.
  • the image to be detected is input into the first convolutional neural network to obtain the first feature of the target object in the image to be detected;
  • the first feature includes: the boundary image feature of the target object in the image to be detected ; Enter the first feature into n second neural network respectively to get n first bounding boxes of the target object;
  • n is an integer greater than or equal to 2; each first bounding box includes the first position parameter value; based on n
  • Each first position parameter value of the first bounding box determines the target bounding box of the target object in the image to be detected.
  • the target bounding box of the target object is obtained based on n second neural network at the same time, the determined target bounding box can be located in a relatively middle region of each first position parameter value of each first bounding box, which can avoid a single second neuron
  • the problem of inaccurate division of the target object caused by the instability of the network can improve the robustness of determining the bounding box of the target object, can effectively avoid the problem of overfitting, and thus can improve the accuracy of the bounding box of the target object .
  • Fig. 2 is a flowchart of a method for determining a bounding box according to an exemplary embodiment. As shown in Fig. 2, the method for determining a bounding box includes the following steps.
  • step S21 the first image sample data is acquired.
  • the first image sample data may be an image captured by a shooting device such as a camera, or may be an image downloaded or received through a network.
  • the first image sample data may include a plurality of different objects or subjects, for example, may include: people, puppies, kittens, and so on. In the embodiments of the present application, this is not specifically limited.
  • the first image sample data may include the above target object.
  • the target object is a puppy
  • the first image sample data may include a puppy. In the embodiments of the present application, this is not specifically limited.
  • the corresponding feature label map may be obtained for the first image sample data, and the feature label map may be the first image sample Figures marked with the first feature of each subject in the data.
  • the feature label map is the image after labeling each pixel in the first image sample data, for example, labeling each Whether the pixel point is the first feature of the puppy, if the pixel belongs to the first feature of the puppy, it can be specifically marked which type of feature in the first feature the pixel is. In the embodiments of the present application, this is not specifically limited.
  • step S22 based on the first image sample data, a first convolutional neural network is trained.
  • the first convolutional neural network can be trained on the first image sample data.
  • the first image sample data can be input into the original first convolutional neural sub-network, and the original first convolutional neural sub-network can be used to extract the feature part corresponding to the first image sample data. It may include local features of each subject in the first image sample data or overall features of the first image sample data, etc. The local features may identify the shape, contour, etc. of each subject in the first image sample data. The overall feature of the image sample data may be the logical structure of each subject in the first image sample data, and so on. For example, people are usually taller than puppies or kittens.
  • the first convolutional neural sub-network may have a corresponding relationship with the target object that acquires the first feature, and during training of the first convolutional neural sub-network, the first image may be mainly concerned Relevant data of the target object in the sample data, thereby improving the pertinence of the first convolutional neural sub-network relative to the target object, thereby improving the accuracy of extracting the first feature for the target object. In the embodiments of the present application, this is not specifically limited.
  • step S23 the image to be detected is input into the first convolutional neural network to obtain the first feature of the target object in the image to be detected; the first feature includes: the boundary image feature of the target object in the image to be detected.
  • step S24 the first image sample data is input into the first convolutional neural sub-network to obtain the third feature of the target object in the first image sample data; the third feature includes: the boundary of the target object in the first image sample data Image features.
  • the first image sample data may be input into the trained first convolutional neural sub-network to obtain the third feature of the target object in the first image sample data.
  • the third feature may include : The boundary image feature of the target object in the first image sample data.
  • the first image sample data may include a puppy.
  • the first image sample data is input into the first trained convolutional neural network to obtain the third feature of the puppy in the first image sample data.
  • the third feature may include: the puppy in the first image sample data Boundary image features, such as the edge image feature and shape image feature of the puppy in the first image sample data.
  • step S25 based on the third feature, n second neural sub-networks are trained.
  • the n second neural sub-networks may be trained for the third feature.
  • the above-mentioned third feature may be input into each original second neural sub-network separately, and each original second neural sub-network may be used to determine the original target object in the first image sample data based on the above-mentioned third feature Bounding box.
  • a corresponding boundary label map may be obtained, and the boundary label map may be the first image sample
  • the diagram of the actual bounding box of the target object in the data For example, for the above example, if the first image sample data includes: people, kittens, puppies, etc., the boundary label image is the image after labeling each pixel in the first image sample data, for example, labeling each Whether the pixel is the actual bounding box of the puppy, etc. In the embodiments of the present application, this is not specifically limited.
  • each second neural sub-network may have a corresponding relationship with the target object that obtains the bounding box, and during the training of each second neural sub-network, the main focus may be on the first image sample data. Relevant data of the target object, thereby improving the pertinence of each second neural network with respect to the target object, thereby improving the recognition accuracy of determining the bounding box for the target object. In the embodiments of the present application, this is not specifically limited.
  • step S26 if the n second neural sub-networks include: m sets of third neural sub-networks, each group of third neural sub-networks includes: y identical fourth neural sub-networks, and the first feature is differentiated Processing to obtain y different second features.
  • m is an integer greater than 0 and less than n.
  • n is 8, m may be 1, 2, 3, or the like.
  • y is an integer greater than or equal to 2 and less than or equal to n. If n is 8, y can be 2, 3, etc.
  • n second neural sub-networks include: m sets of third neural sub-networks, each set of third neural sub-networks includes: y identical fourth neural sub-networks, y is an integer greater than or equal to 2 and less than or equal to n , N is an integer greater than or equal to 2, and the first feature is differentiated to obtain y different second features.
  • the 8 second neural sub-networks may include: 3 groups of third neural sub-networks, then m may be 3, where the first group of third neural sub-networks includes 3 identical fourth Neural network, such as: A1, A2, A3, then y1 can be 3, the second group of third neural network can include 2 identical fourth neural network, for example, B1, B2, then y2 can be 2,
  • the third group of third neural sub-networks may include two identical fourth neural sub-networks, for example, C1, C2, then y3 may be 2, and one different from the above-mentioned third neural sub-network or fourth neural sub-network
  • the second neural network for example, D1.
  • the differentiation processing includes at least one of cropping processing, scaling processing, mirror image processing, rotation processing, enhancement processing, and sharpening processing.
  • the cropping process may be to crop a part of the image features from the above-mentioned first features, the zooming process may be to enlarge or reduce the above-mentioned first features, etc.
  • the mirroring process may be to perform the mirroring process to the above-mentioned first features, and the rotation process may be to The first feature is rotated at a preset angle, and the enhancement process may be denoising or enhancing the edge of the first feature, and the sharpening process may be contour compensation for the first feature, enhancing the edge and gray transition , Etc., that is, by performing the above-mentioned differentiation processing on the above-mentioned image processing to be detected, y different second characteristics can be obtained. In the embodiments of the present application, this is not specifically limited.
  • the above first feature can be performed Differentiated processing to obtain 3 different second features;
  • the second group of third neural sub-networks includes two identical fourth neural sub-networks, for example, B1, B2, then y2 can be 2, then the first Differentiate the features to obtain 2 different second features;
  • the third group of third neural subnetworks can include two identical fourth neural subnetworks, for example, C1, C2, then y3 can be 2, you can The first feature is differentiated to obtain two different second features.
  • the second group of third neural sub-networks includes two identical fourth neural sub-networks, such as B1 and B2, then y2 can be 2, and the third group of third neural sub-networks can include two identical Of the fourth neural network, such as C1 and C2, then y3 can be 2.
  • the number of the fourth neural network included in the second neural network of the second group and the third neural network of the third group is the same, both Is 2, in order to reduce the workload, you can only differentiate the number of the fourth neural sub-network in a group of third neural sub-networks, for example, differentiate the first feature above to get 2 different
  • the second feature of, the second feature obtained by differentiation is directly used as the second feature corresponding to the third group of third neural network, which is equivalent to, for the second group of third neural network and the third group of third neural network
  • only the first feature is differentiated to obtain 2 different second features, not 2 different second features. In order to reduce the amount of calculation and so on, improve efficiency.
  • the above-mentioned differentiation processing on the image processing is also possible to perform the above-mentioned differentiation processing on the image processing to be detected only for the group with the largest number of the fourth neural sub-networks among the m sets of the third neural sub-networks to obtain multiple second features ,
  • the third neural network of the remaining groups can be directly selected from the multiple second features described above.
  • the eight second neural sub-networks may include: three groups of third neural sub-networks, and the first group of third neural sub-networks, including three identical fourth neural sub-networks , Such as: A1, A2, A3, then y1 can be 3, the second group of third neural subnetworks includes 2 identical fourth neural subnetworks, such as B1, B2, then y2 can be 2, the third group
  • the three-neural network may include two identical fourth neural networks, for example, C1 and C2, then y3 may be 2.
  • the first group of third neural networks includes the fourth The number of neural sub-networks is the largest, three.
  • the second group of third neural network and the third group of third neural network can be directly selected from the above three second features. Equivalent to, for the three sets of the third neural network, only the first feature is differentiated to obtain 3 different second features, not 7 different second features, thereby reducing the amount of calculation and improving efficiency .
  • the second neural network is: any combination of a average pooling layer, b maximum pooling layers, c convolutional layers, and d fully connected layers; a, b , D is an integer greater than or equal to 0; c is an integer greater than or equal to 1.
  • the second neural sub-network may be: any combination of a average pooling layer, b maximum pooling layers, c convolutional layers, and d fully connected layers, where a, b, d can be an integer greater than or equal to 0; c can be an integer greater than or equal to 1.
  • the second neural sub-network may be: 10 convolutional layers are connected to 5 average pooling layers, and then 1 fully connected layer is connected.
  • step S27 y different second features are input to each fourth neural sub-network in each group of third neural sub-networks to obtain k first bounding boxes of the target object.
  • each first bounding box includes a first position parameter value.
  • each fourth neural sub-network in each group of third neural sub-networks may be input into each fourth neural sub-network in each group of third neural sub-networks to determine the boundary of the target object to obtain k target objects
  • k is an integer greater than 0 and less than or equal to n
  • each first bounding box may include a first position parameter value.
  • the 8 second neural sub-networks may include: 3 groups of third neural sub-networks, then m may be 3, where the first group of third neural sub-networks includes 3
  • the same fourth neural network such as: A1, A2, A3, then y1 can be 3
  • the second group of third neural network can include two identical fourth neural networks, such as B1, B2, then y2 It can be 2
  • the third group of third neural subnetworks can include 2 identical fourth neural subnetworks, for example, C1, C2, then y3 can be 2, and 1 is the same as the above third neural subnetwork or fourth neural network
  • a second neural subnet with a different subnet for example, D1.
  • the above three different second features can be input into the third fourth neural sub-network in the first group of third neural sub-networks to determine the boundary of the target object to obtain three first bounding boxes of the target object.
  • the above two different second features are input into the second fourth neural network of the second group of third neural networks to determine the boundary of the target object to obtain two first bounding boxes of the target object.
  • Different second features input 2 fourth neural sub-networks in the third group of third neural sub-networks, determine the boundary of the target object, and obtain two first bounding boxes of the target object. A total of 7 first bounding boxes of the above target objects are obtained.
  • step S28 the first feature is input into x second neural sub-networks respectively to obtain x first bounding boxes of the target object.
  • x second neural sub-networks are n second neural sub-networks, except for the third neural sub-network or the fourth neural sub-network, x is an integer greater than or equal to 0 and less than n.
  • a bounding box includes the first position parameter value.
  • the above first features may be input into x second neural network to determine the boundary of the target object to obtain x first bounding boxes of the target object, and the above x second neurons
  • the network may be a neural network other than the third neural network or the fourth neural network among the n second neural network.
  • x can be an integer greater than or equal to 0 and less than n.
  • x may be the difference between n and k.
  • Each first bounding box may include a first position parameter value.
  • the 8 second neural sub-networks may include: 3 groups of third neural sub-networks, then m may be 3, where the first group of third neural sub-networks includes 3
  • the same fourth neural network such as: A1, A2, A3, then y1 can be 3
  • the second group of third neural network can include two identical fourth neural networks, such as B1, B2, then y2 It can be 2
  • the third group of third neural subnetworks can include 2 identical fourth neural subnetworks, for example, C1, C2, then y3 can be 2, and 1 is the same as the above third neural subnetwork or fourth neural network
  • a second neural subnet with different subnets for example, D1, x may be 1.
  • the first feature may be input into the second neural network D1 to determine the boundary of the target object to obtain a first bounding box of the target object.
  • n second neural sub-networks include: m sets of third neural sub-networks, the m is an integer greater than 0 and less than n, and each of the third neural sub-networks includes: y
  • the above y is an integer greater than or equal to 2 and less than or equal to n
  • the first feature is differentiated to obtain y different second features
  • the above y different second features are respectively Input each fourth neural sub-network in each group of third neural sub-networks to determine the boundary of the target object to obtain k first bounding boxes of the target object, k is an integer greater than 0 and equal to or less than n
  • Characteristics respectively input x second neural network to determine the boundary of the above target object, get x first bounding box of the target object, x second neural network is the above n second neural network, except In the neural network other than the third neural network or the fourth neural network, x is an integer greater than or equal to 0 and less than n.
  • the first feature is differentiated, so that the second features input by the same sub-network are different, so that the same
  • the sub-network can pay attention to the different boundary image features of the target object in the image to be detected, and to a certain extent, it can improve the accuracy of the subsequent determination of the target boundary frame of the target object.
  • step S29 the first position parameter values of the n first bounding boxes are weighted and summed according to preset weights to obtain the second position parameter value of the target object, and the image to be detected is determined based on the second position parameter value The target bounding box of the target object.
  • the first position parameter values of the n first bounding boxes may be weighted and summed according to preset weights to obtain the second position parameter value of the target object, and based on the second position parameter The value determines the target bounding box of the target object in the image to be detected.
  • the preset weight may be set according to actual needs, etc. In the embodiments of the present application, this is not specifically limited.
  • the second position parameter value of the target object may be: the coordinates of the vertex of the upper left corner are (3 ⁇ 0.5+3 ⁇ 0.5, 9 ⁇ 0.5+7 ⁇ 0.5), the coordinates of the apex of the upper right corner are (6 ⁇ 0.5+4 ⁇ 0.5, 2 ⁇ 0.5+4 ⁇ 0.5), that is, the second position parameter value of the target object can be: the coordinates of the apex of the upper left corner are (3, 8), The vertex coordinates of the upper right corner are (5,
  • the determined target bounding box may be located in a relatively middle region of the first position parameter values of the above two first bounding boxes, which can avoid the problem of inaccurate division of the target object caused by the instability of the single second neural network , Can improve the accuracy of the target object's bounding box.
  • step S30 the actual bounding box of the target object in the image to be detected is obtained.
  • the actual bounding box of the target object in the image to be detected may be obtained.
  • the process may be manually determining the actual frame of the target object in the image to be detected, or the process may be: inputting the image to be detected into the first bounding box input model to obtain the target object in the image to be detected Actual bounding box.
  • the first bounding box model may be a model for determining the bounding box of the target object with high confidence for the target object, which is not specifically limited in the embodiment of the present application.
  • step S31 based on the actual bounding box and the target bounding box, the evaluation value of the target bounding box is determined.
  • the evaluation value of the target bounding box may be determined based on the actual bounding box and the target bounding box.
  • the evaluation value may be a parameter value that can judge the similarity between the target bounding box and the actual bounding box.
  • the evaluation value may be the intersection ratio of the actual bounding box and the target bounding box, and the accuracy of the target bounding box may be evaluated by the intersection ratio of the actual bounding box and the target bounding box. For example, the higher the intersection ratio of the actual bounding box and the target bounding box, the higher the accuracy of the target bounding box.
  • the actual bounding box of the target object in the image to be detected is obtained, and the evaluation value of the target bounding box is determined based on the actual bounding box and the target bounding box.
  • the above target bounding box can be accurately evaluated.
  • the image to be detected is input into the first convolutional neural network to obtain the first feature of the target object in the image to be detected;
  • the first feature may include: the boundary image feature of the target object in the image to be detected;
  • the first feature is to input n second neural network respectively to obtain n first bounding boxes of the target object; n is an integer greater than or equal to 2; each first bounding box includes a first position parameter value; based on n first
  • Each first position parameter value of the bounding box determines the target bounding box of the target object in the image to be detected.
  • the intersection ratio of the actual bounding box and the target bounding box can be increased by about 2%.
  • the image to be detected is input into the first convolutional neural network to obtain the first feature of the target object in the image to be detected;
  • the first feature includes: the boundary image feature of the target object in the image to be detected ; Enter the first feature into n second neural network respectively to get n first bounding boxes of the target object;
  • n is an integer greater than or equal to 2; each first bounding box includes the first position parameter value; based on n
  • Each first position parameter value of the first bounding box determines the target bounding box of the target object in the image to be detected.
  • the target bounding box of the target object is obtained based on n second neural network at the same time, the determined target bounding box can be located in a relatively middle region of each first position parameter value of each first bounding box, which can avoid a single second neuron
  • the problem of inaccurate division of the target object caused by the instability of the network can improve the robustness of determining the bounding box of the target object, can effectively avoid the problem of overfitting, and thus can improve the accuracy of the bounding box of the target object .
  • Fig. 3 is a block diagram of a device for determining a bounding box according to an exemplary embodiment.
  • the device 300 includes a first feature acquisition module 303, a first bounding box determination module 307 and a target bounding box determination module 308.
  • the first feature obtaining module 303 is configured to input the image to be detected into the first convolutional neural sub-network to obtain the first feature of the target object in the image to be detected; the first feature includes: the image to be detected Boundary image features of the target object;
  • the first bounding box determination module 307 is configured to input the first features into n second neural sub-networks respectively to obtain n first bounding boxes of the target object; the n is an integer greater than or equal to 2 ; Each of the first bounding boxes includes a first position parameter value;
  • the target bounding box determining module 308 is configured to determine the target bounding box of the target object in the image to be detected based on the first position parameter values of the n first bounding boxes.
  • FIG. 4 is a block diagram of another apparatus for determining a bounding box according to an exemplary embodiment.
  • n second neural sub-networks include: m sets of third neural sub-networks, m is an integer greater than 0 and less than n, and each group of third neural sub-networks includes: y identical fourth Neural network, where y is an integer greater than or equal to 2 and less than or equal to n, and the device may further include:
  • the differential processing module 306 is configured to perform differential processing on the first feature to obtain y different second features
  • the first bounding box determination module 307 may include:
  • the first bounding box first determining unit 3071 is configured to input the y different second features into the fourth neural sub-networks in each group of third neural sub-networks to obtain k target objects
  • the first bounding box; the k is an integer greater than 0 and less than or equal to n;
  • the first bounding box second determining unit 3072 is configured to input the first features into x second neural subnetworks respectively to obtain x first bounding boxes of the target object; the x second nerves
  • the sub-network is a neural network other than the third neural network or the fourth neural network among the n second neural network; the x is an integer greater than or equal to 0 and smaller than n.
  • the differentiation processing may include at least one of cropping processing, scaling processing, mirror image processing, rotation processing, enhancement processing, and sharpening processing.
  • the target bounding box determination module 308 may include:
  • the target bounding box determining unit 3081 is configured to weight and sum the first position parameter values of the n first bounding boxes according to a predetermined weight to obtain the second position parameter value of the target object, and based on the The second position parameter value determines the target bounding box of the target object in the image to be detected.
  • the second neural sub-network is: any combination of a average pooling layer, b maximum pooling layers, c convolutional layers, and d fully connected layers; the a, b, and d are: An integer greater than or equal to 0; and c is an integer greater than or equal to 1.
  • the apparatus 300 for determining a bounding box may further include:
  • the first image sample data obtaining module 301 is configured to obtain first image sample data
  • the first convolutional neural network training module 302 is configured to train the first convolutional neural network based on the first image sample data
  • the third feature obtaining module 304 is configured to input the first image sample data into the first convolutional neural sub-network to obtain the third feature of the target object in the first image sample data; the third Features include: boundary image features of the target object in the first image sample data;
  • the second convolutional neural network training module 305 is configured to train the n second neural network based on the third feature.
  • the apparatus 300 for determining a bounding box may further include:
  • the actual bounding box obtaining module 309 is configured to obtain the actual bounding box of the target object in the image to be detected
  • the evaluation module 310 is configured to determine the evaluation value of the target bounding box based on the actual bounding box and the target bounding box.
  • Fig. 5 is a block diagram of an electronic device 500 according to an exemplary embodiment.
  • the electronic device 500 may be a mobile phone, computer, digital broadcasting terminal, messaging device, game console, tablet device, medical device, fitness device, personal digital assistant, or the like.
  • the electronic device 500 may include one or more of the following components: a processing component 502, a memory 504, a power supply component 506, a multimedia component 508, an audio component 510, an input/output (I/O) interface 512, and a sensor component 514 , ⁇ 516.
  • the processing component 502 generally controls the overall operations of the electronic device 500, such as operations associated with display, telephone calls, data communications, camera operations, and recording operations.
  • the processing component 502 may include one or more processors 520 to execute instructions to complete all or part of the steps of the above method.
  • the processing component 502 may include one or more modules to facilitate interaction between the processing component 502 and other components.
  • the processing component 502 may include a multimedia module to facilitate interaction between the multimedia component 508 and the processing component 502.
  • the memory 504 is configured to store various types of data to support operation at the device 500. Examples of these data include instructions for any application or method for operating on the electronic device 500, contact data, phone book data, messages, pictures, videos, etc.
  • the memory 504 may be implemented by any type of volatile or nonvolatile storage device or a combination thereof, such as static random access memory (SRAM), electrically erasable programmable read only memory (EEPROM), erasable and removable Programmable read only memory (EPROM), programmable read only memory (PROM), read only memory (ROM), magnetic memory, flash memory, magnetic disk or optical disk.
  • SRAM static random access memory
  • EEPROM electrically erasable programmable read only memory
  • EPROM erasable and removable Programmable read only memory
  • PROM programmable read only memory
  • ROM read only memory
  • magnetic memory flash memory
  • flash memory magnetic disk or optical disk.
  • the power supply component 506 provides power to various components of the electronic device 500.
  • the power supply component 506 may include a power management system, one or more power supplies, and other components associated with generating, managing, and distributing power for the electronic device 500.
  • the multimedia component 508 includes a screen that provides an output interface between the electronic device 500 and the user.
  • the screen may include a liquid crystal display (LCD) and a touch panel (TP). If the screen includes a touch panel, the screen may be implemented as a touch screen to receive input signals from the user.
  • the touch panel includes one or more touch sensors to sense touch, swipe, and gestures on the touch panel. The touch sensor may not only sense the boundary of the touch or sliding action, but also detect the duration and pressure related to the touch or sliding operation.
  • the multimedia component 508 includes a front camera and/or a rear camera. When the device 500 is in an operation mode, such as a shooting mode or a video mode, the front camera and/or the rear camera may receive external multimedia data. Each front camera and rear camera can be a fixed optical lens system or have focal length and optical zoom capabilities.
  • the audio component 510 is configured to output and/or input audio signals.
  • the audio component 510 includes a microphone (MIC).
  • the microphone is configured to receive an external audio signal.
  • the received audio signal may be further stored in the memory 504 or transmitted via the communication component 516.
  • the audio component 510 further includes a speaker for outputting audio signals.
  • the I/O interface 512 provides an interface between the processing component 502 and a peripheral interface module.
  • the peripheral interface module may be a keyboard, a click wheel, or a button. These buttons may include, but are not limited to: home button, volume button, start button, and lock button.
  • the sensor component 514 includes one or more sensors for providing the electronic device 500 with various aspects of status assessment.
  • the sensor component 514 can detect the on/off state of the device 500, and the relative positioning of the components, for example, the component is the display and keypad of the electronic device 500, and the sensor component 514 can also detect the electronic device 500 or a component of the electronic device 500 , The location of the user changes, the presence or absence of user contact with the electronic device 500, the orientation or acceleration/deceleration of the electronic device 500, and the temperature of the electronic device 500 change.
  • the sensor assembly 514 may include a proximity sensor configured to detect the presence of nearby objects without any physical contact.
  • the sensor component 514 may also include a light sensor, such as a CMOS or CCD image sensor, for use in imaging applications.
  • the sensor component 514 may further include an acceleration sensor, a gyro sensor, a magnetic sensor, a pressure sensor, or a temperature sensor.
  • the communication component 516 is configured to facilitate wired or wireless communication between the electronic device 500 and other devices.
  • the electronic device 500 can access a wireless network based on a communication standard, such as WiFi, an operator network (such as 2G, 3G, 4G, or 5G), or a combination thereof.
  • the communication component 516 receives a broadcast signal or broadcast related information from an external broadcast management system via a broadcast channel.
  • the communication component 516 further includes a near field communication (NFC) module to facilitate short-range communication.
  • the NFC module can be implemented based on radio frequency identification (RFID) technology, infrared data association (IrDA) technology, ultra-wideband (UWB) technology, Bluetooth (BT) technology, and other technologies.
  • RFID radio frequency identification
  • IrDA infrared data association
  • UWB ultra-wideband
  • Bluetooth Bluetooth
  • the electronic device 500 may be one or more application specific integrated circuits (ASICs), digital signal processors (DSPs), digital signal processing devices (DSPDs), programmable logic devices (PLDs), field Programmable gate array (FPGA), controller, microcontroller, microprocessor or other electronic components, used to perform the above method of determining the bounding box of Figures 1 to 2, and can achieve the same technical effect, in order to avoid repeating this I will not repeat them here.
  • ASICs application specific integrated circuits
  • DSPs digital signal processors
  • DSPDs digital signal processing devices
  • PLDs programmable logic devices
  • FPGA field Programmable gate array
  • controller microcontroller, microprocessor or other electronic components
  • a non-transitory computer-readable storage medium including instructions, for example, a memory 504 including instructions, the above instructions may be executed by the processor 520 of the device 500 to complete the above-mentioned FIG. 1 to FIG. 2
  • the method of determining the bounding box can achieve the same technical effect, and will not be repeated here in order to avoid repetition.
  • the non-transitory computer-readable storage medium may be ROM, random access memory (RAM), CD-ROM, magnetic tape, floppy disk, optical data storage device, or the like.
  • Fig. 6 is a block diagram of another electronic device 600 according to an exemplary embodiment.
  • the electronic device 600 may be provided as a server.
  • the electronic device 600 includes a processing component 622, which further includes one or more processors, and memory resources represented by the memory 632, for storing instructions executable by the processing component 622, such as application programs.
  • the application programs stored in the memory 632 may include one or more modules each corresponding to a set of instructions.
  • the processing component 622 is configured to execute instructions to perform the above-mentioned bounding box determination method of FIG. 1 to FIG. 2, and can achieve the same technical effect, and will not be repeated here to avoid repetition.
  • the electronic device 600 may also include a power component 626 configured to perform power management of the electronic device 600, a wired or wireless network interface 650 configured to connect the electronic device 600 to the network, and an input output (I/O) interface 658 .
  • the electronic device 600 can operate based on an operating system stored in the memory 632, such as Windows ServerTM, Mac OS XTM, UnixTM, LinuxTM, FreeBSDTM, or the like.
  • a non-transitory computer-readable storage medium is provided, and when instructions in the storage medium are executed by a processor of a mobile terminal, the mobile terminal can execute the above-mentioned FIGS. 1 to 2
  • the method of determining the bounding box can achieve the same technical effect, and will not be repeated here in order to avoid repetition.

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Multimedia (AREA)
  • Computational Linguistics (AREA)
  • Evolutionary Computation (AREA)
  • Artificial Intelligence (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • Health & Medical Sciences (AREA)
  • Data Mining & Analysis (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • General Health & Medical Sciences (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • General Engineering & Computer Science (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Image Analysis (AREA)

Abstract

一种边界框确定方法、装置、电子设备及存储介质,涉及图像处理技术领域,所述边界框确定方法包括:将待检测图像输入第一卷积神经子网络,获取所述待检测图像中目标物体的第一特征(S11);所述第一特征包括:所述待检测图像中所述目标物体的边界图像特征;将所述第一特征,分别输入n个第二神经子网络,得到所述目标物体的n个第一边界框(S12);所述n为大于等于2的整数;所述各个第一边界框包括第一位置参数值;基于所述n个第一边界框的各个第一位置参数值,确定所述待检测图像中目标物体的目标边界框(S13)。该方法能够提升确定目标物体的边界框的鲁棒性,有效避免过拟合,能够提升目标物体的边界框的准确性。

Description

一种边界框确定方法、装置、电子设备及存储介质
本申请要求于2018年12月4日提交中国专利局、申请号为201811475816.4发明名称为“边界框确定方法、装置、电子设备及存储介质”的中国专利申请的优先权,其全部内容通过引用结合在本申请中。
技术领域
本申请涉及图像处理技术领域,尤其涉及一种边界框确定方法、装置、电子设备及存储介质。
背景技术
边界框为物体检测时界定目标物体边缘的界限。目前主要通过边界框对应的第一位置参数值对目标物体的界限进行划分。
目前主要基于卷积神经子网络,将待检测的图像作为输入,通过卷积神经子网络进行图像特征提取后,获得待检测图像的图像特征后,将该待检测图像的图像特征通过第二神经子网络,输出待检测图像中的目标物体的边界框。
相关技术中,由于第二神经子网络本身的不稳定性等,导致基于该第二神经子网络得到的边界框存在识别不准的问题。
发明内容
为克服相关技术中存在的问题,本申请提供一种边界框确定方法及装置。
根据本申请实施例的第一方面,提供一种边界框确定方法,包括:
将待检测图像输入第一卷积神经子网络,获取所述待检测图像中目标物体的第一特征;所述第一特征包括:所述待检测图像中所述目标物体的边界图像特征;
将所述第一特征,分别输入n个第二神经子网络,得到所述目标物体的n个第一边界框;所述n为大于等于2的整数;所述各个第一边界框包括第一位置参数值;
基于所述n个第一边界框的各个第一位置参数值,确定所述待检测图像 中目标物体的目标边界框。
根据本申请实施例的第二方面,提供一种边界框确定装置,包括:
第一特征获取模块,被配置为将待检测图像输入第一卷积神经子网络,获取所述待检测图像中目标物体的第一特征;所述第一特征包括:所述待检测图像中所述目标物体的边界图像特征;
第一边界框确定模块,被配置为将所述第一特征,分别输入n个第二神经子网络,得到所述目标物体的n个第一边界框;所述n为大于等于2的整数;所述各个第一边界框包括第一位置参数值;
目标边界框确定模块,被配置为基于所述n个第一边界框的各个第一位置参数值,确定所述待检测图像中目标物体的目标边界框。
根据本申请实施例的第三方面,提供一种电子设备,包括:处理器;用于存储处理器可执行指令的存储器;其中,所述处理器被配置为:执行上述的边界框确定方法。
根据本申请实施例的第四方面,提供一种非临时性计算机可读存储介质,当所述存储介质中的指令由移动终端的处理器执行时,使得移动终端能够执行上述的边界框确定方法。
根据本申请实施例的第五方面,提供一种计算机程序产品,当所述计算机程序产品中的指令由移动终端的处理器执行时,使得移动终端能够执行上述的边界框确定方法。
基于本申请实施例提供的技术方案,将待检测图像输入第一卷积神经子网络,获取待检测图像中目标物体的第一特征;第一特征包括:待检测图像中目标物体的边界图像特征;将第一特征,分别输入n个第二神经子网络,得到目标物体的n个第一边界框;n为大于等于2的整数;各个第一边界框包括第一位置参数值;基于n个第一边界框的各个第一位置参数值,确定待检测图像中目标物体的目标边界框。该目标物体的目标边界框同时基于n种第二神经子网络获得,确定的该目标边界框可以位于各个第一边界框的各个第一位置参数值的相对中间区域,能够避免单个第二神经子网络的不稳定性导致的对目标物体的划分不准的问题,能够提升确定目标物体的边界框的鲁棒性,能够有效避免过拟合的问题,进而能够提升目标物体的边界框的准确性。
应当理解的是,以上的一般描述和后文的细节描述仅是示例性和解释性的,并不能限制本申请。
附图说明
图1是根据一示例性实施例示出的一种边界框确定方法的流程图。
图2是根据一示例性实施例示出的另一种边界框确定方法的流程图。
图3是根据一示例性实施例示出的一种边界框确定装置框图。
图4是根据一示例性实施例示出的另一种边界框确定装置框图。
图5是根据一示例性实施例示出的一种电子设备的框图。
图6是根据一示例性实施例示出的另一种电子设备的框图。
具体实施方式
图1是根据一示例性实施例示出的一种边界框确定方法的流程图,如图1所示,边界框确定方法用于终端中,可以包括以下步骤。
在步骤S11中,将待检测图像输入第一卷积神经子网络,获取待检测图像中目标物体的第一特征。
其中,第一特征可以包括:待检测图像中目标物体的边界图像特征。
在本申请实施例中,上述待检测图像可以为RGB格式、YUV格式等。在本申请实施例中,对此不作具体限定。
在本申请实施例中,上述待检测图像可以为拍摄的图像,或基于网络下载的图像,或者可以为视频中的一帧或多帧图像等,在本申请实施例中,对此不作具体限定。
在本申请实施例中,该待检测图像中的目标物体可以为待检测图像中的任意一个主体,例如,若该待检测图像中包括:小狗、小猫、花等,该待检测图像中的目标物体可以为该待检测图像中的小狗,或者,该待检测图像中的目标物体可以为该待检测图像中的小猫,或该待检测图像中的目标物体可以为该待检测图像中的花等。
在本申请实施例中,该第一卷积神经子网络可以为针对该目标物体的第一特征提取的卷积神经子网络。例如,针对上述例子,若目标物体为该待检测图像中的小狗,则该第一卷积神经子网络可以为针对小狗的第一特征提取 的卷积神经子网络。在本申请实施例中,对此不作具体限定。
在本申请实施例中,该第一卷积神经子网络可以包括:输入层、卷积层、池化层等,在本申请实施例中,对此不作具体限定。第一卷积神经子网络还可以包括:密集卷积网络。该密集卷积网络中每一个第一网络层的输入包括该密集卷积网络中所有第二网络层的输入,其中,第一网络层可以为该密集卷积网络中,任一网络层,该第二网络层可以为该密集卷积网络中,该第一网络层之前的各个网络层。例如,若第一网络层若为该密集网络层中的第4层,则该第四层的输入可以包括:该密集卷积网络中,第一层的输入、第二层的输入、第三层的输入,以及第三层的输出。相比于任意网络层的输入只包括其上一层的输出而言,密集卷积网络可以减轻梯度消失的现象,同时使得待检测图像中目标物体的第一特征传递更加有效,便于网络训练,且能够从一定程度上抑制过拟合的问题,进而使得,基于该第一卷积神经子网络,提取的待检测图像中目标物体的第一特征更加准确。
在本申请实施例中,待检测图像中目标物体的第一特征可以包括:待检测图像中目标物体的边界图像特征。待检测图像中目标物体的边界图像特征可以为:待检测图像中目标物体的轮廓图像特征、形状图像特征、空间位置图像特征等。在本申请实施例中,对此不作具体限定。
在本申请实施例中,将待检测图像输入上述第一卷积神经子网络,获取该待检测图像中目标物体的第一特征,该第一特征可以包括:该待检测图像中该目标物体的边界图像特征。
在步骤S12中,将第一特征,分别输入n个第二神经子网络,得到目标物体的n个第一边界框。
其中,n为大于等于2的整数,各个第一边界框包括第一位置参数值。
在本申请实施例中,上述n为大于等于2的整数,具体可以根据实际需要进行设定。例如,n可以为2,n可以为8、n可以为20等,在本申请实施例中,对此不作具体限定。
在本申请实施例中,上述各个第二神经子网络可以基于上述待检测图像中目标物体的第一特征,输出该目标物体的边界,进而可以得到该目标物体的各个第一边界框。该第二神经子网络可以包括:卷积层、池化层、全连接 层中的至少一种等,在本申请实施例中,对此不作具体限定。
在本申请实施例中,目标物体的第一边界框,可以为在表征待检测图像中目标物体的轮廓的边界框,该边界框可以为矩形,该第一边界框可以包括第一位置参数值,该第一边界框的第一位置参数值可以为能够在待检测图像中限定目标物体的轮廓的位置参数,例如,该第一边界框的第一位置参数可以为:该第一边界框的中心坐标,以及该边界框的长和宽等,或者,第一边界框的第一位置参数可以为:该第一边界框至少一条对角线上的顶点坐标,例如,第一边界框的第一位置参数可以为:该边界框的左上角的顶点坐标和右下角的顶点坐标等。在本申请实施例中,对此不作具体限定。
在本申请实施例中,n个第二神经子网络可以为n个相同的第二神经子网络,或者,n个第二神经子网络可以为n个不同的第二神经子网络,或者,n个第二神经子网络可以包括m组第三神经子网络,m为大于0小于n的整数,各组第三神经子网络分别可以包括:y个相同的第四神经子网络,y为大于等于2小于等于n的整数,以及x个与上述第三神经子网络或第四神经子网络不同的第二神经子网络等。在本申请实施例中,对此不作具体限定。
例如,若n为8,该8个第二神经子网络中可以包括:8个相同的第二神经子网络,如,A1、A2、A3、A4、A5、A6、A7、A8,或者,该8个第二神经子网络中可以包括:8个互不相同的第二神经子网络,如,A1、B1、C1、D1、E1、F1、G1、H1,或者,该8个第二神经子网络中可以包括:3组第三神经子网络,则m可以为3,其中,第1组第三神经子网络包括3个相同的第四神经子网络,如:A1、A2、A3,则y1可以为3,第2组第三神经子网络可以包括2个相同的第四神经子网络,如,B1、B2,则y2可以为2,第3组第三神经子网络可以包括2个相同的第四神经子网络,如,C1、C2,则y3可以为2,以及1个与上述第三神经子网络或第四神经子网络不同的第二神经子网络,如,D1,则x可以为1。
在本申请实施例中,可以将上述第一特征,分别输入上述n个第二神经子网络,确定该目标物体的边界,得到该目标物体的n个第一边界框,各个第一边界框包括第一位置参数值。例如,针对上述例子,可以将上述第一特征,输入上述A1、B1、C1、D1、E1、F1、G1、H1,8个第二神经子网络, 参照下表1,表1示出了上述8个第二神经子网络的一种模型结构。
表1
Figure PCTCN2019111645-appb-000001
表1中,Mobile_unit可以为e个第一卷积层、f个第一池化层、p个第一全连接层组成的第五神经子网络,e、f、p、s、w可以为大于0的整数,s*w可以为上述第一特征的尺寸,input可以为上述第一特征通道数,第一特征的尺寸、通道数可以根据第一特征实际情况进行确定,例如,s*w可以为:16*16,通道数可以为1,或者,通道数可以为3,或者通道数可以为4等,AvgPool可以为平均池化层,conv_layer可以为卷积层,MaxPool可以为最大池化层。表1中,conv_layer1、conv_layer9、conv_layer10、conv_layer12可以相同或不同,在本申请实施例中,对此不作具体限定。
表1中,各个第二神经子网络中,各个层的数量可以根据实际情况进行设定。在本申请实施例中,对此不作具体限定。
在本申请实施例中,可以将上述第一特征输入上述A1第二神经子网络,进而得到第1个第一边界框,将上述第一特征输入上述B1第二神经子网络,进而得到第2个第一边界框,将上述第一特征输入上述C1第二神经子网络,进而得到第3个第一边界框,将上述第一特征输入上述D1第二神经子网络,进而得到第4个第一边界框,将上述第一特征输入上述E1第二神经子网络,进而得到第5个第一边界框,将上述第一特征输入上述F1第二神经子网络, 进而得到第6个第一边界框,将上述第一特征输入上述G1第二神经子网络,进而得到第7个第一边界框,将上述第一特征输入上述H1第二神经子网络,进而得到第8个第一边界框。上述8个第一边界框,各个第一边界框包括第一位置参数值。
在步骤S13中,基于n个第一边界框的各个第一位置参数值,确定待检测图像中目标物体的目标边界框。
在本申请实施例中,基于上述n个第一边界框的各个第一位置参数值,确定上述待检测图像中目标物体的目标边界框。
一种实现方式中,可以将上述n个第一边界框的各个第一位置参数值进行均值计算等,进而确定上述待检测图像中目标物体的目标边界框。该均值计算可以为几何均值计算、算术均值计算、加权均值计算、均方根平均值计算、调和平均值计算等,在本申请实施例中,对此不作具体限定。例如,针对上述例子,将上述第一特征,输入上述A1、B1、C1、D1、E1、F1、G1、H1,8个第二神经子网络,得到8个第一边界框,各个第一边界框包括第一位置参数值,基于上述8个第一边界框的第一位置参数值,确定该待检测图像中目标物体的目标边界框。确定的该目标边界框可以位于各个第一边界框的各个第一位置参数值的相对中间区域,能够避免单个第二神经子网络的不稳定性导致的对目标物体的划分不准的问题,能够提升目标物体的边界框的准确性。
基于本申请实施例提供的技术方案,将待检测图像输入第一卷积神经子网络,获取待检测图像中目标物体的第一特征;第一特征包括:待检测图像中目标物体的边界图像特征;将第一特征,分别输入n个第二神经子网络,得到目标物体的n个第一边界框;n为大于等于2的整数;各个第一边界框包括第一位置参数值;基于n个第一边界框的各个第一位置参数值,确定待检测图像中目标物体的目标边界框。该目标物体的目标边界框同时基于n种第二神经子网络获得,确定的该目标边界框可以位于各个第一边界框的各个第一位置参数值的相对中间区域,能够避免单个第二神经子网络的不稳定性导致的对目标物体的划分不准的问题,能够提升确定目标物体的边界框的鲁棒性,能够有效避免过拟合的问题,进而能够提升目标物体的边界框的准确性。
图2是根据一示例性实施例示出的一种边界框确定方法的流程图,如图2所示,边界框确定方法包括以下步骤。
在步骤S21中,获取第一图像样本数据。
在本申请实施例中,第一图像样本数据可以是通过相机等拍摄装置拍摄的图像,或者可以为通过网络下载或接收的图像等。该第一图像样本数据中可以包括多个不同的物体或主体,例如,可以包括:人、小狗、小猫等。在本申请实施例中,对此不作具体限定。
在本申请实施例中,该第一图像样本数据可以包括上述目标物体。例如,上述目标物体若为小狗,则上述第一图像样本数据中可以包括小狗。在本申请实施例中,对此不作具体限定。
在本申请实施例中,在获取到第一图像样本数据的同时或之后,针对上述第一图像样本数据,可以获取其对应的特征标签图,该特征标签图可以是对上述各个第一图像样本数据中各个主体的第一特征进行标注的图。例如,针对上述例子,若第一图像样本数据包括:人、小猫、小狗等,该特征标签图即为将该第一图像样本数据中各个像素进行标注后的图,例如,标注出各个像素点是否为小狗的第一特征,若该像素属于小狗的第一特征,可以具体标注出该像素具体是该第一特征中的哪一类特征等。在本申请实施例中,对此不作具体限定。
在步骤S22中,基于第一图像样本数据,训练第一卷积神经子网络。
在本申请实施例中,可以针对上述第一图像样本数据,训练第一卷积神经子网络。
一种实现方式中,可以将上述第一图像样本数据,输入原始第一卷积神经子网络,采用该原始第一卷积神经子网络,提取第一图像样本数据对应的特征部分,该特征部分可以包括该第一图像样本数据中各个主体的局部特征或该第一图像样本数据的整体特征等,该局部特征可以识别出该第一图像样本数据中各个主体的形状、轮廓等,该第一图像样本数据的整体特征可以为该第一图像样本数据中各个主体的逻辑结构等。例如,人通常比小狗或小猫高等。
对上述第一图像样本数据对应的特征部分进行相应的运算,获得上述第 一图像样本数据中各个主体的第一特征,主要是获取目标物体的第一特征,将上述第一图像样本数据的整体特征、局部特征、边缘特征、第一特征等进行相应运算,同时,参照上述第一图像样本数据对应的特征标签图,对上述第一卷积神经子网络中的各项参数等进行调整,直至根据该原始第一卷积神经子网络输出的第一特征,与上述第一图像样本数据对应的特征标签图中标注的目标物体的第一特征的一致性达到预设阈值的情况下,训练得到该第一卷积神经子网络。在本申请实施例中,对此不作具体限定。
在本申请实施例中,该第一卷积神经子网络可以与获取第一特征的目标物体具有对应关系,则在训练上述第一卷积神经子网络的过程中,可以主要关注上述第一图像样本数据中该目标物体的相关数据,进而提高该第一卷积神经子网络相对该目标物体的针对性,进而提高针对该目标物体第一特征提取的准确性。在本申请实施例中,对此不作具体限定。
在步骤S23中,将待检测图像输入第一卷积神经子网络,获取待检测图像中目标物体的第一特征;第一特征包括:待检测图像中目标物体的边界图像特征。
在本申请实施例中,上述步骤可以参照上述步骤S11的具体记载,为了避免重复,此处不再赘述。
在步骤S24中,将第一图像样本数据,输入第一卷积神经子网络,获取第一图像样本数据中目标物体的第三特征;第三特征包括:第一图像样本数据中目标物体的边界图像特征。
在本申请实施例中,可以将上述第一图像样本数据,输入上述训练好的第一卷积神经子网络,获取上述第一图像样本数据中目标物体的第三特征,该第三特征可以包括:该第一图像样本数据中该目标物体的边界图像特征。
例如,若上述目标物体若为小狗,则上述第一图像样本数据中可以包括小狗。将上述第一图像样本数据,输入上述训练好的第一卷积神经子网络,获取上述第一图像样本数据中小狗的第三特征,该第三特征可以包括:该第一图像样本数据中小狗的边界图像特征,如,第一图像样本数据中小狗的边缘图像特征、形状图像特征等。
在步骤S25中,基于第三特征,训练n个第二神经子网络。
在本申请实施例中,可以针对上述第三特征,训练上述n个第二神经子网络。一种实现方式中,可以将上述第三特征,分别输入各个原始第二神经子网络,采用各个原始第二神经子网络,基于上述第三特征,确定上述第一图像样本数据中目标物体的原始边界框。
在本申请实施例中,在获取到第一图像样本数据的同时或之后,针对上述第一图像样本数据,可以获取其对应的边界标签图,该边界标签图可以是对上述各个第一图像样本数据中目标物体的实际边界框进行标注的图。例如,针对上述例子,若第一图像样本数据包括:人、小猫、小狗等,该边界标签图即为将该第一图像样本数据中各个像素进行标注后的图,例如,标注出各个像素点是否为小狗的实际边界框等。在本申请实施例中,对此不作具体限定。同时,可以参照上述第一图像样本数据对应的边界标签图中标注的目标物体的实际边界框,调整各个原始第二神经子网络的各项参数,使得各个原始第二神经子网络得出的该第一图像样本数据中目标物体的边界框,与上述第一图像样本数据对应的边界标签图中标注的目标物体的实际边界框的交并比等达到预设值,训练得到各个第二神经子网络。在本申请实施例中,对此不作具体限定。
在本申请实施例中,各个第二神经子网络可以与获取边界框的目标物体具有对应关系,则在训练上述各个第二神经子网络的过程中,可以主要关注上述第一图像样本数据中该目标物体的相关数据,进而提高各个第二神经子网络相对该目标物体的针对性,进而提高针对该目标物体确定边界框的识别准确性。在本申请实施例中,对此不作具体限定。
在步骤S26中,若n个第二神经子网络中包括:m组第三神经子网络,各组第三神经子网络分别包括:y个相同的第四神经子网络,对第一特征进行差异化处理,得到y种不同的第二特征。
在本申请实施例中,m为大于0小于n的整数,例如,若n为8,m可以为1、2、3等。y为大于等于2小于等于n的整数。若n为8,y可以为2、3等。如果上述n个第二神经子网络中包括:m组第三神经子网络,各组第三神经子网络分别包括:y个相同的第四神经子网络,y为大于等于2小于等于n的整数,n为大于等于2的整数,对第一特征进行差异化处理,得到y种不 同的第二特征。
例如,若n为8,该8个第二神经子网络中可以包括:3组第三神经子网络,则m可以为3,其中,第1组第三神经子网络包括3个相同的第四神经子网络,如:A1、A2、A3,则y1可以为3,第2组第三神经子网络可以包括2个相同的第四神经子网络,如,B1、B2,则y2可以为2,第3组第三神经子网络可以包括2个相同的第四神经子网络,如,C1、C2,则y3可以为2,以及1个与上述第三神经子网络或第四神经子网络不同的第二神经子网络,如,D1。
在本申请实施例中,针对各组第三神经子网络,由于其包括了y个相同的第四神经子网络,可以将上述第一特征进行差异化处理,以得到y种不同的第二特征。在本申请实施例中,可选的,差异化处理包括:剪裁处理、缩放处理、镜像处理、旋转处理、增强处理、锐化处理中的至少一种。剪裁处理可以是从上述第一特征中,剪裁一部分图像特征,缩放处理可以为将上述第一特征的进行放大缩小处理等,镜像处理可以为将上述第一特征进行镜像处理,旋转处理可以为将上述第一特征进行预设角度的旋转等,增强处理可以为将上述第一特征进行去噪或增强边缘等,锐化处理可以为将上述第一特征进行轮廓补偿,增强边缘及灰度跳变的部分等,即可以通过对上述待检测图像处理进行上述差异化处理,进而得到y种不同的第二特征。在本申请实施例中,对此不作具体限定。
例如,针对上述例子,针对第1组第三神经子网络,其包括3个相同的第四神经子网络,如:A1、A2、A3,则y1可以为3,则可以对上述第一特征进行差异化处理,得到3种不同的第二特征;第2组第三神经子网络包括2个相同的第四神经子网络,如,B1、B2,则y2可以为2,则可以对上述第一特征进行差异化处理,得到2种不同的第二特征;第3组第三神经子网络可以包括2个相同的第四神经子网络,如,C1、C2,则y3可以为2,则可以对上述第一特征进行差异化处理,得到2种不同的第二特征。
另外,如果某几组第三神经子网络中,第四神经子网络的个数相同,为了减少工作量,可以只针对一组第三神经子网络中的第四神经子网络的个数,待检测图像处理进行上述进行差异化处理,然后,将差异化得到的第二特征 直接作为另外1组第三神经子网络对应的第二特征。例如,针对上述例子,第2组第三神经子网络包括2个相同的第四神经子网络,如,B1、B2,则y2可以为2,第3组第三神经子网络可以包括2个相同的第四神经子网络,如,C1、C2,则y3可以为2,上述第2组第三神经子网络和第3组第三神经子网络包括的第四神经子网络的个数相同,均为2,为了减少工作量,可以只针对一组第三神经子网络中的第四神经子网络的个数,进行差异化处理,例如,对上述第一特征进行差异化处理,得到2种不同的第二特征,将差异化得到的第二特征直接作为第3组第三神经子网络对应的第二特征,相当于,针对第2组第三神经子网络和第3组第三神经子网络,一共只针对第一特征进行差异化处理,得到2种不同的第二特征,并不是得到2种不同的第二特征,进而为了减少运算量等,提升效率。
在本申请实施例中,还可以只针对上述m组第三神经子网络中,第四神经子网络个数最多的一组,针对待检测图像处理进行上述差异化处理,得到多个第二特征,其余各组第三神经子网络,直接从上述多个第二特征中任选即可。例如,针对上述例子,若n为8,该8个第二神经子网络中可以包括:3组第三神经子网络,第1组第三神经子网络,包括3个相同的第四神经子网络,如:A1、A2、A3,则y1可以为3,第2组第三神经子网络包括2个相同的第四神经子网络,如,B1、B2,则y2可以为2,第3组第三神经子网络可以包括2个相同的第四神经子网络,如,C1、C2,则y3可以为2,上述3组第三神经子网络中,第1组第三神经子网络包括的第四神经子网络的个数最多,为3个,为了减少工作量,可以只针对第1组第三神经子网络中的第四神经子网络的个数,进行差异化处理,例如,对上述第一特征进行差异化处理,得到3种不同的第二特征,第2组第三神经子网络、第3组第三神经子网络,直接从上述3个第二特征中任选即可。相当于,针对3组第三神经子网络,一共只针对第一特征进行差异化处理,得到3种不同的第二特征,并不是得到7种不同的第二特征,进而减少运算量,提升效率。
在本申请实施例中,可选的,第二神经子网络为:a个平均池化层、b个最大池化层、c个卷积层、d个全连接层的任意组合;a、b、d为大于等于0的整数;c为大于等于1的整数。一种实现方式中,该第二神经子网络可以为: a个平均池化层、b个最大池化层、c个卷积层、d个全连接层的任意组合,其中,a、b、d可以为大于等于0的整数;c可以为大于等于1的整数。例如,第二神经子网络可以为:10个卷积层连接5个平均池化层,再连接1个全连接层。
在步骤S27中,将y种不同的第二特征,分别输入各组第三神经子网络中的各个第四神经子网络,得到目标物体的k个第一边界框。
其中,k为大于0小于等于n的整数,各个第一边界框包括第一位置参数值。
在本申请实施例中,可以将上述y种不同的第二特征,分别输入各组第三神经子网络中的各个第四神经子网络,确定上述目标物体的边界,得到上述目标物体的k个第一边界框,k为大于0小于等于n的整数,各个第一边界框可以包括第一位置参数值。
例如,针对上述例子,若n为8,该8个第二神经子网络中可以包括:3组第三神经子网络,则m可以为3,其中,第1组第三神经子网络包括3个相同的第四神经子网络,如:A1、A2、A3,则y1可以为3,第2组第三神经子网络可以包括2个相同的第四神经子网络,如,B1、B2,则y2可以为2,第3组第三神经子网络可以包括2个相同的第四神经子网络,如,C1、C2,则y3可以为2,以及1个与上述第三神经子网络或第四神经子网络不同的第二神经子网络,如,D1。则可以将上述3种不同的第二特征,输入第1组第三神经子网络中的3第四神经子网络,确定上述目标物体的边界,得到上述目标物体的3个第一边界框,将上述2种不同的第二特征,输入第2组第三神经子网络中的2第四神经子网络,确定上述目标物体的边界,得到上述目标物体的2个第一边界框,将上述2种不同的第二特征,输入第3组第三神经子网络中的2第四神经子网络,确定上述目标物体的边界,得到上述目标物体的2个第一边界框。一共得到上述目标物体的7个第一边界框。则上述k可以为:3+2+2=7。
在步骤S28中,将第一特征,分别输入x个第二神经子网络,得到目标物体的x个第一边界框。
其中,x个第二神经子网络为n个第二神经子网络中,除第三神经子网络 或第四神经子网络之外的神经子网络,x为大于等于0小于n的整数,各个第一边界框包括第一位置参数值。
在本申请实施例中,可以将上述第一特征,分别输入x个第二神经子网络,确定该目标物体的边界,得到该目标物体的x个第一边界框,上述x个第二神经子网络可以为上述n个第二神经子网络中,除上述第三神经子网络或第四神经子网络之外的神经子网络。x可以为大于等于0小于n的整数。具体的,x可以为n与上述k的差值。各个第一边界框可以包括第一位置参数值。
例如,针对上述例子,若n为8,该8个第二神经子网络中可以包括:3组第三神经子网络,则m可以为3,其中,第1组第三神经子网络包括3个相同的第四神经子网络,如:A1、A2、A3,则y1可以为3,第2组第三神经子网络可以包括2个相同的第四神经子网络,如,B1、B2,则y2可以为2,第3组第三神经子网络可以包括2个相同的第四神经子网络,如,C1、C2,则y3可以为2,以及1个与上述第三神经子网络或第四神经子网络不同的第二神经子网络,如,D1,则x可以为1。则可以将上述第一特征,输入上述第二神经子网络D1中,确定该目标物体的边界,得到该目标物体的1个第一边界框。
在本申请实施例中,若上述n个第二神经子网络中包括:m组第三神经子网络,上述m为大于0小于n的整数,上述各组第三神经子网络分别包括:y个相同的第四神经子网络,上述y为大于等于2小于等于n的整数,对上述第一特征进行差异化处理,得到y种不同的第二特征;将上述y种不同的第二特征,分别输入各组第三神经子网络中的各个第四神经子网络,确定上述目标物体的边界,得到上述目标物体的k个第一边界框,k为大于0小于等于n的整数;将上述第一特征,分别输入x个第二神经子网络,确定上述目标物体的边界,得到该目标物体的x个第一边界框,x个第二神经子网络为上述n个第二神经子网络中,除上述第三神经子网络或第四神经子网络之外的神经子网络,x为大于等于0小于n的整数。即在上述n个第二神经子网络中存在相同子网络的情况下,对上述第一特征进行了差异化处理,进而使得各个相同子网络输入的第二特征,各不相同,进而使得各个相同子网络能够关注,该待检测图像中目标物体不同的边界图像特征,能够从一定程度上,提升后 续确定目标物体的目标边界框的准确性。
在步骤S29中,将n个第一边界框的各个第一位置参数值,按照预设权重加权求和,得到目标物体的第二位置参数值,并基于第二位置参数值确定待检测图像中目标物体的目标边界框。
在本申请实施例中,可以将上述n个第一边界框的各个第一位置参数值,按照预设权重加权求和,得到上述目标物体的第二位置参数值,并基于上述第二位置参数值确定上述待检测图像中目标物体的目标边界框。该预设权重可以根据实际需要进行设定等,在本申请实施例中,对此不作具体限定。
例如,若n为2,若第1个第一边界框左上角的顶点坐标为:(3,9),右下角的的顶点坐标为:(6,2),第2个第一边界框左上角的顶点坐标为:(3,7),右下角的的顶点坐标为:(4,4),第1个第一边界框的第一位置参数值对应的权重为:0.5,若第2个第一边界框的第一位置参数值对应的权重为:0.5,则得到该目标物体的第二位置参数值可以为:左上角顶点坐标为(3×0.5+3×0.5,9×0.5+7×0.5),右上角顶点坐标为(6×0.5+4×0.5,2×0.5+4×0.5),即目标物体的第二位置参数值可以为:左上角顶点坐标为(3,8),右上角顶点坐标为(5,3),进而基于上述第二位置参数值确定该待检测图像中目标物体的目标边界框为:待检测图像中左上角顶点坐标为(3,8),右上角顶点坐标为(5,3)的矩形框。确定的该目标边界框可以位于上述2个第一边界框的各个第一位置参数值的相对中间区域,能够避免单个第二神经子网络的不稳定性导致的对目标物体的划分不准的问题,能够提升目标物体的边界框的准确性。
在步骤S30中,获取待检测图像中目标物体的实际边界框。
在本申请实施例中,可以获取上述待检测图像中目标物体的实际边界框。该过程可以由人工在上述待检测图像中,对上述目标物体的实际边框进行确定,或者,该过程可以为:将上述待检测图像输入第一边界框输入模型,得到的待检测图像中目标物体的实际边界框。该第一边界框模型可以为针对该目标物体的,置信度较高的确定目标物体边界框的模型等,在本申请实施例中,对此不作具体限定。
在步骤S31中,基于实际边界框和目标边界框,确定目标边界框的评价 值。
在本申请实施例中,可以基于上述实际边界框和该目标边界框,确定该目标边界框的评价值。该评价值可以为能够评判该目标边界框与实际边界框相似性的参数值等。例如,该评价值可以为上述实际边界框和该目标边界框的交并比等,进而通过上述实际边界框和该目标边界框的交并比等,对该目标边界框的准确性进行评价。例如,上述实际边界框和该目标边界框的交并比越高,可以认为该目标边界框的准确性越高。
在本申请实施例中,获取上述待检测图像中目标物体的实际边界框,基于上述实际边界框和上述目标边界框,确定上述目标边界框的评价值。可以准确的对上述目标边界框进行评价。
在本申请实施例中,将待检测图像输入第一卷积神经子网络,获取待检测图像中目标物体的第一特征;第一特征可以包括:待检测图像中目标物体的边界图像特征;将第一特征,分别输入n个第二神经子网络,得到目标物体的n个第一边界框;n为大于等于2的整数;各个第一边界框包括第一位置参数值;基于n个第一边界框的各个第一位置参数值,确定待检测图像中目标物体的目标边界框。可以使得上述实际边界框和该目标边界框的交并比提升2%左右。
基于本申请实施例提供的技术方案,将待检测图像输入第一卷积神经子网络,获取待检测图像中目标物体的第一特征;第一特征包括:待检测图像中目标物体的边界图像特征;将第一特征,分别输入n个第二神经子网络,得到目标物体的n个第一边界框;n为大于等于2的整数;各个第一边界框包括第一位置参数值;基于n个第一边界框的各个第一位置参数值,确定待检测图像中目标物体的目标边界框。该目标物体的目标边界框同时基于n种第二神经子网络获得,确定的该目标边界框可以位于各个第一边界框的各个第一位置参数值的相对中间区域,能够避免单个第二神经子网络的不稳定性导致的对目标物体的划分不准的问题,能够提升确定目标物体的边界框的鲁棒性,能够有效避免过拟合的问题,进而能够提升目标物体的边界框的准确性。
图3是根据一示例性实施例示出的一种边界框确定装置框图。参照图3,该装置300包括第一特征获取模块303,第一边界框确定模块307和目标边界 框确定模块308。
第一特征获取模块303,被配置为将待检测图像输入第一卷积神经子网络,获取所述待检测图像中目标物体的第一特征;所述第一特征包括:所述待检测图像中所述目标物体的边界图像特征;
第一边界框确定模块307,被配置为将所述第一特征,分别输入n个第二神经子网络,得到所述目标物体的n个第一边界框;所述n为大于等于2的整数;所述各个第一边界框包括第一位置参数值;
目标边界框确定模块308,被配置为基于所述n个第一边界框的各个第一位置参数值,确定所述待检测图像中目标物体的目标边界框。
可选的,参见图4,图4是根据一示例性实施例示出的另一种边界框确定装置框图。
若所述n个第二神经子网络中包括:m组第三神经子网络,所述m为大于0小于n的整数,所述各组第三神经子网络分别包括:y个相同的第四神经子网络,所述y为大于等于2小于等于n的整数,所述装置,还可以包括:
差异化处理模块306,被配置为对所述第一特征进行差异化处理,得到y种不同的第二特征;
所述第一边界框确定模块307可以包括:
第一边界框第一确定单元3071,被配置为将所述y种不同的第二特征,分别输入各组第三神经子网络中的各个第四神经子网络,得到所述目标物体的k个第一边界框;所述k为大于0小于等于n的整数;
第一边界框第二确定单元3072,被配置为将所述第一特征,分别输入x个第二神经子网络,得到所述目标物体的x个第一边界框;所述x个第二神经子网络为所述n个第二神经子网络中,除所述第三神经子网络或第四神经子网络之外的神经子网络;所述x为大于等于0小于n的整数。
可选的,所述差异化处理可以包括:剪裁处理、缩放处理、镜像处理、旋转处理、增强处理、锐化处理中的至少一种。
可选的,所述目标边界框确定模块308可以包括:
目标边界框确定单元3081,被配置为将所述n个第一边界框的各个第一位置参数值,按照预设权重加权求和,得到所述目标物体的第二位置参数值, 并基于所述第二位置参数值确定所述待检测图像中目标物体的目标边界框。
可选的,所述第二神经子网络为:a个平均池化层、b个最大池化层、c个卷积层、d个全连接层的任意组合;所述a、b、d为大于等于0的整数;所述c为大于等于1的整数。
可选的,所述的边界框确定装置300还可以包括:
第一图像样本数据获取模块301,被配置为获取第一图像样本数据;
第一卷积神经子网络训练模块302,被配置为基于所述第一图像样本数据,训练所述第一卷积神经子网络;
第三特征获取模块304,被配置为将所述第一图像样本数据,输入所述第一卷积神经子网络,获取所述第一图像样本数据中目标物体的第三特征;所述第三特征包括:所述第一图像样本数据中所述目标物体的边界图像特征;
第二卷积神经子网络训练模块305,被配置为基于所述第三特征,训练所述n个第二神经子网络。
可选的,所述边界框确定装置300还可以包括:
实际边界框获取模块309,被配置为获取所述待检测图像中目标物体的实际边界框;
评价模块310,被配置为基于所述实际边界框和所述目标边界框,确定所述目标边界框的评价值。
关于上述实施例中的装置,其中各个单元执行操作的具体方式已经在有关该方法的实施例中进行了详细描述,此处将不做详细阐述说明。
图5是根据一示例性实施例示出的一种电子设备500的框图。例如,电子设备500可以是移动电话,计算机,数字广播终端,消息收发设备,游戏控制台,平板设备,医疗设备,健身设备,个人数字助理等。
参照图5,电子设备500可以包括以下一个或多个组件:处理组件502,存储器504,电源组件506,多媒体组件508,音频组件510,输入/输出(I/O)的接口512,传感器组件514,以及通信组件516。
处理组件502通常控制电子设备500的整体操作,诸如与显示,电话呼叫,数据通信,相机操作和记录操作相关联的操作。处理组件502可以包括一个或多个处理器520来执行指令,以完成上述的方法的全部或部分步骤。 此外,处理组件502可以包括一个或多个模块,便于处理组件502和其他组件之间的交互。例如,处理组件502可以包括多媒体模块,以方便多媒体组件508和处理组件502之间的交互。
存储器504被配置为存储各种类型的数据以支持在设备500的操作。这些数据的示例包括用于在电子设备500上操作的任何应用程序或方法的指令,联系人数据,电话簿数据,消息,图片,视频等。存储器504可以由任何类型的易失性或非易失性存储设备或者它们的组合实现,如静态随机存取存储器(SRAM),电可擦除可编程只读存储器(EEPROM),可擦除可编程只读存储器(EPROM),可编程只读存储器(PROM),只读存储器(ROM),磁存储器,快闪存储器,磁盘或光盘。
电源组件506为电子设备500的各种组件提供电力。电源组件506可以包括电源管理系统,一个或多个电源,及其他与为电子设备500生成、管理和分配电力相关联的组件。
多媒体组件508包括在所述电子设备500和用户之间的提供一个输出接口的屏幕。在一些实施例中,屏幕可以包括液晶显示器(LCD)和触摸面板(TP)。如果屏幕包括触摸面板,屏幕可以被实现为触摸屏,以接收来自用户的输入信号。触摸面板包括一个或多个触摸传感器以感测触摸、滑动和触摸面板上的手势。所述触摸传感器可以不仅感测触摸或滑动动作的边界,而且还检测与所述触摸或滑动操作相关的持续时间和压力。在一些实施例中,多媒体组件508包括一个前置摄像头和/或后置摄像头。当设备500处于操作模式,如拍摄模式或视频模式时,前置摄像头和/或后置摄像头可以接收外部的多媒体数据。每个前置摄像头和后置摄像头可以是一个固定的光学透镜系统或具有焦距和光学变焦能力。
音频组件510被配置为输出和/或输入音频信号。例如,音频组件510包括一个麦克风(MIC),当电子设备500处于操作模式,如呼叫模式、记录模式和语音识别模式时,麦克风被配置为接收外部音频信号。所接收的音频信号可以被进一步存储在存储器504或经由通信组件516发送。在一些实施例中,音频组件510还包括一个扬声器,用于输出音频信号。
I/O接口512为处理组件502和外围接口模块之间提供接口,上述外围接 口模块可以是键盘,点击轮,按钮等。这些按钮可包括但不限于:主页按钮、音量按钮、启动按钮和锁定按钮。
传感器组件514包括一个或多个传感器,用于为电子设备500提供各个方面的状态评估。例如,传感器组件514可以检测到设备500的打开/关闭状态,组件的相对定位,例如所述组件为电子设备500的显示器和小键盘,传感器组件514还可以检测电子设备500或电子设备500一个组件的位置改变,用户与电子设备500接触的存在或不存在,电子设备500方位或加速/减速和电子设备500的温度变化。传感器组件514可以包括接近传感器,被配置用来在没有任何的物理接触时检测附近物体的存在。传感器组件514还可以包括光传感器,如CMOS或CCD图像传感器,用于在成像应用中使用。在一些实施例中,该传感器组件514还可以包括加速度传感器,陀螺仪传感器,磁传感器,压力传感器或温度传感器。
通信组件516被配置为便于电子设备500和其他设备之间有线或无线方式的通信。电子设备500可以接入基于通信标准的无线网络,如WiFi,运营商网络(如2G、3G、4G或5G),或它们的组合。在一个示例性实施例中,通信组件516经由广播信道接收来自外部广播管理系统的广播信号或广播相关信息。在一个示例性实施例中,所述通信组件516还包括近场通信(NFC)模块,以促进短程通信。例如,在NFC模块可基于射频识别(RFID)技术,红外数据协会(IrDA)技术,超宽带(UWB)技术,蓝牙(BT)技术和其他技术来实现。
在示例性实施例中,电子设备500可以被一个或多个应用专用集成电路(ASIC)、数字信号处理器(DSP)、数字信号处理设备(DSPD)、可编程逻辑器件(PLD)、现场可编程门阵列(FPGA)、控制器、微控制器、微处理器或其他电子元件实现,用于执行上述图1至图2的边界框确定方法,且能达到相同的技术效果,为了避免重复此处不再赘述。
在示例性实施例中,还提供了一种包括指令的非临时性计算机可读存储介质,例如包括指令的存储器504,上述指令可由装置500的处理器520执行以完成上述图1至图2的边界框确定方法,且能达到相同的技术效果,为了避免重复此处不再赘述。例如,所述非临时性计算机可读存储介质可以是ROM、 随机存取存储器(RAM)、CD-ROM、磁带、软盘和光数据存储设备等。
图6是根据一示例性实施例示出的另一种电子设备600的框图。例如,电子设备600可以被提供为一服务器。参照图6,电子设备600包括处理组件622,其进一步包括一个或多个处理器,以及由存储器632所代表的存储器资源,用于存储可由处理组件622的执行的指令,例如应用程序。存储器632中存储的应用程序可以包括一个或一个以上的每一个对应于一组指令的模块。此外,处理组件622被配置为执行指令,以执行上述图1至图2的边界框确定方法,且能达到相同的技术效果,为了避免重复此处不再赘述。
电子设备600还可以包括一个电源组件626被配置为执行电子设备600的电源管理,一个有线或无线网络接口650被配置为将电子设备600连接到网络,和一个输入输出(I/O)接口658。电子设备600可以操作基于存储在存储器632的操作系统,例如Windows ServerTM,Mac OS XTM,UnixTM,LinuxTM,FreeBSDTM或类似。
根据本申请实施例的一个方面,提供一种非临时性计算机可读存储介质,当所述存储介质中的指令由移动终端的处理器执行时,使得移动终端能够执行上述图1至图2的边界框确定方法,且能达到相同的技术效果,为了避免重复此处不再赘述。本领域技术人员在考虑说明书及实践这里公开的申请后,将容易想到本申请的其它实施方案。本申请旨在涵盖本申请的任何变型、用途或者适应性变化,这些变型、用途或者适应性变化遵循本申请的一般性原理并包括本申请未公开的本技术领域中的公知常识或惯用技术手段。说明书和实施例仅被视为示例性的,本申请的真正范围和精神由下面的权利要求指出。
应当理解的是,本申请并不局限于上面已经描述并在附图中示出的精确结构,并且可以在不脱离其范围进行各种修改和改变。本申请的范围仅由所附的权利要求来限制。

Claims (16)

  1. 一种边界框确定方法,包括:
    将待检测图像输入第一卷积神经子网络,获取所述待检测图像中目标物体的第一特征;所述第一特征包括:所述待检测图像中所述目标物体的边界图像特征;
    将所述第一特征,分别输入n个第二神经子网络,得到所述目标物体的n个第一边界框;所述n为大于等于2的整数;所述各个第一边界框包括第一位置参数值;
    基于所述n个第一边界框的各个第一位置参数值,确定所述待检测图像中目标物体的目标边界框。
  2. 根据权利要求1所述的边界框确定方法,若所述n个第二神经子网络中包括:m组第三神经子网络,所述m为大于0小于n的整数,所述各组第三神经子网络分别包括:y个相同的第四神经子网络,所述y为大于等于2小于等于n的整数,所述将所述第一特征,分别输入n个第二神经子网络,得到所述目标物体的n个第一边界框之前,还包括:
    对所述第一特征进行差异化处理,得到y种不同的第二特征;
    所述将所述第一特征,分别输入n个第二神经子网络,得到所述目标物体的n个第一边界框,包括:
    将所述y种不同的第二特征,分别输入各组第三神经子网络中的各个第四神经子网络,得到所述目标物体的k个第一边界框;所述k为大于0小于等于n的整数;
    将所述第一特征,分别输入x个第二神经子网络,得到所述目标物体的x个第一边界框;所述x个第二神经子网络为所述n个第二神经子网络中,除所述第三神经子网络或第四神经子网络之外的神经子网络;所述x为大于等于0小于n的整数。
  3. 根据权利要求2所述的边界框确定方法,所述差异化处理包括:剪裁处理、缩放处理、镜像处理、旋转处理、增强处理、锐化处理中的至少一种。
  4. 根据权利要求1所述的边界框确定方法,所述基于所述n个第一边界框的各个第一位置参数值,确定所述待检测图像中目标物体的目标边界框, 包括:
    将所述n个第一边界框的各个第一位置参数值,按照预设权重加权求和,得到所述目标物体的第二位置参数值,并基于所述第二位置参数值确定所述待检测图像中目标物体的目标边界框。
  5. 根据权利要求1所述的边界框确定方法,所述第二神经子网络为:a个平均池化层、b个最大池化层、c个卷积层、d个全连接层的任意组合;所述a、b、d为大于等于0的整数;所述c为大于等于1的整数。
  6. 根据权利要求1所述的边界框确定方法,所述将待检测图像输入第一卷积神经子网络之前,还包括:
    获取第一图像样本数据;
    基于所述第一图像样本数据,训练所述第一卷积神经子网络;
    所述将所述第一特征,分别输入n个第二神经子网络,确定所述目标物体的边界之前,还包括:
    将所述第一图像样本数据,输入所述第一卷积神经子网络,获取所述第一图像样本数据中目标物体的第三特征;所述第三特征包括:所述第一图像样本数据中所述目标物体的边界图像特征;
    基于所述第三特征,训练所述n个第二神经子网络。
  7. 根据权利要求1所述的边界框确定方法,所述基于所述n个第一边界框的各个第一位置参数值,确定所述待检测图像中目标物体的目标边界框之后,还包括:
    获取所述待检测图像中目标物体的实际边界框;
    基于所述实际边界框和所述目标边界框,确定所述目标边界框的评价值。
  8. 一种边界框确定装置,包括:
    第一特征获取模块,被配置为将待检测图像输入第一卷积神经子网络,获取所述待检测图像中目标物体的第一特征;所述第一特征包括:所述待检测图像中所述目标物体的边界图像特征;
    第一边界框确定模块,被配置为将所述第一特征,分别输入n个第二神经子网络,得到所述目标物体的n个第一边界框;所述n为大于等于2的整数;所述各个第一边界框包括第一位置参数值;
    目标边界框确定模块,被配置为基于所述n个第一边界框的各个第一位置参数值,确定所述待检测图像中目标物体的目标边界框。
  9. 根据权利要求8所述的边界框确定装置,若所述n个第二神经子网络中包括:m组第三神经子网络,所述m为大于0小于n的整数,所述各组第三神经子网络分别包括:y个相同的第四神经子网络,所述y为大于等于2小于等于n的整数,所述装置,还包括:
    差异化处理模块,被配置为对所述第一特征进行差异化处理,得到y种不同的第二特征;
    所述第一边界框确定模块包括:
    第一边界框第一确定单元,被配置为将所述y种不同的第二特征,分别输入各组第三神经子网络中的各个第四神经子网络,得到所述目标物体的k个第一边界框;所述k为大于0小于等于n的整数;
    第一边界框第二确定单元,被配置为将所述第一特征,分别输入x个第二神经子网络,得到所述目标物体的x个第一边界框;所述x个第二神经子网络为所述n个第二神经子网络中,除所述第三神经子网络或第四神经子网络之外的神经子网络;所述x为大于等于0小于n的整数。
  10. 根据权利要求9所述的边界框确定装置,所述差异化处理包括:剪裁处理、缩放处理、镜像处理、旋转处理、增强处理、锐化处理中的至少一种。
  11. 根据权利要求8所述的边界框确定装置,所述目标边界框确定模块包括:
    目标边界框确定单元,被配置为将所述n个第一边界框的各个第一位置参数值,按照预设权重加权求和,得到所述目标物体的第二位置参数值,并基于所述第二位置参数值确定所述待检测图像中目标物体的目标边界框。
  12. 根据权利要求8所述的边界框确定装置,所述第二神经子网络为:a个平均池化层、b个最大池化层、c个卷积层、d个全连接层的任意组合;所述a、b、d为大于等于0的整数;所述c为大于等于1的整数。
  13. 根据权利要求8所述的边界框确定装置,还包括:
    第一图像样本数据获取模块,被配置为获取第一图像样本数据;
    第一卷积神经子网络训练模块,被配置为基于所述第一图像样本数据,训练所述第一卷积神经子网络;
    第三特征获取模块,被配置为将所述第一图像样本数据,输入所述第一卷积神经子网络,获取所述第一图像样本数据中目标物体的第三特征;所述第三特征包括:所述第一图像样本数据中所述目标物体的边界图像特征;
    第二卷积神经子网络训练模块,被配置为基于所述第三特征,训练所述n个第二神经子网络。
  14. 根据权利要求8所述的边界框确定装置,还包括:
    实际边界框获取模块,被配置为获取所述待检测图像中目标物体的实际边界框;
    评价模块,被配置为基于所述实际边界框和所述目标边界框,确定所述目标边界框的评价值。
  15. 一种电子设备,包括:
    处理器;
    用于存储处理器可执行指令的存储器;
    其中,所述处理器被配置为:执行如权利要求1至7中任一项所述的边界框确定方法。
  16. 一种非临时性计算机可读存储介质,当所述存储介质中的指令由移动终端的处理器执行时,使得移动终端能够执行如权利要求1至7中任一项所述的边界框确定方法。
PCT/CN2019/111645 2018-12-04 2019-10-17 一种边界框确定方法、装置、电子设备及存储介质 WO2020114097A1 (zh)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN201811475816.4A CN109784327B (zh) 2018-12-04 2018-12-04 边界框确定方法、装置、电子设备及存储介质
CN201811475816.4 2018-12-04

Publications (1)

Publication Number Publication Date
WO2020114097A1 true WO2020114097A1 (zh) 2020-06-11

Family

ID=66496667

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2019/111645 WO2020114097A1 (zh) 2018-12-04 2019-10-17 一种边界框确定方法、装置、电子设备及存储介质

Country Status (2)

Country Link
CN (1) CN109784327B (zh)
WO (1) WO2020114097A1 (zh)

Families Citing this family (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109784327B (zh) * 2018-12-04 2021-04-16 北京达佳互联信息技术有限公司 边界框确定方法、装置、电子设备及存储介质
CN111126587B (zh) * 2019-12-30 2021-02-02 上海安路信息科技有限公司 交并比电路
CN112288712B (zh) * 2020-10-28 2022-07-22 山东黄金矿业(莱州)有限公司三山岛金矿 一种基于实景建模的黄金矿山溜井入料可视化检测方法

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2016179808A1 (en) * 2015-05-13 2016-11-17 Xiaoou Tang An apparatus and a method for face parts and face detection
CN107527053A (zh) * 2017-08-31 2017-12-29 北京小米移动软件有限公司 目标检测方法及装置
CN108009543A (zh) * 2017-11-29 2018-05-08 深圳市华尊科技股份有限公司 一种车牌识别方法及装置
CN108921159A (zh) * 2018-07-26 2018-11-30 北京百度网讯科技有限公司 用于检测安全帽的佩戴情况的方法和装置
CN109784327A (zh) * 2018-12-04 2019-05-21 北京达佳互联信息技术有限公司 边界框确定方法、装置、电子设备及存储介质

Family Cites Families (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107851195B (zh) * 2015-07-29 2022-02-11 诺基亚技术有限公司 利用神经网络进行目标检测
CN106548127B (zh) * 2015-09-18 2022-11-04 松下电器(美国)知识产权公司 图像识别方法
CN109643388A (zh) * 2016-06-21 2019-04-16 维凯睿思Fpc公司 用于递归皮层网络的系统和方法
US10452899B2 (en) * 2016-08-31 2019-10-22 Siemens Healthcare Gmbh Unsupervised deep representation learning for fine-grained body part recognition
US10657424B2 (en) * 2016-12-07 2020-05-19 Samsung Electronics Co., Ltd. Target detection method and apparatus
CN108171222B (zh) * 2018-02-11 2020-08-25 清华大学 一种基于多流神经网络的实时视频分类方法及装置
CN108520197A (zh) * 2018-02-28 2018-09-11 中国航空工业集团公司洛阳电光设备研究所 一种遥感图像目标检测方法及装置
CN108416327B (zh) * 2018-03-28 2022-04-29 京东方科技集团股份有限公司 一种目标检测方法、装置、计算机设备及可读存储介质
CN108520229B (zh) * 2018-04-04 2020-08-07 北京旷视科技有限公司 图像检测方法、装置、电子设备和计算机可读介质
CN108509920B (zh) * 2018-04-04 2022-04-26 南京信息工程大学 基于CNN的多patch多通道联合特征选择学习的人脸识别方法

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2016179808A1 (en) * 2015-05-13 2016-11-17 Xiaoou Tang An apparatus and a method for face parts and face detection
CN107527053A (zh) * 2017-08-31 2017-12-29 北京小米移动软件有限公司 目标检测方法及装置
CN108009543A (zh) * 2017-11-29 2018-05-08 深圳市华尊科技股份有限公司 一种车牌识别方法及装置
CN108921159A (zh) * 2018-07-26 2018-11-30 北京百度网讯科技有限公司 用于检测安全帽的佩戴情况的方法和装置
CN109784327A (zh) * 2018-12-04 2019-05-21 北京达佳互联信息技术有限公司 边界框确定方法、装置、电子设备及存储介质

Also Published As

Publication number Publication date
CN109784327B (zh) 2021-04-16
CN109784327A (zh) 2019-05-21

Similar Documents

Publication Publication Date Title
US11114130B2 (en) Method and device for processing video
CN111641778B (zh) 一种拍摄方法、装置与设备
WO2019109801A1 (zh) 拍摄参数的调整方法、装置、存储介质及移动终端
WO2020108291A1 (zh) 人脸美化方法、装置、计算机设备及存储介质
WO2016011747A1 (zh) 肤色调整方法和装置
EP3125158B1 (en) Method and device for displaying images
US10007841B2 (en) Human face recognition method, apparatus and terminal
WO2020042826A1 (zh) 视频流降噪方法和装置、电子设备及存储介质
US11030733B2 (en) Method, electronic device and storage medium for processing image
WO2020007241A1 (zh) 图像处理方法和装置、电子设备以及计算机可读存储介质
WO2020114097A1 (zh) 一种边界框确定方法、装置、电子设备及存储介质
US11308692B2 (en) Method and device for processing image, and storage medium
WO2020114236A1 (zh) 关键点检测方法、装置、电子设备及存储介质
WO2022110837A1 (zh) 图像处理方法及装置
CN109784164B (zh) 前景识别方法、装置、电子设备及存储介质
WO2022121577A1 (zh) 图像处理方法及装置
KR102273059B1 (ko) 얼굴 이미지를 강화하는 방법 및 장치, 전자 기기
WO2021035485A1 (zh) 拍摄防抖方法、装置、终端及存储介质
CN109509195B (zh) 前景处理方法、装置、电子设备及存储介质
KR20210053121A (ko) 이미지 처리 모델의 훈련 방법, 장치 및 매체
US9665925B2 (en) Method and terminal device for retargeting images
CN106469446B (zh) 深度图像的分割方法和分割装置
WO2023240452A1 (zh) 图像处理方法、装置、电子设备和存储介质
CN116416141A (zh) 图像处理方法及装置、设备、存储介质
CN115482309A (zh) 图像处理方法、计算机设备及存储介质

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 19892259

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 19892259

Country of ref document: EP

Kind code of ref document: A1