WO2021147563A1 - Object detection method and apparatus, electronic device, and computer readable storage medium - Google Patents

Object detection method and apparatus, electronic device, and computer readable storage medium Download PDF

Info

Publication number
WO2021147563A1
WO2021147563A1 PCT/CN2020/135967 CN2020135967W WO2021147563A1 WO 2021147563 A1 WO2021147563 A1 WO 2021147563A1 CN 2020135967 W CN2020135967 W CN 2020135967W WO 2021147563 A1 WO2021147563 A1 WO 2021147563A1
Authority
WO
WIPO (PCT)
Prior art keywords
corner point
corner
image
point
detected
Prior art date
Application number
PCT/CN2020/135967
Other languages
French (fr)
Chinese (zh)
Inventor
王飞
钱晨
Original Assignee
上海商汤临港智能科技有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 上海商汤临港智能科技有限公司 filed Critical 上海商汤临港智能科技有限公司
Priority to JP2021557733A priority Critical patent/JP2022526548A/en
Priority to KR1020217030884A priority patent/KR20210129189A/en
Publication of WO2021147563A1 publication Critical patent/WO2021147563A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/44Local feature extraction by analysis of parts of the pattern, e.g. by detecting edges, contours, loops, corners, strokes or intersections; Connectivity analysis, e.g. of connected components
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/20Scenes; Scene-specific elements in augmented reality scenes
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/70Determining position or orientation of objects or cameras
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/20Image preprocessing
    • G06V10/25Determination of region of interest [ROI] or a volume of interest [VOI]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/77Processing image or video features in feature spaces; using data integration or data reduction, e.g. principal component analysis [PCA] or independent component analysis [ICA] or self-organising maps [SOM]; Blind source separation
    • G06V10/774Generating sets of training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/82Arrangements for image or video recognition or understanding using pattern recognition or machine learning using neural networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20081Training; Learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V2201/00Indexing scheme relating to image or video recognition or understanding
    • G06V2201/07Target detection

Definitions

  • the embodiments of the present disclosure relate to the field of image recognition technology, and in particular, to a target detection method, device, electronic equipment, and computer-readable storage medium.
  • Target detection is an important basic problem of computer vision. Many computer vision applications rely on target detection, such as autonomous driving, video surveillance, and mobile entertainment.
  • the main task is to use the detection frame to mark the location of the object in the image.
  • This process can determine the location of the object in the image based on the target detection algorithm of the key point of the object, and determine all the key points of the object in the image Then, the key points of the objects belonging to the same object are matched to obtain the detection frame of the object.
  • the matching degree between the key points of the objects corresponding to the similar objects is high, which is easy to cause wrong detection results.
  • the detection result is that the same detection frame contains Multiple objects, therefore, the detection accuracy of current target detection methods is low.
  • the embodiments of the present disclosure provide at least one target detection solution.
  • embodiments of the present disclosure provide a target detection method, including:
  • the target object in the image to be detected is determined based on the corner position information of each corner point in the image to be detected and the heart offset tensor corresponding to each corner point.
  • the corner points in the image to be detected can characterize the position of each target object in the image to be detected.
  • the corner points can include the upper left corner point and the lower right corner point, where the upper left corner point refers to the corresponding The intersection of the straight line of the upper contour of the target object and the straight line corresponding to the left contour of the target object.
  • the lower right corner point refers to the intersection of the straight line corresponding to the lower contour of the target object and the straight line corresponding to the right contour of the target object.
  • the positions pointed by the centripetal offset tensors corresponding to the upper left corner points and the lower right corner points should be relatively close. Therefore, the target detection method proposed in the embodiments of the present disclosure is based on The corner position information of each corner point in the image to be detected and the centripetal offset tensor corresponding to each corner point can determine the corner point belonging to the same target object, and then the same target can be detected based on the determined corner point Object.
  • the determining, based on the to-be-detected image, the corner position information of each corner point in the to-be-detected image and the centripetal offset tensor corresponding to each corner point includes:
  • the corner point position information of each corner point in the image to be detected and the centripetal offset tensor corresponding to each corner point are determined.
  • the method provided by the embodiment of the present disclosure obtains an initial feature map by extracting features from the image to be detected, and performs corner pooling processing on the initial feature map, so as to obtain a convenient extraction of the corner points and the centripetal offset corresponding to the corner points.
  • Feature map that is, the feature map after corner point pooling.
  • the determining the corner position information of each corner point in the image to be detected based on the feature map after the corner point pooling includes:
  • the local offset information is used to indicate that the real physical point represented by the corresponding corner point is at Position offset information in the corner point heat map;
  • the local offset information corresponding to each corner point, and the size ratio between the corner point heat map and the image to be detected determine each The position information of the corner point in the image to be detected.
  • the embodiment of the present disclosure provides a method for determining the position information of each corner point in the image to be detected.
  • This process introduces a corner point heat map, and determines that the corner point can be used as a corner point by the probability value of each feature point as a corner point.
  • the feature point of the point after the corner point is selected, the position information of the corner point in the corner point heat map is corrected to determine the corner point position information of the corner point in the image to be detected.
  • This method can obtain accuracy
  • the corner position information of the higher corner point facilitates subsequent detection of the position of the target object in the image to be detected based on the corner point.
  • the determining the centripetal offset tensor corresponding to each corner point based on the feature map after the corner point pooling includes:
  • the steering offset tensor corresponding to each feature point in the corner point pooling feature map is determined, and the steering offset tensor corresponding to each feature point is represented by The offset tensor of the feature point pointing to the center point of the target object in the image to be detected;
  • the offset domain information includes multiple initial feature points associated with the feature point respectively pointing to their corresponding offsets The offset tensor of the feature point after the shift;
  • the feature data of the feature points in the corner point pooled feature map Make adjustments to obtain the adjusted feature map
  • centripetal offset tensor corresponding to each corner point is determined.
  • the process of determining the centripetal offset tensor considers the target object information, such as introducing the steering offset tensor corresponding to the corner point, and the offset domain information of the feature point.
  • the feature data of the feature points in the feature map are adjusted so that the feature data of the feature points in the adjusted feature map can contain richer target object information, so that a more accurate orientation corresponding to each corner point can be determined.
  • the central offset tensor, through accurate centripetal offset tensor can accurately obtain the position information of the center point pointed by the corner point, so as to accurately detect the position of the target object in the image to be detected.
  • the corner heat map corresponding to the image to be detected includes a corner heat map corresponding to multiple channels, and each channel of the multiple channels corresponds to a preset object category; After determining the probability value of each feature point in the corner heat map as a corner point based on the corner heat map, the detection method further includes:
  • the corner point exists in the corner point heat map corresponding to the channel, it is determined that the image to be detected contains the target object of the preset object category corresponding to the channel.
  • the corner heat map containing the preset number of channels can be obtained, and the corner heat map corresponding to each channel can be obtained. Whether there is a corner point in the image, it can be determined whether there is a target object corresponding to the channel in the image to be detected.
  • the target object in the image to be detected is determined based on the position information of each corner point in the image to be detected and the centripetal offset tensor corresponding to each corner point ,include:
  • the detection frame of the target object in the image to be detected is determined.
  • the method provided by the embodiments of the present disclosure can determine the detection frame of each target object based on the corner position information of each corner point in the image to be detected and the centripetal offset tensor corresponding to each corner point.
  • the position information of the target object in the image to be detected can be determined.
  • the target object in the image to be detected is determined based on the position information of each corner point in the image to be detected and the centripetal offset tensor corresponding to each corner point
  • the detection box includes:
  • the detection frame of the target object is determined in the candidate detection frame based on the position information of the center point pointed to by each corner point in each candidate corner point pair and the center area information corresponding to the candidate corner point pair.
  • the corner point position information of the corner points is used to first determine the candidate corner point pairs that can constitute the candidate detection frame, and then based on each corner point in the candidate corner point pair.
  • the corresponding centripetal offset tensor is used to determine whether the target object surrounded by the candidate detection frame is the same target object, so that the detection frame of all target objects in the image to be detected can be detected more accurately.
  • the determining the center region information corresponding to the candidate corner point pair based on the corner point position information of each corner point in the candidate corner point pair in the image to be detected includes :
  • the coordinate range of the central area frame corresponding to the candidate corner point pair is determined.
  • Determining the detection frame of the target object includes:
  • the central area information corresponding to the valid candidate corner point pair, and the probability corresponding to each corner point in the valid candidate corner point pair Value determine the score of the candidate detection frame corresponding to each valid candidate corner point; the probability value corresponding to each corner point is used to indicate the probability value of the corresponding feature point of the corner point in the corner heat map as the corner point ;
  • the detection frame of the target object is determined in the candidate detection frame.
  • the method provided by the embodiment of the present disclosure effectively screens the candidate corner points that constitute the candidate detection frame, and determines that the candidate detection frame that only represents one target object can be screened out, and then performs detection on these candidate detection frames that only represent one target object.
  • Soft non-maximum suppression screening so as to obtain an accurate detection frame that characterizes the target object.
  • the target detection method further includes:
  • the instance information of the target object in the image to be detected is determined based on the detection frame of the target object and the initial feature map obtained by feature extraction of the image to be detected.
  • the method provided by the embodiments of the present disclosure can determine the instance information of the target object.
  • the instance here means that after the instance segmentation of the target object in the image, the pixel of each target object is given at the pixel level, and the instance segmentation can be accurate to the object. To obtain more accurate position information of the target object in the image to be detected.
  • the determining the instance information of the target object in the image to be detected is based on the detection frame of the target object and the initial feature map obtained by feature extraction of the image to be detected, include:
  • the instance information of the target object in the image to be detected is determined.
  • the target detection method is implemented by a neural network, and the neural network is obtained by training using sample pictures containing labeled target sample objects.
  • the neural network is obtained by training using the following steps:
  • the network parameter value of the neural network is adjusted based on the predicted target sample object in the sample image and the labeled target sample object in the sample image.
  • the neural network training method obtains a sample image, and based on the sample image, determines the position information of each sample corner point in the sample image, and the centripetal offset corresponding to each sample corner point. Based on the corner position information of each sample corner point in the sample image and the centripetal offset tensor corresponding to each sample corner point, the target sample object is detected in the sample image, because the sample corner point refers to the main Feature points, such as sample corner points may include upper left sample corner points and lower right sample corner points, where the upper left sample corner point refers to the intersection of a line corresponding to the upper contour of the target sample object and a line corresponding to the left contour of the target sample object.
  • the lower right sample corner point refers to the intersection of the straight line corresponding to the lower contour of the target sample object and the straight line corresponding to the right contour of the target sample object.
  • the upper left sample corner and the lower right sample corner belong to the detection frame of the same target sample object.
  • the centripetal offset tensor corresponding to the upper left sample corner point and the lower right sample corner point should be relatively close to each other.
  • the neural network training method proposed in the embodiment of the present disclosure is based on characterizing the target sample object in the sample
  • the corner position information of the corner points of the position in the image, and the centripetal offset tensor corresponding to each sample corner point determine the sample corner points belonging to the same target sample object, and then based on the determined sample corner points can be detected
  • the same target sample object is extracted, and then the neural network parameters are continuously adjusted based on the target object in the sample image, so as to obtain a neural network with higher accuracy.
  • the target object can be Perform accurate detection.
  • a target detection device including:
  • the obtaining part is configured to obtain the image to be detected
  • the determining part is configured to determine, based on the image to be detected, the corner position information of each corner point in the image to be detected and the centripetal offset tensor corresponding to each corner point, where the corner points represent the image to be detected.
  • the detection part is configured to determine the target object in the image to be detected based on the position information of each corner point in the image to be detected and the centripetal offset tensor corresponding to each corner point.
  • the determining part is configured to:
  • the corner point position information of each corner point in the image to be detected and the centripetal offset tensor corresponding to each corner point are determined.
  • the method when the determining part is configured to determine the corner position information of each corner point in the image to be detected based on the feature map after the corner point pooling, the method includes :
  • the local offset information is used to indicate that the real physical point represented by the corresponding corner point is at Position offset information in the corner point heat map;
  • the local offset information corresponding to each corner point, and the size ratio between the corner point heat map and the image to be detected determine The position information of each corner point in the image to be detected.
  • the method includes:
  • the steering offset tensor corresponding to each feature point in the corner point pooling feature map is determined, and the steering offset tensor corresponding to each feature point is represented by The offset tensor of the feature point pointing to the center point of the target object in the image to be detected;
  • the offset domain information includes multiple initial feature points associated with the feature point respectively pointing to their corresponding offsets The offset tensor of the feature point after the shift;
  • the feature data of the feature points in the corner point pooled feature map Make adjustments to obtain the adjusted feature map
  • centripetal offset tensor corresponding to each corner point is determined.
  • the corner heat map corresponding to the image to be detected includes a corner heat map corresponding to multiple channels, and each channel of the multiple channels corresponds to a preset object category;
  • the determining part is configured to determine the probability value of each feature point in the corner heat map as a corner point based on the corner heat map, it is further configured to:
  • the corner point exists in the corner point heat map corresponding to the channel, it is determined that the image to be detected contains the target object of the preset object category corresponding to the channel.
  • the detection part is configured to:
  • the detection frame of the target object in the image to be detected is determined.
  • the detection part is configured to determine the to-be-detected image based on the position information of each corner point in the image to be detected and the centripetal offset tensor corresponding to each corner point.
  • the detection frame of the target object in the image it includes:
  • the detection frame of the target object is determined in the candidate detection frame based on the position information of the center point pointed to by each corner point in each candidate corner point pair and the center area information corresponding to the candidate corner point pair.
  • the detection part is configured to determine the candidate corner point pair based on the corner point position information of each corner point of each candidate corner point pair in the image to be detected.
  • the corresponding central area information includes:
  • the coordinate range of the central area frame corresponding to the candidate corner point pair is determined.
  • the detection part is configured to be based on the position information of the center point pointed to by each corner point in each candidate corner point pair, and the center area information corresponding to the candidate corner point pair,
  • the method includes:
  • the central area information corresponding to the valid candidate corner point pair, and the probability corresponding to each corner point in the valid candidate corner point pair Value determine the score of the candidate detection frame corresponding to each valid candidate corner point; the probability value corresponding to each corner point is used to indicate the probability value of the corresponding feature point of the corner point in the corner heat map as the corner point ;
  • the detection frame of the target object is determined in the candidate detection frame.
  • the detection part is further configured to:
  • the target object in the image to be detected is determined based on the detection frame of the target object and the initial feature map obtained by feature extraction of the image to be detected Instance information.
  • the detection part is configured to determine the image in the image to be detected based on the detection frame of the target object and the initial feature map obtained by feature extraction of the image to be detected In the case of the instance information of the target object, it includes:
  • the instance information of the target object in the image to be detected is determined.
  • the target detection device further includes a neural network training part, and the neural network training part is configured to:
  • Training a neural network for target detection the neural network is obtained by training using sample pictures containing labeled target sample objects.
  • the neural network training part is configured to train the neural network according to the following steps:
  • the network parameter value of the neural network is adjusted based on the predicted target sample object in the sample image and the labeled target sample object in the sample image.
  • an embodiment of the present disclosure provides an electronic device, including a processor, a memory, and a bus.
  • the memory stores machine-readable instructions executable by the processor.
  • the electronic device is running, the The processor and the memory communicate through a bus, and when the machine-readable instructions are executed by the processor, the steps of the target detection method as described in the first aspect are executed.
  • embodiments of the present disclosure provide a computer-readable storage medium having a computer program stored on the computer-readable storage medium, and the computer program executes the target detection method as described in the first aspect when the computer program is run by a processor. step.
  • the embodiments of the present disclosure provide a computer program, including computer-readable code.
  • the processor in the electronic device executes the following On the one hand, the steps of the target detection method.
  • Figure 1 shows a schematic diagram of a result obtained when detecting an image to be detected
  • Fig. 2 shows a flow chart of an exemplary target detection method provided by an embodiment of the present disclosure
  • FIG. 3 shows a flowchart of a process for determining the position information of corner points and the centripetal offset tensor corresponding to the corner points provided by an embodiment of the present disclosure
  • FIG. 4 shows a flowchart for determining the position information of a corner point and the centripetal offset tensor corresponding to the corner point provided by an embodiment of the present disclosure
  • FIG. 5 shows a flow chart of determining the centripetal offset tensor corresponding to a corner point provided by an embodiment of the present disclosure
  • FIG. 6 shows a schematic flow chart of an exemplary feature adjustment network provided by an embodiment of the present disclosure for adjusting a feature map after corner pooling
  • FIG. 7 shows a schematic diagram of a process for determining the category of a target object provided by an embodiment of the present disclosure
  • FIG. 8 shows a schematic flow chart of determining a detection frame of a target object provided by an embodiment of the present disclosure
  • FIG. 9 shows a schematic flowchart of determining a detection frame of a target object based on each candidate corner point pair provided by an embodiment of the present disclosure
  • FIG. 10 shows a schematic flowchart corresponding to an exemplary target detection method provided by an embodiment of the present disclosure
  • FIG. 11 shows a schematic flowchart of a neural network training method provided by an embodiment of the present disclosure
  • FIG. 12 shows a schematic structural diagram of a target detection device provided by an embodiment of the present disclosure
  • FIG. 13 shows a schematic structural diagram of an electronic device provided by an embodiment of the present disclosure.
  • the embodiment of the present disclosure provides a target detection method, which can improve the accuracy of the detection result.
  • the embodiments of the present disclosure provide a target detection method. After acquiring the image to be detected, first determine the corner position information of each corner point in the image to be detected and the centripetal offset corresponding to each corner point. Because the corner point refers to the main feature point in the image, the position information of the corner point in the image to be detected can characterize the position of each target object in the image to be detected. For example, the corner point can include the upper left corner and the lower right corner.
  • the target detection method proposed in the embodiment can determine the corner points belonging to the same target object based on the corner position information of each corner point in the image to be detected and the centripetal offset tensor corresponding to each corner point.
  • the corner point of can detect the same target object.
  • the execution subject of the target detection method provided in the embodiment of the present disclosure is generally a computer device with a certain computing capability.
  • the equipment includes, for example, terminal equipment or servers or other processing equipment.
  • the target detection method can be implemented by a processor invoking a computer-readable instruction stored in a memory.
  • the method includes steps S201 to S203, and the steps are as follows:
  • the image to be detected here can be an image to be detected in a specific environment.
  • a camera can be installed at the traffic intersection, and the video stream of the traffic intersection in a certain period of time can be collected by the camera, and then Frame the video stream to obtain the image to be detected; or to detect animals in a zoo, a camera can be installed in the zoo, and the video stream of the zoo in a certain period of time can be collected by the camera, and then the video The stream undergoes framing processing to obtain the image to be detected.
  • the image to be detected can contain the target object.
  • the target object here refers to the object to be detected in a specific environment, such as a vehicle at a traffic intersection mentioned above, and an animal in a zoo, or it may not contain the target.
  • Object if the target object is not included, the detection result is empty, and the implementation of the present disclosure will describe the image to be detected that contains the target object.
  • S202 Based on the image to be detected, determine the corner position information of each corner point in the image to be detected and the centripetal offset tensor corresponding to each corner point, where the corner point represents the position of the target object in the image to be detected.
  • the position of the target object in the image to be detected can be represented by a detection frame.
  • the embodiment of the present disclosure uses corner points to characterize the position of the target object in the image to be detected, that is, the corner points here may be the corner points of the detection frame, for example,
  • the position of the target object in the image to be detected is characterized by the upper left corner point and the lower right corner point.
  • the upper left corner point is the upper left corner point of the detection frame
  • the lower right corner point is the lower right corner point of the detection frame, where the upper left corner point is Refers to the intersection of the line corresponding to the upper contour of the target object and the line corresponding to the left contour of the target object.
  • the lower right corner point refers to the intersection of the line corresponding to the lower contour of the target object and the line corresponding to the right contour of the target object.
  • the position of the target object is not limited to the upper left corner point and the lower right corner point.
  • the position of the target object can also be characterized by the upper right corner point and the lower left corner point.
  • the embodiment of the present disclosure uses the upper left corner point and the lower right corner point. Take an example for illustration.
  • the centripetal offset tensor here refers to the offset tensor from the corner point to the center position of the target object. Because the image to be detected is a two-dimensional image, the centripetal offset tensor here includes the offset in two directions. When the two directions are the X-axis direction and the Y-axis direction, the centripetal offset tensor includes the offset value in the X-axis direction and the offset value in the Y-axis direction. Through the centripetal offset tensor corresponding to the corner point and the corner point, the center position of the corner point can be determined.
  • the center position of the point should be the same , Or relatively close, so the corner points belonging to the same target object can be determined based on the centripetal offset tensor corresponding to each corner point, and then the detection frame of the target object can be determined based on the determined corner points.
  • the embodiment of the present disclosure uses a neural network to determine the corner point and the centripetal offset tensor corresponding to the corner point, which will be described in conjunction with the following embodiments.
  • S203 Determine a target object in the image to be detected based on the corner position information of each corner point in the image to be detected and the centripetal offset tensor corresponding to each corner point.
  • the corner position information of each corner in the image to be detected refers to the corner position information of each of the multiple corner points in the image to be detected
  • the centripetal offset tensor corresponding to each corner refers to the Each of the multiple corner points corresponds to the centripetal offset tensor.
  • the detection of the target object in the image to be detected can include the location of the detected target object, such as determining the detection frame of the target object in the image to be detected, or determining the instance information of the target object in the image to be detected, or at the same time determining the target object to be detected.
  • the detection frame and instance information of the target object in the image, and how to determine the target object in the image to be detected will be explained in detail later.
  • the target detection method proposed in the above steps S201 to S203 after acquiring the image to be detected, first determine the corner position information of each corner point in the image to be detected, and the centripetal offset tensor corresponding to each corner point, because the angle Point refers to the main feature point in the image.
  • the position information of the corner point in the image to be detected can characterize the position of each target object in the image to be detected.
  • the corner point can include the upper left corner point and the lower right corner point, where the upper left corner The corner point refers to the intersection point of the line corresponding to the upper contour of the target object and the line corresponding to the left contour of the target object.
  • the lower right corner point refers to the intersection point of the line corresponding to the lower contour of the target object and the line corresponding to the right contour of the target object.
  • the positions of the centripetal offset tensor corresponding to the upper left corner point and the lower right corner point should be relatively close. Therefore, the embodiment of the present disclosure proposes
  • the target detection method can determine the corner points belonging to the same target object based on the corner position information of each corner point in the image to be detected and the centripetal offset tensor corresponding to each corner point, and then based on the determined corner point The same target object is detected.
  • the corner position information of the corner point in the image to be detected and the centripetal offset tensor corresponding to each corner point are determined, as shown in FIG. 3 It may include the following steps S301 to S303:
  • S301 Perform feature extraction on the image to be detected to obtain an initial feature map corresponding to the image to be detected;
  • S301 Perform corner pooling processing on the initial feature map to obtain a feature map after corner pooling
  • S303 Based on the feature map after corner pooling, determine the corner position information of each corner in the image to be detected, and the centripetal offset tensor corresponding to each corner.
  • the size of the image to be detected is fixed, for example, the size is H*W, where H and W represent the pixel values in the length and width directions of the image to be detected respectively, and then input the image to be detected into the pre-trained hourglass convolution
  • the neural network performs feature extraction, such as texture feature extraction, color feature extraction, edge feature extraction, etc., and the initial feature map corresponding to the image to be detected can be obtained.
  • the input end of the hourglass convolutional neural network has requirements for the received image size, that is, it receives a set size of the image to be inspected. If the size of the image to be inspected does not meet the set size, it needs to be treated first. Adjust the size of the detected image, and then input the adjusted image to be detected into the hourglass convolutional neural network for feature extraction and size compression, that is, an initial feature map with a size of h*w*c can be obtained, where c Indicates the number of channels of the initial feature map, and h and w represent the size of the initial feature map on each channel.
  • the initial feature map contains multiple feature points, and each feature point has feature data. These feature data can represent the global information of the image to be detected.
  • the embodiment of the present disclosure proposes to The graph performs corner pooling processing to obtain the feature map after corner pooling. Compared with the initial feature map, the feature map after corner pooling enhances the semantic information of the target object contained in the corner points, so it is based on The feature map after corner pooling can more accurately determine the corner position information of each corner point in the image to be detected, and the centripetal offset tensor corresponding to each corner point.
  • an initial feature map is obtained, and corner pooling is performed on the initial feature map to obtain a feature map that can facilitate the extraction of the corner points and the centripetal offset corresponding to the corner points, that is, the corner points Feature map after pooling.
  • the corner point pooling feature map and the pre-trained neural network can be used to determine whether there is a corner point. If there is a corner point, determine whether each corner point is The position information of the corner points in the image to be detected.
  • the position of the target object in the image to be detected is characterized by the upper left corner point and the lower right corner point, that is, the corner point position information of each corner point in the image to be detected is determined
  • the process can be the process of determining the corner position information of the upper left corner point in the image to be detected, and the process of determining the corner position information of the lower right corner point in the image to be detected, where the upper left corner can be detected through the upper left corner point detection network
  • the corner position information of the point in the image to be detected, and the corner position information of the lower right corner point in the image to be detected through the lower right corner point detection network because the corner position information of the upper left corner point in the image to be detected and
  • the method for determining the corner position information of the lower right corner point in the image to be detected is similar, and the embodiment of the present disclosure determines the corner position information of the upper left corner point in the image to be detected as an example for detailed description.
  • the upper left corner point detection network may include the upper left corner point heat map prediction network and the upper left corner point local offset prediction network. Based on the feature map after corner point pooling, it is determined that each corner point is in the image to be detected. In the case of the corner position information in, as shown in FIG. 4, the following steps S401 to S404 may be included:
  • the corner point heat map here can be obtained by the upper left corner point heat map prediction network in the upper left corner point detection network, and the corner point pooled
  • the feature map is input into the upper left corner point heat map prediction network, and the upper left corner point heat map corresponding to the image to be detected can be obtained.
  • the upper left corner point heat map contains multiple feature points, and each feature point has the feature data corresponding to the feature point.
  • the probability value of the feature point as the upper left corner point can be determined.
  • the upper left corner heat map can be used to determine the upper left corner point in the image to be detected, and can also be used to determine that the upper left corner point is in the image to be detected.
  • the category of the target object represented in the image, and the process of how to determine the category of the target object will be explained in detail later.
  • the probability that the feature point is the upper left corner point can be determined, so that the feature point with the probability value greater than the set threshold is taken as the upper left corner point.
  • S403 Obtain the position information of the selected corner points in the corner point heat map and the local offset information corresponding to each corner point.
  • the local offset information is used to indicate the position offset information of the real physical point represented by the corresponding corner point in the corner heat map.
  • the local offset information corresponding to each upper left corner point is used to indicate the position of the upper left corner point.
  • the local offset information here can be represented by a local offset tensor.
  • the local offset tensor can also represent the offset values in two directions in the upper left corner heat map, such as the coordinates in the upper left corner heat map
  • the system includes two directions, namely the x-axis direction and the y-axis direction.
  • the local offset tensor includes the offset value in the x-axis direction and the offset value in the y-axis direction.
  • the position information of each feature point in the upper left corner point heat map in the upper left corner point heat map can be obtained, taking into account the obtained position information of the upper left corner point and the upper left corner point There may be errors between the position information of the real physical points represented.
  • the position information of a certain upper left corner point can be obtained by detecting the position of the upper left corner point heat map, and the position information of the real physical point represented by the upper left corner point is different from the position information of the real physical point.
  • the detected position information of the upper left corner point should have a certain deviation, and the local offset information is used to indicate the deviation.
  • the acquired position information of each upper left corner point in the upper left corner point heat map may include the coordinate value x in the x-axis direction in the upper left corner point heat map, and the coordinate value y in the y-axis direction, to be detected
  • the corner position information of each upper left corner point in the image may include the coordinate value X in the X-axis direction and the coordinate value Y in the Y-axis direction.
  • the corner position information of the i-th upper left corner point in the image to be detected can be determined according to the following formula (1) and formula (2):
  • tl x(i) represents the coordinate value of the i-th upper left corner point in the X-axis direction of the image to be detected
  • tl y(i) represents the coordinate value of the i-th upper left corner point in the Y-axis direction of the image to be detected Value
  • n represents the size ratio between the upper left corner point heat map and the image to be detected
  • x l(i) represents the coordinate value of the i-th upper left corner point in the x-axis direction of the upper left corner point heat map
  • y l(i ) Represents the coordinate value of the i-th upper left corner point in the y-axis direction of the corner heat map
  • ⁇ lx(i) represents the real physical point represented by the i-th upper left corner point in the x-axis direction of the corner heat map
  • the offset value of ⁇ ly(i) represents the offset value of the real physical point represented by the i-th upper left corner point in the y-axis
  • the above process is the process of determining the corner position information of the upper left corner point in the image to be detected, and the process of determining the corner point position information of the lower right corner point in the image to be detected is the same, that is, the feature map after the corner point pooling is input to the right
  • the lower right corner point heat map prediction network in the lower corner point prediction network obtains the lower right corner point heat map, and then determines the probability value of each feature point in the lower right corner point heat map as the lower right corner point, selects the lower right corner point from it, and combines them at the same time
  • the position information of the corner point of each lower right corner point in the image to be detected is determined by the local offset information corresponding to the lower right corner point determined by the lower right corner point local offset network in the lower right corner point prediction network, which will not be repeated here.
  • the corner position information of the j-th lower right corner point in the image to be detected can be determined according to the following formula (3) and formula (4):
  • br x(j) n*(x r(j) + ⁇ rx(j) ); (3);
  • br x(j) represents the coordinate value of the j-th lower right corner point in the X-axis direction of the image to be detected
  • br y(j) represents the coordinate value of the j-th lower right corner point in the Y-axis direction of the image to be detected Value
  • n represents the size ratio between the lower right corner point heat map and the image to be detected
  • x r(j) represents the coordinate value of the j-th lower right corner point in the x-axis direction of the lower right corner point heat map
  • y r(j ) Represents the coordinate value of the j-th lower right corner point in the y-axis direction of the corner heat map
  • ⁇ rx(j) represents the real physical point represented by the j-th lower right corner point in the x-axis direction of the corner heat map
  • the offset value of ⁇ ry(j) represents the offset value of the real physical point represented by the j-th lower right corner point in the y-axis direction of the corner
  • the above steps S401 to S404 are a way to determine the corner position information of each corner in the image to be detected according to the embodiment of the present disclosure.
  • This process introduces a corner heat map and passes each feature point as the probability of a corner point. The value determines the feature point that can be used as the corner point.
  • the position information of the corner point in the corner point heat map is corrected to determine the corner point position information of the corner point in the image to be detected.
  • This method can obtain the corner point position information of the corner point with higher accuracy, thereby facilitating the subsequent detection of the position of the target object in the image to be detected based on the corner point.
  • centripetal offset tensor corresponding to each corner point determines the centripetal offset tensor corresponding to the upper left corner point as an example for detailed description, the centripetal offset tensor corresponding to the lower right corner point and the upper left corner point.
  • the method for determining the corresponding centripetal offset tensor is similar, and will not be repeated in the embodiment of the present disclosure.
  • a feature adjustment process is introduced to adjust the feature map after corner pooling, and then the direction is determined.
  • the central offset tensor where, in the case of determining the centripetal offset tensor corresponding to each corner point based on the feature map after the corner point pooling, as shown in FIG. 5, the following steps S501 to S504 may be included:
  • S501 Determine a steering offset tensor corresponding to each feature point in the feature map after corner point pooling based on the feature map after corner point pooling.
  • the steering offset tensor corresponding to each feature point represents the offset tensor from the feature point to the center point of the target object in the image to be detected.
  • the position of the target object in the image to be detected is related to the target object information, that is, it is hoped that the feature data of the corner points of the feature map after the corner point pooling can contain richer target object information, so each feature point can be considered here.
  • the corner point feature vector can contain richer target object information, so based on the steering offset tensor corresponding to each feature point, the feature map after the corner point pooling can be adjusted to make the adjusted feature map
  • Each feature point, especially the corner point can contain richer target object information.
  • the corner point pooled feature map can be convolved to obtain the steering offset tensor corresponding to each feature point in the corner point pooled feature map.
  • the steering offset tensor includes the direction along x The offset value in the axis direction and the offset value along the y-axis direction.
  • the convolution operation is performed on the feature map after the corner point pooling, and the feature point is mainly obtained as the steering offset tensor corresponding to the upper left corner point.
  • S502 Determine the offset domain information of each feature point based on the steering offset tensor corresponding to each feature point.
  • the offset domain information includes a plurality of initial feature points associated with the feature point and respectively point to the offset tensors of the respective offset feature points.
  • a convolution operation is performed based on the steering offset tensor corresponding to each feature point to obtain the offset domain information of the feature point.
  • centripetal offset tensor corresponding to the upper left corner point As an example, after obtaining the corresponding steering offset tensor with each feature point as the upper left corner point, then use each feature point as the upper left corner point.
  • the corresponding steering offset tensor is subjected to convolution operation to obtain the offset domain information when the feature point is used as the upper left corner point.
  • S503 Based on the feature map after corner point pooling and the offset domain information of the feature points in the feature map after corner point pooling, adjust the feature data of the feature points in the feature map after corner point pooling , Get the adjusted feature map.
  • the feature map after the corner point pooling After the feature point of the feature map after corner point pooling is obtained as the offset domain information in the case of the upper left corner point, the feature map after the corner point pooling can be pooled, and the feature map after the corner point pooling Each feature point of is used as the offset domain information of the upper left corner point, and the deformable convolution operation is performed at the same time to obtain the adjusted feature map corresponding to the upper left corner point.
  • steps S501 to S503 can be determined through the feature adjustment network as shown in FIG. 6:
  • the tensor performs convolution operation to obtain offset domain information.
  • the offset domain information here is explained as follows:
  • the feature map after corner pooling can be used to include feature point A.
  • the feature data of the 9 initial feature points represented by the solid line frame are obtained by convolution operation. After considering the offset domain information, it is hoped that the feature point A can be adjusted by the feature data containing more abundant target object information. For example, the feature points used for feature adjustment of feature point A can be offset based on the steering offset vector corresponding to each feature point.
  • the offset feature points can be pooled by corner points as shown in Figure 6.
  • the latter feature map is represented by the 9 dashed boxes, so that the feature data of the 9 offset feature points can be used to perform the convolution operation, and the feature data of feature point A can be adjusted.
  • the offset domain information can be passed
  • the offset tensor in Figure 6 is represented.
  • Each offset tensor in the offset tensor is the offset tensor of each initial feature point pointing to the offset feature point corresponding to the initial feature point, representing the initial feature After the point is offset in the x-axis direction and the y direction, the offset feature point corresponding to the initial feature point is obtained.
  • centripetal offset tensor corresponding to each upper left corner point, a more accurate centripetal offset tensor can be obtained.
  • the feature point after feature adjustment contains richer target object information, which is convenient for the subsequent adjustment based on the target object information.
  • centripetal offset tensor corresponding to each lower right corner of the feature map
  • a convolution operation is performed on the feature data corresponding to the corner points in the adjusted feature map, and the centripetal offset tensor corresponding to each corner point is determined.
  • the adjusted feature map may include the adjusted feature map corresponding to the upper left corner point, and the adjusted feature map corresponding to the lower right corner point, and each upper left corner is determined based on the adjusted feature map corresponding to the upper left corner point
  • the centripetal offset tensor corresponding to the point it can be determined by the centripetal offset prediction network corresponding to the upper left corner point.
  • the centripetal offset prediction network corresponding to the lower right corner point determine the direction corresponding to each lower right corner point.
  • the centripetal offset prediction network corresponding to the lower right corner point In the case of the heart offset tensor, it can be determined by the centripetal offset prediction network corresponding to the lower right corner point.
  • the above process of S501 to S504 is the process of determining the centripetal offset tensor provided by the embodiments of the present disclosure, by considering the target object information, such as introducing the steering offset tensor corresponding to the corner point, and the offset domain information of the feature point , Adjust the feature data of the feature points in the feature map after the corner point pooling, so that the feature data of the feature points in the adjusted feature map can contain richer target object information, so that each A more accurate centripetal offset tensor corresponding to each corner point.
  • the position information of the center point pointed by the corner point can be accurately obtained, so as to accurately detect the position of the target object in the image to be detected .
  • the category of the target object contained in the image to be detected can be determined through the corner heat map.
  • the corner heat map is how to determine the category of the target object based on the corner heat map. From the above, we know the corner heat of the image to be detected.
  • the figure includes the corner heat maps corresponding to multiple channels, and each channel corresponds to a preset object category; in the above-mentioned corner heat map, each feature point in the corner heat map is determined as a corner point
  • the detection method provided by the embodiment of the present disclosure further includes the following steps S701 to S702:
  • S701 For each channel of the multiple channels, determine whether there is a corner point in the corner heat map corresponding to the channel based on the probability value of each feature point as the corner point in the corner heat map corresponding to the channel.
  • the probability value of each feature point in the corner heat map corresponding to each channel as a corner point can be determined whether there is a corner point in the corner heat map of the channel, for example,
  • the corner feature map of a channel contains multiple feature points with a corresponding probability value greater than the set threshold, it means that the corner feature map of the channel contains corner points with a high probability, and the corner points are used to represent the target object
  • the position in the image to be detected so that it can be explained that the image to be detected contains the target object of the preset object category corresponding to the channel.
  • the number of channels to 100, that is, the obtained corner heat map is h*w*100, and each channel corresponds to a preset object category, for a certain type of object to be detected Image, among the 100 channels of the corner heat map corresponding to the image to be detected, only the corner heat maps in the first and second channels contain corner points, and the first channel corresponds to the pre- Assuming that the object category is 01, and the preset object category corresponding to the second channel is 02, it can be explained that the image to be detected contains target objects of the categories 01 and 02.
  • the embodiment of the present disclosure proposes that by inputting the feature map after the corner point pooling into the corner heat map prediction network, the corner heat map containing the preset number of channels can be obtained, and whether the corner heat map corresponding to each channel is There are corner points, and then it can be determined whether there is a target object corresponding to the channel in the image to be detected.
  • the centripetal offset tensor corresponding to the corner point can be determined, so as to determine the position of the target object corresponding to each channel in the image to be detected.
  • the category of each target object in the image to be detected in combination with the category of the target object corresponding to the channel.
  • the detection frame of the target object in the image to be detected is determined.
  • the embodiment of the present disclosure uses an upper left corner point and a lower right corner point to determine the detection frame as an example for description.
  • the upper left corner and the lower right corner can be judged first Whether the points belong to the same target object category, in the case of determining that any upper left corner point and lower right corner point belong to the same target object category, continue to determine the corner position of any upper left corner point and lower right corner point in the image to be detected Whether the information constitutes the same candidate detection frame.
  • the upper left corner point should be located at the upper left of the lower right corner point in the image to be detected, and the position information of the corner point based on the upper left corner point and the lower right corner point, such as the position coordinates of the upper left corner point in the image to be detected, and the right If the position coordinates of the lower corner point in the image to be detected cannot be such that the upper left corner point is located at the upper left corner of the lower right corner point, the upper left corner point and the lower right corner point cannot constitute a candidate corner point pair.
  • a coordinate system can be established in the image to be detected, the coordinate system includes X axis and Y axis, and the corner position information of each corner point in the coordinate system includes the abscissa value in the X axis direction and the Y axis.
  • the ordinate value in the axis direction, and then in the coordinate system, according to the corresponding coordinate value of each corner point in the coordinate system, the upper left corner point and the lower right corner point that can constitute the candidate detection frame are filtered.
  • S802 Determine the position information of the center point to which the corner point points based on the corner position information of each corner point in the image to be detected in each candidate corner point pair and the centripetal offset tensor corresponding to the corner point.
  • the position information of the center point pointed to by the upper left corner point in each candidate corner point pair can be determined according to the following formula (5), and the center point pointed to by the lower right corner point in each candidate corner point pair can be determined according to the following formula (6) location information:
  • the central area information here can be preset, which is defined as the coordinate range of the central area frame that coincides with the center of the detection frame of the target object. Through the coordinate range of the central area frame, it is possible to detect whether the candidate detection frame contains a unique target.
  • the position information of the center point pointed to by the upper left corner point and the center point position information pointed to by the lower right corner point are located within the coordinate range of the central area frame, in the case where the coordinate range of the central area frame is small. Then, it can be considered that the position information of the center point pointed to by the upper left corner point is relatively close to the position information of the center point pointed to by the lower right corner point, so as to determine that the candidate detection frame formed by the candidate corner point pair contains a unique target object.
  • the center area information corresponding to the candidate corner point pair based on the corner point position information of each corner point in the candidate corner point pair in the image to be detected it may include:
  • the m-th candidate corner point pair is composed of the i-th upper left corner point and the j-th lower right corner point
  • the m-th candidate corner point pair can be determined according to the following formulas (7) to (10) Corresponding corner position information of the central area frame:
  • the coordinate range of the center area frame can be determined according to the following formula (11):
  • R central(m) represents the coordinate range of the central area frame corresponding to the m-th candidate corner point pair.
  • the coordinate range of the central area frame passes through the x(m) value in the X-axis direction and the Y-axis direction.
  • y(m) value where the range of x(m) satisfies The range of y(m) satisfies
  • S804 Determine the detection frame of the target object in the candidate detection frame based on the position information of the center point pointed to by each corner point in each candidate corner point pair and the center area information corresponding to the candidate corner point pair.
  • the center area information corresponding to each candidate corner point pair is used to restrict the proximity between the center point position information pointed to by each corner point in the candidate corner point pair, and each corner point in a certain candidate corner point pair
  • the position information of the pointed center point is located in the center area frame corresponding to the candidate corner point pair
  • the center point of each corner point in the candidate corner point pair is relatively close
  • the candidate corner point pair constitutes
  • the target object contained in the candidate detection frame of is the only target object.
  • the corner point position information of the corner points is used to first determine the candidate corner point pairs that can constitute the candidate detection frame, and then based on each corner point in the candidate corner point pair.
  • the centripetal offset tensor of the object is used to determine whether the target object surrounded by the candidate detection frame is the same target object, so that the detection frame of all target objects in the image to be detected can be detected more accurately.
  • the detection frame of the target object is determined in the candidate detection frame based on the position information of the center point pointed to by each corner point in each candidate corner point pair and the center area information corresponding to the candidate corner point pair, As shown in Figure 9, the following steps S901 to S903 may be included:
  • S901 Determine a valid candidate corner point pair based on the position information of the center point pointed to by each corner point in each candidate corner point pair and the center area information corresponding to the candidate corner point pair.
  • the candidate corner point pair is regarded as a valid candidate corner point pair.
  • the following formula (12) can be used to determine whether the candidate corner point pair formed by the i-th upper left corner point and the j-th lower right corner point is a valid candidate corner point pair, that is, the i-th upper left corner point and the j-th corner point pair are judged Whether the coordinate range of the m-th central area frame corresponding to the candidate detection frame formed by the lower right corner point and the center point position information pointed to by the i-th upper left corner point and the j-th lower right corner point respectively meet the following formula (12):
  • the coordinate range of the m-th central area frame corresponding to the candidate detection frame formed by the i-th upper-left corner point and the j-th lower-right corner point, and the center point respectively pointed to by the i-th upper-left corner point and the j-th lower-right corner point When the position information satisfies the above formula (12), it means that the candidate corner point pair formed by the i-th upper left corner point and the j-th lower right corner point is a valid candidate corner point pair, and then continue to perform S902 on the valid candidate corner point Otherwise, if the candidate corner point pair formed by the i-th upper left corner point and the j-th lower right corner point is an invalid candidate corner point pair, continue to determine whether the i-th upper left corner point and other lower right corner points are A valid candidate corner point pair can be formed, and the subsequent steps can be executed after a valid candidate corner point pair is obtained.
  • the probability value corresponding to each corner point is used to indicate the probability value of the corresponding feature point of the corner point in the corner point heat map as the corner point.
  • the candidate detection frame corresponding to each valid candidate corner point pair such as the area formed by the center point of each corner point in the valid candidate corner point pair and the center area corresponding to the valid candidate corner point pair
  • the area relationship between the frames and the probability value corresponding to each corner point in the effective candidate corner point pair represent the score value of the candidate detection frame corresponding to each effective candidate corner point pair.
  • the candidate detection frame with the higher score is taken as The probability of the detection frame of the target object is relatively large, and the candidate detection frame is screened through this.
  • the score of the candidate detection frame corresponding to the valid candidate corner point pair can be determined according to the following formula (13):
  • s represents the score of the candidate detection frame corresponding to the valid candidate corner point pair formed by the i-th upper left corner point and the j-th lower right corner point;
  • s tl(i) represents the i-th upper left corner point at the upper left corner point
  • the corresponding feature point in the heat map is used as the probability value of the upper left corner point;
  • s br(j) represents the probability value of the j-th lower right corner point in the lower right corner point in the heat map as the lower right corner point.
  • S903 Determine the detection frame of the target object in the candidate detection frame based on the score of the candidate detection frame corresponding to each valid candidate corner point and the size of the overlapping area between adjacent candidate detection frames.
  • the overlap area can be determined by the size of the overlap area in the image to be detected. The following describes how to base each valid candidate corner point on the corresponding candidate detection frame score and the overlap area between adjacent candidate detection frames , To filter the detection frame of the target object.
  • the detection frame of the target object can be screened in multiple candidate detection frames by soft non-maximum suppression.
  • the candidate detection frame with the highest corresponding score can be used as For the detection frame of the target object, delete other candidate detection frames in the multiple candidate detection frames, so that the detection frame of the target object in the image to be detected can be obtained.
  • the instance information of the target object in the detection frame can be determined.
  • the object to be detected can be determined based on the detection frame of the target object and the initial feature map obtained by feature extraction of the image to be detected Instance information of the target object in the image.
  • the instance information here can be represented by a mask.
  • the mask here means that after instance segmentation of the target object in the image, the pixels of each target object are given at the pixel level, so the mask can be accurate to the edge of the object. In this way, a more accurate position of the target object in the image to be detected can be obtained; in addition, the shape of the target object can also be represented based on the mask, so that the determination of the target object's category can be verified based on the shape, and based on The shape of the target object represented by the mask is subjected to subsequent action analysis on the target object, which is not described in the embodiment of the present disclosure.
  • the instance information of the target object in the image to be detected based on the detection frame of the target object and the initial feature map obtained by feature extraction of the image to be detected, it may include:
  • the detection frame of the target object and the initial feature map corresponding to the image to be detected are input to the region of interest extraction network.
  • the region of interest extraction network can first extract the region of interest matching the size of the initial feature map, and then pass the alignment pool of interest
  • the feature data of the feature points of the initial feature map in the detection frame (that is, the region of interest) is obtained by transformation processing, and then the feature data of the feature points of the initial feature map in the detection frame are input into the mask prediction network.
  • Generate instance information of the target object the instance information can be expressed in the form of a mask, and then the mask of the target object can be expanded to the same size as the target object in the image to be detected, that is, the target object of the image to be detected can be obtained Instance information.
  • the initial feature map f can be corner points Pooling process to obtain the corner point pooled feature map p, and then perform the upper left corner point detection and feature adjustment on the corner point pooled feature map p, and the direction corresponding to the upper left corner point and the upper left corner point can be obtained.
  • Heart offset tensor the process of obtaining the upper left corner point is determined by the upper left corner point detection network.
  • the upper left corner point detection network includes the upper left corner point heat map prediction network and the upper left corner point local offset prediction network (none in Figure 10).
  • the feature adjustment network is first used to adjust the feature map p after the corner point pooling. This process includes determining the steering offset tensor corresponding to the upper left corner point. Then, based on the deformable convolution operation, the feature map p after the corner point pooling is adjusted to obtain the adjusted feature map g, and then through the convolution operation, the centripetal corresponding to the upper left corner point is determined The offset tensor.
  • the lower right corner point is determined by the lower right corner point detection network.
  • the centripetal offset tensor corresponding to the lower right corner point is obtained by feature adjustment and convolution operation.
  • the process is the same as the centripetal offset tensor corresponding to the upper left corner and the upper left corner point.
  • the determination process is similar, and then the detection frame of the target object is determined based on the centripetal offset tensor corresponding to the upper left corner point and the upper left corner point, and the centripetal offset tensor corresponding to the lower right corner point and the lower right corner point.
  • the region of interest is extracted based on the detection frame of the target object and the initial feature map f, and then the region of interest is aligned and pooled to obtain the feature of the region of interest (ie, the initial feature
  • the feature of the region of interest ie, the initial feature
  • the mask of the target object can be obtained, and then the size of the mask is enlarged, and the image to be detected is obtained
  • the mask image of the same size ie, the instance information of the target object).
  • the detection frame of the target object, the mask of the target object, and the target object category can be output, and the required results can be obtained according to the preset requirements, such as outputting the detection of the target object
  • the frame, or the output of the mask image of the target object, or both the detection frame of the target object and the mask image of the target object, and the category of the target object are output at the same time, which are not limited in the embodiment of the present disclosure.
  • the target detection method in the embodiments of the present disclosure may be implemented by a neural network, which is obtained by training using sample pictures containing labeled target sample objects.
  • the neural network of the target detection method proposed in the embodiment of the present disclosure can be obtained by training using the following steps, including steps S1101 to S1104:
  • the sample image here may include a positive sample that annotates the target sample object, and a negative sample that does not include the target sample object, and the target object contained in the positive sample may include multiple categories.
  • the positive samples labeled with the target sample objects can be divided into the target sample objects labeled with the detection frame and the target sample objects labeled with the mask.
  • the process of determining the corner position information of the sample corner point in the sample image and the centripetal offset tensor corresponding to each sample corner point is the same as the process of determining the corner point in the image to be detected as mentioned above.
  • the corner position information in and the centripetal offset tensor corresponding to each corner are similar, so I won’t repeat them here.
  • S1103 Predict the target sample object in the sample image based on the corner position information of each sample corner point in the sample image and the centripetal offset tensor corresponding to each sample corner point.
  • the process of predicting the target sample object in the sample image is the same as the method of determining the target object in the image to be detected as mentioned above, and will not be repeated here.
  • S1104 Adjust network parameter values of the neural network based on the predicted target sample object in the sample image and the labeled target sample object in the sample image.
  • a loss function can be introduced to determine the loss value corresponding to the target sample object prediction.
  • the network parameter value of the neural network can be adjusted through the loss value, for example, when the loss value is less than the set threshold, that is You can stop training to get the network parameter values of the neural network.
  • the detection frame of the target sample object, the mask of the target sample object, and the determination process of the target sample object's category are the same as the detection frame of the target object, the mask of the target object, and the category of the target object described above. The process is similar, so I won't repeat it here.
  • the neural network training method obtains a sample image, and based on the sample image, determines the corner position information of each sample corner point in the sample image, and the centripetal offset corresponding to each sample corner point. Based on the corner position information of each sample corner point in the sample image and the centripetal offset tensor corresponding to each sample corner point, the target sample object is detected in the sample image, because the sample corner point refers to the main Feature points, for example, the sample corner points can include the upper left sample corner point and the lower right sample corner point, where the upper left sample corner point refers to the intersection of the line corresponding to the upper contour of the target sample object and the line corresponding to the left contour of the target sample object.
  • the lower sample corner point refers to the intersection of the straight line corresponding to the lower contour of the target sample object and the straight line corresponding to the right contour of the target sample object.
  • the upper left sample corner and the lower right sample corner belong to the detection frame of the same target sample object.
  • the positions of the centripetal offset tensor corresponding to the upper left sample corner point and the lower right sample corner point should be relatively close. Therefore, the training method of the neural network proposed in the embodiment of the present disclosure is based on the representation of the target sample object being trained
  • the corner point position information of the position in the sample image, and the centripetal offset tensor corresponding to each sample corner point determine the sample corner point belonging to the same target sample object, and then based on the determined sample corner point, the sample corner point can be detected.
  • the embodiment of the present disclosure also provides a target detection device corresponding to the target detection method. Since the technical principle of the device in the embodiment of the disclosure is similar to the target detection method described in the embodiment of the disclosure, the implementation of the device can be referred to The implementation of the method will not repeat the repetition.
  • FIG. 12 it is a schematic diagram of a target detection device 1200 provided by an embodiment of the present disclosure.
  • the device includes: an acquisition part 1201, a determination part 1202, and a detection part 1203.
  • the acquiring part 1201 is configured to acquire the image to be detected
  • the determining part 1202 is configured to determine, based on the image to be detected, the corner position information of each corner point in the image to be detected and the centripetal offset tensor corresponding to each corner point, where the corner points represent the target object in the image to be detected Location;
  • the detection part 1203 is configured to determine the target object in the image to be detected based on the position information of each corner point in the image to be detected and the centripetal offset tensor corresponding to each corner point.
  • the determining part 1202 is configured to:
  • the corner position information of each corner point in the image to be detected and the centripetal offset tensor corresponding to each corner point are determined.
  • the method includes:
  • the local offset information is used to indicate that the real physical point represented by the corresponding corner point is in the corner heat map Position offset information in;
  • each corner point is in the to-be-detected image.
  • the position information of the corner points in the image is determined.
  • the method includes:
  • the offset domain information contains multiple initial feature points associated with the feature point respectively pointing to their corresponding offset feature points The offset tensor;
  • the feature data of the feature points in the corner point pooled feature map are adjusted to obtain The adjusted feature map
  • centripetal offset tensor corresponding to each corner point is determined.
  • the corner heat map corresponding to the image to be detected includes a corner heat map corresponding to multiple channels, and each channel of the multiple channels corresponds to a preset object category; the determining part 1202 is in After being configured to determine the probability value of each feature point in the corner heat map as a corner point based on the corner heat map, it is also configured to:
  • the image to be detected contains the target object of the preset object category corresponding to the channel.
  • the detection part 1203 is configured to:
  • the detection part 1203 is configured to determine the target object in the image to be detected based on the position information of each corner point in the image to be detected and the centripetal offset tensor corresponding to each corner point.
  • the case of the detection box includes:
  • the detection frame of the target object is determined in the candidate detection frame.
  • the detecting part 1203 is configured to be based on the corner position information of each corner point in each candidate corner point pair in the image to be detected.
  • the central region information corresponding to the candidate corner point pair it includes:
  • the coordinate range of the central area frame corresponding to the candidate corner point pair is determined.
  • the detection part 1203 is configured to be based on the position information of the center point pointed to by each corner point in each candidate corner point pair and the center area information corresponding to the candidate corner point pair.
  • the detection frame determines the detection frame of the target object it includes:
  • the detection frame of the target object is determined in the candidate detection frame.
  • the detection part 1203 is further configured to:
  • the instance information of the target object in the image to be detected is determined based on the detection frame of the target object and the initial feature map obtained by feature extraction of the image to be detected.
  • the detection part 1203 is configured to determine the instance information of the target object in the image to be detected based on the detection frame of the target object and the initial feature map obtained by feature extraction of the image to be detected, including :
  • the instance information of the target object in the image to be detected is determined.
  • the target detection device 1200 further includes a neural network training part 1204, and the neural network training part 1204 is configured to:
  • the neural network is trained using sample images that contain labeled target sample objects.
  • the neural network training part 1204 is configured to train the neural network according to the following steps:
  • the network parameter values of the neural network are adjusted.
  • parts may be parts of circuits, parts of processors, parts of programs or software, etc., of course, may also be units, modules, or non-modular.
  • an embodiment of the present disclosure further provides an electronic device 1300.
  • a schematic structural diagram of the electronic device 1300 provided by the embodiment of the present disclosure includes:
  • the target object in the image to be detected is determined.
  • the embodiments of the present disclosure also provide a computer-readable storage medium having a computer program stored on the computer-readable storage medium, and the computer program executes the steps of the target detection method in the foregoing method embodiment when the computer program is run by a processor.
  • the storage medium may be a volatile or non-volatile computer readable storage medium.
  • the embodiments of the present disclosure also provide a computer program, including computer-readable code, and when the computer-readable code runs in an electronic device, the processor in the electronic device executes the same as described in the first aspect.
  • the computer program product of the target detection method provided by the embodiment of the present disclosure includes a computer-readable storage medium storing program code, and the program code includes instructions that can be used to execute the steps of the target detection method described in the above method embodiment
  • the program code includes instructions that can be used to execute the steps of the target detection method described in the above method embodiment
  • the embodiments of the present disclosure also provide a computer program, which, when executed by a processor, implements any one of the methods in the foregoing embodiments.
  • the computer program product can be specifically implemented by hardware, software, or a combination thereof.
  • the computer program product is specifically embodied as a computer storage medium.
  • the computer program product is specifically embodied as a software product, such as a software development kit (SDK) and so on.
  • SDK software development kit
  • the working process of the system and device described above can refer to the corresponding process in the foregoing method embodiment, which will not be repeated here.
  • the disclosed system, device, and method may be implemented in other ways.
  • the device embodiments described above are merely illustrative.
  • the division of the units is only a logical function division, and there may be other divisions in actual implementation.
  • multiple units or components may be combined or It can be integrated into another system, or some features can be ignored or not implemented.
  • the displayed or discussed mutual coupling or direct coupling or communication connection may be through some communication interfaces, indirect coupling or communication connection of devices or units, and may be in electrical, mechanical or other forms.
  • the units described as separate components may or may not be physically separated, and the components displayed as units may or may not be physical units, that is, they may be located in one place, or they may be distributed on multiple network units. Some or all of the units may be selected according to actual needs to achieve the objectives of the solutions of the embodiments.
  • the functional units in the various embodiments of the present disclosure may be integrated into one processing unit, or each unit may exist alone physically, or two or more units may be integrated into one unit.
  • the function is implemented in the form of a software functional unit and sold or used as an independent product, it can be stored in a non-volatile computer readable storage medium executable by a processor.
  • the technical solution of the present disclosure essentially or the part that contributes to the prior art or the part of the technical solution can be embodied in the form of a software product, and the computer software product is stored in a storage medium, including Several instructions are used to make a computer device (which may be a personal computer, a server, or a network device, etc.) execute all or part of the steps of the methods described in the various embodiments of the present disclosure.
  • the aforementioned storage media include: U disk, mobile hard disk, read-only memory (Read-Only Memory, ROM), random access memory (Random Access Memory, RAM), magnetic disk or optical disk and other media that can store program code .
  • the embodiments of the present disclosure provide a target detection method, device, electronic equipment, and computer-readable storage medium.
  • the target detection method includes: acquiring an image to be detected; and determining that each corner point is in the image based on the image to be detected.
  • the corner points represent the position of the target object in the image to be detected; based on the angle of each corner point in the image to be detected
  • the point position information and the centripetal offset tensor corresponding to each corner point are used to determine the target object in the image to be detected.
  • the target detection method proposed in the embodiments of the present disclosure can determine the corner points belonging to the same target object based on the corner position information of each corner point in the image to be detected and the centripetal offset tensor corresponding to each corner point, and then based on The determined corner point can detect the same target object.

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Multimedia (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Evolutionary Computation (AREA)
  • Health & Medical Sciences (AREA)
  • Artificial Intelligence (AREA)
  • Computing Systems (AREA)
  • Databases & Information Systems (AREA)
  • General Health & Medical Sciences (AREA)
  • Medical Informatics (AREA)
  • Software Systems (AREA)
  • Image Analysis (AREA)

Abstract

An object detection method and apparatus, an electronic device, and a computer readable storage medium. The object detection method comprises: acquiring an image to be inspected (S201); on the basis of said image, determining corner position information of each corner in said image and a centripetal shift tensor corresponding to each corner, the corner representing the position of a target object in said image (S202); and on the basis of the corner position information of each corner in said image and the centripetal shift tensor corresponding to each corner, determining the target object in said image (S203).

Description

目标检测方法、装置、电子设备及计算机可读存储介质Target detection method, device, electronic equipment and computer readable storage medium
相关申请的交叉引用Cross-references to related applications
本申请基于申请号为202010073142.6、申请日为2020年01月22日的中国专利申请提出,并要求该中国专利申请的优先权,该中国专利申请的全部内容在此引入本申请作为参考。This application is filed based on a Chinese patent application with an application number of 202010073142.6 and an application date of January 22, 2020, and claims the priority of the Chinese patent application. The entire content of the Chinese patent application is hereby incorporated by reference into this application.
技术领域Technical field
本公开实施例涉及图像识别技术领域,具体而言,涉及一种目标检测方法、装置、电子设备及计算机可读存储介质。The embodiments of the present disclosure relate to the field of image recognition technology, and in particular, to a target detection method, device, electronic equipment, and computer-readable storage medium.
背景技术Background technique
目标检测是计算机视觉重要的基础问题,许多计算机视觉的应用都依赖于目标检测,比如自动驾驶、视频监控和移动娱乐。Target detection is an important basic problem of computer vision. Many computer vision applications rely on target detection, such as autonomous driving, video surveillance, and mobile entertainment.
在进行目标检测的情况下,主要任务是用检测框标出图像中物体所在位置,该过程可以基于物体关键点的目标检测算法确定物体在图像中的位置,通过确定出图像中所有物体关键点后,将属于相同物体的物体关键点进行匹配,从而得到物体的检测框。In the case of target detection, the main task is to use the detection frame to mark the location of the object in the image. This process can determine the location of the object in the image based on the target detection algorithm of the key point of the object, and determine all the key points of the object in the image Then, the key points of the objects belonging to the same object are matched to obtain the detection frame of the object.
在图像中包含多个外观相似的物体的情况下,因外观相似的物体对应的物体关键点之间的匹配度较高,容易造成错误的检测结果,比如检测结果为同一个检测框中包含了多个物体,因此,目前的目标检测方法的检测精度准确度较低。In the case that the image contains multiple objects with similar appearance, the matching degree between the key points of the objects corresponding to the similar objects is high, which is easy to cause wrong detection results. For example, the detection result is that the same detection frame contains Multiple objects, therefore, the detection accuracy of current target detection methods is low.
发明内容Summary of the invention
本公开实施例至少提供一种目标检测方案。The embodiments of the present disclosure provide at least one target detection solution.
第一方面,本公开实施例提供了一种目标检测方法,包括:In the first aspect, embodiments of the present disclosure provide a target detection method, including:
获取待检测图像;Obtain the image to be detected;
基于所述待检测图像,确定各个角点在所述待检测图像中的角点位置信息以及各个角点对应的向心偏移张量,角点表征所述待检测图像中的目标对象的位置;Based on the image to be detected, determine the position information of each corner point in the image to be detected and the centripetal offset tensor corresponding to each corner point, and the corner point represents the position of the target object in the image to be detected ;
基于各个角点在所述待检测图像中的角点位置信息及各个角点对应的心偏移张量,确定所述待检测图像中的目标对象。The target object in the image to be detected is determined based on the corner position information of each corner point in the image to be detected and the heart offset tensor corresponding to each corner point.
本公开实施例提供的方法,在获取到待检测图像后,首先确定角点在待检测图像中的角点位置信息以及各个角点对应的向心偏移张量,因为角点是指图像中的主要特征点,其在待检测图像中的角点位置信息能够表征每个目标对象在待检测图像中的位置,比如角点可以包括左上角点和右下角点,其中左上角点是指对应目标对象上侧轮廓的直线与对应目标对象左侧轮廓的直线的交点,右下角点是指对应目标对象下侧轮廓的直线与对应目标对象右侧轮廓的直线的交点,在左上角点和右下角点属于同一个目标对象的检测框的情况下,左上角点和右下角点分别对应的向心偏移张量指向的位置应该比较接近,因此,本公开实施例提出的目标检测方法,基于各个角点在待检测图像中的角点位置信息及各个角点对应的向心偏移张量,能够确定出属于同一目标对象的角点,进而基于确定出的角点可以检测出该同一目标对象。In the method provided by the embodiments of the present disclosure, after acquiring the image to be detected, first determine the corner position information of the corner point in the image to be detected and the centripetal offset tensor corresponding to each corner point, because the corner point refers to the image in the image The main feature points of the corner points in the image to be detected can characterize the position of each target object in the image to be detected. For example, the corner points can include the upper left corner point and the lower right corner point, where the upper left corner point refers to the corresponding The intersection of the straight line of the upper contour of the target object and the straight line corresponding to the left contour of the target object. The lower right corner point refers to the intersection of the straight line corresponding to the lower contour of the target object and the straight line corresponding to the right contour of the target object. In the case where the lower corner points belong to the detection frame of the same target object, the positions pointed by the centripetal offset tensors corresponding to the upper left corner points and the lower right corner points should be relatively close. Therefore, the target detection method proposed in the embodiments of the present disclosure is based on The corner position information of each corner point in the image to be detected and the centripetal offset tensor corresponding to each corner point can determine the corner point belonging to the same target object, and then the same target can be detected based on the determined corner point Object.
在一种可能的实施方式中,所述基于所述待检测图像,确定各个角点在待检测图像中的角点位置信息以及各个角点对应的向心偏移张量,包括:In a possible implementation manner, the determining, based on the to-be-detected image, the corner position information of each corner point in the to-be-detected image and the centripetal offset tensor corresponding to each corner point includes:
对所述待检测图像进行特征提取,得到所述待检测图像对应的初始特征图;Performing feature extraction on the image to be detected to obtain an initial feature map corresponding to the image to be detected;
对所述初始特征图进行角点池化处理,得到角点池化后的特征图;Performing corner pooling processing on the initial feature map to obtain a feature map after corner pooling;
基于所述角点池化后的特征图,确定各个角点在所述待检测图像中的角点位置信息,以及各个角点对应的向心偏移张量。Based on the feature map after the corner point pooling, the corner point position information of each corner point in the image to be detected and the centripetal offset tensor corresponding to each corner point are determined.
本公开实施例提供的方法,通过对待检测图像进行特征提取,得到初始特征图,并对初始特征图进行角点池化处理,得到能够便于提取角点以及角点对应的向心偏移量的特征图,即角点池化后的特征图。The method provided by the embodiment of the present disclosure obtains an initial feature map by extracting features from the image to be detected, and performs corner pooling processing on the initial feature map, so as to obtain a convenient extraction of the corner points and the centripetal offset corresponding to the corner points. Feature map, that is, the feature map after corner point pooling.
在一种可能的实施方式中,所述基于所述角点池化后的特征图,确定各个角点在所述待检测图像中的角点位置信息,包括:In a possible implementation manner, the determining the corner position information of each corner point in the image to be detected based on the feature map after the corner point pooling includes:
基于所述角点池化后的特征图,生成所述待检测图像对应的角点热力图;Generating a corner heat map corresponding to the image to be detected based on the feature map after corner pooling;
基于所述角点热力图,确定所述角点热力图中每个特征点作为角点的概率值,并基于每个特征点作为角点的概率值,从所述特征点中筛选出所述角点;Based on the corner point heat map, determine the probability value of each feature point in the corner point heat map as a corner point, and filter out the feature points based on the probability value of each feature point as a corner point corner;
获取筛选出的各个角点在所述角点热力图中的位置信息、以及各个角点对应的局部偏移信息,所述局部偏移信息用于表示对应的角点所表征的真实物理点在所述角点热力图中的位置偏移信息;Obtain the position information of the selected corner points in the corner point heat map and the local offset information corresponding to each corner point. The local offset information is used to indicate that the real physical point represented by the corresponding corner point is at Position offset information in the corner point heat map;
基于获取到的各个角点在所述角点热力图中的位置信息、各个角点对应的局部偏移信息、以及所述角点热力图和所述待检测图像之间的尺寸比例,确定各个角点在所述待检测图像中的角点位置信息。Based on the acquired position information of each corner point in the corner point heat map, the local offset information corresponding to each corner point, and the size ratio between the corner point heat map and the image to be detected, determine each The position information of the corner point in the image to be detected.
本公开实施例提供的一种确定各个角点在待检测图像中的角点位置信息的方式,该过程通过引入角点热力图,通过每个特征点作为角点的概率值确定出可以作为角点的特征点,在选择出角点后,通过对角点在角点热力图中的位置信息进行修正后,确定出角点在待检测图像中的角点位置信息,该方式能够得到准确度较高的角点的角点位置信息,从而便于后续基于该角点检测目标对象在待检测图像中的位置。The embodiment of the present disclosure provides a method for determining the position information of each corner point in the image to be detected. This process introduces a corner point heat map, and determines that the corner point can be used as a corner point by the probability value of each feature point as a corner point. The feature point of the point, after the corner point is selected, the position information of the corner point in the corner point heat map is corrected to determine the corner point position information of the corner point in the image to be detected. This method can obtain accuracy The corner position information of the higher corner point facilitates subsequent detection of the position of the target object in the image to be detected based on the corner point.
在一种可能的实施方式中,所述基于所述角点池化后的特征图,确定各个角点对应的向心偏移张量,包括:In a possible implementation manner, the determining the centripetal offset tensor corresponding to each corner point based on the feature map after the corner point pooling includes:
基于所述角点池化后的特征图,确定所述角点池化后的特征图中的每个特征点对应的导向偏移张量,每个特征点对应的导向偏移张量表征由该特征点指向所述待检测图像中的目标对象中心点的偏移张量;Based on the feature map after corner point pooling, the steering offset tensor corresponding to each feature point in the corner point pooling feature map is determined, and the steering offset tensor corresponding to each feature point is represented by The offset tensor of the feature point pointing to the center point of the target object in the image to be detected;
基于每个特征点对应的所述导向偏移张量,确定该特征点的偏移域信息;所述偏移域信息中包含与该特征点关联的多个初始特征点分别指向各自对应的偏移后特征点的偏移张量;Based on the steering offset tensor corresponding to each feature point, determine the offset domain information of the feature point; the offset domain information includes multiple initial feature points associated with the feature point respectively pointing to their corresponding offsets The offset tensor of the feature point after the shift;
基于所述角点池化后的特征图,以及该角点池化后的特征图中的特征点的偏移域信息,对所述角点池化后的特征图中的特征点的特征数据进行调整,得到调整后的特征图;Based on the corner point pooled feature map and the offset domain information of the feature points in the corner point pooled feature map, the feature data of the feature points in the corner point pooled feature map Make adjustments to obtain the adjusted feature map;
基于所述调整后的特征图,确定各个角点对应的向心偏移张量。Based on the adjusted feature map, the centripetal offset tensor corresponding to each corner point is determined.
本公开实施例提供的确定向心偏移张量的过程,通过考虑目标对象信息,比如引入角点对应的导向偏移张量,以及特征点的偏移域信息,对角点池化后的特征图中的特征点的特征数据进行调整,使得得到的调整后的特征图中的特征点的特征数据能够包含更丰富的目标对象信息,从而能够确定出每个角点对应的更加准确的向心偏移张量,通过准确的向心偏移张量,能够准确得到角点指向的中心点位置信息,从而准确地检测目标对象在待检测图像中的位置。The process of determining the centripetal offset tensor provided by the embodiments of the present disclosure considers the target object information, such as introducing the steering offset tensor corresponding to the corner point, and the offset domain information of the feature point. The feature data of the feature points in the feature map are adjusted so that the feature data of the feature points in the adjusted feature map can contain richer target object information, so that a more accurate orientation corresponding to each corner point can be determined. The central offset tensor, through accurate centripetal offset tensor, can accurately obtain the position information of the center point pointed by the corner point, so as to accurately detect the position of the target object in the image to be detected.
在一种可能的实施方式中,所述待检测图像对应的角点热力图包括多个通道分别对应的角点热力图,所述多个通道中的每个通道对应一种预设对象类别;所述基于所述角点热力图,确定所述角点热力图中每个特征点作为角点的概率值之后,所述检测方法还包括:In a possible implementation manner, the corner heat map corresponding to the image to be detected includes a corner heat map corresponding to multiple channels, and each channel of the multiple channels corresponds to a preset object category; After determining the probability value of each feature point in the corner heat map as a corner point based on the corner heat map, the detection method further includes:
针对所述多个通道中的每个通道,基于该通道对应的角点热力图中每个特征点作为角点的概率值,确定该通道对应的角点热力图中是否存在所述角点;For each channel of the plurality of channels, determine whether the corner point exists in the corner heat map corresponding to the channel based on the probability value of each feature point in the corner heat map corresponding to the channel as a corner point;
在该通道对应的角点热力图中存在所述角点的情况下,确定所述待检测图像中包含该通道对应的预设对象类别的目标对象。When the corner point exists in the corner point heat map corresponding to the channel, it is determined that the image to be detected contains the target object of the preset object category corresponding to the channel.
本公开实施例提供的方法,通过将角点池化后的特征图输入角点热力图预测网络,即可以得到包含预设通道个数的角点热力图,通过每个通道对应的角点热力图中是否存在角点,进而能够确定待检测图像中是否存在该通道对应类别的目标对象。In the method provided by the embodiment of the present disclosure, by inputting the feature map after the corner point pooling into the corner point heat map prediction network, the corner heat map containing the preset number of channels can be obtained, and the corner heat map corresponding to each channel can be obtained. Whether there is a corner point in the image, it can be determined whether there is a target object corresponding to the channel in the image to be detected.
在一种可能的实施方式中,所述基于各个角点在所述待检测图像中的角点位置信息及各个角点对应的向心偏移张量,确定所述待检测图像中的目标对象,包括:In a possible implementation manner, the target object in the image to be detected is determined based on the position information of each corner point in the image to be detected and the centripetal offset tensor corresponding to each corner point ,include:
基于各个角点在所述待检测图像中的角点位置信息和各个角点对应的向心偏移张量,确定所述待检测图像中目标对象的检测框。Based on the corner position information of each corner point in the image to be detected and the centripetal offset tensor corresponding to each corner point, the detection frame of the target object in the image to be detected is determined.
本公开实施例提供的方法,可以基于各个角点在待检测图像中的角点位置信息和各个角点对应的向心偏移张量,确定出每个目标对象的检测框,即得到每个目标对象在待检测图像中的位置信息。The method provided by the embodiments of the present disclosure can determine the detection frame of each target object based on the corner position information of each corner point in the image to be detected and the centripetal offset tensor corresponding to each corner point. The position information of the target object in the image to be detected.
在一种可能的实施方式中,所述基于各个角点在所述待检测图像中的角点位置信息和各个角点对应的向心偏移张量,确定成所述待检测图像中目标对象的检测框,包括:In a possible implementation manner, the target object in the image to be detected is determined based on the position information of each corner point in the image to be detected and the centripetal offset tensor corresponding to each corner point The detection box includes:
基于各个角点在所述待检测图像中的角点位置信息,筛选能够构成候选检测框的候选角点对;Based on the corner point position information of each corner point in the image to be detected, filter candidate corner point pairs that can form a candidate detection frame;
基于每个候选角点对中每个角点在所述待检测图像中的角点位置信息和该角点对应的向心偏移张量,确定该角点指向的中心点位置信息;Determine the position information of the center point pointed to by the corner point based on the corner position information of each corner point in the image to be detected in each candidate corner point pair and the centripetal offset tensor corresponding to the corner point;
基于每个候选角点对中每个角点在所述待检测图像中的角点位置信息,确定该候选角点对所对应的中心区域信息;Determine the center area information corresponding to the candidate corner point pair based on the corner point position information of each corner point in the candidate corner point pair in the image to be detected;
基于每个候选角点对中每个角点指向的中心点位置信息以及该候选角点对所对应的中心区域信息,在所述候选检测框中确定所述目标对象的检测框。The detection frame of the target object is determined in the candidate detection frame based on the position information of the center point pointed to by each corner point in each candidate corner point pair and the center area information corresponding to the candidate corner point pair.
本公开实施例提出的确定目标对象的检测框的方式,通过角点的角点位置信息首先确定能够构成候选检测框的候选角点对,然后再基于该候选角点对中每个角点分别对应的向心偏移张量,来确定候选检测框包围的目标对象是否为同一个目标对象,从而能够较为准确地检测出待检测图像中的所有目标对象的检测框。In the method of determining the detection frame of the target object proposed by the embodiment of the present disclosure, the corner point position information of the corner points is used to first determine the candidate corner point pairs that can constitute the candidate detection frame, and then based on each corner point in the candidate corner point pair. The corresponding centripetal offset tensor is used to determine whether the target object surrounded by the candidate detection frame is the same target object, so that the detection frame of all target objects in the image to be detected can be detected more accurately.
在一种可能的实施方式中,所述基于每个候选角点对中每个角点在所述待检测图像中的角点位置信 息,确定该候选角点对所对应的中心区域信息,包括:In a possible implementation manner, the determining the center region information corresponding to the candidate corner point pair based on the corner point position information of each corner point in the candidate corner point pair in the image to be detected includes :
基于该候选角点对的每个角点的角点位置信息,确定表征该候选角点对所对应的中心区域框的角点位置信息;Based on the corner point position information of each corner point of the candidate corner point pair, determine the corner point position information that characterizes the center area frame corresponding to the candidate corner point pair;
基于所述中心区域框的角点位置信息,确定该候选角点对所对应的中心区域框的坐标范围。Based on the corner point position information of the central area frame, the coordinate range of the central area frame corresponding to the candidate corner point pair is determined.
在一种可能的实施方式中,所述基于每个候选角点对中每个角点指向的中心点位置信息,以及该候选角点对所对应的中心区域信息,在所述候选检测框中确定所述目标对象的检测框,包括:In a possible implementation manner, based on the position information of the center point pointed to by each corner point in each candidate corner point pair, and the center area information corresponding to the candidate corner point pair, in the candidate detection frame Determining the detection frame of the target object includes:
基于每个候选角点对中每个角点指向的中心点位置信息,以及该候选角点对所对应的中心区域信息,确定有效候选角点对;Determine a valid candidate corner point pair based on the position information of the center point pointed to by each corner point in each candidate corner point pair, and the center area information corresponding to the candidate corner point pair;
基于所述有效候选角点对中每个角点指向的中心点位置信息、所述有效候选角点对所对应的中心区域信息、以及所述有效候选角点对中每个角点对应的概率值,确定每个有效候选角点对所对应的候选检测框的分值;每个角点对应的概率值用于表示该角点在角点热力图中对应的特征点作为角点的概率值;Based on the position information of the center point pointed to by each corner point in the valid candidate corner point pair, the central area information corresponding to the valid candidate corner point pair, and the probability corresponding to each corner point in the valid candidate corner point pair Value, determine the score of the candidate detection frame corresponding to each valid candidate corner point; the probability value corresponding to each corner point is used to indicate the probability value of the corresponding feature point of the corner point in the corner heat map as the corner point ;
基于每个有效候选角点对所对应的候选检测框的分值、以及相邻候选检测框之间的重叠区域大小,在所述候选检测框中确定所述目标对象的检测框。Based on the score of the candidate detection frame corresponding to each valid candidate corner point and the size of the overlapping area between adjacent candidate detection frames, the detection frame of the target object is determined in the candidate detection frame.
本公开实施例提供的方法,通过对构成候选检测框的候选角点对进行有效筛选,确定能够筛选出只表征一个目标对象的候选检测框,然后对这些只表征一个目标对象的候选检测框进行软式非极大抑制筛选,从而得到准确的表征目标对象的检测框。The method provided by the embodiment of the present disclosure effectively screens the candidate corner points that constitute the candidate detection frame, and determines that the candidate detection frame that only represents one target object can be screened out, and then performs detection on these candidate detection frames that only represent one target object. Soft non-maximum suppression screening, so as to obtain an accurate detection frame that characterizes the target object.
在一种可能的实施方式中,所述确定所述待检测图像中目标对象的检测框之后,所述目标检测方法还包括:In a possible implementation manner, after the determination of the detection frame of the target object in the image to be detected, the target detection method further includes:
基于所述目标对象的检测框和对所述待检测图像进行特征提取得到的初始特征图,确定所述待检测图像中所述目标对象的实例信息。The instance information of the target object in the image to be detected is determined based on the detection frame of the target object and the initial feature map obtained by feature extraction of the image to be detected.
本公开实施例提供的方法,可以确定目标对象的实例信息,这里的实例是指对图像中的目标对象进行实例分割后,在像素层面给出每个目标对象的像素,实例分割可以精确到物体的边缘,从而得到目标对象在待检测图像中更加准确的位置信息。The method provided by the embodiments of the present disclosure can determine the instance information of the target object. The instance here means that after the instance segmentation of the target object in the image, the pixel of each target object is given at the pixel level, and the instance segmentation can be accurate to the object. To obtain more accurate position information of the target object in the image to be detected.
在一种可能的实施方式中,所述基于所述目标对象的检测框和对所述待检测图像进行特征提取得到的初始特征图,确定所述待检测图像中所述目标对象的实例信息,包括:In a possible implementation manner, the determining the instance information of the target object in the image to be detected is based on the detection frame of the target object and the initial feature map obtained by feature extraction of the image to be detected, include:
基于所述目标对象的检测框以及所述初始特征图,提取所述初始特征图在所述检测框内的特征点的特征数据;Extracting feature data of feature points of the initial feature map in the detection frame based on the detection frame of the target object and the initial feature map;
基于所述初始特征图在所述检测框内的特征点的特征数据,确定所述待检测图像中所述目标对象的实例信息。Based on the feature data of the feature points of the initial feature map in the detection frame, the instance information of the target object in the image to be detected is determined.
在一种可能的实施方式中,所述目标检测方法是由神经网络实现,所述神经网络利用包含了标注目标样本对象的样本图片训练得到。In a possible implementation manner, the target detection method is implemented by a neural network, and the neural network is obtained by training using sample pictures containing labeled target sample objects.
在一种可能的实施方式中,所述神经网络采用以下步骤训练得到:In a possible implementation manner, the neural network is obtained by training using the following steps:
获取样本图像;Obtain sample images;
基于所述样本图像,确定各个样本角点在样本图像中的角点位置信息及各个样本角点对应的向心偏移张量,所述样本角点表征所述样本图像中的目标样本对象的位置;Based on the sample image, determine the corner position information of each sample corner point in the sample image and the centripetal offset tensor corresponding to each sample corner point, and the sample corner point represents the target sample object in the sample image. Location;
基于各个样本角点在所述样本图像中的角点位置信息及各个样本角点对应的向心偏移张量,预测所述样本图像中的目标样本对象;Predicting the target sample object in the sample image based on the corner position information of each sample corner point in the sample image and the centripetal offset tensor corresponding to each sample corner point;
基于预测的所述样本图像中的目标样本对象和所述样本图像中的标注目标样本对象,对所述神经网络的网络参数值进行调整。The network parameter value of the neural network is adjusted based on the predicted target sample object in the sample image and the labeled target sample object in the sample image.
本公开实施例提供的神经网络的训练方法,通过获取样本图像,并基于该样本图像确定各个样本角点在样本图像中的角点位置信息,及各个样本角点分别对应的向心偏移张量,从而基于各个样本角点在样本图像中的角点位置信息和各个样本角点对应的向心偏移张量,在样本图像中检测目标样本对象,因为样本角点是指图像中的主要特征点,比如样本角点可以包括左上样本角点和右下样本角点,其中,左上样本角点是指对应目标样本对象上侧轮廓的直线与对应目标样本对象左侧轮廓的直线的交点,右下样本角点是指对应目标样本对象下侧轮廓的直线与对应目标样本对象右侧轮廓的直线的交点,在左上样本角点和右下样本角点属于同一个目标样本对象的检测框的情况下,左上样本角点和右下样本角点分别对应的向心偏移张量指向的位置应该比较接近,因此,本公开实施例提出的神经网络的训练方法,基于表征目标样本对象在样本图像中的位置的角点的角点位置信息,以及每个样本角点对应的向心偏移张量,确定出属于同一目标样本对象的样本角点,进而基于确定出的样本角点可以检测出该同一目标样本对象,然后通过不断地基于样本图像中的标注目标对象,不断调整神经网络参数,从而得到准确度较高的神经网络,基于该准确度较高的神经网络即可以对目标对象进行准确检测。The neural network training method provided by the embodiments of the present disclosure obtains a sample image, and based on the sample image, determines the position information of each sample corner point in the sample image, and the centripetal offset corresponding to each sample corner point. Based on the corner position information of each sample corner point in the sample image and the centripetal offset tensor corresponding to each sample corner point, the target sample object is detected in the sample image, because the sample corner point refers to the main Feature points, such as sample corner points may include upper left sample corner points and lower right sample corner points, where the upper left sample corner point refers to the intersection of a line corresponding to the upper contour of the target sample object and a line corresponding to the left contour of the target sample object. The lower right sample corner point refers to the intersection of the straight line corresponding to the lower contour of the target sample object and the straight line corresponding to the right contour of the target sample object. The upper left sample corner and the lower right sample corner belong to the detection frame of the same target sample object. In this case, the centripetal offset tensor corresponding to the upper left sample corner point and the lower right sample corner point should be relatively close to each other. Therefore, the neural network training method proposed in the embodiment of the present disclosure is based on characterizing the target sample object in the sample The corner position information of the corner points of the position in the image, and the centripetal offset tensor corresponding to each sample corner point, determine the sample corner points belonging to the same target sample object, and then based on the determined sample corner points can be detected The same target sample object is extracted, and then the neural network parameters are continuously adjusted based on the target object in the sample image, so as to obtain a neural network with higher accuracy. Based on the neural network with higher accuracy, the target object can be Perform accurate detection.
第二方面,本公开实施例提供了一种目标检测装置,包括:In a second aspect, embodiments of the present disclosure provide a target detection device, including:
获取部分,被配置为获取待检测图像;The obtaining part is configured to obtain the image to be detected;
确定部分,被配置为基于所述待检测图像,确定各个角点在所述待检测图像中的角点位置信息以及各个角点对应的向心偏移张量,角点表征所述待检测图像中的目标对象的位置;The determining part is configured to determine, based on the image to be detected, the corner position information of each corner point in the image to be detected and the centripetal offset tensor corresponding to each corner point, where the corner points represent the image to be detected The location of the target object in
检测部分,被配置为基于各个角点在所述待检测图像中的角点位置信息及各个角点对应的向心偏移张量,确定所述待检测图像中的目标对象。The detection part is configured to determine the target object in the image to be detected based on the position information of each corner point in the image to be detected and the centripetal offset tensor corresponding to each corner point.
在一种可能的实施方式中,所述确定部分被配置为:In a possible implementation manner, the determining part is configured to:
对所述待检测图像进行特征提取,得到所述待检测图像对应的初始特征图;Performing feature extraction on the image to be detected to obtain an initial feature map corresponding to the image to be detected;
对所述初始特征图进行角点池化处理,得到角点池化后的特征图;Performing corner pooling processing on the initial feature map to obtain a feature map after corner pooling;
基于所述角点池化后的特征图,确定各个角点在所述待检测图像中的角点位置信息,以及各个角点对应的向心偏移张量。Based on the feature map after the corner point pooling, the corner point position information of each corner point in the image to be detected and the centripetal offset tensor corresponding to each corner point are determined.
在一种可能的实施方式中,所述确定部分在被配置为基于所述角点池化后的特征图,确定各个角点在所述待检测图像中的角点位置信息的情况下,包括:In a possible implementation manner, when the determining part is configured to determine the corner position information of each corner point in the image to be detected based on the feature map after the corner point pooling, the method includes :
基于所述角点池化后的特征图,生成所述待检测图像对应的角点热力图;Generating a corner heat map corresponding to the image to be detected based on the feature map after corner pooling;
基于所述角点热力图,确定所述角点热力图中每个特征点作为角点的概率值,并基于每个特征点作为角点的概率值,从所述特征点中筛选出所述角点;Based on the corner point heat map, determine the probability value of each feature point in the corner point heat map as a corner point, and filter out the feature points based on the probability value of each feature point as a corner point corner;
获取筛选出的各个角点在所述角点热力图中的位置信息、以及各个角点对应的局部偏移信息,所述局部偏移信息用于表示对应的角点所表征的真实物理点在所述角点热力图中的位置偏移信息;Obtain the position information of the selected corner points in the corner point heat map and the local offset information corresponding to each corner point. The local offset information is used to indicate that the real physical point represented by the corresponding corner point is at Position offset information in the corner point heat map;
基于获取到的每个角点在所述角点热力图中的位置信息、各个角点对应的局部偏移信息、以及所述角点热力图和所述待检测图像之间的尺寸比例,确定各个角点在所述待检测图像中的角点位置信息。Based on the acquired position information of each corner point in the corner point heat map, the local offset information corresponding to each corner point, and the size ratio between the corner point heat map and the image to be detected, determine The position information of each corner point in the image to be detected.
在一种可能的实施方式中,所述确定部分在被配置为基于所述角点池化后的特征图,确定各个角点对应的向心偏移张量的情况下,包括:In a possible implementation manner, when the determining part is configured to determine the centripetal offset tensor corresponding to each corner point based on the feature map after the corner point pooling, the method includes:
基于所述角点池化后的特征图,确定所述角点池化后的特征图中的每个特征点对应的导向偏移张量,每个特征点对应的导向偏移张量表征由该特征点指向所述待检测图像中的目标对象中心点的偏移张量;Based on the feature map after corner point pooling, the steering offset tensor corresponding to each feature point in the corner point pooling feature map is determined, and the steering offset tensor corresponding to each feature point is represented by The offset tensor of the feature point pointing to the center point of the target object in the image to be detected;
基于每个特征点对应的所述导向偏移张量,确定该特征点的偏移域信息;所述偏移域信息中包含与该特征点关联的多个初始特征点分别指向各自对应的偏移后特征点的偏移张量;Based on the steering offset tensor corresponding to each feature point, determine the offset domain information of the feature point; the offset domain information includes multiple initial feature points associated with the feature point respectively pointing to their corresponding offsets The offset tensor of the feature point after the shift;
基于所述角点池化后的特征图,以及该角点池化后的特征图中的特征点的偏移域信息,对所述角点池化后的特征图中的特征点的特征数据进行调整,得到调整后的特征图;Based on the corner point pooled feature map and the offset domain information of the feature points in the corner point pooled feature map, the feature data of the feature points in the corner point pooled feature map Make adjustments to obtain the adjusted feature map;
基于所述调整后的特征图,确定各个角点对应的向心偏移张量。Based on the adjusted feature map, the centripetal offset tensor corresponding to each corner point is determined.
在一种可能的实施方式中,所述待检测图像对应的角点热力图包括多个通道分别对应的角点热力图,所述多个通道中的每个通道对应一种预设对象类别;所述确定部分在被配置为基于所述角点热力图,确定所述角点热力图中每个特征点作为角点的概率值之后,还被配置为:In a possible implementation manner, the corner heat map corresponding to the image to be detected includes a corner heat map corresponding to multiple channels, and each channel of the multiple channels corresponds to a preset object category; After the determining part is configured to determine the probability value of each feature point in the corner heat map as a corner point based on the corner heat map, it is further configured to:
针对所述多个通道中的每个通道,基于该通道对应的角点热力图中每个特征点作为角点的概率值,确定该通道对应的角点热力图中是否存在所述角点;For each channel of the plurality of channels, determine whether the corner point exists in the corner heat map corresponding to the channel based on the probability value of each feature point in the corner heat map corresponding to the channel as a corner point;
在该通道对应的角点热力图中存在所述角点的情况下,确定所述待检测图像中包含该通道对应的预设对象类别的目标对象。When the corner point exists in the corner point heat map corresponding to the channel, it is determined that the image to be detected contains the target object of the preset object category corresponding to the channel.
在一种可能的实施方式中,所述检测部分被配置为:In a possible implementation manner, the detection part is configured to:
基于各个角点在所述待检测图像中的角点位置信息和各个角点对应的向心偏移张量,确定所述待检测图像中目标对象的检测框。Based on the corner position information of each corner point in the image to be detected and the centripetal offset tensor corresponding to each corner point, the detection frame of the target object in the image to be detected is determined.
在一种可能的实施方式中,所述检测部分在被配置为基于各个角点在所述待检测图像中的角点位置信息和各个角点对应的向心偏移张量,确定所述待检测图像中目标对象的检测框的情况下,包括:In a possible implementation manner, the detection part is configured to determine the to-be-detected image based on the position information of each corner point in the image to be detected and the centripetal offset tensor corresponding to each corner point. In the case of detecting the detection frame of the target object in the image, it includes:
基于各个角点在所述待检测图像中的角点位置信息,筛选能够构成候选检测框的候选角点对;Based on the corner point position information of each corner point in the image to be detected, filter candidate corner point pairs that can form a candidate detection frame;
基于每个候选角点对中每个角点在所述待检测图像中的角点位置信息和该角点对应的向心偏移张量,确定该角点指向的中心点位置信息;Determine the position information of the center point pointed to by the corner point based on the corner position information of each corner point in the image to be detected in each candidate corner point pair and the centripetal offset tensor corresponding to the corner point;
基于每个候选角点对中每个角点在所述待检测图像的角点位置信息,确定该候选角点对所对应的中心区域信息;Determine the central region information corresponding to the candidate corner point pair based on the corner point position information of each corner point in the candidate corner point pair in the image to be detected;
基于每个候选角点对中每个角点指向的中心点位置信息以及该候选角点对所对应的中心区域信息,在所述候选检测框中确定所述目标对象的检测框。The detection frame of the target object is determined in the candidate detection frame based on the position information of the center point pointed to by each corner point in each candidate corner point pair and the center area information corresponding to the candidate corner point pair.
在一种可能的实施方式中,所述检测部分在被配置为所述基于每个候选角点对中每个角点在所述待检测图像中的角点位置信息,确定该候选角点对所对应的中心区域信息的情况下,包括:In a possible implementation manner, the detection part is configured to determine the candidate corner point pair based on the corner point position information of each corner point of each candidate corner point pair in the image to be detected. The corresponding central area information includes:
基于该候选角点对的每个角点的角点位置信息,确定表征该候选角点对所对应的中心区域框的角点位置信息;Based on the corner point position information of each corner point of the candidate corner point pair, determine the corner point position information that characterizes the center area frame corresponding to the candidate corner point pair;
基于所述中心区域框的角点位置信息,确定该候选角点对所对应的中心区域框的坐标范围。Based on the corner point position information of the central area frame, the coordinate range of the central area frame corresponding to the candidate corner point pair is determined.
在一种可能的实施方式中,所述检测部分在被配置为基于每个候选角点对中每个角点指向的中心点位置信息,以及该候选角点对所对应的中心区域信息,在所述候选检测框中确定所述目标对象的检测框的情况下,包括:In a possible implementation manner, the detection part is configured to be based on the position information of the center point pointed to by each corner point in each candidate corner point pair, and the center area information corresponding to the candidate corner point pair, In the case that the detection frame of the target object is determined in the candidate detection frame, the method includes:
基于每个候选角点对中每个角点指向的中心点位置信息,以及该候选角点对所对应的中心区域信息,确定有效候选角点对;Determine a valid candidate corner point pair based on the position information of the center point pointed to by each corner point in each candidate corner point pair, and the center area information corresponding to the candidate corner point pair;
基于所述有效候选角点对中每个角点指向的中心点位置信息、所述有效候选角点对所对应的中心区域信息、以及所述有效候选角点对中每个角点对应的概率值,确定每个有效候选角点对所对应的候选检测框的分值;每个角点对应的概率值用于表示该角点在角点热力图中对应的特征点作为角点的概率值;Based on the position information of the center point pointed to by each corner point in the valid candidate corner point pair, the central area information corresponding to the valid candidate corner point pair, and the probability corresponding to each corner point in the valid candidate corner point pair Value, determine the score of the candidate detection frame corresponding to each valid candidate corner point; the probability value corresponding to each corner point is used to indicate the probability value of the corresponding feature point of the corner point in the corner heat map as the corner point ;
基于每个有效候选角点对所对应的候选检测框的分值、以及相邻候选检测框之间的重叠区域大小,在所述候选检测框中确定所述目标对象的检测框。Based on the score of the candidate detection frame corresponding to each valid candidate corner point and the size of the overlapping area between adjacent candidate detection frames, the detection frame of the target object is determined in the candidate detection frame.
在一种可能的实施方式中,所述检测部分还被配置为:In a possible implementation manner, the detection part is further configured to:
在确定所述待检测图像中目标对象的检测框之后,基于所述目标对象的检测框和对所述待检测图像进行特征提取得到的初始特征图,确定所述待检测图像中所述目标对象的实例信息。After the detection frame of the target object in the image to be detected is determined, the target object in the image to be detected is determined based on the detection frame of the target object and the initial feature map obtained by feature extraction of the image to be detected Instance information.
在一种可能的实施方式中,所述检测部分在被配置为基于所述目标对象的检测框和对所述待检测图像进行特征提取得到的初始特征图,确定所述待检测图像中所述目标对象的实例信息的情况下,包括:In a possible implementation manner, the detection part is configured to determine the image in the image to be detected based on the detection frame of the target object and the initial feature map obtained by feature extraction of the image to be detected In the case of the instance information of the target object, it includes:
基于所述目标对象的检测框以及所述初始特征图,提取所述初始特征图在所述检测框内的特征点的特征数据;Extracting feature data of feature points of the initial feature map in the detection frame based on the detection frame of the target object and the initial feature map;
基于所述初始特征图在所述检测框内的特征点的特征数据,确定所述待检测图像中所述目标对象的实例信息。Based on the feature data of the feature points of the initial feature map in the detection frame, the instance information of the target object in the image to be detected is determined.
在一种可能的实施方式中,所述目标检测装置还包括神经网络训练部分,所述神经网络训练部分被配置为:In a possible implementation manner, the target detection device further includes a neural network training part, and the neural network training part is configured to:
训练用于进行目标检测的神经网络,所述神经网络利用包含了标注目标样本对象的样本图片训练得到。Training a neural network for target detection, the neural network is obtained by training using sample pictures containing labeled target sample objects.
在一种可能的实施方式中,所述神经网络训练部分被配置为按照以下步骤训练所述神经网络:In a possible implementation manner, the neural network training part is configured to train the neural network according to the following steps:
获取样本图像;Obtain sample images;
基于所述样本图像,确定各个样本角点在样本图像中的角点位置信息以及各个样本角点对应的向心偏移张量,所述样本角点表征所述样本图像中的目标样本对象的位置;Based on the sample image, determine the corner position information of each sample corner point in the sample image and the centripetal offset tensor corresponding to each sample corner point, and the sample corner point represents the target sample object in the sample image. Location;
基于各个样本角点在所述样本图像中的角点位置信息及各个样本角点对应的向心偏移张量,预测所述样本图像中的目标样本对象;Predicting the target sample object in the sample image based on the corner position information of each sample corner point in the sample image and the centripetal offset tensor corresponding to each sample corner point;
基于预测的所述样本图像中的目标样本对象和所述样本图像中的标注目标样本对象,对所述神经网络的网络参数值进行调整。The network parameter value of the neural network is adjusted based on the predicted target sample object in the sample image and the labeled target sample object in the sample image.
第三方面,本公开实施例提供了一种电子设备,包括:处理器、存储器和总线,所述存储器存储有所述处理器可执行的机器可读指令,在电子设备运行的情况下,所述处理器与所述存储器之间通过总线通信,所述机器可读指令被所述处理器执行的情况下执行如第一方面所述的目标检测方法的步骤。In a third aspect, an embodiment of the present disclosure provides an electronic device, including a processor, a memory, and a bus. The memory stores machine-readable instructions executable by the processor. When the electronic device is running, the The processor and the memory communicate through a bus, and when the machine-readable instructions are executed by the processor, the steps of the target detection method as described in the first aspect are executed.
第四方面,本公开实施例提供了一种计算机可读存储介质,该计算机可读存储介质上存储有计算机程序,该计算机程序被处理器运行时执行如第一方面所述的目标检测方法的步骤。In a fourth aspect, embodiments of the present disclosure provide a computer-readable storage medium having a computer program stored on the computer-readable storage medium, and the computer program executes the target detection method as described in the first aspect when the computer program is run by a processor. step.
第五方面,本公开实施例提供了一种计算机程序,包括计算机可读代码,在所述计算机可读代码在电子设备中运行的情况下,所述电子设备中的处理器执行时实现如第一方面所述的目标检测方法的步骤。In a fifth aspect, the embodiments of the present disclosure provide a computer program, including computer-readable code. When the computer-readable code runs in an electronic device, the processor in the electronic device executes the following On the one hand, the steps of the target detection method.
为使本公开的上述特征和优点能更明显易懂,下文特举较佳实施例,并配合所附附图,作详细说明如下。In order to make the above-mentioned features and advantages of the present disclosure more comprehensible, preferred embodiments accompanied with accompanying drawings are described in detail below.
附图说明Description of the drawings
为了更清楚地说明本公开实施例的技术方案,下面将对实施例中所需要使用的附图作简单地介绍,此处的附图被并入说明书中并构成本说明书中的一部分,这些附图示出了符合本公开的实施例,并与说明书一起用于说明本公开的技术方案。应当理解,以下附图仅示出了本公开的某些实施例,因此不应被看作是对范围的限定,对于本领域普通技术人员来讲,在不付出创造性劳动的前提下,还可以根据这些附图获得其他相关的附图。In order to explain the technical solutions of the embodiments of the present disclosure more clearly, the following will briefly introduce the drawings that need to be used in the embodiments. The drawings here are incorporated into the specification and constitute a part of the specification. The figure shows an embodiment conforming to the present disclosure, and is used together with the description to explain the technical solution of the present disclosure. It should be understood that the following drawings only show certain embodiments of the present disclosure, and therefore should not be regarded as limiting the scope. For those of ordinary skill in the art, they can also Obtain other related drawings based on these drawings.
图1示出了一种针对待检测图像进行检测时得到的结果示意图;Figure 1 shows a schematic diagram of a result obtained when detecting an image to be detected;
图2示出了本公开实施例所提供的示例性的一种目标检测方法的流程流程图;Fig. 2 shows a flow chart of an exemplary target detection method provided by an embodiment of the present disclosure;
图3示出了本公开实施例所提供的一种确定角点位置信息以及角点对应的向心偏移张量的流程流程图;FIG. 3 shows a flowchart of a process for determining the position information of corner points and the centripetal offset tensor corresponding to the corner points provided by an embodiment of the present disclosure;
图4示出了本公开实施例所提供的一种确定角点位置信息以及角点对应的向心偏移张量的流程图;FIG. 4 shows a flowchart for determining the position information of a corner point and the centripetal offset tensor corresponding to the corner point provided by an embodiment of the present disclosure;
图5示出了本公开实施例所提供的一种确定角点对应的向心偏移张量的流程图;FIG. 5 shows a flow chart of determining the centripetal offset tensor corresponding to a corner point provided by an embodiment of the present disclosure;
图6示出了本公开实施例所提供的示例性的一种特征调整网络针对角点池化后的特征图进行调节的流程示意图;FIG. 6 shows a schematic flow chart of an exemplary feature adjustment network provided by an embodiment of the present disclosure for adjusting a feature map after corner pooling;
图7示出了本公开实施例所提供的一种确定目标对象的类别的流程示意图;FIG. 7 shows a schematic diagram of a process for determining the category of a target object provided by an embodiment of the present disclosure;
图8示出了本公开实施例所提供的一种确定目标对象的检测框的流程示意图;FIG. 8 shows a schematic flow chart of determining a detection frame of a target object provided by an embodiment of the present disclosure;
图9示出了本公开实施例所提供的基于每个候选角点对确定目标对象的检测框的流程示意图;FIG. 9 shows a schematic flowchart of determining a detection frame of a target object based on each candidate corner point pair provided by an embodiment of the present disclosure;
图10示出了本公开实施例所提供的示例性的一种目标检测方法对应的流程示意图;FIG. 10 shows a schematic flowchart corresponding to an exemplary target detection method provided by an embodiment of the present disclosure; FIG.
图11示出了本公开实施例所提供的一种神经网络的训练方法流程示意图;FIG. 11 shows a schematic flowchart of a neural network training method provided by an embodiment of the present disclosure;
图12示出了本公开实施例所提供的一种目标检测装置的结构示意图;FIG. 12 shows a schematic structural diagram of a target detection device provided by an embodiment of the present disclosure;
图13示出了本公开实施例所提供的一种电子设备的结构示意图。FIG. 13 shows a schematic structural diagram of an electronic device provided by an embodiment of the present disclosure.
具体实施方式Detailed ways
为使本公开实施例的技术方案和优点更加清楚,下面将结合本公开实施例中附图,对本公开实施例中的技术方案进行清楚、完整地描述,显然,所描述的实施例仅仅是本公开一部分实施例,而不是全部的实施例。通常在此处附图中描述和示出的本公开实施例的组件可以以各种不同的配置来布置和设计。因此,以下对在附图中提供的本公开的实施例的详细描述并非旨在限制要求保护的本公开的范围,而是仅仅表示本公开的选定实施例。基于本公开的实施例,本领域技术人员在没有做出创造性劳动的前提下所获得的所有其他实施例,都属于本公开实施例保护的范围。In order to make the technical solutions and advantages of the embodiments of the present disclosure clearer, the technical solutions in the embodiments of the present disclosure will be clearly and completely described below in conjunction with the accompanying drawings in the embodiments of the present disclosure. Obviously, the described embodiments are only the present invention. A part of the embodiments is disclosed, but not all of the embodiments are disclosed. The components of the embodiments of the present disclosure generally described and illustrated in the drawings herein may be arranged and designed in various different configurations. Therefore, the following detailed description of the embodiments of the present disclosure provided in the accompanying drawings is not intended to limit the scope of the claimed present disclosure, but merely represents selected embodiments of the present disclosure. Based on the embodiments of the present disclosure, all other embodiments obtained by those skilled in the art without creative work shall fall within the protection scope of the embodiments of the present disclosure.
在针对图像进行目标检测的情况下,在该图像中包含多个相似的目标对象的情况下,如图1所示,在该图像中存在多个相似的飞机的情况下,在采用基于飞机的关键点对图像中的飞机进行检测的情况下,容易出现图1中检测框(1)和检测框(2)所示的情况,即同一个检测框中包含多个飞机的情况,即检测出现失误,即目前在对图像中的目标对象进行检测的情况下,检测结果的准确度较低,针对此,本公开实施例提供了一种目标检测方法,能够提高检测结果的准确度。In the case of target detection for an image, in the case that the image contains multiple similar target objects, as shown in Figure 1, in the case of multiple similar aircraft in the image, the use of aircraft-based When the key point detects the aircraft in the image, the situation shown in the detection frame (1) and the detection frame (2) in Figure 1 is easy to occur, that is, the same detection frame contains multiple aircraft, that is, the detection occurs Mistake, that is, in the current situation of detecting the target object in the image, the accuracy of the detection result is low. For this, the embodiment of the present disclosure provides a target detection method, which can improve the accuracy of the detection result.
基于上述研究,本公开实施例提供了一种目标检测方法,在获取到待检测图像后,首先确定各个角点在待检测图像中的角点位置信息以及各个角点对应的向心偏移张量,因为角点是指图像中的主要特征点,其在待检测图像中的角点位置信息能够表征每个目标对象在待检测图像中的位置,比如角点可以包括左上角点和右下角点,其中左上角点是指对应目标对象上侧轮廓的直线与对应目标对象左侧轮廓的直线的交点,右下角点是指对应目标对象下侧轮廓的直线与对应目标对象右侧轮廓的直线的交点,在左上角点和右下角点属于同一个目标对象的检测框的情况下,左上角点和右下角点分别对应的向心偏移张量指向的位置应该比较接近,因此,本公开实施例提出的目标检测方法,基于各个角点在待检测图像中的角点位置信息及各个角点对应的向心偏移张量,能够确定出属于同一目标对象的角点,进而基于确定出的角点可以检测出该同一目标对象。Based on the above research, the embodiments of the present disclosure provide a target detection method. After acquiring the image to be detected, first determine the corner position information of each corner point in the image to be detected and the centripetal offset corresponding to each corner point. Because the corner point refers to the main feature point in the image, the position information of the corner point in the image to be detected can characterize the position of each target object in the image to be detected. For example, the corner point can include the upper left corner and the lower right corner. Point, where the upper left point refers to the intersection of the line corresponding to the upper profile of the target object and the line corresponding to the left profile of the target object, and the lower right point refers to the line corresponding to the lower profile of the target object and the line corresponding to the right profile of the target object In the case where the upper left and lower right corners belong to the detection frame of the same target object, the centripetal offset tensor corresponding to the upper left and lower right points should be relatively close to each other. Therefore, the present disclosure The target detection method proposed in the embodiment can determine the corner points belonging to the same target object based on the corner position information of each corner point in the image to be detected and the centripetal offset tensor corresponding to each corner point. The corner point of can detect the same target object.
下面将结合本公开实施例中的附图,对本公开实施例中的技术方案进行清楚、完整地描述,显然,所描述的实施例仅仅是本公开一部分实施例,而不是全部的实施例。通常在此处附图中描述和示出的本公开实施例的组件可以以各种不同的配置来布置和设计。因此,以下对在附图中提供的本公开的实施例的详细描述并非旨在限制要求保护的本公开实施例的范围,而是仅仅表示本公开的选定实施例。基于本公开的实施例,本领域技术人员在没有做出创造性劳动的前提下所获得的所有其他实施例,都属于本公开实施例保护的范围。The technical solutions in the embodiments of the present disclosure will be clearly and completely described below in conjunction with the accompanying drawings in the embodiments of the present disclosure. Obviously, the described embodiments are only a part of the embodiments of the present disclosure, rather than all the embodiments. The components of the embodiments of the present disclosure generally described and illustrated in the drawings herein may be arranged and designed in various different configurations. Therefore, the following detailed description of the embodiments of the present disclosure provided in the accompanying drawings is not intended to limit the scope of the claimed embodiments of the present disclosure, but merely represents selected embodiments of the present disclosure. Based on the embodiments of the present disclosure, all other embodiments obtained by those skilled in the art without creative work shall fall within the protection scope of the embodiments of the present disclosure.
应注意到:相似的标号和字母在下面的附图中表示类似项,因此,一旦某一项在一个附图中被定义,则在随后的附图中不需要对其进行定义和解释。It should be noted that similar reference numerals and letters indicate similar items in the following drawings. Therefore, once a certain item is defined in one drawing, it does not need to be defined and explained in the subsequent drawings.
为便于对本实施例进行理解,首先对本公开实施例所公开的一种目标检测方法进行详细介绍,本公开实施例所提供的目标检测方法的执行主体一般为具有一定计算能力的计算机设备,该计算机设备例如包括:终端设备或服务器或其它处理设备。在一些可能的实现方式中,该目标检测方法可以通过处理器调用存储器中存储的计算机可读指令的方式来实现。In order to facilitate the understanding of this embodiment, a target detection method disclosed in the embodiment of the present disclosure is first introduced in detail. The execution subject of the target detection method provided in the embodiment of the present disclosure is generally a computer device with a certain computing capability. The equipment includes, for example, terminal equipment or servers or other processing equipment. In some possible implementations, the target detection method can be implemented by a processor invoking a computer-readable instruction stored in a memory.
参见图2所示,为本公开实施例提供的目标检测方法的流程图,所述方法包括步骤S201~S203,步骤如下:Referring to FIG. 2, which is a flowchart of a target detection method provided by an embodiment of the present disclosure, the method includes steps S201 to S203, and the steps are as follows:
S201,获取待检测图像。S201: Acquire an image to be detected.
这里的待检测图像可以是针对特定环境下的待检测图像,比如针对某交通路口的车辆进行检测,可 以在该交通路口安装摄像机,通过摄像机采集该交通路口在一定时间段内的视频流,然后对该视频流进行分帧处理,得到待检测图像;或者针对某动物园中的动物进行检测,可以在该动物园内安装摄像机,通过摄像机采集该动物园在一定时间段内的视频流,然后对该视频流进行分帧处理,得到待检测图像。The image to be detected here can be an image to be detected in a specific environment. For example, to detect a vehicle at a certain traffic intersection, a camera can be installed at the traffic intersection, and the video stream of the traffic intersection in a certain period of time can be collected by the camera, and then Frame the video stream to obtain the image to be detected; or to detect animals in a zoo, a camera can be installed in the zoo, and the video stream of the zoo in a certain period of time can be collected by the camera, and then the video The stream undergoes framing processing to obtain the image to be detected.
这里,待检测图像中可以包含目标对象,这里的目标对象是指在特定环境下要检测的对象,比如上文提到的某交通路口的车辆,以及某动物园中的动物,也可以不包含目标对象,在不包括目标对象的情况下,检测结果即为空,本公开实施针对包含目标对象的待检测图像进行说明。Here, the image to be detected can contain the target object. The target object here refers to the object to be detected in a specific environment, such as a vehicle at a traffic intersection mentioned above, and an animal in a zoo, or it may not contain the target. Object, if the target object is not included, the detection result is empty, and the implementation of the present disclosure will describe the image to be detected that contains the target object.
S202,基于待检测图像,确定各个角点在待检测图像中的角点位置信息以及各个角点对应的向心偏移张量,其中,角点表征待检测图像中的目标对象的位置。S202: Based on the image to be detected, determine the corner position information of each corner point in the image to be detected and the centripetal offset tensor corresponding to each corner point, where the corner point represents the position of the target object in the image to be detected.
目标对象在待检测图像中的位置可以通过检测框来表示,本公开实施例通过角点来表征目标对象在待检测图像中的位置,即这里的角点可以为检测框的角点,比如可以通过左上角点和右下角点来表征目标对象在待检测图像中的位置,左上角点即为检测框的左上角点,右下角点即为检测框的右下角点,其中,左上角点是指对应目标对象上侧轮廓的直线与对应目标对象左侧轮廓的直线的交点,右下角点是指对应目标对象下侧轮廓的直线与对应目标对象右侧轮廓的直线的交点。The position of the target object in the image to be detected can be represented by a detection frame. The embodiment of the present disclosure uses corner points to characterize the position of the target object in the image to be detected, that is, the corner points here may be the corner points of the detection frame, for example, The position of the target object in the image to be detected is characterized by the upper left corner point and the lower right corner point. The upper left corner point is the upper left corner point of the detection frame, and the lower right corner point is the lower right corner point of the detection frame, where the upper left corner point is Refers to the intersection of the line corresponding to the upper contour of the target object and the line corresponding to the left contour of the target object. The lower right corner point refers to the intersection of the line corresponding to the lower contour of the target object and the line corresponding to the right contour of the target object.
当然,并不局限于通过左上角点和右下角点来表征目标对象的位置,还可以通过右上角点和左下角点来表征目标对象的位置,本公开实施例以左上角点和右下角点为例进行说明。Of course, the position of the target object is not limited to the upper left corner point and the lower right corner point. The position of the target object can also be characterized by the upper right corner point and the lower left corner point. The embodiment of the present disclosure uses the upper left corner point and the lower right corner point. Take an example for illustration.
这里的向心偏移张量是指角点向目标对象中心位置的偏移张量,因为待检测图像是二维图像,则这里的向心偏移张量包括向两个方向上的偏移值,在这两个方向分别为X轴方向和Y轴方向的情况下,该向心偏移张量包括在X轴方向上的偏移值,以及在Y轴方向上的偏移值。通过角点和该角点对应的向心偏移张量,可以确定该角点指向的中心位置,在左上角点和右下角点位于同一个检测框的情况下,其指向的中心位置应该相同,或者比较接近,故可以基于每个角点对应的向心偏移张量,确定出属于同一目标对象的角点,进而基于确定出的角点可以确定目标对象的检测框。The centripetal offset tensor here refers to the offset tensor from the corner point to the center position of the target object. Because the image to be detected is a two-dimensional image, the centripetal offset tensor here includes the offset in two directions. When the two directions are the X-axis direction and the Y-axis direction, the centripetal offset tensor includes the offset value in the X-axis direction and the offset value in the Y-axis direction. Through the centripetal offset tensor corresponding to the corner point and the corner point, the center position of the corner point can be determined. In the case where the upper left corner point and the lower right corner point are in the same detection frame, the center position of the point should be the same , Or relatively close, so the corner points belonging to the same target object can be determined based on the centripetal offset tensor corresponding to each corner point, and then the detection frame of the target object can be determined based on the determined corner points.
本公开实施例通过神经网络来确定角点以及角点对应的向心偏移张量,将结合下文实施例进行阐述。The embodiment of the present disclosure uses a neural network to determine the corner point and the centripetal offset tensor corresponding to the corner point, which will be described in conjunction with the following embodiments.
S203,基于各个角点在待检测图像中的角点位置信息及各个角点对应的向心偏移张量,确定待检测图像中的目标对象。S203: Determine a target object in the image to be detected based on the corner position information of each corner point in the image to be detected and the centripetal offset tensor corresponding to each corner point.
其中,各个角点在待检测图像中的角点位置信息是指多个角点中各个角点在待检测图像中的角点位置信息,各个角点对应的向心偏移张量是指该多个角点中各个角点分别对应的向心偏移张量。Among them, the corner position information of each corner in the image to be detected refers to the corner position information of each of the multiple corner points in the image to be detected, and the centripetal offset tensor corresponding to each corner refers to the Each of the multiple corner points corresponds to the centripetal offset tensor.
这里在待检测图像中检测出目标对象,可以包含检测出目标对象的位置,比如确定待检测图像中目标对象的检测框,也可以确定待检测图像中目标对象的实例信息,或者同时确定待检测图像中目标对象的检测框和实例信息,如何确定待检测图像中的目标对象,将在后文进行详细解释。The detection of the target object in the image to be detected here can include the location of the detected target object, such as determining the detection frame of the target object in the image to be detected, or determining the instance information of the target object in the image to be detected, or at the same time determining the target object to be detected The detection frame and instance information of the target object in the image, and how to determine the target object in the image to be detected will be explained in detail later.
上述步骤S201~S203提出的目标检测方法,在获取到待检测图像后,首先确定各个角点在待检测图像中的角点位置信息,以及各个角点对应的向心偏移张量,因为角点是指图像中的主要特征点,其在待检测图像中的角点位置信息能够表征每个目标对象在待检测图像中的位置,比如角点可以包括左上角点和右下角点,其中左上角点是指对应目标对象上侧轮廓的直线与对应目标对象左侧轮廓的直线的交点,右下角点是指对应目标对象下侧轮廓的直线与对应目标对象右侧轮廓的直线的交点,在左上角点和右下角点属于同一个目标对象的检测框的情况下,左上角点和右下角点分别对应的向心偏移张量指向的位置应该比较接近,因此,本公开实施例提出的目标检测方法,基于各个角点在待检测图像中的角点位置信息及各个角点对应的向心偏移张量,能够确定出属于同一目标对象的角点,进而基于确定出的角点可以检测出该同一目标对象。The target detection method proposed in the above steps S201 to S203, after acquiring the image to be detected, first determine the corner position information of each corner point in the image to be detected, and the centripetal offset tensor corresponding to each corner point, because the angle Point refers to the main feature point in the image. The position information of the corner point in the image to be detected can characterize the position of each target object in the image to be detected. For example, the corner point can include the upper left corner point and the lower right corner point, where the upper left corner The corner point refers to the intersection point of the line corresponding to the upper contour of the target object and the line corresponding to the left contour of the target object. The lower right corner point refers to the intersection point of the line corresponding to the lower contour of the target object and the line corresponding to the right contour of the target object. In the case where the upper left corner point and the lower right corner point belong to the detection frame of the same target object, the positions of the centripetal offset tensor corresponding to the upper left corner point and the lower right corner point should be relatively close. Therefore, the embodiment of the present disclosure proposes The target detection method can determine the corner points belonging to the same target object based on the corner position information of each corner point in the image to be detected and the centripetal offset tensor corresponding to each corner point, and then based on the determined corner point The same target object is detected.
下面将对上述S201~S203继续进行说明。The description of S201 to S203 above will be continued below.
针对上述S202,在一种实施方式中,在基于待检测图像,确定角点在待检测图像中的角点位置信息以及各个角点对应的向心偏移张量的情况下,如图3所示,可以包括以下步骤S301~S303:Regarding the above S202, in an embodiment, based on the image to be detected, the corner position information of the corner point in the image to be detected and the centripetal offset tensor corresponding to each corner point are determined, as shown in FIG. 3 It may include the following steps S301 to S303:
S301,对待检测图像进行特征提取,得到待检测图像对应的初始特征图;S301: Perform feature extraction on the image to be detected to obtain an initial feature map corresponding to the image to be detected;
S301,对初始特征图进行角点池化处理,得到角点池化后的特征图;S301: Perform corner pooling processing on the initial feature map to obtain a feature map after corner pooling;
S303,基于角点池化后的特征图,确定各个角点在待检测图像中的角点位置信息,以及各个角点对应的向心偏移张量。S303: Based on the feature map after corner pooling, determine the corner position information of each corner in the image to be detected, and the centripetal offset tensor corresponding to each corner.
这里待检测图像的尺寸是一定的,比如尺寸为H*W,这里的H和W分别表示待检测图像中长和宽方向上的像素值,然后将该待检测图像输入预先训练的沙漏卷积神经网络进行特征提取,比如进行纹理特征提取、颜色特征提取、边缘特征提取等,即可以得到该待检测图像对应的初始特征图。Here the size of the image to be detected is fixed, for example, the size is H*W, where H and W represent the pixel values in the length and width directions of the image to be detected respectively, and then input the image to be detected into the pre-trained hourglass convolution The neural network performs feature extraction, such as texture feature extraction, color feature extraction, edge feature extraction, etc., and the initial feature map corresponding to the image to be detected can be obtained.
当然,因为沙漏卷积神经网络的输入端对接收的图像尺寸是有要求的,即接收设定尺寸的待检测图像,在待检测图像的尺寸不符合该设定尺寸的情况下,需要先对待检测图像的尺寸进行调节,然后再将尺寸调节后的待检测图像输入该沙漏卷积神经网络,进行特征提取以及尺寸压缩,即可以得到大小为h*w*c的初始特征图,这里的c表示初始特征图的通道个数,h和w表示每个通道上的初始特征图的尺 寸。Of course, because the input end of the hourglass convolutional neural network has requirements for the received image size, that is, it receives a set size of the image to be inspected. If the size of the image to be inspected does not meet the set size, it needs to be treated first. Adjust the size of the detected image, and then input the adjusted image to be detected into the hourglass convolutional neural network for feature extraction and size compression, that is, an initial feature map with a size of h*w*c can be obtained, where c Indicates the number of channels of the initial feature map, and h and w represent the size of the initial feature map on each channel.
初始特征图中包含多个特征点,每个特征点都具有特征数据,这些特征数据能够表示待检测图像的全局信息,为了便于在这些特征点中提取角点,本公开实施例提出对初始特征图进行(Corner Pooling)角点池化处理,得到角点池化后的特征图,角点池化后的特征图相比初始特征图,增强了角点所蕴含的目标对象语义信息,因此基于角点池化后的特征图,能够更准确地确定出在待检测图像中每个角点的角点位置信息,以及每个角点对应的向心偏移张量。The initial feature map contains multiple feature points, and each feature point has feature data. These feature data can represent the global information of the image to be detected. In order to facilitate the extraction of corner points from these feature points, the embodiment of the present disclosure proposes to The graph performs corner pooling processing to obtain the feature map after corner pooling. Compared with the initial feature map, the feature map after corner pooling enhances the semantic information of the target object contained in the corner points, so it is based on The feature map after corner pooling can more accurately determine the corner position information of each corner point in the image to be detected, and the centripetal offset tensor corresponding to each corner point.
这里,通过对待检测图像进行特征提取,得到初始特征图,并对初始特征图进行角点池化处理,得到能够便于提取角点以及角点对应的向心偏移量的特征图,即角点池化后的特征图。Here, by performing feature extraction on the image to be detected, an initial feature map is obtained, and corner pooling is performed on the initial feature map to obtain a feature map that can facilitate the extraction of the corner points and the centripetal offset corresponding to the corner points, that is, the corner points Feature map after pooling.
在得到角点池化后的特征图后,可以根据该角点池化后的特征图以及预先训练的神经网络,来确定是否存在角点,在存在角点的情况下,确定各个角点在待检测图像中的角点位置信息,本公开实施例通过左上角点和右下角点来表征目标对象在待检测图像中的位置,即确定每个角点在待检测图像中的角点位置信息的过程,可以是确定左上角点在待检测图像中的角点位置信息,以及右下角点在待检测图像中的角点位置信息的过程,其中,可以通过左上角点检测网络来检测左上角点在待检测图像中的角点位置信息,以及通过右下角点检测网络来检测右下角点在待检测图像中的角点位置信息,因为左上角点在待检测图像中的角点位置信息和右下角点在待检测图像中的角点位置信息的确定方式相似,本公开实施例以确定左上角点在待检测图像中的角点位置信息为例进行详细说明。After the corner point pooled feature map is obtained, the corner point pooling feature map and the pre-trained neural network can be used to determine whether there is a corner point. If there is a corner point, determine whether each corner point is The position information of the corner points in the image to be detected. In the embodiments of the present disclosure, the position of the target object in the image to be detected is characterized by the upper left corner point and the lower right corner point, that is, the corner point position information of each corner point in the image to be detected is determined The process can be the process of determining the corner position information of the upper left corner point in the image to be detected, and the process of determining the corner position information of the lower right corner point in the image to be detected, where the upper left corner can be detected through the upper left corner point detection network The corner position information of the point in the image to be detected, and the corner position information of the lower right corner point in the image to be detected through the lower right corner point detection network, because the corner position information of the upper left corner point in the image to be detected and The method for determining the corner position information of the lower right corner point in the image to be detected is similar, and the embodiment of the present disclosure determines the corner position information of the upper left corner point in the image to be detected as an example for detailed description.
在一种实施方式中,左上角点检测网络可以包括左上角点热力图预测网络和左上角点局部偏移预测网络,在基于角点池化后的特征图,确定各个角点在待检测图像中的角点位置信息的情况下,如图4所示,可以包括以下步骤S401~S404:In an embodiment, the upper left corner point detection network may include the upper left corner point heat map prediction network and the upper left corner point local offset prediction network. Based on the feature map after corner point pooling, it is determined that each corner point is in the image to be detected. In the case of the corner position information in, as shown in FIG. 4, the following steps S401 to S404 may be included:
S401,基于角点池化后的特征图,生成待检测图像对应的角点热力图。S401: Based on the feature map after corner pooling, generate a corner heat map corresponding to the image to be detected.
在预测左上角点在待检测图像中的角点位置信息的情况下,这里的角点热力图可以通过左上角点检测网络中的左上角点热力图预测网络得到,将角点池化后的特征图输入左上角点热力图预测网络,即可以得到待检测图像对应的左上角点热力图,该左上角点热力图中包含多个特征点,每个特征点具有该特征点对应的特征数据,基于角点热力图中特征点的特征数据,能够确定该特征点作为左上角点的概率值。In the case of predicting the position information of the corner point of the upper left corner point in the image to be detected, the corner point heat map here can be obtained by the upper left corner point heat map prediction network in the upper left corner point detection network, and the corner point pooled The feature map is input into the upper left corner point heat map prediction network, and the upper left corner point heat map corresponding to the image to be detected can be obtained. The upper left corner point heat map contains multiple feature points, and each feature point has the feature data corresponding to the feature point. , Based on the feature data of the feature point in the corner point heat map, the probability value of the feature point as the upper left corner point can be determined.
这里得到的左上角点热力图大小为h*w*m,其中h*w表示角点热力图在每个通道上的尺寸,m表示预设通道数量,每个预设通道对应一种预设对象类别,比如m=40,即表示有40种预设对象类别,该左上角点热力图除了能够用于确定待检测图像中的左上角点,还可以用于确定该左上角点在待检测图像中表征的目标对象的类别,如何确定目标对象的类别的过程将在后文进行详细解释。The size of the upper left corner heat map obtained here is h*w*m, where h*w represents the size of the corner heat map on each channel, m represents the number of preset channels, and each preset channel corresponds to a preset Object category, such as m=40, which means there are 40 preset object categories. The upper left corner heat map can be used to determine the upper left corner point in the image to be detected, and can also be used to determine that the upper left corner point is in the image to be detected. The category of the target object represented in the image, and the process of how to determine the category of the target object will be explained in detail later.
S402,基于角点热力图,确定角点热力图中每个特征点作为角点的概率值,并基于每个特征点作为角点的概率值,从角点热力图的特征点中筛选出角点。S402, based on the corner heat map, determine the probability value of each feature point in the corner heat map as a corner point, and filter out the corners from the feature points of the corner heat map based on the probability value of each feature point as a corner point point.
基于每个特征点作为左上角点的概率值,可以确定该特征点是左上角点的概率,从而将概率值大于设定阈值的特征点作为左上角点。Based on the probability value of each feature point as the upper left corner point, the probability that the feature point is the upper left corner point can be determined, so that the feature point with the probability value greater than the set threshold is taken as the upper left corner point.
S403,获取筛选出的各个角点在角点热力图中的位置信息、以及各个角点对应的局部偏移信息。S403: Obtain the position information of the selected corner points in the corner point heat map and the local offset information corresponding to each corner point.
局部偏移信息用于表示对应的角点所表征的真实物理点在角点热力图中的位置偏移信息,其中,每个左上角点对应的局部偏移信息用于表示该左上角点所表征的真实物理点在左上角点热力图中的位置偏移信息。The local offset information is used to indicate the position offset information of the real physical point represented by the corresponding corner point in the corner heat map. The local offset information corresponding to each upper left corner point is used to indicate the position of the upper left corner point. The position offset information of the characterization of the real physical point in the upper left corner of the heat map.
这里的局部偏移信息可以通过局部偏移张量来表示,该局部偏移张量在左上角点热力图中同样可以表示两个方向上的偏移值,比如左上角点热力图中的坐标系包括两个方向,分别为x轴方向和y轴方向,则该局部偏移张量包括在x轴方向上的偏移值,以及在y轴方向上的偏移值。The local offset information here can be represented by a local offset tensor. The local offset tensor can also represent the offset values in two directions in the upper left corner heat map, such as the coordinates in the upper left corner heat map The system includes two directions, namely the x-axis direction and the y-axis direction. The local offset tensor includes the offset value in the x-axis direction and the offset value in the y-axis direction.
基于左上角点热力图对应的坐标系,可以得到左上角点热力图中每个特征点在该左上角点热力图中的位置信息,考虑到得到的左上角点的位置信息与该左上角点表征的真实物理点的位置信息之间可能存在误差,比如,通过对左上角点热力图进行位置检测得到某个左上角点的位置信息,而该左上角点表征的真实物理点的位置信息与检测到的该左上角点的位置信息该具有一定的偏差,则该局部偏移信息用于表示该偏差。Based on the coordinate system corresponding to the upper left corner point heat map, the position information of each feature point in the upper left corner point heat map in the upper left corner point heat map can be obtained, taking into account the obtained position information of the upper left corner point and the upper left corner point There may be errors between the position information of the real physical points represented. For example, the position information of a certain upper left corner point can be obtained by detecting the position of the upper left corner point heat map, and the position information of the real physical point represented by the upper left corner point is different from the position information of the real physical point. The detected position information of the upper left corner point should have a certain deviation, and the local offset information is used to indicate the deviation.
为了提高目标对象检测的准确度,这里可以通过引入提前训练的左上角点局部偏移预测网络,然后将池化后的特征图输入左上角点预测网络中的左上角点局部偏移预测网络,确定出左上角点热力图中每个特征点对应的局部偏移信息,然后基于该局部偏移信息对特征点在角点热力图中的位置信息进行修正,然后再基于修正后的位置信息,确定左上角点在待检测图像中的角点位置信息。In order to improve the accuracy of target object detection, we can introduce the pre-trained upper left corner point local offset prediction network, and then input the pooled feature map into the upper left corner point local offset prediction network in the upper left corner point prediction network. Determine the local offset information corresponding to each feature point in the upper left corner heat map, and then modify the position information of the feature point in the corner heat map based on the local offset information, and then based on the corrected position information, Determine the corner position information of the upper left corner point in the image to be detected.
S404,基于获取到的各个角点在角点热力图中的位置信息、各个角点对应的局部偏移信息、以及角点热力图和待检测图像之间的尺寸比例,确定各个角点在待检测图像中的角点位置信息。S404: Based on the acquired position information of each corner point in the corner point heat map, the local offset information corresponding to each corner point, and the size ratio between the corner point heat map and the image to be detected, it is determined that each corner point is in the to-be-detected image. Detect the position information of the corner points in the image.
这里,获取到的每个左上角点在左上角点热力图中的位置信息可以包括在左上角点热力图中x轴方向上的坐标值x,以及y轴方向上的坐标值y,待检测图像中每个左上角点的角点位置信息可以包括X 轴方向上的坐标值X,以及Y轴方向上的坐标值Y。Here, the acquired position information of each upper left corner point in the upper left corner point heat map may include the coordinate value x in the x-axis direction in the upper left corner point heat map, and the coordinate value y in the y-axis direction, to be detected The corner position information of each upper left corner point in the image may include the coordinate value X in the X-axis direction and the coordinate value Y in the Y-axis direction.
这里,可以根据以下公式(1)和公式(2)来确定第i个左上角点在待检测图像中的角点位置信息:Here, the corner position information of the i-th upper left corner point in the image to be detected can be determined according to the following formula (1) and formula (2):
tl x(i)=n*(x l(i)lx(i));          (1); tl x(i) =n*(x l(i)lx(i) ); (1);
tl y(i)=n*(y l(i)ly(i));           (2); tl y(i) = n*(y l(i)ly(i) ); (2);
其中,tl x(i)表示第i个左上角点在待检测图像中X轴方向上的坐标值,tl y(i)表示第i个左上角点在待检测图像中Y轴方向上的坐标值;n表示左上角点热力图和待检测图像之间的尺寸比例;x l(i)表示第i个左上角点在左上角点热力图中x轴方向上的坐标值,y l(i)表示第i个左上角点在角点热力图中y轴方向上的坐标值;σ lx(i)表示第i个左上角点所表征的真实物理点在角点热力图中x轴方向上的偏移值,σ ly(i)表示第i个左上角点所表征的真实物理点在角点热力图中y轴方向上的偏移值。 Among them, tl x(i) represents the coordinate value of the i-th upper left corner point in the X-axis direction of the image to be detected, and tl y(i) represents the coordinate value of the i-th upper left corner point in the Y-axis direction of the image to be detected Value; n represents the size ratio between the upper left corner point heat map and the image to be detected; x l(i) represents the coordinate value of the i-th upper left corner point in the x-axis direction of the upper left corner point heat map, y l(i ) Represents the coordinate value of the i-th upper left corner point in the y-axis direction of the corner heat map; σ lx(i) represents the real physical point represented by the i-th upper left corner point in the x-axis direction of the corner heat map The offset value of σ ly(i) represents the offset value of the real physical point represented by the i-th upper left corner point in the y-axis direction of the corner heat map.
以上过程为左上角点在待检测图像中的角点位置信息的确定过程,右下角点在待检测图像中的角点位置信息的确定过程同理,即将角点池化后的特征图输入右下角点预测网络中的右下角点热力图预测网络,得到右下角点热力图,进而确定右下角点热力图中每个特征点作为右下角点的概率值,从中筛选出右下角点,同时结合通过右下角点预测网络中的右下角点局部偏移网络确定的右下角点对应的局部偏移信息,确定每个右下角点在待检测图像中的角点位置信息,在此不进行赘述。The above process is the process of determining the corner position information of the upper left corner point in the image to be detected, and the process of determining the corner point position information of the lower right corner point in the image to be detected is the same, that is, the feature map after the corner point pooling is input to the right The lower right corner point heat map prediction network in the lower corner point prediction network obtains the lower right corner point heat map, and then determines the probability value of each feature point in the lower right corner point heat map as the lower right corner point, selects the lower right corner point from it, and combines them at the same time The position information of the corner point of each lower right corner point in the image to be detected is determined by the local offset information corresponding to the lower right corner point determined by the lower right corner point local offset network in the lower right corner point prediction network, which will not be repeated here.
同样,可以根据以下公式(3)和公式(4)来确定第j个右下角点在待检测图像中的角点位置信息:Similarly, the corner position information of the j-th lower right corner point in the image to be detected can be determined according to the following formula (3) and formula (4):
br x(j)=n*(x r(j)rx(j));           (3); br x(j) =n*(x r(j)rx(j) ); (3);
br y(j)=n*(y r(j)ry(j));         (4); br y(j) = n*(y r(j)ry(j) ); (4);
其中,br x(j)表示第j个右下角点在待检测图像中X轴方向上的坐标值,br y(j)表示第j个右下角点在待检测图像中Y轴方向上的坐标值;n表示右下角点热力图和待检测图像之间的尺寸比例;x r(j)表示第j个右下角点在右下角点热力图中x轴方向上的坐标值,y r(j)表示第j个右下角点在角点热力图中y轴方向上的坐标值;σ rx(j)表示第j个右下角点所表征的真实物理点在角点热力图中x轴方向上的偏移值,σ ry(j)表示第j个右下角点所表征的真实物理点在角点热力图中y轴方向上的偏移值。 Among them, br x(j) represents the coordinate value of the j-th lower right corner point in the X-axis direction of the image to be detected, and br y(j) represents the coordinate value of the j-th lower right corner point in the Y-axis direction of the image to be detected Value; n represents the size ratio between the lower right corner point heat map and the image to be detected; x r(j) represents the coordinate value of the j-th lower right corner point in the x-axis direction of the lower right corner point heat map, y r(j ) Represents the coordinate value of the j-th lower right corner point in the y-axis direction of the corner heat map; σ rx(j) represents the real physical point represented by the j-th lower right corner point in the x-axis direction of the corner heat map The offset value of σ ry(j) represents the offset value of the real physical point represented by the j-th lower right corner point in the y-axis direction of the corner heat map.
以上步骤S401~S404是本公开实施例提供的一种确定各个角点在待检测图像中的角点位置信息的方式,该过程通过引入角点热力图,通过每个特征点作为角点的概率值确定出可以作为角点的特征点,在选择出角点后,通过对角点在角点热力图中的位置信息进行修正后,确定出角点在待检测图像中的角点位置信息,该方式能够得到准确度较高的角点的角点位置信息,从而便于后续基于该角点检测目标对象在待检测图像中的位置。The above steps S401 to S404 are a way to determine the corner position information of each corner in the image to be detected according to the embodiment of the present disclosure. This process introduces a corner heat map and passes each feature point as the probability of a corner point. The value determines the feature point that can be used as the corner point. After the corner point is selected, the position information of the corner point in the corner point heat map is corrected to determine the corner point position information of the corner point in the image to be detected. This method can obtain the corner point position information of the corner point with higher accuracy, thereby facilitating the subsequent detection of the position of the target object in the image to be detected based on the corner point.
下面介绍确定各个角点对应的向心偏移张量的过程,在角点分为左上角点和右下角点的情况下,同样需要分别确认左上角点对应的向心偏移张量,以及右下角点对应的向心偏移张量,本公开实施例以确定左上角点对应的向心偏移张量为例进行详细说明,右下角点对应的向心偏移张量与左上角点对应的向心偏移张量的确定方式相似,本公开实施例不再进行赘述。The following describes the process of determining the centripetal offset tensor corresponding to each corner point. When the corner points are divided into the upper left corner point and the lower right corner point, it is also necessary to confirm the centripetal offset tensor corresponding to the upper left corner point, and The centripetal offset tensor corresponding to the lower right corner point, the embodiment of the present disclosure determines the centripetal offset tensor corresponding to the upper left corner point as an example for detailed description, the centripetal offset tensor corresponding to the lower right corner point and the upper left corner point The method for determining the corresponding centripetal offset tensor is similar, and will not be repeated in the embodiment of the present disclosure.
在一种实施方式中,为了得到更加准确的向心偏移张量,在确定向心偏移张量之前,引入特征调整过程对角点池化后的特征图进行调整后,再进行确定向心偏移张量,其中,在基于角点池化后的特征图,确定各个角点对应的向心偏移张量的情况下,如图5所示,可以包括以下步骤S501~S504:In one embodiment, in order to obtain a more accurate centripetal offset tensor, before determining the centripetal offset tensor, a feature adjustment process is introduced to adjust the feature map after corner pooling, and then the direction is determined. The central offset tensor, where, in the case of determining the centripetal offset tensor corresponding to each corner point based on the feature map after the corner point pooling, as shown in FIG. 5, the following steps S501 to S504 may be included:
S501,基于角点池化后的特征图,确定角点池化后的特征图中的每个特征点对应的导向偏移张量。S501: Determine a steering offset tensor corresponding to each feature point in the feature map after corner point pooling based on the feature map after corner point pooling.
其中,每个特征点对应的导向偏移张量表征由该特征点指向待检测图像中的目标对象中心点的偏移张量。Wherein, the steering offset tensor corresponding to each feature point represents the offset tensor from the feature point to the center point of the target object in the image to be detected.
考虑到目标对象在待检测图像中的位置与目标对象信息有关,即希望角点池化后特征图的角点的特征数据能够包含更加丰富的目标对象信息,故这里可以考虑每个特征点,尤其是表示角点特征向量能够包含更加丰富的目标对象信息,因此基于每个特征点对应的导向偏移张量,可以对角点池化后的特征图 进行特征调整,使得调整后的特征图中,每个特征点尤其是角点能够包含更加丰富的目标对象信息。Considering that the position of the target object in the image to be detected is related to the target object information, that is, it is hoped that the feature data of the corner points of the feature map after the corner point pooling can contain richer target object information, so each feature point can be considered here. In particular, it means that the corner point feature vector can contain richer target object information, so based on the steering offset tensor corresponding to each feature point, the feature map after the corner point pooling can be adjusted to make the adjusted feature map Each feature point, especially the corner point, can contain richer target object information.
这里,可以通过对角点池化后的特征图进行卷积运算,得到角点池化后的特征图中的每个特征点对应的导向偏移张量,该导向偏移张量包括沿x轴方向上的偏移值以及沿y轴方向上的偏移值。Here, the corner point pooled feature map can be convolved to obtain the steering offset tensor corresponding to each feature point in the corner point pooled feature map. The steering offset tensor includes the direction along x The offset value in the axis direction and the offset value along the y-axis direction.
以确定左上角点对应的向心偏移张量为例,这里对角点池化后的特征图进行卷积运算,主要得到特征点作为左上角点对应的导向偏移张量。To determine the centripetal offset tensor corresponding to the upper left corner point as an example, the convolution operation is performed on the feature map after the corner point pooling, and the feature point is mainly obtained as the steering offset tensor corresponding to the upper left corner point.
S502,基于每个特征点对应的导向偏移张量,确定该特征点的偏移域信息。S502: Determine the offset domain information of each feature point based on the steering offset tensor corresponding to each feature point.
其中,偏移域信息中包含与该特征点关联的多个初始特征点分别指向各自对应的偏移后特征点的偏移张量。Wherein, the offset domain information includes a plurality of initial feature points associated with the feature point and respectively point to the offset tensors of the respective offset feature points.
在得到每个特征点对应的导向偏移张量后,基于每个特征点对应的导向偏移张量进行卷积运算,得到该特征点的偏移域信息。After the steering offset tensor corresponding to each feature point is obtained, a convolution operation is performed based on the steering offset tensor corresponding to each feature point to obtain the offset domain information of the feature point.
以确定左上角点对应的向心偏移张量为例,在得到每个特征点作为左上角点的情况下对应的导向偏移张量后,再对每个特征点作为左上角点的情况下对应的导向偏移张量进行卷积运算,从而得到该特征点作为左上角点的情况下的偏移域信息。To determine the centripetal offset tensor corresponding to the upper left corner point as an example, after obtaining the corresponding steering offset tensor with each feature point as the upper left corner point, then use each feature point as the upper left corner point The corresponding steering offset tensor is subjected to convolution operation to obtain the offset domain information when the feature point is used as the upper left corner point.
S503,基于角点池化后的特征图,以及该角点池化后的特征图中的特征点的偏移域信息,对角点池化后的特征图中的特征点的特征数据进行调整,得到调整后的特征图。S503: Based on the feature map after corner point pooling and the offset domain information of the feature points in the feature map after corner point pooling, adjust the feature data of the feature points in the feature map after corner point pooling , Get the adjusted feature map.
在得到角点池化后的特征图的特征点作为左上角点的情况下的偏移域信息后,即可以将角点池化后的特征图,以及该角点池化后的特征图中的每个特征点作为左上角点时的偏移域信息,同时进行可变形卷积运算,得到左上角点对应的调整后的特征图。After the feature point of the feature map after corner point pooling is obtained as the offset domain information in the case of the upper left corner point, the feature map after the corner point pooling can be pooled, and the feature map after the corner point pooling Each feature point of is used as the offset domain information of the upper left corner point, and the deformable convolution operation is performed at the same time to obtain the adjusted feature map corresponding to the upper left corner point.
这里,步骤S501至S503的过程可以通过如图6所示的特征调整网络确定:Here, the process of steps S501 to S503 can be determined through the feature adjustment network as shown in FIG. 6:
对角点池化后的特征图进行卷积运算,得到该角点池化后的特征图中每个特征点作为左上角点的情况下对应的导向偏移张量,然后对该导向偏移张量进行卷积运算,得到偏移域信息,针对这里的偏移域信息做如下解释:Perform a convolution operation on the corner point pooled feature map to obtain the corresponding steering offset tensor when each feature point in the corner point pooled feature map is used as the upper left corner point, and then the steering offset tensor The tensor performs convolution operation to obtain offset domain information. The offset domain information here is explained as follows:
在对角点池化后的特征图中的特征点的特征数据进行卷积运算,比如针对角点池化后的特征图中的特征点A的特征数据进行卷积运算的情况下,在不存在偏移域信息的情况下,在按照3*3卷积对特征点A的特征数据进行卷积运算的情况下,则可以通过角点池化后的特征图中包含特征点A在内的9个通过实线框表示的初始特征点的特征数据进行卷积运算得到,考虑偏移域信息后,希望通过包含更加丰富的目标对象信息的特征点的特征数据,对特征点A进行特征调节,比如可以基于每个特征点对应的导向偏移向量,来对用于对特征点A进行特征调节的特征点进行偏移,比如偏移后的特征点可以通过如图6中角点池化后的特征图中9个虚线框来表示,这样可以通过这9个偏移后的特征点的特征数据进行卷积运算,对特征点A的特征数据进行特征调整,其中偏移域信息可以通过图6中的偏移张量进行表示,偏移张量中的每个偏移张量,即为每个初始特征点指向该初始特征点对应的偏移后特征点的偏移张量,表示初始特征点在x轴方向上和y方向上偏移后,得到该初始特征点对应的偏移后特征点。Perform the convolution operation on the feature data of the feature points in the feature map after the corner point pooling. For example, if the feature data of the feature point A in the feature map after the corner point pooling is subjected to the convolution operation, the When there is offset domain information, when the feature data of feature point A is convolved according to 3*3 convolution, the feature map after corner pooling can be used to include feature point A. The feature data of the 9 initial feature points represented by the solid line frame are obtained by convolution operation. After considering the offset domain information, it is hoped that the feature point A can be adjusted by the feature data containing more abundant target object information. For example, the feature points used for feature adjustment of feature point A can be offset based on the steering offset vector corresponding to each feature point. For example, the offset feature points can be pooled by corner points as shown in Figure 6. The latter feature map is represented by the 9 dashed boxes, so that the feature data of the 9 offset feature points can be used to perform the convolution operation, and the feature data of feature point A can be adjusted. The offset domain information can be passed The offset tensor in Figure 6 is represented. Each offset tensor in the offset tensor is the offset tensor of each initial feature point pointing to the offset feature point corresponding to the initial feature point, representing the initial feature After the point is offset in the x-axis direction and the y direction, the offset feature point corresponding to the initial feature point is obtained.
考虑每个特征点作为左上角点的情况下对应的导向偏移张量,使得特征调整后的特征点中的特征数据包含的目标对象信息更加丰富,从而便于在后期基于调整后的特征图确定每个左上角点对应的向心偏移张量的情况下,能够得到更加准确的向心偏移张量。Considering each feature point as the corresponding steering offset tensor in the case of the upper left corner point, so that the feature data in the feature point after feature adjustment contains richer target object information, which is convenient for later determination based on the adjusted feature map In the case of the centripetal offset tensor corresponding to each upper left corner point, a more accurate centripetal offset tensor can be obtained.
同理,考虑每个特征点作为右下角点的情况下相对目标对象的中心点的导向偏移张量,使得特征调整后的特征点包含的目标对象信息更加丰富,从而便于在后期基于调整后的特征图确定每个右下角点对应的向心偏移张量的情况下,能够得到更加准确的向心偏移张量。In the same way, considering each feature point as the lower right corner point of the guide offset tensor relative to the center point of the target object, the feature point after feature adjustment contains richer target object information, which is convenient for the subsequent adjustment based on the target object information. In the case of determining the centripetal offset tensor corresponding to each lower right corner of the feature map, a more accurate centripetal offset tensor can be obtained.
S504,基于调整后的特征图,确定各个角点对应的向心偏移张量。S504, based on the adjusted feature map, determine the centripetal offset tensor corresponding to each corner point.
这里,对调整后的特征图中的角点对应的特征数据进行卷积运算,确定各个角点对应的向心偏移张量。Here, a convolution operation is performed on the feature data corresponding to the corner points in the adjusted feature map, and the centripetal offset tensor corresponding to each corner point is determined.
这里,调整后的特征图中可以包含左上角点对应的调整后的特征图,以及右下角点对应的调整后的特征图,在基于左上角点对应的调整后的特征图确定每个左上角点对应的向心偏移张量的情况下,可以通过左上角点对应的向心偏移预测网络来确定,在基于右下角点对应的调整后的特征图确定每个右下角点对应的向心偏移张量的情况下,可以通过右下角点对应的向心偏移预测网络来确定。Here, the adjusted feature map may include the adjusted feature map corresponding to the upper left corner point, and the adjusted feature map corresponding to the lower right corner point, and each upper left corner is determined based on the adjusted feature map corresponding to the upper left corner point In the case of the centripetal offset tensor corresponding to the point, it can be determined by the centripetal offset prediction network corresponding to the upper left corner point. Based on the adjusted feature map corresponding to the lower right corner point, determine the direction corresponding to each lower right corner point. In the case of the heart offset tensor, it can be determined by the centripetal offset prediction network corresponding to the lower right corner point.
以上S501~S504的过程,为本公开实施例提供的确定向心偏移张量的过程,通过考虑目标对象信息,比如引入角点对应的导向偏移张量,以及特征点的偏移域信息,对角点池化后的特征图中的特征点的特征数据进行调整,使得得到的调整后的特征图中的特征点的特征数据,能够包含更丰富的目标对象信息,从而能够确定出每个角点对应的更加准确的向心偏移张量,通过准确的向心偏移张量,能够准确得到角点指向的中心点位置信息,从而准确地检测目标对象在待检测图像中的位置。The above process of S501 to S504 is the process of determining the centripetal offset tensor provided by the embodiments of the present disclosure, by considering the target object information, such as introducing the steering offset tensor corresponding to the corner point, and the offset domain information of the feature point , Adjust the feature data of the feature points in the feature map after the corner point pooling, so that the feature data of the feature points in the adjusted feature map can contain richer target object information, so that each A more accurate centripetal offset tensor corresponding to each corner point. Through accurate centripetal offset tensor, the position information of the center point pointed by the corner point can be accurately obtained, so as to accurately detect the position of the target object in the image to be detected .
上文提到可以通过角点热力图确定待检测图像中包含的目标对象的类别,这里介绍如何根据角点热力图来确定目标对象的类别,从上文得知待检测图像对应的角点热力图包括多个通道分别对应的角点热 力图,每个通道对应一种预设对象类别;则在上文提到的基于角点热力图,确定角点热力图中每个特征点作为角点的概率值之后,如图7所示,本公开实施例提供的检测方法还包括以下步骤S701~S702:As mentioned above, the category of the target object contained in the image to be detected can be determined through the corner heat map. Here is how to determine the category of the target object based on the corner heat map. From the above, we know the corner heat of the image to be detected. The figure includes the corner heat maps corresponding to multiple channels, and each channel corresponds to a preset object category; in the above-mentioned corner heat map, each feature point in the corner heat map is determined as a corner point After the probability value of, as shown in FIG. 7, the detection method provided by the embodiment of the present disclosure further includes the following steps S701 to S702:
S701,针对多个通道中的每个通道,基于该通道对应的角点热力图中每个特征点作为角点的概率值,确定该通道对应的角点热力图中是否存在角点。S701: For each channel of the multiple channels, determine whether there is a corner point in the corner heat map corresponding to the channel based on the probability value of each feature point as the corner point in the corner heat map corresponding to the channel.
S702,在该通道对应的角点热力图中存在角点的情况下,确定待检测图像中包含该通道对应的预设对象类别的目标对象。S702: When there are corner points in the corner point heat map corresponding to the channel, determine that the image to be detected contains the target object of the preset object category corresponding to the channel.
针对待检测图像对应的角点热力图中,每个通道分别对应的角点热力图中各个特征点作为角点的概率值,可以确定该通道的角点热力图中是否存在角点,比如在某个通道的角点特征图中包含多个对应概率值大于设定阈值的特征点的情况下,则说明该通道的角点特征图中大概率包含角点,而角点用于表征目标对象在待检测图像中的位置,这样即可以说明待检测图像中包含该通道对应的预设对象类别的目标对象。For the corner heat map corresponding to the image to be detected, the probability value of each feature point in the corner heat map corresponding to each channel as a corner point can be determined whether there is a corner point in the corner heat map of the channel, for example, When the corner feature map of a channel contains multiple feature points with a corresponding probability value greater than the set threshold, it means that the corner feature map of the channel contains corner points with a high probability, and the corner points are used to represent the target object The position in the image to be detected, so that it can be explained that the image to be detected contains the target object of the preset object category corresponding to the channel.
比如,针对某动物园中的动物进行检测,可以设定通道个数为100,即得到的角点热力图为h*w*100,每个通道对应一种预设对象类别,针对某种待检测图像,在得到的该待检测图像对应的角点热力图的100个通道中,只有第1个通道和第2个通道中的角点热力图中包含角点,且第1个通道对应的预设对象类别为01,第2个通道对应的预设对象类别为02的情况下,则可以说明待检测图像中包含类别为01和02的目标对象。For example, to detect animals in a zoo, you can set the number of channels to 100, that is, the obtained corner heat map is h*w*100, and each channel corresponds to a preset object category, for a certain type of object to be detected Image, among the 100 channels of the corner heat map corresponding to the image to be detected, only the corner heat maps in the first and second channels contain corner points, and the first channel corresponds to the pre- Assuming that the object category is 01, and the preset object category corresponding to the second channel is 02, it can be explained that the image to be detected contains target objects of the categories 01 and 02.
本公开实施例提出通过将角点池化后的特征图输入角点热力图预测网络,即可以得到包含预设通道个数的角点热力图,通过每个通道对应的角点热力图中是否存在角点,进而能够确定待检测图像中是否存在该通道对应类别的目标对象。The embodiment of the present disclosure proposes that by inputting the feature map after the corner point pooling into the corner heat map prediction network, the corner heat map containing the preset number of channels can be obtained, and whether the corner heat map corresponding to each channel is There are corner points, and then it can be determined whether there is a target object corresponding to the channel in the image to be detected.
另外,这里在检测出每个通道上的角点热力图包含的角点后,可以确定角点对应的向心偏移张量,从而确定每个通道对应的目标对象在待检测图像中的位置,从而结合该通道对应的目标对象的类别,确定待检测图像中各个目标对象的类别。In addition, after detecting the corner points contained in the corner heat map on each channel, the centripetal offset tensor corresponding to the corner point can be determined, so as to determine the position of the target object corresponding to each channel in the image to be detected. , In order to determine the category of each target object in the image to be detected in combination with the category of the target object corresponding to the channel.
针对上述S203,即在基于各个角点在待检测图像中的角点位置信息及各个角点对应的向心偏移张量,确定待检测图像中的目标对象的情况下,可以包括:Regarding the above S203, that is, in the case of determining the target object in the image to be detected based on the corner position information of each corner point in the image to be detected and the centripetal offset tensor corresponding to each corner point, it may include:
基于各个角点在待检测图像中的角点位置信息和各个角点对应的向心偏移张量,确定待检测图像中目标对象的检测框。Based on the corner position information of each corner point in the image to be detected and the centripetal offset tensor corresponding to each corner point, the detection frame of the target object in the image to be detected is determined.
这里,在生成待检测图像中目标对象的检测框的情况下,需要至少一个左上角点和右下角点的角点位置信息确定,或者需要至少一个右上角点和左下角点的角点位置信息确定,本公开实施例以通过一个左上角点和一个右下角点来确定检测框为例进行说明。Here, in the case of generating the detection frame of the target object in the image to be detected, it is necessary to determine the corner position information of at least one upper left corner point and the lower right corner point, or at least one corner point position information of the upper right corner point and the lower left corner point is required To be sure, the embodiment of the present disclosure uses an upper left corner point and a lower right corner point to determine the detection frame as an example for description.
这里,在基于各个角点在待检测图像中的角点位置信息和各个角点对应的向心偏移张量,确定待检测图像中目标对象的检测框的情况下,如图8所示,可以包括:Here, in the case of determining the detection frame of the target object in the image to be detected based on the corner position information of each corner point in the image to be detected and the centripetal offset tensor corresponding to each corner point, as shown in Figure 8, Can include:
S801,基于各个角点在待检测图像中的角点位置信息,筛选构成候选检测框的候选角点对。S801: Based on the corner position information of each corner point in the image to be detected, filter candidate corner point pairs that constitute a candidate detection frame.
以包含左上角点和右下角点的候选角点对为例,在筛选能够构成候选检测框的左上角点和右下角点的情况下,为了提高筛选速度,可以先判断左上角点和右下角点是否属于同一目标对象类别,在确定任一左上角点和右下角点属于同一目标对象类别的情况下,再继续判断该任一左上角点和右下角点在待检测图像中的角点位置信息能否构成同一个候选检测框。Take the candidate corner point pair containing the upper left corner and the lower right corner as an example. In the case of screening the upper left and lower right corner points that can constitute the candidate detection frame, in order to improve the screening speed, the upper left corner and the lower right corner can be judged first Whether the points belong to the same target object category, in the case of determining that any upper left corner point and lower right corner point belong to the same target object category, continue to determine the corner position of any upper left corner point and lower right corner point in the image to be detected Whether the information constitutes the same candidate detection frame.
比如,左上角点在待检测图像中应该位于右下角点的左上方,在根据左上角点和右下角点的角点位置信息,比如根据左上角点在待检测图像中的位置坐标,以及右下角点在待检测图像中的位置坐标,无法使得左上角点位于右下角点的左上方的情况下,则该左上角点和右下角点无法构成候选角点对。For example, the upper left corner point should be located at the upper left of the lower right corner point in the image to be detected, and the position information of the corner point based on the upper left corner point and the lower right corner point, such as the position coordinates of the upper left corner point in the image to be detected, and the right If the position coordinates of the lower corner point in the image to be detected cannot be such that the upper left corner point is located at the upper left corner of the lower right corner point, the upper left corner point and the lower right corner point cannot constitute a candidate corner point pair.
这里,可以在待检测图像中建立坐标系,该坐标系包括X轴和Y轴,每个角点在该坐标系中的角点位置信息包括在X轴方向上的横坐标值,以及在Y轴方向上的纵坐标值,然后在该坐标系中,根据各个角点在该坐标系中对应的坐标值,筛选能够构成候选检测框的左上角点和右下角点。Here, a coordinate system can be established in the image to be detected, the coordinate system includes X axis and Y axis, and the corner position information of each corner point in the coordinate system includes the abscissa value in the X axis direction and the Y axis. The ordinate value in the axis direction, and then in the coordinate system, according to the corresponding coordinate value of each corner point in the coordinate system, the upper left corner point and the lower right corner point that can constitute the candidate detection frame are filtered.
S802,基于每个候选角点对中每个角点在待检测图像中的角点位置信息和该角点对应的向心偏移张量,确定该角点指向的中心点位置信息。S802: Determine the position information of the center point to which the corner point points based on the corner position information of each corner point in the image to be detected in each candidate corner point pair and the centripetal offset tensor corresponding to the corner point.
这里,可以根据以下公式(5)来确定每个候选角点对中左上角点指向的中心点位置信息,根据以下公式(6)来确定每个候选角点对中右下角点指向的中心点位置信息:Here, the position information of the center point pointed to by the upper left corner point in each candidate corner point pair can be determined according to the following formula (5), and the center point pointed to by the lower right corner point in each candidate corner point pair can be determined according to the following formula (6) location information:
Figure PCTCN2020135967-appb-000001
Figure PCTCN2020135967-appb-000001
Figure PCTCN2020135967-appb-000002
Figure PCTCN2020135967-appb-000002
其中,
Figure PCTCN2020135967-appb-000003
表示第i个左上角点指向的中心点位置信息中X轴方向上对应的横坐标值,
Figure PCTCN2020135967-appb-000004
表示第i个左上角点指向的中心点位置信息中Y轴方向上对应的纵坐标值;tl x(i)表示第i个左上角点在待检测图像中的角点位置信息中X轴方向上对应的横坐标值;tl y(i)表示左上角点在待检测图像中的角点位置信息中Y轴方向上对应的纵坐标值;
Figure PCTCN2020135967-appb-000005
表示第i个左上角点的向心偏移张量中向X轴方向的偏移值,
Figure PCTCN2020135967-appb-000006
表示第i个左上角点的向心偏移张量中向Y轴方向的偏移值。
in,
Figure PCTCN2020135967-appb-000003
Represents the corresponding abscissa value in the X-axis direction in the position information of the center point pointed to by the i-th upper left corner point,
Figure PCTCN2020135967-appb-000004
Indicates the corresponding ordinate value in the Y-axis direction in the center point position information pointed to by the i-th upper left corner point; tl x(i) indicates the X-axis direction in the corner point position information of the i-th upper left corner point in the image to be detected The abscissa value corresponding to above; tl y(i) represents the ordinate value corresponding to the Y axis direction of the upper left corner point in the corner point position information in the image to be detected;
Figure PCTCN2020135967-appb-000005
Represents the offset value to the X axis in the centripetal offset tensor of the i-th upper left corner point,
Figure PCTCN2020135967-appb-000006
Represents the Y-axis offset value in the centripetal offset tensor of the i-th upper left corner point.
其中,
Figure PCTCN2020135967-appb-000007
表示第j个右下角点指向的中心点位置信息中X轴方向上对应的横坐标值,
Figure PCTCN2020135967-appb-000008
表示第j个右下角点指向的中心点位置信息中Y轴方向上对应的纵坐标值;br x(j)表示第j个右下角点在待检测图像中的角点位置信息中X轴方向上对应的横坐标值;br y(j)表示第j个右下角点在待检测图像中的角点位置信息中Y轴方向上对应的纵坐标值;
Figure PCTCN2020135967-appb-000009
表示第j个右下角点的向心偏移张量中向X轴方向的偏移值,
Figure PCTCN2020135967-appb-000010
表示第j个右下角点的向心偏移张量中向Y轴方向的偏移值。
in,
Figure PCTCN2020135967-appb-000007
Represents the corresponding abscissa value in the X-axis direction in the position information of the center point pointed to by the j-th lower right corner point,
Figure PCTCN2020135967-appb-000008
Represents the corresponding ordinate value in the Y axis direction in the center point position information pointed to by the jth lower right corner point; br x(j) represents the X axis direction in the corner point position information of the jth lower right corner point in the image to be detected The abscissa value corresponding to above; br y(j) represents the ordinate value corresponding to the Y-axis direction of the j-th lower right corner point in the corner position information of the image to be detected;
Figure PCTCN2020135967-appb-000009
Represents the offset value to the X axis in the centripetal offset tensor of the j-th lower right corner point,
Figure PCTCN2020135967-appb-000010
Represents the offset value to the Y axis in the centripetal offset tensor of the jth lower right corner point.
S803,基于每个候选角点对中每个角点在待检测图像中的角点位置信息,确定该候选角点对所对应的中心区域信息。S803: Based on the corner position information of each corner point in each candidate corner point pair in the image to be detected, determine the center area information corresponding to the candidate corner point pair.
这里的中心区域信息可以是预先设定好的,其定义为与目标对象的检测框中心重合的中心区域框的坐标范围,通过该中心区域框的坐标范围能够检测候选检测框是否包含了唯一的目标对象。The central area information here can be preset, which is defined as the coordinate range of the central area frame that coincides with the center of the detection frame of the target object. Through the coordinate range of the central area frame, it is possible to detect whether the candidate detection frame contains a unique target.
比如,在候选角点对中,在左上角点指向的中心点位置信息和右下角点指向的中心点位置信息位于中心区域框的坐标范围内,在该中心区域框的坐标范围较小的情况下,则可以认为左上角点指向的中心点位置信息和右下角点指向的中心点位置信息比较接近,从而确定该候选角点对构成的候选检测框为包含了唯一的目标对象。For example, in the candidate corner point pair, the position information of the center point pointed to by the upper left corner point and the center point position information pointed to by the lower right corner point are located within the coordinate range of the central area frame, in the case where the coordinate range of the central area frame is small Then, it can be considered that the position information of the center point pointed to by the upper left corner point is relatively close to the position information of the center point pointed to by the lower right corner point, so as to determine that the candidate detection frame formed by the candidate corner point pair contains a unique target object.
这里,在所述基于每个候选角点对中每个角点在所述待检测图像中的角点位置信息,确定该候选角点对所对应的中心区域信息的情况下,可以包括:Here, in the case of determining the center area information corresponding to the candidate corner point pair based on the corner point position information of each corner point in the candidate corner point pair in the image to be detected, it may include:
(1)基于该候选角点对的每个角点的角点位置信息,确定表征该候选角点对所对应的中心区域框的角点位置信息;(1) Based on the corner position information of each corner point of the candidate corner point pair, determine the corner point position information that characterizes the center area frame corresponding to the candidate corner point pair;
(2)基于中心区域框的角点位置信息,确定该候选角点对所对应的中心区域框的坐标范围。(2) Based on the corner position information of the central area frame, determine the coordinate range of the central area frame corresponding to the candidate corner point pair.
在第m个候选角点对由第i个左上角点和第j个右下角点构成的情况下,则可以按照以下公式(7)~公式(10)来确定该第m个候选角点对所对应的中心区域框的角点位置信息:In the case that the m-th candidate corner point pair is composed of the i-th upper left corner point and the j-th lower right corner point, the m-th candidate corner point pair can be determined according to the following formulas (7) to (10) Corresponding corner position information of the central area frame:
Figure PCTCN2020135967-appb-000011
Figure PCTCN2020135967-appb-000011
Figure PCTCN2020135967-appb-000012
Figure PCTCN2020135967-appb-000012
Figure PCTCN2020135967-appb-000013
Figure PCTCN2020135967-appb-000013
Figure PCTCN2020135967-appb-000014
Figure PCTCN2020135967-appb-000014
其中,
Figure PCTCN2020135967-appb-000015
表示第m个候选角点对所对应的中心区域框的左上角点在待检测图像中X轴方向上的横坐标值;
Figure PCTCN2020135967-appb-000016
表示第m个候选角点对所对应的中心区域框的左上角点在待检测图像中Y轴方向上的横坐标值;
Figure PCTCN2020135967-appb-000017
表示第m个候选角点对所对应的中心区域框的右下角点在待检测图像中X轴方向上的横坐标值;
Figure PCTCN2020135967-appb-000018
表示第m个候选角点对所对应的中心区域框的右下角点在待检测图像中Y轴方向上的横坐标值;μ表示中心区域框的长宽与候选检测框的长宽的比例,该比例为预先设定好的, 且μ∈(0,1)。
in,
Figure PCTCN2020135967-appb-000015
Represents the abscissa value of the upper left corner point of the center area frame corresponding to the m-th candidate corner point pair in the X-axis direction of the image to be detected;
Figure PCTCN2020135967-appb-000016
Represents the abscissa value of the upper left corner of the center area frame corresponding to the m-th candidate corner point pair in the Y-axis direction of the image to be detected;
Figure PCTCN2020135967-appb-000017
Represents the abscissa value of the lower right corner of the center area frame corresponding to the m-th candidate corner point pair in the X-axis direction of the image to be detected;
Figure PCTCN2020135967-appb-000018
Represents the abscissa value of the lower right corner of the central area frame corresponding to the m-th candidate corner point pair in the Y-axis direction of the image to be detected; μ represents the ratio of the length and width of the central area frame to the length and width of the candidate detection frame, The ratio is preset, and μ∈(0,1).
在确定出第m个候选角点对所对应的中心区域框的角点位置信息后,可以按照以下公式(11)来确定该中心区域框的坐标范围:After determining the corner position information of the center area frame corresponding to the m-th candidate corner point pair, the coordinate range of the center area frame can be determined according to the following formula (11):
Figure PCTCN2020135967-appb-000019
Figure PCTCN2020135967-appb-000019
其中,R central(m)表示第m个候选角点对所对应的中心区域框的坐标范围,该中心区域框的坐标范围通过X轴方向上的x(m)值,以及Y轴方向上的y(m)值来表示,其中x(m)的范围满足
Figure PCTCN2020135967-appb-000020
y(m)的范围满足
Figure PCTCN2020135967-appb-000021
Among them, R central(m) represents the coordinate range of the central area frame corresponding to the m-th candidate corner point pair. The coordinate range of the central area frame passes through the x(m) value in the X-axis direction and the Y-axis direction. y(m) value, where the range of x(m) satisfies
Figure PCTCN2020135967-appb-000020
The range of y(m) satisfies
Figure PCTCN2020135967-appb-000021
S804,基于每个候选角点对中每个角点指向的中心点位置信息,以及该候选角点对所对应的中心区域信息,在候选检测框中确定目标对象的检测框。S804: Determine the detection frame of the target object in the candidate detection frame based on the position information of the center point pointed to by each corner point in each candidate corner point pair and the center area information corresponding to the candidate corner point pair.
每个候选角点对所对应的中心区域信息用于制约该候选角点对中,每个角点指向的中心点位置信息之间的接近程度,在某个候选角点对中每个角点指向的中心点位置信息位于该候选角点对所对应的中心区域框内的情况下,则可以说明候选角点对中每个角点指向的中心点比较接近,则可以说明候选角点对构成的候选检测框中包含的目标对象是唯一的目标对象。The center area information corresponding to each candidate corner point pair is used to restrict the proximity between the center point position information pointed to by each corner point in the candidate corner point pair, and each corner point in a certain candidate corner point pair When the position information of the pointed center point is located in the center area frame corresponding to the candidate corner point pair, it can be explained that the center point of each corner point in the candidate corner point pair is relatively close, and it can be explained that the candidate corner point pair constitutes The target object contained in the candidate detection frame of is the only target object.
本公开实施例提出的确定目标对象的检测框的方式,通过角点的角点位置信息首先确定能够构成候选检测框的候选角点对,然后再基于该候选角点对中每个角点分别对象的向心偏移张量,来确定候选检测框包围的目标对象是否为同一个目标对象,从而能够较为准确地检测出待检测图像中的所有目标对象的检测框。In the method of determining the detection frame of the target object proposed by the embodiment of the present disclosure, the corner point position information of the corner points is used to first determine the candidate corner point pairs that can constitute the candidate detection frame, and then based on each corner point in the candidate corner point pair. The centripetal offset tensor of the object is used to determine whether the target object surrounded by the candidate detection frame is the same target object, so that the detection frame of all target objects in the image to be detected can be detected more accurately.
这里,在基于每个候选角点对中每个角点指向的中心点位置信息,以及该候选角点对所对应的中心区域信息,在候选检测框中确定目标对象的检测框的情况下,如图9所示,可以包括以下步骤S901~S903:Here, in the case that the detection frame of the target object is determined in the candidate detection frame based on the position information of the center point pointed to by each corner point in each candidate corner point pair and the center area information corresponding to the candidate corner point pair, As shown in Figure 9, the following steps S901 to S903 may be included:
S901,基于每个候选角点对中每个角点指向的中心点位置信息,以及该候选角点对所对应的中心区域信息,确定有效候选角点对。S901: Determine a valid candidate corner point pair based on the position information of the center point pointed to by each corner point in each candidate corner point pair and the center area information corresponding to the candidate corner point pair.
这里,在某个候选角点对中每个角点指向的中心点位置信息位于该候选角点对所对应的中心区域框内的情况下,则将该候选角点对作为有效候选角点对。Here, in a case where the position information of the center point pointed to by each corner point in a certain candidate corner point pair is located in the center area frame corresponding to the candidate corner point pair, the candidate corner point pair is regarded as a valid candidate corner point pair. .
这里,可以通过以下公式(12)来判断第i个左上角点和第j个右下角点构成的候选角点对是否为有效候选角点对,即判断第i个左上角点和第j个右下角点构成的候选检测框对应的第m个中心区域框的坐标范围,与第i个左上角点和第j个右下角点各自指向的中心点位置信息是否满足以下公式(12):Here, the following formula (12) can be used to determine whether the candidate corner point pair formed by the i-th upper left corner point and the j-th lower right corner point is a valid candidate corner point pair, that is, the i-th upper left corner point and the j-th corner point pair are judged Whether the coordinate range of the m-th central area frame corresponding to the candidate detection frame formed by the lower right corner point and the center point position information pointed to by the i-th upper left corner point and the j-th lower right corner point respectively meet the following formula (12):
Figure PCTCN2020135967-appb-000022
Figure PCTCN2020135967-appb-000022
在第i个左上角点和第j个右下角点构成的候选检测框对应的第m个中心区域框的坐标范围,与第i个左上角点和第j个右下角点各自指向的中心点位置信息满足以上公式(12)的情况下,则说明第i个左上角点和第j个右下角点构成的候选角点对为有效候选角点对,然后继续对该有效候选角点执行S902的步骤,否则,在第i个左上角点和第j个右下角点构成的候选角点对为无效候选角点对的情况下,继续判断该第i个左上角点和其它右下角点是否能够构成有效候选角点对,直至得到有效候选角点对后,再执行后续步骤。The coordinate range of the m-th central area frame corresponding to the candidate detection frame formed by the i-th upper-left corner point and the j-th lower-right corner point, and the center point respectively pointed to by the i-th upper-left corner point and the j-th lower-right corner point When the position information satisfies the above formula (12), it means that the candidate corner point pair formed by the i-th upper left corner point and the j-th lower right corner point is a valid candidate corner point pair, and then continue to perform S902 on the valid candidate corner point Otherwise, if the candidate corner point pair formed by the i-th upper left corner point and the j-th lower right corner point is an invalid candidate corner point pair, continue to determine whether the i-th upper left corner point and other lower right corner points are A valid candidate corner point pair can be formed, and the subsequent steps can be executed after a valid candidate corner point pair is obtained.
S902,基于有效候选角点对中每个角点指向的中心点位置信息、有效候选角点对所对应的中心区域信息、以及有效候选角点对中每个角点对应的概率值,确定每个有效候选角点对所对应的候选检测框的分值。S902, based on the position information of the center point pointed to by each corner point in the valid candidate corner point pair, the central area information corresponding to the valid candidate corner point pair, and the probability value corresponding to each corner point in the valid candidate corner point pair, determine each The score of the candidate detection frame corresponding to the valid candidate corner points.
其中,每个角点对应的概率值用于表示该角点在角点热力图中对应的特征点作为角点的概率值。Wherein, the probability value corresponding to each corner point is used to indicate the probability value of the corresponding feature point of the corner point in the corner point heat map as the corner point.
在对待检测图像进行检测的情况下,可能存在针对同一个目标对象出现多个候选检测框的情况,有些候选检测框表示的目标对象在待检测图像中的位置的准确度可能较低,这里,引入每个有效候选角点对所对应的候选检测框的分值,比如可以通过有效候选角点对中每个角点指向的中心点所构成的区域和有效候选角点对所对应的中心区域框之间的面积关系,以及有效候选角点对中每个角点对应的概率值,来表示每个有效候选角点对所对应的候选检测框的分值,分值高的候选检测框作为目标对象的检测框的概率较大,通过此对候选检测框进行筛选。In the case of detecting the image to be detected, there may be multiple candidate detection frames for the same target object. The position of the target object represented by some candidate detection frames in the image to be detected may be less accurate. Here, Introduce the score of the candidate detection frame corresponding to each valid candidate corner point pair, such as the area formed by the center point of each corner point in the valid candidate corner point pair and the center area corresponding to the valid candidate corner point pair The area relationship between the frames and the probability value corresponding to each corner point in the effective candidate corner point pair represent the score value of the candidate detection frame corresponding to each effective candidate corner point pair. The candidate detection frame with the higher score is taken as The probability of the detection frame of the target object is relatively large, and the candidate detection frame is screened through this.
这里,针对第i个左上角点和第j个右下角点构成的有效候选角点对,可以按照以下公式(13)来确定该有效候选角点对所对应的候选检测框的分值:Here, for the valid candidate corner point pair formed by the i-th upper left corner point and the j-th lower right corner point, the score of the candidate detection frame corresponding to the valid candidate corner point pair can be determined according to the following formula (13):
Figure PCTCN2020135967-appb-000023
Figure PCTCN2020135967-appb-000023
其中,s表示第i个左上角点和第j个右下角点构成的有效候选角点对所对应的候选检测框的分值;s tl(i)表示第i个左上角点在左上角点热力图中对应的特征点作为左上角点的概率值;s br(j)表示第j个右下角点在右下角点热力图中对应的特征点作为右下角点的概率值。 Among them, s represents the score of the candidate detection frame corresponding to the valid candidate corner point pair formed by the i-th upper left corner point and the j-th lower right corner point; s tl(i) represents the i-th upper left corner point at the upper left corner point The corresponding feature point in the heat map is used as the probability value of the upper left corner point; s br(j) represents the probability value of the j-th lower right corner point in the lower right corner point in the heat map as the lower right corner point.
S903,基于每个有效候选角点对所对应的候选检测框的分值、以及相邻候选检测框之间的重叠区域大小,在候选检测框中确定目标对象的检测框。S903: Determine the detection frame of the target object in the candidate detection frame based on the score of the candidate detection frame corresponding to each valid candidate corner point and the size of the overlapping area between adjacent candidate detection frames.
这里,重叠区域可以通过重叠区域在待检测图像中的尺寸来确定,下面介绍如何基于每个有效候选角点对所对应的候选检测框的分值、以及相邻候选检测框之间的重叠区域,来筛选目标对象的检测框。Here, the overlap area can be determined by the size of the overlap area in the image to be detected. The following describes how to base each valid candidate corner point on the corresponding candidate detection frame score and the overlap area between adjacent candidate detection frames , To filter the detection frame of the target object.
这里,可以通过软式非极大抑制在多个候选检测框中筛选目标对象的检测框,这里,针对有超过重叠区域阈值的多个候选检测框,可以将对应分值最高的候选检测框作为目标对象的检测框,将多个候选检测框中其它的候选检测框删除,这样即可得到待检测图像中目标对象的检测框。Here, the detection frame of the target object can be screened in multiple candidate detection frames by soft non-maximum suppression. Here, for multiple candidate detection frames that exceed the overlap area threshold, the candidate detection frame with the highest corresponding score can be used as For the detection frame of the target object, delete other candidate detection frames in the multiple candidate detection frames, so that the detection frame of the target object in the image to be detected can be obtained.
以上步骤S901~S903通过对构成候选检测框的候选角点对进行有效筛选,确定能够筛选出只表征一个目标对象的候选检测框,然后对这些只表征一个目标对象的候选检测框进行软式非极大抑制筛选,从而准确地得到表征目标对象的检测框。In the above steps S901 to S903, through effective screening of the candidate corner points constituting the candidate detection frame, it is determined that the candidate detection frame that only characterizes one target object can be screened out, and then the candidate detection frames that only represent one target object are softly non-selected. The screening is greatly suppressed, so as to accurately obtain the detection frame that characterizes the target object.
在得到待检测图像中目标对象的检测框后,可以确定该检测框内目标对象的实例信息,这里,可以基于目标对象的检测框和对待检测图像进行特征提取得到的初始特征图,确定待检测图像中所述目标对象的实例信息。After the detection frame of the target object in the image to be detected is obtained, the instance information of the target object in the detection frame can be determined. Here, the object to be detected can be determined based on the detection frame of the target object and the initial feature map obtained by feature extraction of the image to be detected Instance information of the target object in the image.
这里的实例信息可以通过掩膜来表示,这里的掩膜是指对图像中的目标对象进行实例分割后,在像素层面给出每个目标对象的像素,因此掩膜可以精确到物体的边缘,从而得到目标对象在待检测图像中更加准确的位置;除此之外,基于掩膜还可以表示目标对象的形貌,从而可以基于该形貌来验证目标对象的类别的确定是否准确,以及基于掩膜表示的目标对象的形貌,对目标对象进行后续动作分析,在本公开实施例中不对此进行阐述。The instance information here can be represented by a mask. The mask here means that after instance segmentation of the target object in the image, the pixels of each target object are given at the pixel level, so the mask can be accurate to the edge of the object. In this way, a more accurate position of the target object in the image to be detected can be obtained; in addition, the shape of the target object can also be represented based on the mask, so that the determination of the target object's category can be verified based on the shape, and based on The shape of the target object represented by the mask is subjected to subsequent action analysis on the target object, which is not described in the embodiment of the present disclosure.
这里,在基于目标对象的检测框和对待检测图像进行特征提取得到的初始特征图,确定待检测图像中目标对象的实例信息的情况下,可以包括:Here, in the case of determining the instance information of the target object in the image to be detected based on the detection frame of the target object and the initial feature map obtained by feature extraction of the image to be detected, it may include:
(1)基于目标对象的检测框以及初始特征图,提取初始特征图在检测框内的特征点的特征数据;(1) Based on the detection frame and the initial feature map of the target object, extract the feature data of the feature points of the initial feature map in the detection frame;
(2)基于初始特征图在检测框内的特征点的特征数据,确定待检测图像中目标对象的实例信息。(2) Based on the feature data of the feature points of the initial feature map in the detection frame, determine the instance information of the target object in the image to be detected.
这里将目标对象的检测框和待检测图像对应的初始特征图输入感兴趣区域提取网络,该感兴趣区域提取网络能够首先提取该初始特征图尺寸匹配的感兴趣区域,然后再通过感兴趣对齐池化处理,得到该初始特征图在该检测框内(即感兴趣区域)的特征点的特征数据,然后将该初始特征图在检测框内的特征点的特征数据输入掩膜预测网络,即可以生成目标对象的实例信息,该实例信息可以以掩模的形式表示,然后可以将该目标对象的掩膜扩大至与待检测图像中的目标对象相同的尺寸,即可以得到待检测图像的目标对象的实例信息。Here, the detection frame of the target object and the initial feature map corresponding to the image to be detected are input to the region of interest extraction network. The region of interest extraction network can first extract the region of interest matching the size of the initial feature map, and then pass the alignment pool of interest The feature data of the feature points of the initial feature map in the detection frame (that is, the region of interest) is obtained by transformation processing, and then the feature data of the feature points of the initial feature map in the detection frame are input into the mask prediction network. Generate instance information of the target object, the instance information can be expressed in the form of a mask, and then the mask of the target object can be expanded to the same size as the target object in the image to be detected, that is, the target object of the image to be detected can be obtained Instance information.
下面结合图10针对本公开实施例提出的目标检测方法进行整体说明:The overall description of the target detection method proposed in the embodiment of the present disclosure will be given below with reference to FIG. 10:
将待检测图像输入沙漏卷积神经网络,得到该待检测图像对应的初始特征图f,然后再对该待检测图像中的目标对象进行检测的情况下,可以将该初始特征图f进行角点池化处理,得到角点池化后的特征图p,进而对该角点池化后的特征图p进行左上角点检测以及特征调整,可以得到该左上角点和该左上角点对应的向心偏移张量,得到左上角点的过程通过左上角点检测网络进行确定,该左上角点检测网络包括左上角点热力图预测网络和左上角点局部偏移预测网络(图10中均未示出),得到左上角点对应的向心偏移张量之前,首先通过特征调整网络对角点池化后的特征图p进行特征调整,该过程包括确定左上角点对应的导向偏移张量以及偏移域信息,然后基于可变形卷积运算,对角点池化后的特征图p进行调整,得到调节后的特征图g,进而通过卷积运算,确定左上角点对应的向心偏移张量。Input the image to be detected into the hourglass convolutional neural network to obtain the initial feature map f corresponding to the image to be detected, and then to detect the target object in the image to be detected, the initial feature map f can be corner points Pooling process to obtain the corner point pooled feature map p, and then perform the upper left corner point detection and feature adjustment on the corner point pooled feature map p, and the direction corresponding to the upper left corner point and the upper left corner point can be obtained. Heart offset tensor, the process of obtaining the upper left corner point is determined by the upper left corner point detection network. The upper left corner point detection network includes the upper left corner point heat map prediction network and the upper left corner point local offset prediction network (none in Figure 10). Show), before obtaining the centripetal offset tensor corresponding to the upper left corner point, the feature adjustment network is first used to adjust the feature map p after the corner point pooling. This process includes determining the steering offset tensor corresponding to the upper left corner point. Then, based on the deformable convolution operation, the feature map p after the corner point pooling is adjusted to obtain the adjusted feature map g, and then through the convolution operation, the centripetal corresponding to the upper left corner point is determined The offset tensor.
右下角点通过右下角点检测网络来确定,右下角点对应的向心偏移张量通过特征调整以及卷积运算得到,过程同左上角点以及左上角点对应的向心偏移张量的确定过程相似,然后基于左上角点及左上角点对应的向心偏移张量,以及右下角点及右下角点对应的向心偏移张量共同确定目标对象的检测框。The lower right corner point is determined by the lower right corner point detection network. The centripetal offset tensor corresponding to the lower right corner point is obtained by feature adjustment and convolution operation. The process is the same as the centripetal offset tensor corresponding to the upper left corner and the upper left corner point. The determination process is similar, and then the detection frame of the target object is determined based on the centripetal offset tensor corresponding to the upper left corner point and the upper left corner point, and the centripetal offset tensor corresponding to the lower right corner point and the lower right corner point.
在得到目标对象的检测框后,基于该目标对象的检测框和初始特征图f提取感兴趣区域,进而对该感兴趣区域进行感兴趣区域对齐池化处理,得到感兴趣区域特征(即初始特征图在该检测框内的特征点的特征数据),进而通过掩膜预测网络中的卷积运算,即可以得到目标对象的掩膜,再对该掩膜进行尺寸扩大后,得到与待检测图像同样尺寸的掩膜图像(即目标对象的实例信息)。After the detection frame of the target object is obtained, the region of interest is extracted based on the detection frame of the target object and the initial feature map f, and then the region of interest is aligned and pooled to obtain the feature of the region of interest (ie, the initial feature The feature data of the feature points in the detection frame), and then through the convolution operation in the mask prediction network, the mask of the target object can be obtained, and then the size of the mask is enlarged, and the image to be detected is obtained The mask image of the same size (ie, the instance information of the target object).
通过本公开实施例提出的目标检测方法,可以输出目标对象的检测框、目标对象的掩膜以及目标对象的类别,可以根据预先设定好的要求,得到需要的结果,比如输出目标对象的检测框、或者输出目标对象的掩膜图像、或者既输出目标对象的检测框,也输出目标对象的掩膜图像,且同时输出目标对象的类别,在本公开实施例中不进行限定。Through the target detection method proposed in the embodiments of the present disclosure, the detection frame of the target object, the mask of the target object, and the target object category can be output, and the required results can be obtained according to the preset requirements, such as outputting the detection of the target object The frame, or the output of the mask image of the target object, or both the detection frame of the target object and the mask image of the target object, and the category of the target object are output at the same time, which are not limited in the embodiment of the present disclosure.
本公开实施例的目标检测方法可以由神经网络实现,神经网络利用包含了标注目标样本对象的样本图片训练得到。The target detection method in the embodiments of the present disclosure may be implemented by a neural network, which is obtained by training using sample pictures containing labeled target sample objects.
这里,如图11所示,本公开实施例提出的目标检测方法的神经网络可以采用以下步骤训练得到,包括步骤S1101~S1104:Here, as shown in FIG. 11, the neural network of the target detection method proposed in the embodiment of the present disclosure can be obtained by training using the following steps, including steps S1101 to S1104:
S1101,获取样本图像。S1101. Obtain a sample image.
这里的样本图像可以包括标注目标样本对象的正样本,以及不包含目标样本对象的负样本,且正样本中包含的目标对象可以包括多种类别。The sample image here may include a positive sample that annotates the target sample object, and a negative sample that does not include the target sample object, and the target object contained in the positive sample may include multiple categories.
这里,标注目标样本对象的正样本可以分为通过检测框标注的目标样本对象,以及通过掩膜标注的目标样本对象。Here, the positive samples labeled with the target sample objects can be divided into the target sample objects labeled with the detection frame and the target sample objects labeled with the mask.
S1102,基于样本图像,确定各个样本角点在样本图像中的角点位置信息及各个样本角点对应的向心偏移张量,样本角点表征样本图像中的目标样本对象的位置。S1102, based on the sample image, determine the corner position information of each sample corner point in the sample image and the centripetal offset tensor corresponding to each sample corner point, where the sample corner point represents the position of the target sample object in the sample image.
这里,基于样本图像,确定样本角点在样本图像中的角点位置信息,以及各个样本角点分别对应的向心偏移张量的过程,与上文提到的确定角点在待检测图像中的角点位置信息,以及各个角点分别对应的向心偏移张量的方式相似,在此不再赘述。Here, based on the sample image, the process of determining the corner position information of the sample corner point in the sample image and the centripetal offset tensor corresponding to each sample corner point is the same as the process of determining the corner point in the image to be detected as mentioned above. The corner position information in and the centripetal offset tensor corresponding to each corner are similar, so I won’t repeat them here.
S1103,基于各个样本角点在样本图像中的角点位置信息及各个样本角点对应的向心偏移张量,预测样本图像中的目标样本对象。S1103: Predict the target sample object in the sample image based on the corner position information of each sample corner point in the sample image and the centripetal offset tensor corresponding to each sample corner point.
这里,在样本图像中预测目标样本对象的过程与上文提到的在待检测图像中确定目标对象的方式相同,在此不再赘述。Here, the process of predicting the target sample object in the sample image is the same as the method of determining the target object in the image to be detected as mentioned above, and will not be repeated here.
S1104,基于预测的样本图像中的目标样本对象和样本图像中的标注目标样本对象,对神经网络的网络参数值进行调整。S1104: Adjust network parameter values of the neural network based on the predicted target sample object in the sample image and the labeled target sample object in the sample image.
这里,可以引入损失函数确定目标样本对象预测时对应的损失值,经过多次训练后,通过损失值来对神经网络的网络参数值进行调整,比如使得损失值小于设定阈值的情况下,即可以停止训练,从而得到神经网络的网络参数值。Here, a loss function can be introduced to determine the loss value corresponding to the target sample object prediction. After multiple training, the network parameter value of the neural network can be adjusted through the loss value, for example, when the loss value is less than the set threshold, that is You can stop training to get the network parameter values of the neural network.
另外,目标样本对象的检测框,以及目标样本对象的掩膜,以及目标样本对象的类别的确定过程,与上文介绍的确定目标对象的检测框、目标对象的掩膜以及目标对象的类别的过程相似,在此不再赘述。In addition, the detection frame of the target sample object, the mask of the target sample object, and the determination process of the target sample object's category are the same as the detection frame of the target object, the mask of the target object, and the category of the target object described above. The process is similar, so I won't repeat it here.
本公开实施例提供的神经网络的训练方法,通过获取样本图像,并基于该样本图像确定各个样本角点在样本图像中的角点位置信息,以及各个样本角点分别对应的向心偏移张量,从而基于各个样本角点在样本图像中的角点位置信息和各个样本角点对应的向心偏移张量,在样本图像中检测目标样本对象,因为样本角点是指图像中的主要特征点,比如样本角点可以包括左上样本角点和右下样本角点,其中左上样本角点是指对应目标样本对象上侧轮廓的直线与对应目标样本对象左侧轮廓的直线的交点,右下样本角点是指对应目标样本对象下侧轮廓的直线与对应目标样本对象右侧轮廓的直线的交点,在左上样本角点和右下样本角点属于同一个目标样本对象的检测框的情况下,左上样本角点和右下样本角点分别对应的向心偏移张量指向的位置应该比较接近,因此,本公开实施例提出的神经网络的训练方法,基于表征目标样本对象在待训练样本图像中的位置的角点位置信息,以及每个样本角点对应的向心偏移张量,确定出属于同一目标样本对象的样本角点,进而基于确定出的样本角点可以检测出该同一目标样本对象,然后通过不断地基于样本图像中的标注目标对象,不断调整神经网络参数,从而得到准确度较高的神经网络,基于该准确度较高的神经网络即可以对目标对象进行准确检测。The neural network training method provided by the embodiments of the present disclosure obtains a sample image, and based on the sample image, determines the corner position information of each sample corner point in the sample image, and the centripetal offset corresponding to each sample corner point. Based on the corner position information of each sample corner point in the sample image and the centripetal offset tensor corresponding to each sample corner point, the target sample object is detected in the sample image, because the sample corner point refers to the main Feature points, for example, the sample corner points can include the upper left sample corner point and the lower right sample corner point, where the upper left sample corner point refers to the intersection of the line corresponding to the upper contour of the target sample object and the line corresponding to the left contour of the target sample object. The lower sample corner point refers to the intersection of the straight line corresponding to the lower contour of the target sample object and the straight line corresponding to the right contour of the target sample object. The upper left sample corner and the lower right sample corner belong to the detection frame of the same target sample object. The positions of the centripetal offset tensor corresponding to the upper left sample corner point and the lower right sample corner point should be relatively close. Therefore, the training method of the neural network proposed in the embodiment of the present disclosure is based on the representation of the target sample object being trained The corner point position information of the position in the sample image, and the centripetal offset tensor corresponding to each sample corner point, determine the sample corner point belonging to the same target sample object, and then based on the determined sample corner point, the sample corner point can be detected. The same target sample object, and then continuously adjust the neural network parameters based on the target object in the sample image, so as to obtain a neural network with higher accuracy. Based on the neural network with higher accuracy, the target object can be accurately measured. Detection.
本领域技术人员可以理解,在上述方法中,各步骤的撰写顺序并不意味着严格的执行顺序而对实施过程构成任何限定,各步骤的执行顺序应当以其功能和可能的内在逻辑确定。Those skilled in the art can understand that in the above method, the writing order of the steps does not mean a strict execution order but constitutes any limitation on the implementation process. The execution order of the steps should be determined by their functions and possible internal logic.
基于同一技术构思,本公开实施例中还提供了与目标检测方法对应的目标检测装置,由于本公开实施例中的装置技术原理与本公开实施例上述目标检测方法相似,因此装置的实施可以参见方法的实施,重复之处不再赘述。Based on the same technical concept, the embodiment of the present disclosure also provides a target detection device corresponding to the target detection method. Since the technical principle of the device in the embodiment of the disclosure is similar to the target detection method described in the embodiment of the disclosure, the implementation of the device can be referred to The implementation of the method will not repeat the repetition.
参照图12所示,为本公开实施例提供的一种目标检测装置1200的示意图,装置包括:获取部分1201、确定部分1202、检测部分1203。Referring to FIG. 12, it is a schematic diagram of a target detection device 1200 provided by an embodiment of the present disclosure. The device includes: an acquisition part 1201, a determination part 1202, and a detection part 1203.
其中,获取部分1201,被配置为获取待检测图像;Wherein, the acquiring part 1201 is configured to acquire the image to be detected;
确定部分1202,被配置为基于待检测图像,确定各个角点在待检测图像中的角点位置信息以及各个角点对应的向心偏移张量,角点表征待检测图像中的目标对象的位置;The determining part 1202 is configured to determine, based on the image to be detected, the corner position information of each corner point in the image to be detected and the centripetal offset tensor corresponding to each corner point, where the corner points represent the target object in the image to be detected Location;
检测部分1203,被配置为基于各个角点在待检测图像中的角点位置信息及各个角点对应的向心偏 移张量,确定待检测图像中中的目标对象。The detection part 1203 is configured to determine the target object in the image to be detected based on the position information of each corner point in the image to be detected and the centripetal offset tensor corresponding to each corner point.
在一种可能的实施方式中,确定部分1202被配置为:In a possible implementation manner, the determining part 1202 is configured to:
对待检测图像进行特征提取,得到待检测图像对应的初始特征图;Perform feature extraction on the image to be detected to obtain an initial feature map corresponding to the image to be detected;
对初始特征图进行角点池化处理,得到角点池化后的特征图;Perform corner pooling processing on the initial feature map to obtain a feature map after corner pooling;
基于角点池化后的特征图,确定各个角点在待检测图像中的角点位置信息,以及各个角点对应的向心偏移张量。Based on the feature map after corner pooling, the corner position information of each corner point in the image to be detected and the centripetal offset tensor corresponding to each corner point are determined.
在一种可能的实施方式中,确定部分1202在被配置为基于角点池化后的特征图,确定各个角点在待检测图像中的角点位置信息的情况下,包括:In a possible implementation manner, when the determining part 1202 is configured to determine the corner position information of each corner in the image to be detected based on the feature map after corner pooling, the method includes:
基于角点池化后的特征图,生成待检测图像对应的角点热力图;Based on the feature map after corner pooling, generate the corner heat map corresponding to the image to be detected;
基于角点热力图,确定角点热力图中每个特征点作为角点的概率值,并基于每个特征点作为角点的概率值,从特征点中筛选出角点;Based on the corner heat map, determine the probability value of each feature point in the corner heat map as a corner point, and filter out the corner points from the feature points based on the probability value of each feature point as a corner point;
获取筛选出的各个角点在角点热力图中的位置信息、以及各个角点对应的局部偏移信息,局部偏移信息用于表示对应的角点所表征的真实物理点在角点热力图中的位置偏移信息;Obtain the position information of the selected corner points in the corner point heat map and the local offset information corresponding to each corner point. The local offset information is used to indicate that the real physical point represented by the corresponding corner point is in the corner heat map Position offset information in;
基于获取到的每个角点在角点热力图中的位置信息、各个角点对应的局部偏移信息、以及角点热力图和待检测图像之间的尺寸比例,确定各个角点在待检测图像中的角点位置信息。Based on the acquired position information of each corner point in the corner point heat map, the local offset information corresponding to each corner point, and the size ratio between the corner point heat map and the image to be detected, it is determined that each corner point is in the to-be-detected image. The position information of the corner points in the image.
在一种可能的实施方式中,确定部分1202在被配置为基于角点池化后的特征图,确定各个角点分别对应的向心偏移张量的情况下,包括:In a possible implementation manner, when the determining part 1202 is configured to determine the centripetal offset tensor corresponding to each corner point based on the feature map after corner point pooling, the method includes:
基于角点池化后的特征图,确定角点池化后的特征图中的每个特征点对应的导向偏移张量,每个特征点对应的导向偏移张量表征由该特征点指向待检测图像中的目标对象中心点的偏移张量;Based on the feature map after corner pooling, determine the steering offset tensor corresponding to each feature point in the feature map after corner pooling, and the steering offset tensor corresponding to each feature point is represented by the feature point. The offset tensor of the center point of the target object in the image to be detected;
基于每个特征点对应的导向偏移张量,确定该特征点的偏移域信息;偏移域信息中包含与该特征点关联的多个初始特征点分别指向各自对应的偏移后特征点的偏移张量;Based on the steering offset tensor corresponding to each feature point, determine the offset domain information of the feature point; the offset domain information contains multiple initial feature points associated with the feature point respectively pointing to their corresponding offset feature points The offset tensor;
基于角点池化后的特征图,以及该角点池化后的特征图中的特征点的偏移域信息,对角点池化后的特征图中的特征点的特征数据进行调整,得到调整后的特征图;Based on the corner point pooled feature map and the offset domain information of the feature points in the corner point pooled feature map, the feature data of the feature points in the corner point pooled feature map are adjusted to obtain The adjusted feature map;
基于调整后的特征图,确定各个角点对应的向心偏移张量。Based on the adjusted feature map, the centripetal offset tensor corresponding to each corner point is determined.
在一种可能的实施方式中,待检测图像对应的角点热力图包括多个通道分别对应的角点热力图,多个通道中的每个通道对应一种预设对象类别;确定部分1202在被配置为基于角点热力图,确定角点热力图中每个特征点作为角点的概率值之后,还被配置为:In a possible implementation, the corner heat map corresponding to the image to be detected includes a corner heat map corresponding to multiple channels, and each channel of the multiple channels corresponds to a preset object category; the determining part 1202 is in After being configured to determine the probability value of each feature point in the corner heat map as a corner point based on the corner heat map, it is also configured to:
针对多个通道中的每个通道,基于该通道对应的角点热力图中每个特征点作为角点的概率值,确定该通道对应的角点热力图中是否存在角点;For each channel in the multiple channels, determine whether there is a corner point in the corner heat map corresponding to the channel based on the probability value of each feature point in the corner heat map corresponding to the channel as a corner point;
在该通道对应的角点热力图中存在角点的情况下,确定待检测图像中包含该通道对应的预设对象类别的目标对象。When there are corner points in the corner point heat map corresponding to the channel, it is determined that the image to be detected contains the target object of the preset object category corresponding to the channel.
在一种可能的实施方式中,检测部分1203被配置为:In a possible implementation manner, the detection part 1203 is configured to:
基于各个角点在待检测图像中的角点位置信息和各个角点对应的向心偏移张量,确定待检测图像中目标对象的检测框;Determine the detection frame of the target object in the image to be detected based on the position information of each corner point in the image to be detected and the centripetal offset tensor corresponding to each corner point;
在一种可能的实施方式中,检测部分1203在被配置为基于各个角点在待检测图像中的角点位置信息和各个角点对应的向心偏移张量,确定待检测图像中目标对象的检测框的情况下,包括:In a possible implementation, the detection part 1203 is configured to determine the target object in the image to be detected based on the position information of each corner point in the image to be detected and the centripetal offset tensor corresponding to each corner point. The case of the detection box includes:
基于各个角点在待检测图像中的角点位置信息,筛选能够构成候选检测框的候选角点对;Based on the corner position information of each corner in the image to be detected, filter candidate corner pairs that can form a candidate detection frame;
基于每个候选角点对中每个角点在待检测图像中的角点位置信息和该角点对应的向心偏移张量,确定该角点指向的中心点位置信息;Based on the corner position information of each corner point in the image to be detected in each candidate corner point pair and the centripetal offset tensor corresponding to the corner point, determine the center point position information pointed to by the corner point;
基于每个候选角点对中每个角点在待检测图像中的角点位置信息,确定该候选角点对所对应的中心区域信息;Determine the center area information corresponding to the candidate corner point pair based on the corner point position information of each corner point in each candidate corner point pair in the image to be detected;
基于每个候选角点对中每个角点指向的中心点位置信息以及该候选角点对所对应的中心区域信息,在候选检测框中确定目标对象的检测框。Based on the position information of the center point pointed to by each corner point in each candidate corner point pair and the center area information corresponding to the candidate corner point pair, the detection frame of the target object is determined in the candidate detection frame.
在一种可能的实施方式中,针对每个候选角点对,检测部分1203在被配置为所述基于每个候选角点对中每个角点在所述待检测图像中的角点位置信息,确定该候选角点对所对应的中心区域信息的情况下,包括:In a possible implementation manner, for each candidate corner point pair, the detecting part 1203 is configured to be based on the corner position information of each corner point in each candidate corner point pair in the image to be detected. , In the case of determining the central region information corresponding to the candidate corner point pair, it includes:
基于该候选角点对的每个角点的角点位置信息,确定表征该候选角点对所对应的中心区域框的角点位置信息;Based on the corner point position information of each corner point of the candidate corner point pair, determine the corner point position information that characterizes the center area frame corresponding to the candidate corner point pair;
基于中心区域框的角点位置信息,确定该候选角点对所对应的中心区域框的坐标范围。Based on the corner position information of the central area frame, the coordinate range of the central area frame corresponding to the candidate corner point pair is determined.
在一种可能的实施方式中,检测部分1203在被配置为基于每个候选角点对中每个角点指向的中心点位置信息,以及该候选角点对所对应的中心区域信息,在候选检测框中确定目标对象的检测框的情况下,包括:In a possible implementation manner, the detection part 1203 is configured to be based on the position information of the center point pointed to by each corner point in each candidate corner point pair and the center area information corresponding to the candidate corner point pair. When the detection frame determines the detection frame of the target object, it includes:
基于每个候选角点对中每个角点指向的中心点位置信息,以及该候选角点对所对应的中心区域信息,确定有效候选角点对;Determine a valid candidate corner point pair based on the position information of the center point pointed to by each corner point in each candidate corner point pair, and the center area information corresponding to the candidate corner point pair;
基于有效候选角点对中每个角点指向的中心点位置信息、有效候选角点对所对应的中心区域信息、以及有效候选角点对中每个角点对应的概率值,确定每个有效候选角点对所对应的候选检测框的分值;每个角点对应的概率值用于表示该角点在角点热力图中对应的特征点作为角点的概率值;Based on the position information of the center point pointed to by each corner point in the valid candidate corner point pair, the central area information corresponding to the valid candidate corner point pair, and the probability value corresponding to each corner point in the valid candidate corner point pair, each valid The score value of the candidate detection frame corresponding to the candidate corner point; the probability value corresponding to each corner point is used to indicate the probability value of the corresponding feature point of the corner point in the corner heat map as the corner point;
基于每个有效候选角点对所对应的候选检测框的分值、以及相邻候选检测框之间的重叠区域大小,在候选检测框中确定目标对象的检测框。Based on the score of the candidate detection frame corresponding to each valid candidate corner point and the size of the overlapping area between adjacent candidate detection frames, the detection frame of the target object is determined in the candidate detection frame.
在一种可能的实施方式中,检测部分1203还被配置为:In a possible implementation manner, the detection part 1203 is further configured to:
在确定待检测图像中目标对象的检测框之后,基于目标对象的检测框和对待检测图像进行特征提取得到的初始特征图,确定待检测图像中目标对象的实例信息。After the detection frame of the target object in the image to be detected is determined, the instance information of the target object in the image to be detected is determined based on the detection frame of the target object and the initial feature map obtained by feature extraction of the image to be detected.
在一种可能的实施方式中,检测部分1203在被配置为基于目标对象的检测框和对待检测图像进行特征提取得到的初始特征图,确定待检测图像中目标对象的实例信息的情况下,包括:In a possible implementation, the detection part 1203 is configured to determine the instance information of the target object in the image to be detected based on the detection frame of the target object and the initial feature map obtained by feature extraction of the image to be detected, including :
基于目标对象的检测框以及初始特征图,提取初始特征图在检测框内的特征点的特征数据;Based on the detection frame and the initial feature map of the target object, extract the feature data of the feature points of the initial feature map in the detection frame;
基于初始特征图在检测框内的特征点的特征数据,确定待检测图像中目标对象的实例信息。Based on the feature data of the feature points of the initial feature map in the detection frame, the instance information of the target object in the image to be detected is determined.
在一种可能的实施方式中,目标检测装置1200还包括神经网络训练部分1204,神经网络训练部分1204被配置为:In a possible implementation manner, the target detection device 1200 further includes a neural network training part 1204, and the neural network training part 1204 is configured to:
训练用于进行目标检测的神经网络,神经网络利用包含了标注目标样本对象的样本图片训练得到。Train a neural network for target detection. The neural network is trained using sample images that contain labeled target sample objects.
在一种可能的实施方式中,神经网络训练部分1204被配置为按照以下步骤训练神经网络:In a possible implementation manner, the neural network training part 1204 is configured to train the neural network according to the following steps:
获取样本图像;Obtain sample images;
基于样本图像,确定各个样本角点在样本图像中的角点位置信息以及各个样本角点对应的向心偏移张量,样本角点表征样本图像中的目标样本对象的位置;Based on the sample image, determine the corner position information of each sample corner point in the sample image and the centripetal offset tensor corresponding to each sample corner point, and the sample corner point represents the position of the target sample object in the sample image;
基于各个样本角点在样本图像中的角点位置信息及各个样本角点对应的向心偏移张量,预测样本图像中的目标样本对象;Predict the target sample object in the sample image based on the corner position information of each sample corner point in the sample image and the centripetal offset tensor corresponding to each sample corner point;
基于预测的样本图像中的目标样本对象和样本图像中的标注目标样本对象,对神经网络的网络参数值进行调整。Based on the predicted target sample object in the sample image and the labeled target sample object in the sample image, the network parameter values of the neural network are adjusted.
在本公开实施例以及其他的实施例中,“部分”可以是部分电路、部分处理器、部分程序或软件等等,当然也可以是单元,还可以是模块也可以是非模块化的。In the embodiments of the present disclosure and other embodiments, "parts" may be parts of circuits, parts of processors, parts of programs or software, etc., of course, may also be units, modules, or non-modular.
对应于图2中的目标检测方法,本公开实施例还提供了一种电子设备1300,如图13所示,为本公开实施例提供的电子设备1300结构示意图,包括:Corresponding to the target detection method in FIG. 2, an embodiment of the present disclosure further provides an electronic device 1300. As shown in FIG. 13, a schematic structural diagram of the electronic device 1300 provided by the embodiment of the present disclosure includes:
处理器1301、存储器1302、和总线1303;存储器1302被配置为存储执行指令,包括内存13021和外部存储器13022;这里的内存13021也称内存储器,被配置为暂时存放处理器1301中的运算数据,以及与硬盘等外部存储器13022交换的数据,处理器1301通过内存13021与外部存储器13022进行数据交换,在电子设备1300运行的情况下,处理器1301与存储器1302之间通过总线1303通信,机器可读指令被处理器1301执行时执行如下处理:The processor 1301, the memory 1302, and the bus 1303; the memory 1302 is configured to store execution instructions, including the memory 13021 and the external memory 13022; the memory 13021 here is also called internal memory, and is configured to temporarily store the calculation data in the processor 1301, As well as data exchanged with external storage 13022 such as a hard disk, the processor 1301 exchanges data with the external storage 13022 through the memory 13021, and when the electronic device 1300 is running, the processor 1301 and the storage 1302 communicate through the bus 1303, which is machine readable When the instruction is executed by the processor 1301, the following processing is performed:
获取待检测图像;Obtain the image to be detected;
基于待检测图像,确定各个角点在待检测图像中的角点位置信息以及各个角点对应的向心偏移张量,角点表征待检测图像中的目标对象的位置;Based on the image to be detected, determine the corner position information of each corner point in the image to be detected and the centripetal offset tensor corresponding to each corner point, and the corner point represents the position of the target object in the image to be detected;
基于各个角点在待检测图像中的角点位置信息及各个角点对应的向心偏移张量,确定待检测图像中的目标对象。Based on the corner position information of each corner point in the image to be detected and the centripetal offset tensor corresponding to each corner point, the target object in the image to be detected is determined.
本公开实施例还提供一种计算机可读存储介质,该计算机可读存储介质上存储有计算机程序,该计算机程序被处理器运行时执行上述方法实施例中的目标检测方法的步骤。其中,该存储介质可以是易失性或非易失的计算机可读取存储介质。The embodiments of the present disclosure also provide a computer-readable storage medium having a computer program stored on the computer-readable storage medium, and the computer program executes the steps of the target detection method in the foregoing method embodiment when the computer program is run by a processor. Wherein, the storage medium may be a volatile or non-volatile computer readable storage medium.
本公开实施例还提供了一种计算机程序,包括计算机可读代码,在所述计算机可读代码在电子设备中运行的情况下,所述电子设备中的处理器执行时实现如第一方面所述的目标检测方法的步骤。The embodiments of the present disclosure also provide a computer program, including computer-readable code, and when the computer-readable code runs in an electronic device, the processor in the electronic device executes the same as described in the first aspect. The steps of the target detection method described.
本公开实施例所提供的目标检测方法的计算机程序产品,包括存储了程序代码的计算机可读存储介质,所述程序代码包括的指令可用于执行上述方法实施例中所述的目标检测方法的步骤,可参见上述方法实施例,在此不再赘述。The computer program product of the target detection method provided by the embodiment of the present disclosure includes a computer-readable storage medium storing program code, and the program code includes instructions that can be used to execute the steps of the target detection method described in the above method embodiment For details, please refer to the above method embodiment, which will not be repeated here.
本公开实施例还提供一种计算机程序,该计算机程序被处理器执行时实现前述实施例的任意一种方法。该计算机程序产品可以具体通过硬件、软件或其结合的方式实现。在一些实施例中,所述计算机程序产品具体体现为计算机存储介质,在另一些实施例中,计算机程序产品具体体现为软件产品,例如软件开发包(Software Development Kit,SDK)等等。The embodiments of the present disclosure also provide a computer program, which, when executed by a processor, implements any one of the methods in the foregoing embodiments. The computer program product can be specifically implemented by hardware, software, or a combination thereof. In some embodiments, the computer program product is specifically embodied as a computer storage medium. In other embodiments, the computer program product is specifically embodied as a software product, such as a software development kit (SDK) and so on.
所属领域的技术人员可以清楚地了解到,为描述的方便和简洁,上述描述的系统和装置的工作过程, 可以参考前述方法实施例中的对应过程,在此不再赘述。在本公开所提供的几个实施例中,应该理解到,所揭露的系统、装置和方法,可以通过其它的方式实现。以上所描述的装置实施例仅仅是示意性的,例如,所述单元的划分,仅仅为一种逻辑功能划分,实际实现时可以有另外的划分方式,又例如,多个单元或组件可以结合或者可以集成到另一个系统,或一些特征可以忽略,或不执行。另一点,所显示或讨论的相互之间的耦合或直接耦合或通信连接可以是通过一些通信接口,装置或单元的间接耦合或通信连接,可以是电性,机械或其它的形式。Those skilled in the art can clearly understand that, for convenience and concise description, the working process of the system and device described above can refer to the corresponding process in the foregoing method embodiment, which will not be repeated here. In the several embodiments provided in the present disclosure, it should be understood that the disclosed system, device, and method may be implemented in other ways. The device embodiments described above are merely illustrative. For example, the division of the units is only a logical function division, and there may be other divisions in actual implementation. For example, multiple units or components may be combined or It can be integrated into another system, or some features can be ignored or not implemented. In addition, the displayed or discussed mutual coupling or direct coupling or communication connection may be through some communication interfaces, indirect coupling or communication connection of devices or units, and may be in electrical, mechanical or other forms.
所述作为分离部件说明的单元可以是或者也可以不是物理上分开的,作为单元显示的部件可以是或者也可以不是物理单元,即可以位于一个地方,或者也可以分布到多个网络单元上。可以根据实际的需要选择其中的部分或者全部单元来实现本实施例方案的目的。The units described as separate components may or may not be physically separated, and the components displayed as units may or may not be physical units, that is, they may be located in one place, or they may be distributed on multiple network units. Some or all of the units may be selected according to actual needs to achieve the objectives of the solutions of the embodiments.
另外,在本公开各个实施例中的各功能单元可以集成在一个处理单元中,也可以是各个单元单独物理存在,也可以两个或两个以上单元集成在一个单元中。In addition, the functional units in the various embodiments of the present disclosure may be integrated into one processing unit, or each unit may exist alone physically, or two or more units may be integrated into one unit.
所述功能如果以软件功能单元的形式实现并作为独立的产品销售或使用时,可以存储在一个处理器可执行的非易失的计算机可读取存储介质中。基于这样的理解,本公开的技术方案本质上或者说对现有技术做出贡献的部分或者该技术方案的部分可以以软件产品的形式体现出来,该计算机软件产品存储在一个存储介质中,包括若干指令用以使得一台计算机设备(可以是个人计算机,服务器,或者网络设备等)执行本公开各个实施例所述方法的全部或部分步骤。而前述的存储介质包括:U盘、移动硬盘、只读存储器(Read-Only Memory,ROM)、随机存取存储器(Random Access Memory,RAM)、磁碟或者光盘等各种可以存储程序代码的介质。If the function is implemented in the form of a software functional unit and sold or used as an independent product, it can be stored in a non-volatile computer readable storage medium executable by a processor. Based on this understanding, the technical solution of the present disclosure essentially or the part that contributes to the prior art or the part of the technical solution can be embodied in the form of a software product, and the computer software product is stored in a storage medium, including Several instructions are used to make a computer device (which may be a personal computer, a server, or a network device, etc.) execute all or part of the steps of the methods described in the various embodiments of the present disclosure. The aforementioned storage media include: U disk, mobile hard disk, read-only memory (Read-Only Memory, ROM), random access memory (Random Access Memory, RAM), magnetic disk or optical disk and other media that can store program code .
最后应说明的是:以上所述实施例,仅为本公开的具体实施方式,用以说明本公开的技术方案,而非对其限制,本公开的保护范围并不局限于此,尽管参照前述实施例对本公开进行了详细的说明,本领域的普通技术人员应当理解:任何熟悉本技术领域的技术人员在本公开揭露的技术范围内,其依然可以对前述实施例所记载的技术方案进行修改或可轻易想到变化,或者对其中部分技术特征进行等同替换;而这些修改、变化或者替换,并不使相应技术方案的本质脱离本公开实施例技术方案的精神和范围,都应涵盖在本公开的保护范围之内。因此,本公开的保护范围应所述以权利要求的保护范围为准。Finally, it should be noted that the above-mentioned embodiments are only specific implementations of the present disclosure, which are used to illustrate the technical solutions of the present disclosure, rather than limit it. The protection scope of the present disclosure is not limited to this, although referring to the foregoing The embodiments describe the present disclosure in detail, and those of ordinary skill in the art should understand that any person skilled in the art can still modify the technical solutions described in the foregoing embodiments within the technical scope disclosed in the present disclosure. Or it can be easily conceived of changes, or equivalent replacements of some of the technical features; and these modifications, changes or replacements do not cause the essence of the corresponding technical solutions to deviate from the spirit and scope of the technical solutions of the embodiments of the present disclosure, and should be covered by the present disclosure. Within the scope of protection. Therefore, the protection scope of the present disclosure should be subject to the protection scope of the claims.
工业实用性Industrial applicability
本公开实施例提供了一种目标检测方法、装置、电子设备及计算机可读存储介质,其中,该目标检测方法包括:获取待检测图像;基于所述待检测图像,确定各个角点在所述待检测图像中的角点位置信息以及各个角点对应的向心偏移张量,角点表征所述待检测图像中的目标对象的位置;基于各个角点在所述待检测图像中的角点位置信息及各个角点对应的向心偏移张量,确定所述待检测图像中的目标对象。本公开实施例提出的目标检测方法,基于各个角点在待检测图像中的角点位置信息及各个角点对应的向心偏移张量,能够确定出属于同一目标对象的角点,进而基于确定出的角点可以检测出该同一目标对象。The embodiments of the present disclosure provide a target detection method, device, electronic equipment, and computer-readable storage medium. The target detection method includes: acquiring an image to be detected; and determining that each corner point is in the image based on the image to be detected. The position information of the corner points in the image to be detected and the centripetal offset tensor corresponding to each corner point. The corner points represent the position of the target object in the image to be detected; based on the angle of each corner point in the image to be detected The point position information and the centripetal offset tensor corresponding to each corner point are used to determine the target object in the image to be detected. The target detection method proposed in the embodiments of the present disclosure can determine the corner points belonging to the same target object based on the corner position information of each corner point in the image to be detected and the centripetal offset tensor corresponding to each corner point, and then based on The determined corner point can detect the same target object.

Claims (17)

  1. 一种目标检测方法,包括:A target detection method includes:
    获取待检测图像;Obtain the image to be detected;
    基于所述待检测图像,确定各个角点在所述待检测图像中的角点位置信息以及各个角点对应的向心偏移张量,角点表征所述待检测图像中的目标对象的位置;Based on the image to be detected, determine the position information of each corner point in the image to be detected and the centripetal offset tensor corresponding to each corner point, and the corner point represents the position of the target object in the image to be detected ;
    基于各个角点在所述待检测图像中的角点位置信息及各个角点对应的向心偏移张量,确定所述待检测图像中的目标对象。Based on the corner position information of each corner point in the image to be detected and the centripetal offset tensor corresponding to each corner point, the target object in the image to be detected is determined.
  2. 根据权利要求1所述的目标检测方法,其中,所述基于所述待检测图像,确定各个角点在待检测图像中的角点位置信息以及各个角点对应的向心偏移张量,包括:The target detection method according to claim 1, wherein the determining, based on the image to be detected, the position information of each corner point in the image to be detected and the centripetal offset tensor corresponding to each corner point comprises :
    对所述待检测图像进行特征提取,得到所述待检测图像对应的初始特征图;Performing feature extraction on the image to be detected to obtain an initial feature map corresponding to the image to be detected;
    对所述初始特征图进行角点池化处理,得到角点池化后的特征图;Performing corner pooling processing on the initial feature map to obtain a feature map after corner pooling;
    基于所述角点池化后的特征图,确定各个角点在所述待检测图像中的角点位置信息,以及各个角点对应的向心偏移张量。Based on the feature map after the corner point pooling, the corner point position information of each corner point in the image to be detected and the centripetal offset tensor corresponding to each corner point are determined.
  3. 根据权利要求2所述的目标检测方法,其中,所述基于所述角点池化后的特征图,确定各个角点在所述待检测图像中的角点位置信息,包括:The target detection method according to claim 2, wherein the determining the corner position information of each corner point in the image to be detected based on the feature map after the corner point pooling comprises:
    基于所述角点池化后的特征图,生成所述待检测图像对应的角点热力图;Generating a corner heat map corresponding to the image to be detected based on the feature map after corner pooling;
    基于所述角点热力图,确定所述角点热力图中每个特征点作为角点的概率值,并基于每个特征点作为角点的概率值,从所述角点热力图的特征点中筛选出所述角点;Based on the corner point heat map, determine the probability value of each feature point in the corner point heat map as a corner point, and based on the probability value of each feature point as a corner point, from the feature point of the corner point heat map To filter out the corner points;
    获取筛选出的各个角点在所述角点热力图中的位置信息、以及各个角点对应的局部偏移信息,所述局部偏移信息用于表示对应的角点所表征的真实物理点在所述角点热力图中的位置偏移信息;Obtain the position information of the selected corner points in the corner point heat map and the local offset information corresponding to each corner point. The local offset information is used to indicate that the real physical point represented by the corresponding corner point is at Position offset information in the corner point heat map;
    基于获取到的各个角点在所述角点热力图中的位置信息、各个角点对应的局部偏移信息、以及所述角点热力图和所述待检测图像之间的尺寸比例,确定各个角点在所述待检测图像中的角点位置信息。Based on the acquired position information of each corner point in the corner point heat map, the local offset information corresponding to each corner point, and the size ratio between the corner point heat map and the image to be detected, determine each The position information of the corner point in the image to be detected.
  4. 根据权利要求2或3所述的目标检测方法,其中,所述基于所述角点池化后的特征图,确定各个角点对应的向心偏移张量,包括:The target detection method according to claim 2 or 3, wherein the determining the centripetal offset tensor corresponding to each corner point based on the feature map after the corner point pooling comprises:
    基于所述角点池化后的特征图,确定所述角点池化后的特征图中的每个特征点对应的导向偏移张量,每个特征点对应的导向偏移张量表征由该特征点指向所述待检测图像中的目标对象中心点的偏移张量;Based on the feature map after corner point pooling, the steering offset tensor corresponding to each feature point in the corner point pooling feature map is determined, and the steering offset tensor corresponding to each feature point is represented by The offset tensor of the feature point pointing to the center point of the target object in the image to be detected;
    基于每个特征点对应的所述导向偏移张量,确定该特征点的偏移域信息;所述偏移域信息中包含与该特征点关联的多个初始特征点分别指向各自对应的偏移后特征点的偏移张量;Based on the steering offset tensor corresponding to each feature point, determine the offset domain information of the feature point; the offset domain information includes multiple initial feature points associated with the feature point respectively pointing to their corresponding offsets The offset tensor of the feature point after the shift;
    基于所述角点池化后的特征图,以及该角点池化后的特征图中的特征点的偏移域信息,对所述角点池化后的特征图中的特征点的特征数据进行调整,得到调整后的特征图;Based on the corner point pooled feature map and the offset domain information of the feature points in the corner point pooled feature map, the feature data of the feature points in the corner point pooled feature map Make adjustments to obtain the adjusted feature map;
    基于所述调整后的特征图,确定各个角点对应的向心偏移张量。Based on the adjusted feature map, the centripetal offset tensor corresponding to each corner point is determined.
  5. 根据权利要求3所述的目标检测方法,其中,所述待检测图像对应的角点热力图包括多个通道分别对应的角点热力图,所述多个通道中的每个通道对应一种预设对象类别;所述基于所述角点热力图,确定所述角点热力图中每个特征点作为角点的概率值之后,所述检测方法还包括:The target detection method according to claim 3, wherein the corner heat map corresponding to the image to be detected includes a corner heat map corresponding to a plurality of channels, and each channel of the plurality of channels corresponds to a preset Set the object category; after determining the probability value of each feature point in the corner heat map as a corner point based on the corner heat map, the detection method further includes:
    针对所述多个通道中的每个通道,基于该通道对应的角点热力图中每个特征点作为角点的概率值,确定该通道对应的角点热力图中是否存在所述角点;For each channel of the plurality of channels, determine whether the corner point exists in the corner heat map corresponding to the channel based on the probability value of each feature point in the corner heat map corresponding to the channel as a corner point;
    在该通道对应的角点热力图中存在所述角点的情况下,确定所述待检测图像中包含该通道对应的预设对象类别的目标对象。When the corner point exists in the corner point heat map corresponding to the channel, it is determined that the image to be detected contains the target object of the preset object category corresponding to the channel.
  6. 根据权利要求1所述的目标检测方法,其中,所述基于各个角点在所述待检测图像中的角点位置信息及各个角点对应的向心偏移张量,确定所述待检测图像中的目标对象,包括:The target detection method according to claim 1, wherein said determining the image to be detected based on the position information of each corner point in the image to be detected and the centripetal offset tensor corresponding to each corner point The target audience in includes:
    基于各个角点在所述待检测图像中的角点位置信息和各个角点对应的向心偏移张量,确定所述待检测图像中目标对象的检测框。Based on the corner position information of each corner point in the image to be detected and the centripetal offset tensor corresponding to each corner point, the detection frame of the target object in the image to be detected is determined.
  7. 根据权利要求6所述的目标检测方法,其中,所述基于各个角点在所述待检测图像中的角点位置信息和各个角点对应的向心偏移张量,确定所述待检测图像中目标对象的检测框,包括:The target detection method according to claim 6, wherein said determining the image to be detected based on the position information of each corner point in the image to be detected and the centripetal offset tensor corresponding to each corner point The detection frame of the target object in the middle, including:
    基于各个角点在所述待检测图像中的角点位置信息,筛选构成候选检测框的候选角点对;Based on the corner position information of each corner point in the image to be detected, screening candidate corner point pairs that constitute a candidate detection frame;
    基于每个候选角点对中每个角点在所述待检测图像中的角点位置信息和该角点对应的向心偏移张量,确定该角点指向的中心点位置信息;Determine the position information of the center point pointed to by the corner point based on the corner position information of each corner point in the image to be detected in each candidate corner point pair and the centripetal offset tensor corresponding to the corner point;
    基于每个候选角点对中每个角点在所述待检测图像中的角点位置信息,确定该候选角点对所对应的中心区域信息;Determine the center area information corresponding to the candidate corner point pair based on the corner point position information of each corner point in the candidate corner point pair in the image to be detected;
    基于每个候选角点对中每个角点指向的中心点位置信息,以及该候选角点对所对应的中心区域信息, 在所述候选检测框中确定所述目标对象的检测框。The detection frame of the target object is determined in the candidate detection frame based on the location information of the center point pointed to by each corner point in each candidate corner point pair and the center area information corresponding to the candidate corner point pair.
  8. 根据权利要求7所述的目标检测方法,其中,所述基于每个候选角点对中每个角点在所述待检测图像中的角点位置信息,确定该候选角点对所对应的中心区域信息,包括:8. The target detection method according to claim 7, wherein said determining the center corresponding to each candidate corner point pair based on the corner point position information of each corner point in the image to be detected in each candidate corner point pair Regional information, including:
    基于该候选角点对的每个角点的角点位置信息,确定表征该候选角点对所对应的中心区域框的角点位置信息;Based on the corner point position information of each corner point of the candidate corner point pair, determine the corner point position information that characterizes the center area frame corresponding to the candidate corner point pair;
    基于所述中心区域框的角点位置信息,确定该候选角点对所对应的中心区域框的坐标范围。Based on the corner point position information of the central area frame, the coordinate range of the central area frame corresponding to the candidate corner point pair is determined.
  9. 根据权利要求7或8所述的目标检测方法,其中,所述基于每个候选角点对中每个角点指向的中心点位置信息,以及该候选角点对所对应的中心区域信息,在所述候选检测框中确定所述目标对象的检测框,包括:The target detection method according to claim 7 or 8, wherein the position information of the center point pointed to by each corner point in each candidate corner point pair and the center area information corresponding to the candidate corner point pair are based on The determination of the detection frame of the target object in the candidate detection frame includes:
    基于每个候选角点对中每个角点指向的中心点位置信息,以及该候选角点对所对应的中心区域信息,确定有效候选角点对;Determine a valid candidate corner point pair based on the position information of the center point pointed to by each corner point in each candidate corner point pair, and the center area information corresponding to the candidate corner point pair;
    基于所述有效候选角点对中每个角点指向的中心点位置信息、所述有效候选角点对所对应的中心区域信息、以及所述有效候选角点对中每个角点对应的概率值,确定每个有效候选角点对所对应的候选检测框的分值;每个角点对应的概率值用于表示该角点在角点热力图中对应的特征点作为角点的概率值;Based on the position information of the center point pointed to by each corner point in the valid candidate corner point pair, the central area information corresponding to the valid candidate corner point pair, and the probability corresponding to each corner point in the valid candidate corner point pair Value, determine the score of the candidate detection frame corresponding to each valid candidate corner point; the probability value corresponding to each corner point is used to indicate the probability value of the corresponding feature point of the corner point in the corner heat map as the corner point ;
    基于每个有效候选角点对所对应的候选检测框的分值、以及相邻候选检测框之间的重叠区域大小,在所述候选检测框中确定所述目标对象的检测框。Based on the score of the candidate detection frame corresponding to each valid candidate corner point and the size of the overlapping area between adjacent candidate detection frames, the detection frame of the target object is determined in the candidate detection frame.
  10. 根据权利要求6所述的目标检测方法,其中,所述确定所述待检测图像中目标对象的检测框之后,所述目标检测方法还包括:The target detection method according to claim 6, wherein, after the determination of the detection frame of the target object in the image to be detected, the target detection method further comprises:
    基于所述目标对象的检测框和对所述待检测图像进行特征提取得到的初始特征图,确定所述待检测图像中所述目标对象的实例信息。The instance information of the target object in the image to be detected is determined based on the detection frame of the target object and the initial feature map obtained by feature extraction of the image to be detected.
  11. 根据权利要求10所述的目标检测方法,其中,所述基于所述目标对象的检测框和对所述待检测图像进行特征提取得到的初始特征图,确定所述待检测图像中所述目标对象的实例信息,包括:The target detection method according to claim 10, wherein the target object in the image to be detected is determined based on the detection frame of the target object and the initial feature map obtained by feature extraction of the image to be detected Instance information, including:
    基于所述目标对象的检测框以及所述初始特征图,提取所述初始特征图在所述检测框内的特征点的特征数据;Extracting feature data of feature points of the initial feature map in the detection frame based on the detection frame of the target object and the initial feature map;
    基于所述初始特征图在所述检测框内的特征点的特征数据,确定所述待检测图像中所述目标对象的实例信息。Based on the feature data of the feature points of the initial feature map in the detection frame, the instance information of the target object in the image to be detected is determined.
  12. 根据权利要求1至11任一所述的目标检测方法,其中,所述目标检测方法是由神经网络实现,所述神经网络利用包含了标注目标样本对象的样本图片训练得到。The target detection method according to any one of claims 1 to 11, wherein the target detection method is implemented by a neural network, and the neural network is obtained by training using sample pictures containing labeled target sample objects.
  13. 根据权利要求12所述的目标检测方法,其中,所述神经网络采用以下步骤训练得到:The target detection method according to claim 12, wherein the neural network is obtained by training in the following steps:
    获取样本图像;Obtain sample images;
    基于所述样本图像,确定各个样本角点在样本图像中的角点位置信息及各个样本角点对应的向心偏移张量,所述样本角点表征所述样本图像中的目标样本对象的位置;Based on the sample image, determine the corner position information of each sample corner point in the sample image and the centripetal offset tensor corresponding to each sample corner point, and the sample corner point represents the target sample object in the sample image. Location;
    基于各个样本角点在所述样本图像中的角点位置信息及各个样本角点对应的向心偏移张量,预测所述样本图像中的目标样本对象;Predicting the target sample object in the sample image based on the corner position information of each sample corner point in the sample image and the centripetal offset tensor corresponding to each sample corner point;
    基于预测的所述样本图像中的目标样本对象和所述样本图像中的标注目标样本对象,对所述神经网络的网络参数值进行调整。The network parameter value of the neural network is adjusted based on the predicted target sample object in the sample image and the labeled target sample object in the sample image.
  14. 一种目标检测装置,包括:A target detection device includes:
    获取部分,被配置为获取待检测图像;The obtaining part is configured to obtain the image to be detected;
    确定部分,被配置为基于所述待检测图像,确定各个角点在所述待检测图像中的角点位置信息以及各个角点对应的向心偏移张量,角点表征所述待检测图像中的目标对象的位置;The determining part is configured to determine, based on the image to be detected, the corner position information of each corner point in the image to be detected and the centripetal offset tensor corresponding to each corner point, where the corner points represent the image to be detected The location of the target object in
    检测部分,被配置为基于各个角点在所述待检测图像中的角点位置信息及各个角点对应的向心偏移张量,确定所述待检测图像中的目标对象。The detection part is configured to determine the target object in the image to be detected based on the position information of each corner point in the image to be detected and the centripetal offset tensor corresponding to each corner point.
  15. 一种电子设备,包括:处理器、存储器和总线,所述存储器存储有所述处理器可执行的机器可读指令,在电子设备运行的情况下,所述处理器与所述存储器之间通过总线通信,所述机器可读指令被所述处理器执行时执行如权利要求1至13任一所述的目标检测方法的步骤。An electronic device includes a processor, a memory, and a bus. The memory stores machine-readable instructions executable by the processor. When the electronic device is running, the processor and the memory pass through Bus communication, when the machine-readable instructions are executed by the processor, the steps of the target detection method according to any one of claims 1 to 13 are executed.
  16. 一种计算机可读存储介质,该计算机可读存储介质上存储有计算机程序,该计算机程序被处理器运行时执行如权利要求1至13任一所述的目标检测方法的步骤。A computer-readable storage medium having a computer program stored on the computer-readable storage medium, which executes the steps of the target detection method according to any one of claims 1 to 13 when the computer program is run by a processor.
  17. 一种计算机程序,包括计算机可读代码,在所述计算机可读代码在电子设备中运行的情况下,所述电子设备中的处理器执行时实现权利要求1至13中任意一项所述的方法。A computer program, comprising computer-readable code, when the computer-readable code runs in an electronic device, the processor in the electronic device executes the method described in any one of claims 1 to 13 when executed method.
PCT/CN2020/135967 2020-01-22 2020-12-11 Object detection method and apparatus, electronic device, and computer readable storage medium WO2021147563A1 (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
JP2021557733A JP2022526548A (en) 2020-01-22 2020-12-11 Target detection methods, devices, electronic devices and computer readable storage media
KR1020217030884A KR20210129189A (en) 2020-01-22 2020-12-11 Target detection method, apparatus, electronic device and computer readable storage medium

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN202010073142.6A CN111242088B (en) 2020-01-22 2020-01-22 Target detection method and device, electronic equipment and storage medium
CN202010073142.6 2020-01-22

Publications (1)

Publication Number Publication Date
WO2021147563A1 true WO2021147563A1 (en) 2021-07-29

Family

ID=70870017

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2020/135967 WO2021147563A1 (en) 2020-01-22 2020-12-11 Object detection method and apparatus, electronic device, and computer readable storage medium

Country Status (4)

Country Link
JP (1) JP2022526548A (en)
KR (1) KR20210129189A (en)
CN (1) CN111242088B (en)
WO (1) WO2021147563A1 (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113936458A (en) * 2021-10-12 2022-01-14 中国联合网络通信集团有限公司 Method, device, equipment and medium for judging congestion of expressway
CN115644933A (en) * 2022-11-17 2023-01-31 深圳微创踪影医疗装备有限公司 Catheter flushing control method and device, computer equipment and storage medium

Families Citing this family (14)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111242088B (en) * 2020-01-22 2023-11-28 上海商汤临港智能科技有限公司 Target detection method and device, electronic equipment and storage medium
CN111681284A (en) * 2020-06-09 2020-09-18 商汤集团有限公司 Corner point detection method and device, electronic equipment and storage medium
CN112215840A (en) * 2020-10-30 2021-01-12 上海商汤临港智能科技有限公司 Image detection method, image detection device, driving control method, driving control device, electronic equipment and storage medium
CN112270278A (en) * 2020-11-02 2021-01-26 重庆邮电大学 Key point-based blue top house detection method
CN112348894B (en) * 2020-11-03 2022-07-29 中冶赛迪重庆信息技术有限公司 Method, system, equipment and medium for identifying position and state of scrap steel truck
CN112733653A (en) * 2020-12-30 2021-04-30 智车优行科技(北京)有限公司 Target detection method and device, computer readable storage medium and electronic equipment
CN113822841B (en) * 2021-01-29 2022-05-20 深圳信息职业技术学院 Sewage impurity caking detection method and device and related equipment
CN112699856A (en) * 2021-03-24 2021-04-23 成都新希望金融信息有限公司 Face ornament identification method and device, electronic equipment and storage medium
CN113033539B (en) * 2021-03-30 2022-12-06 北京有竹居网络技术有限公司 Calligraphy practicing lattice detection method and device, readable medium and electronic equipment
CN113095228B (en) * 2021-04-13 2024-04-30 地平线(上海)人工智能技术有限公司 Method and device for detecting target in image and computer readable storage medium
CN113569911A (en) * 2021-06-28 2021-10-29 北京百度网讯科技有限公司 Vehicle identification method and device, electronic equipment and storage medium
CN113920538B (en) * 2021-10-20 2023-04-14 北京多维视通技术有限公司 Object detection method, device, equipment, storage medium and computer program product
CN113850238B (en) * 2021-11-29 2022-03-04 北京世纪好未来教育科技有限公司 Document detection method and device, electronic equipment and storage medium
CN116309587A (en) * 2023-05-22 2023-06-23 杭州百子尖科技股份有限公司 Cloth flaw detection method and device, electronic equipment and storage medium

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20180018503A1 (en) * 2015-12-11 2018-01-18 Tencent Technology (Shenzhen) Company Limited Method, terminal, and storage medium for tracking facial critical area
CN109670503A (en) * 2018-12-19 2019-04-23 北京旷视科技有限公司 Label detection method, apparatus and electronic system
CN110490256A (en) * 2019-08-20 2019-11-22 中国计量大学 A kind of vehicle checking method based on key point thermal map
CN111242088A (en) * 2020-01-22 2020-06-05 上海商汤临港智能科技有限公司 Target detection method and device, electronic equipment and storage medium

Family Cites Families (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20040259667A1 (en) * 2003-06-02 2004-12-23 Simon Berdugo Motorized image rotating target apparatus for all sports
US20100317466A1 (en) * 2009-05-24 2010-12-16 Semple Kerry J Miniature Kick Bag Game and Apparatus Kit
WO2014002554A1 (en) * 2012-06-29 2014-01-03 日本電気株式会社 Image processing device, image processing method, and program
CN106557940B (en) * 2015-09-25 2019-09-17 杭州海康威视数字技术股份有限公司 Information release terminal and method
CN106683091B (en) * 2017-01-06 2019-09-24 北京理工大学 A kind of target classification and attitude detecting method based on depth convolutional neural networks
CN108229307B (en) * 2017-11-22 2022-01-04 北京市商汤科技开发有限公司 Method, device and equipment for object detection
CN108446707B (en) * 2018-03-06 2020-11-24 北方工业大学 Remote sensing image airplane detection method based on key point screening and DPM confirmation
US10872406B2 (en) * 2018-04-13 2020-12-22 Taiwan Semiconductor Manufacturing Company, Ltd. Hot spot defect detecting method and hot spot defect detecting system
CN109801335A (en) * 2019-01-08 2019-05-24 北京旷视科技有限公司 Image processing method, device, electronic equipment and computer storage medium
CN110378891A (en) * 2019-07-24 2019-10-25 广东工业大学 A kind of hazardous material detection method, device and equipment based on terahertz image
CN110532894B (en) * 2019-08-05 2021-09-03 西安电子科技大学 Remote sensing target detection method based on boundary constraint CenterNet
CN110543838A (en) * 2019-08-19 2019-12-06 上海光是信息科技有限公司 Vehicle information detection method and device
CN110647931A (en) * 2019-09-20 2020-01-03 深圳市网心科技有限公司 Object detection method, electronic device, system, and medium

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20180018503A1 (en) * 2015-12-11 2018-01-18 Tencent Technology (Shenzhen) Company Limited Method, terminal, and storage medium for tracking facial critical area
CN109670503A (en) * 2018-12-19 2019-04-23 北京旷视科技有限公司 Label detection method, apparatus and electronic system
CN110490256A (en) * 2019-08-20 2019-11-22 中国计量大学 A kind of vehicle checking method based on key point thermal map
CN111242088A (en) * 2020-01-22 2020-06-05 上海商汤临港智能科技有限公司 Target detection method and device, electronic equipment and storage medium

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113936458A (en) * 2021-10-12 2022-01-14 中国联合网络通信集团有限公司 Method, device, equipment and medium for judging congestion of expressway
CN115644933A (en) * 2022-11-17 2023-01-31 深圳微创踪影医疗装备有限公司 Catheter flushing control method and device, computer equipment and storage medium
CN115644933B (en) * 2022-11-17 2023-08-22 深圳微创踪影医疗装备有限公司 Catheter flushing control method, catheter flushing control device, computer equipment and storage medium

Also Published As

Publication number Publication date
CN111242088A (en) 2020-06-05
JP2022526548A (en) 2022-05-25
CN111242088B (en) 2023-11-28
KR20210129189A (en) 2021-10-27

Similar Documents

Publication Publication Date Title
WO2021147563A1 (en) Object detection method and apparatus, electronic device, and computer readable storage medium
US10885365B2 (en) Method and apparatus for detecting object keypoint, and electronic device
US11443445B2 (en) Method and apparatus for depth estimation of monocular image, and storage medium
JP5952001B2 (en) Camera motion estimation method and apparatus using depth information, augmented reality system
US20180225527A1 (en) Method, apparatus, storage medium and device for modeling lane line identification, and method, apparatus, storage medium and device for identifying lane line
US9600898B2 (en) Method and apparatus for separating foreground image, and computer-readable recording medium
JP7051267B2 (en) Image detection methods, equipment, electronic equipment, storage media, and programs
WO2018025831A1 (en) People flow estimation device, display control device, people flow estimation method, and recording medium
WO2018054329A1 (en) Object detection method and device, electronic apparatus, computer program and storage medium
JP6345147B2 (en) Method for detecting an object in a pair of stereo images
CN106845338B (en) Pedestrian detection method and system in video stream
CN108229494B (en) Network training method, processing method, device, storage medium and electronic equipment
US20240037898A1 (en) Method for predicting reconstructabilit, computer device and storage medium
CN114511661A (en) Image rendering method and device, electronic equipment and storage medium
CN115797735A (en) Target detection method, device, equipment and storage medium
CN113343981A (en) Visual feature enhanced character recognition method, device and equipment
CN108229281B (en) Neural network generation method, face detection device and electronic equipment
JP7014005B2 (en) Image processing equipment and methods, electronic devices
CN111027551B (en) Image processing method, apparatus and medium
CN113255700B (en) Image feature map processing method and device, storage medium and terminal
CN110135474A (en) A kind of oblique aerial image matching method and system based on deep learning
CN115375742A (en) Method and system for generating depth image
JP2015219756A (en) Image comparison method, image comparison device, and program
Shiratori et al. Detection of pointing position by omnidirectional camera
CN113379838B (en) Method for generating roaming path of virtual reality scene and storage medium

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 20915404

Country of ref document: EP

Kind code of ref document: A1

ENP Entry into the national phase

Ref document number: 20217030884

Country of ref document: KR

Kind code of ref document: A

ENP Entry into the national phase

Ref document number: 2021557733

Country of ref document: JP

Kind code of ref document: A

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 20915404

Country of ref document: EP

Kind code of ref document: A1

32PN Ep: public notification in the ep bulletin as address of the adressee cannot be established

Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205A DATED 19/05/2023)

122 Ep: pct application non-entry in european phase

Ref document number: 20915404

Country of ref document: EP

Kind code of ref document: A1