CN110033424A - Method, apparatus, electronic equipment and the computer readable storage medium of image procossing - Google Patents
Method, apparatus, electronic equipment and the computer readable storage medium of image procossing Download PDFInfo
- Publication number
- CN110033424A CN110033424A CN201910313968.2A CN201910313968A CN110033424A CN 110033424 A CN110033424 A CN 110033424A CN 201910313968 A CN201910313968 A CN 201910313968A CN 110033424 A CN110033424 A CN 110033424A
- Authority
- CN
- China
- Prior art keywords
- target object
- candidate frame
- candidate
- preset threshold
- frame
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
- 238000000034 method Methods 0.000 title claims abstract description 47
- 238000012545 processing Methods 0.000 claims abstract description 52
- 230000001629 suppression Effects 0.000 claims abstract description 47
- 238000001514 detection method Methods 0.000 claims description 7
- 238000004040 coloring Methods 0.000 claims description 5
- 238000003672 processing method Methods 0.000 claims description 2
- 238000005516 engineering process Methods 0.000 abstract description 4
- 238000007781 pre-processing Methods 0.000 description 6
- 238000010586 diagram Methods 0.000 description 4
- 238000012549 training Methods 0.000 description 4
- 230000003287 optical effect Effects 0.000 description 3
- 230000008901 benefit Effects 0.000 description 2
- 230000008878 coupling Effects 0.000 description 2
- 238000010168 coupling process Methods 0.000 description 2
- 238000005859 coupling reaction Methods 0.000 description 2
- 230000006870 function Effects 0.000 description 2
- 238000012986 modification Methods 0.000 description 2
- 230000004048 modification Effects 0.000 description 2
- 230000008569 process Effects 0.000 description 2
- 230000003068 static effect Effects 0.000 description 2
- 241000282575 Gorilla Species 0.000 description 1
- 238000013528 artificial neural network Methods 0.000 description 1
- 230000006835 compression Effects 0.000 description 1
- 238000007906 compression Methods 0.000 description 1
- 238000004590 computer program Methods 0.000 description 1
- 238000013135 deep learning Methods 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 238000009432 framing Methods 0.000 description 1
- 238000003702 image correction Methods 0.000 description 1
- 238000007689 inspection Methods 0.000 description 1
- 239000011159 matrix material Substances 0.000 description 1
- 239000000126 substance Substances 0.000 description 1
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T5/00—Image enhancement or restoration
- G06T5/50—Image enhancement or restoration using two or more images, e.g. averaging or subtraction
Landscapes
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Image Analysis (AREA)
Abstract
The embodiment of the present application provides method, apparatus, electronic equipment and the computer readable storage medium of a kind of image procossing, is related to field of computer technology.This method comprises: determining the corresponding location information of each candidate frame in image to be processed, candidate frame is the candidate frame for target object, it is then based on image and the corresponding location information of each candidate frame to be processed, and determine the attribute information of the corresponding target object of each candidate frame respectively by preset model, then by non-maxima suppression algorithm and the attribute information based on the corresponding location information of each candidate frame and the corresponding target object of each candidate frame, processing is merged to each candidate frame.The embodiment of the present application realizes the accuracy for improving and merging processing to each candidate frame, promotes user experience.
Description
Technical field
This application involves field of computer technology, specifically, this application involves a kind of method, apparatus of image procossing,
Electronic equipment and computer readable storage medium.
Background technique
With the development of information technology, image processing techniques also develops therewith, is related to a kind of figure in the image processing arts
As Processing Algorithm, i.e. non-maxima suppression algorithm (Non-maximum suppression, NMS), the essence of NMS is search office
Portion's maximum inhibits non-maximum element.In deep learning target detection, usually it is directed to output result (for same
There may be several very close candidate frames to be corresponding to it for one target object) non-maxima suppression processing is carried out, by phase
As candidate frame merge.
It is handled in the prior art based on non-maxima suppression, in a manner of merging similar candidate frame: being based on appointing
The corresponding location information of each candidate frame, merges candidate frame in one image, the corresponding position of any candidate frame
Confidence breath includes: center position, the upper left corner location of candidate frame, the length of candidate frame and/or the candidate frame of candidate frame
Width etc., for example, successively merging the candidate frame that distance between center position is less than preset threshold.
But non-maxima suppression method in the prior art is only the location information based on each candidate frame to candidate frame
It merges, merges into a time so as to cause to be closer in the picture and belong to the candidate frame of different target object
The case where selecting frame, or the multiple candidate frames for belonging to same target object be merged into more than one frame, and then cause to image
In each candidate frame merge processing accuracy it is lower, user experience is poor.
Summary of the invention
This application provides a kind of method, apparatus of image procossing, electronic equipment and computer readable storage mediums, are used for
Solve the technical problem that the accuracy for merging processing to candidate frame each in image is lower and user experience is poor.
In a first aspect, a kind of method of image procossing is provided, this method comprises:
Determine that the corresponding location information of each candidate frame in image to be processed, candidate frame are for target object
Candidate frame;
Based on image to be processed and the corresponding location information of each candidate frame, and distinguished by preset model true
Determine the attribute information of the corresponding target object of each candidate frame;
Pass through non-maxima suppression algorithm and is based on the corresponding location information of each candidate frame and each candidate frame
The attribute information of corresponding target object merges processing to each candidate frame.
In one possible implementation, it for any two candidate frame, by non-maxima suppression algorithm and is based on
The attribute information of each corresponding location information of candidate frame and the corresponding target object of each candidate frame, to each candidate
Frame merges processing, comprising:
The friendship of any two candidate frame is determined based on the corresponding location information of each candidate frame and compares IoU;
By non-maxima suppression algorithm and based on each candidate in the IoU of any two candidate frames and any two candidate frames
The attribute information of the corresponding target object of frame, determines whether any two candidate frames belong to pending combined candidate frame;
If belonging to pending combined candidate frame, processing is merged to any two candidate frame.
In one possible implementation, the corresponding position letter of each candidate frame in image to be processed is determined
Breath, before further include:
Multimedia messages are pre-processed, image to be processed is obtained;
Determine the corresponding location information of each candidate frame in image to be processed, comprising:
By image to be processed by detection network, the corresponding position of each candidate frame in image to be processed is exported
Information.
In one possible implementation, the attribute information of the corresponding target object of any candidate frame include it is following at least
One:
The colouring information of target object;The density information of target object;The shape information of target object;Belonging to target object
Classification information.
In alternatively possible implementation, if candidate frame is the candidate frame for face information, any candidate frame
The attribute information of corresponding target object includes at least one of the following:
The age information of target object;The gender information of target object;The skin color information of target object;Target object
Towards radian information.
In one possible implementation, if candidate frame is the candidate frame for face information,
By non-maxima suppression algorithm, and based on each time in the IoU of any two candidate frames and any two candidate frames
The attribute information for selecting the corresponding target object of frame, determines whether any two candidate frames belong to pending combined candidate frame,
Include:
By non-maxima suppression algorithm, the IoU of any two candidate frames is determined, each candidate frame point in any two candidate frames
Whether the attribute information of not corresponding face information meets preset rules, to determine whether any two candidate frames belong to pending conjunction
And candidate frame;
Preset rules include at least one of the following:
IoU is greater than the first attribute information phase of the first preset threshold and the corresponding target object of any two candidate frames
Together;
IoU is not greater than the first attribute information of the second preset threshold and the corresponding target object of any two candidate frames
Together, the first preset threshold is less than the second preset threshold;
IoU is greater than the first preset threshold, and the first attribute information of the corresponding target object of any two candidate frames is
The probability difference of preset attribute information is away from no more than third predetermined threshold value;
IoU is greater than the second preset threshold, and the first attribute information of the corresponding target object of any two candidate frames is
The probability difference of preset attribute information is away from not less than third predetermined threshold value;
IoU be greater than the first preset threshold, and the second attribute information of the corresponding target object of any two candidate frames it
Difference is not more than the 4th preset threshold;
IoU be greater than the second preset threshold, and the second attribute information of the corresponding target object of any two candidate frames it
Difference is greater than the 4th preset threshold;
IoU is greater than the 5th preset threshold, and the 5th preset threshold is that the first preset threshold is respectively corresponded with any two candidate frames
Target object the first attribute information difference between weighted information;
IoU is greater than the 6th preset threshold, and the 6th preset threshold is that the first preset threshold is respectively corresponded with any two candidate frames
Target object the first attribute information difference and the corresponding target object of any two candidate frames the second attribute information
For the weighted information between the probability difference of preset attribute information;
IoU is greater than the 7th preset threshold, and the 7th preset threshold is that the first preset threshold is respectively corresponded with any two candidate frames
Third attribute information difference between weighted information.
In alternatively possible implementation, preset rules are included at least one of the following:
IoU is greater than the first preset threshold and the gender information of the corresponding target object of any two candidate frames is identical;
IoU is greater than the second preset threshold and the gender information of the corresponding target object of any two candidate frames is different, the
One preset threshold is less than the second preset threshold;
IoU is greater than the first preset threshold, and the gender of the corresponding target object of any two candidate frames is that the same sex is other
Probability difference is away from no more than third predetermined threshold value;
IoU is greater than the second preset threshold, and the gender of the corresponding target object of any two candidate frames is that the same sex is other
Probability difference is away from greater than third predetermined threshold value;
IoU is greater than the first preset threshold, and the age gap of the corresponding target object of any two candidate frames is not more than
4th preset threshold;
IoU is greater than the second preset threshold, and the age gap of the corresponding target object of any two candidate frames is greater than the
Four preset thresholds;
IoU is greater than the 5th preset threshold, the 5th preset threshold=first the+the first parameter value of preset threshold × any two candidates
The age gap of the corresponding target object of frame;
IoU is greater than the 6th preset threshold, the 6th preset threshold=first the+the first parameter value of preset threshold × any two candidates
The+the second parameter value of age gap of the corresponding target object of frame × any two candidate frames corresponding target object property
Not Wei the other probability difference of the same sex away from;
IoU is greater than the 7th preset threshold, the 7th preset threshold=first the+the second parameter value of preset threshold × any two candidates
Difference between the radian value of the corresponding target object direction of frame.
Second aspect provides a kind of image processing apparatus, which includes:
First determining module, it is described for determining the corresponding location information of each candidate frame in image to be processed
Candidate frame is the candidate frame for target object;
Second determining module, each time for being determined based on the image to be processed and first determining module
The corresponding location information of frame is selected, and determines the category of the corresponding target object of each candidate frame respectively by preset model
Property information;
Merging treatment module, for by non-maxima suppression algorithm and based on the corresponding position letter of each candidate frame
The attribute information of breath and the corresponding target object of each candidate frame, merges processing to each candidate frame.
In one possible implementation, for any two candidate frame, merging treatment module, comprising: first determines
Unit, the second determination unit and merge processing unit, wherein
First determination unit, for determining any two candidate frame based on the corresponding location information of each candidate frame
It hands over and compares IoU;
Second determination unit, any two times for passing through non-maxima suppression algorithm and being determined based on the first determination unit
The attribute information for selecting the corresponding target object of each candidate frame in the IoU and any two candidate frames of frame, determines any two
Whether candidate frame belongs to pending combined candidate frame;
Merge processing unit, for determining that any two candidate frames belong to pending combined candidate when the second determination unit
Frame then merges processing to any two candidate frame.
In one possible implementation, device further include: preprocessing module, wherein
Preprocessing module obtains image to be processed for pre-processing to multimedia messages;
First determining module, specifically for by detection network, determining to be processed the image to be processed
The corresponding location information of each candidate frame in image.
In one possible implementation, the attribute information of the corresponding target object of any candidate frame include it is following at least
One:
The colouring information of target object;The density information of target object;The shape information of target object;Belonging to target object
Classification information.
In one possible implementation, if candidate frame is the candidate frame for face information, any candidate frame pair
The attribute information for the target object answered includes at least one of the following:
The age information of target object;The gender information of target object;The skin color information of target object;Target object
Towards radian information.
In alternatively possible implementation, when candidate frame is the candidate frame for face information, second determines list
Member is specifically used for determining by non-maxima suppression algorithm the IoU of any two candidate frames, each candidate in any two candidate frames
Whether the attribute information of the corresponding face information of frame meets preset rules, with determine any two candidate frames whether belong to into
The combined candidate frame of row;
Preset rules include at least one of the following:
IoU is greater than the first attribute information phase of the first preset threshold and the corresponding target object of any two candidate frames
Together;
IoU is not greater than the first attribute information of the second preset threshold and the corresponding target object of any two candidate frames
Together, the first preset threshold is less than the second preset threshold;
IoU is greater than the first preset threshold, and the first attribute information of the corresponding target object of any two candidate frames is
The probability difference of preset attribute information is away from no more than third predetermined threshold value;
IoU is greater than the second preset threshold, and the first attribute information of the corresponding target object of any two candidate frames is
The probability difference of preset attribute information is away from no more than third predetermined threshold value;
IoU be greater than the first preset threshold, and the second attribute information of the corresponding target object of any two candidate frames it
Difference is not more than the 4th preset threshold;
IoU be greater than the second preset threshold, and the second attribute information of the corresponding target object of any two candidate frames it
Difference is greater than the 4th preset threshold;
IoU is greater than the 5th preset threshold, and the 5th preset threshold is the of the corresponding target object of any two candidate frames
The weighted information of one attribute information difference;
IoU is greater than the 6th preset threshold, and the 6th preset threshold is the of the corresponding target object of any two candidate frames
Second attribute information of one attribute information difference and the corresponding target object of any two candidate frames is preset attribute information
Probability difference between weighted information;
IoU is greater than the 7th preset threshold, and the 7th preset threshold is the corresponding third attribute information of any two candidate frames
The weighted information of difference.
In alternatively possible implementation, preset rules are included at least one of the following:
IoU is greater than the first preset threshold and the gender information of the corresponding target object of any two candidate frames is identical;
IoU is greater than the second preset threshold and the gender information of the corresponding target object of any two candidate frames is different, the
One preset threshold is less than the second preset threshold;
IoU is greater than the first preset threshold, and the gender of the corresponding target object of any two candidate frames is that the same sex is other
Probability difference is away from no more than third predetermined threshold value;
IoU is greater than the second preset threshold, and the gender of the corresponding target object of any two candidate frames is that the same sex is other
Probability difference is away from greater than third predetermined threshold value;
IoU is greater than the first preset threshold, and the age gap of the corresponding target object of any two candidate frames is not more than
4th preset threshold;
IoU is greater than the second preset threshold, and the age gap of the corresponding target object of any two candidate frames is greater than the
Four preset thresholds;
IoU is greater than the 5th preset threshold, the 5th preset threshold=first the+the first parameter value of preset threshold × any two candidates
The age gap of the corresponding target object of frame;
IoU is greater than the 6th preset threshold, the 6th preset threshold=first the+the first parameter value of preset threshold × any two candidates
The+the second parameter value of age gap of the corresponding target object of frame × any two candidate frames corresponding target object property
Not Wei the other probability difference of the same sex away from;
IoU is greater than the 7th preset threshold, the 7th preset threshold=first the+the second parameter value of preset threshold × any two candidates
Difference between the radian value of the corresponding target object direction of frame.
The third aspect provides a kind of electronic equipment, which includes:
One or more processors;
Memory;
One or more application program, wherein one or more application programs be stored in memory and be configured as by
One or more processors execute, and one or more programs are configured to: executing times according to first aspect or first aspect
The corresponding operation of method of image procossing shown in one possible implementation.
Fourth aspect, provides a kind of computer readable storage medium, and storage medium is stored at least one instruction, at least
One Duan Chengxu, code set or instruction set, at least one instruction, an at least Duan Chengxu, code set or instruction set are loaded by processor
And the method executed to realize the image procossing as shown in first aspect or first aspect any possible implementation.
Technical solution provided by the present application has the benefit that
It is and existing this application provides a kind of method, apparatus of image procossing, electronic equipment and computer readable storage medium
Non-maxima suppression method in technology is only that the location information based on each candidate frame merges candidate frame and compares, this
The corresponding location information of each candidate frame in image to be processed is determined in application, candidate frame is the time for target object
Frame is selected, is then based on image and the corresponding location information of each candidate frame to be processed, and distinguish by preset model
It determines the attribute information of the corresponding target object of each candidate frame, then pass through non-maxima suppression algorithm and is based on each candidate
The attribute information of the corresponding location information of frame and the corresponding target object of each candidate frame, closes each candidate frame
And it handles.It when i.e. the application merges processing for the candidate frame in image to be processed, is distinguished based on each candidate frame
The attribute information of corresponding location information and the corresponding target object of each candidate frame, will be in image so as to reduce
Be closer and the candidate frame for belonging to different target object be incorporated into the probability of a candidate frame, and can also reduce will belong to it is same
The case where multiple candidate frames of one target object are merged into more than one frame, and then can be improved and each candidate frame is merged
The accuracy of processing, and then user experience can be promoted.
Detailed description of the invention
In order to more clearly explain the technical solutions in the embodiments of the present application, institute in being described below to the embodiment of the present application
Attached drawing to be used is needed to be briefly described.
Fig. 1 is a kind of flow diagram of the method for image procossing provided by the embodiments of the present application;
Fig. 2 is a kind of apparatus structure schematic diagram of image procossing provided by the embodiments of the present application;
Fig. 3 is a kind of structural schematic diagram of the electronic equipment of image procossing provided by the embodiments of the present application;
Fig. 4 is a kind of overall schematic of the method for image procossing provided by the embodiments of the present application;
Fig. 5 is the form of expression exemplary diagram of candidate frame in the embodiment of the present application.
Specific embodiment
Embodiments herein is described below in detail, examples of the embodiments are shown in the accompanying drawings, wherein from beginning to end
Same or similar label indicates same or similar element or element with the same or similar functions.Below with reference to attached
The embodiment of figure description is exemplary, and is only used for explaining the application, and is not construed as limiting the claims.
Those skilled in the art of the present technique are appreciated that unless expressly stated, singular " one " used herein, " one
It is a ", " described " and "the" may also comprise plural form.It is to be further understood that being arranged used in the description of the present application
Diction " comprising " refer to that there are the feature, integer, step, operation, element and/or component, but it is not excluded that in the presence of or addition
Other one or more features, integer, step, operation, element, component and/or their group.It should be understood that when we claim member
Part is " connected " or when " coupled " to another element, it can be directly connected or coupled to other elements, or there may also be
Intermediary element.In addition, " connection " used herein or " coupling " may include being wirelessly connected or wirelessly coupling.It is used herein to arrange
Diction "and/or" includes one or more associated wholes for listing item or any cell and all combinations.
To keep the purposes, technical schemes and advantages of the application clearer, below in conjunction with attached drawing to the application embodiment party
Formula is described in further detail.
How the technical solution of the application and the technical solution of the application are solved with specifically embodiment below above-mentioned
Technical problem is described in detail.These specific embodiments can be combined with each other below, for the same or similar concept
Or process may repeat no more in certain embodiments.Below in conjunction with attached drawing, embodiments herein is described.
The embodiment of the present application provides a kind of method of image procossing, as shown in Figure 1, this method comprises:
Step S101, the corresponding location information of each candidate frame in image to be processed is determined.
Wherein, candidate frame is the candidate frame for target object.
For the embodiment of the present application, it can also include: that multimedia messages are pre-processed before step S101, obtain
Image to be processed.
For the embodiment of the present application, candidate circle favored area is region shared by the target object that identifies.
For the embodiment of the present application, candidate frame can be the mark letter for indicating the region of target object in the picture
Breath, is identified by arbitrary shape, for example, rectangle, square etc..For example, as shown in figure 5, the square-shaped frame in image is
Candidate frame, for indicating the face area of gorilla in the images.
Step S102, based on image to be processed and the corresponding location information of each candidate frame, and by default
Model determines the attribute information of the corresponding target object of each candidate frame respectively.
For the embodiment of the present application, after obtaining the corresponding location information of each candidate frame, by each candidate frame
Corresponding location information and image to be processed be input to network model, export the corresponding target of each candidate frame
The attribute information of object.
The category of the corresponding target object of each candidate frame is determined respectively by preset model for the embodiment of the present application
It can also include: to be trained to preset model before property information.Specifically, training sample is obtained, and passes through training sample,
Preset model is trained, training sample includes: image, the candidate frame for target object that magnanimity includes target object
The attribute information of location information and target object.
Step S103, by non-maxima suppression algorithm and based on the corresponding location information of each candidate frame and respectively
The attribute information of the corresponding target object of a candidate frame, merges processing to each candidate frame.
For the embodiment of the present application, by non-maxima suppression algorithm and it is based on the corresponding location information of each candidate frame
And the attribute information of the corresponding target object of each candidate frame, determine the candidate that merging is able to carry out in each candidate frame
Frame, and merging treatment is carried out, the location information of the candidate frame after being merged.
For the embodiment of the present application, the corresponding location information of any candidate frame may include at least one of following: this is any
The center position information of candidate frame;The corresponding location information in each vertex;Any candidate frame and current candidate frame it
Between hand over and compare IoU, wherein hand over and than (Intersection-over-Union, IoU), used in target detection one it is general
It reads, is the overlapping rate of the candidate frame generated and former indicia framing, the i.e. ratio of their intersection and union, most ideally completely
Overlapping, i.e. ratio are 1.
For example, the IoU between any candidate frame A and current candidate frame B isIts
In, area (A) is the corresponding region candidate frame A, the region of the corresponding current candidate frame B of area (B).
For the embodiment of the present application, the attribute information of the corresponding target object of any candidate frame can by matrix or to
The form of amount or discrete data indicates.In the embodiment of the present application without limitation.
For the embodiment of the present application, through the above technical solutions, being merged to multiple candidate frames for target object
Processing, can carry out the tracking etc. of target image based on the candidate frame after merging, merge place to each candidate frame improving
When the accuracy of reason, it can also be improved the target object such as to be tracked and handled based on the candidate frame for target object
The accuracy of mode, and then promote user experience.
The embodiment of the present application provides a kind of method of image procossing, only with non-maxima suppression method in the prior art
It is that the location information based on each candidate frame merges candidate frame and compares, image to be processed is determined in the embodiment of the present application
In the corresponding location information of each candidate frame, candidate frame is the candidate frame for target object, is then based on to be processed
Image and the corresponding location information of each candidate frame, and the corresponding mesh of each candidate frame is determined by preset model respectively
Mark object attribute information, then by non-maxima suppression algorithm and based on the corresponding location information of each candidate frame with
And the attribute information of the corresponding target object of each candidate frame, processing is merged to each candidate frame.That is the embodiment of the present application
When merging processing for the candidate frame in image to be processed, be based on the corresponding location information of each candidate frame with
And the attribute information of the corresponding target object of each candidate frame, it will be closer and belong to not in image so as to reduce
It is incorporated into the probability of a candidate frame with the candidate frame of target object, and can also reduce and will belong to the multiple of same target object
Candidate frame is merged into the case where more than one frame, and then can be improved and merge the accurate of processing to candidate frame each in image
Degree, and then user experience can be promoted.
A kind of possible implementation of the embodiment of the present application, for any two candidate frame, step S103 specifically can be with
It include: step S1031 (not shown), step S1032 (not shown) and step S1033 (not shown),
In,
Step S1031, friendship and the ratio of any two candidate frame are determined based on the corresponding location information of each candidate frame
IoU。
Step S1032, by non-maxima suppression algorithm and based on the IoU of any two candidate frames and any two candidate frames
In the corresponding target object of each candidate frame attribute information, determine whether any two candidate frames belong to pending merging
Candidate frame.
If step S1033, belonging to pending combined candidate frame, processing is merged to any two candidate frame.
For the embodiment of the present application, by non-maxima suppression algorithm and based on each time in IoU and any two candidate frames
The attribute information for selecting the corresponding target object of frame merges processing to any two candidate frame, can reduce image
In be closer and belong to the candidate frame of different target object and be incorporated into the probability of a candidate frame, and can also reduce and will belong to
The case where multiple candidate frames of same target object are merged into more than one frame, and then can be improved to candidate frame each in image
The accuracy of processing is merged, and then user experience can be promoted.
For the embodiment of the present application, when merging processing for multiple candidate frames, one can be first chosen with reference to time
Select frame, determine any one other candidate frame and this with reference to the friendship between candidate frame and than IoU, and determine other any candidates
The attribute information of frame and the corresponding target object of reference candidate frame, however, it is determined that can merge, then merge, and will be after merging
Candidate frame be used as and refer to candidate frame, determine the IoU with reference between candidate frame and other any candidate frames, and determine this its
The attribute information of its any candidate frame and the corresponding target object of reference candidate frame, however, it is determined that it can merge, then merge,
Circulation is executed using the candidate frame after merging as candidate frame is referred to, and determines that this is referred between candidate frame and other any candidate frames
IoU, and determine other any candidate frames and the attribute information with reference to the corresponding target object of candidate frame, however, it is determined that
The step of can merging, then merge, until meeting preset condition.
For example, for target object include 4 candidate frames, candidate frame 1, candidate frame 2, candidate frame 3 and candidate frame 4,
Middle candidate frame 1 of first choosing is with reference to candidate frame and to determine the IoU between candidate frame 1 and candidate frame 2 and candidate frame 1 and rear choosing
The attribute information of the corresponding target object of frame 2 merges candidate frame 1 and candidate frame 2 if can merge, then
Determine the IoU between the candidate frame and candidate frame 3 after merging, and the candidate frame after merging selects the corresponding mesh of frame 3 with after
The attribute information of mark object merges if can merge, and then determines between candidate frame and candidate frame 4 after merging
IoU, and candidate frame after merging select the attribute information of the corresponding target object of frame 4 will conjunction if can merge with after
Candidate frame after and is merged with candidate frame 4, the location information of candidate frame after finally being merged.
For the embodiment of the present application, if finally there are the candidate frame that can not merge processing, from can not merge
Target candidate frame is determined in the candidate frame of processing.In the embodiment of the present application, the location information of target candidate frame is used for final table
Levy the location information of target object in the picture.
For example, the candidate frame after merging in above-mentioned example (is merged by candidate frame 1, candidate frame 2 and candidate frame 3 and is post-processed
Candidate frame afterwards) with candidate frame 4 can not merge processing, then from the candidate frame and candidate frame 4 after merging determination target
Candidate frame (can be the candidate frame or candidate frame 4 after merging).
For the embodiment of the present application, the corresponding IoU of any two candidate frame and any two can also be determined respectively
The attribute information of the corresponding target object of each candidate frame, determines candidate frame to be combined, and closed in a candidate frame
And it handles, the candidate frame that obtains that treated (each candidate frame that merges that treated and/or the candidate for not merging processing
Frame), and determine any two treated the corresponding IoU of candidate frame and any two treated and is each in candidate frame
The attribute information of a treated corresponding target object of candidate frame is then based on any two treated candidate frame point
Each treated corresponding target object of candidate frame in not corresponding IoU and any two treated candidate frame
Attribute information, determine candidate frame to be combined, and carry out merging treatment, circulation executes that treated is candidate based on any two
Each treated corresponding target of candidate frame in the corresponding IoU of frame and any two treated candidate frame
The attribute information of object determines candidate frame to be combined, and carries out merging treatment, until meeting preset condition.
For example, including 4 candidate frames for target object, candidate frame 1, candidate frame 2, candidate frame 3 and candidate frame 4 are first
The first corresponding IoU between determining candidate frame 1 and candidate frame 2 and candidate frame 1 and the corresponding target of candidate frame 2
The attribute information of object, corresponding IoU and candidate frame 3 and candidate frame 4 between candidate frame 3 and candidate frame 4 distinguish
The attribute information of corresponding target object, corresponding IoU and the candidate frame being then based between candidate frame 1 and candidate frame 2
1 and the corresponding target object of candidate frame 2 attribute information, determine whether candidate frame 1 and candidate frame 2 can merge,
And merging treatment is carried out, and based on the corresponding IoU and candidate frame 3 and candidate frame 4 between candidate frame 3 and candidate frame 4
The attribute information of corresponding target object, determines whether candidate frame 3 and candidate frame 4 can merge, and merges place
Reason, if then candidate frame 1 and candidate frame 2 merge, candidate frame 3 and candidate frame 4 merge, it is determined that two candidates after merging
The attribute letter of the corresponding target object of each candidate frame in corresponding IoU between frame, and two candidate frames after merging
Breath, and each candidate frame is right respectively in two candidate frames after corresponding IoU between two candidate frames after merging, and merging
The attribute information for the target object answered, determines whether two candidate frames after merging merge, and carries out merging treatment.Certainly, if
Candidate frame 1 and candidate frame 2 can merge processing, and candidate frame 3 and candidate frame 4 can not merge processing, then will close
And treated that candidate frame successively successively merges processing with candidate frame 3 and candidate frame 4, until meeting preset condition.
For the embodiment of the present application, above-mentioned preset condition may include at least one of following:
There is no the candidate frames that can merge processing;
The candidate frame of target object is directed to comprising one;
It include a candidate frame for either objective object;
The condition of user's input.
The alternatively possible implementation of the embodiment of the present application, the attribute information of the corresponding target object of any candidate frame
It includes at least one of the following:
The colouring information of target object;The density information of target object;The shape information of target object;Belonging to target object
Classification information.
If candidate frame is the candidate frame for face information, the attribute information packet of the corresponding target object of any candidate frame
It includes at least one of following:
The age information of target object;The gender information of target object;The skin color information of target object;Target object
Towards radian information.
The alternatively possible implementation of the embodiment of the present application, if candidate frame is the candidate frame for face information,
Step S1032 can specifically include: step S10321 (not shown), wherein
Step S10321, by non-maxima suppression algorithm, the IoU of any two candidate frames is determined, in any two candidate frames
Whether the attribute information of the corresponding face information of each candidate frame meets preset rules, whether to determine any two candidate frames
Belong to pending combined candidate frame.
Wherein, preset rules include at least one of the following:
IoU is greater than the first attribute information phase of the first preset threshold and the corresponding target object of any two candidate frames
Together;
IoU is not greater than the first attribute information of the second preset threshold and the corresponding target object of any two candidate frames
Together, the first preset threshold is less than the second preset threshold;
IoU is greater than the first preset threshold, and the first attribute information of the corresponding target object of any two candidate frames is
The probability difference of preset attribute information is away from no more than third predetermined threshold value;
IoU is greater than the second preset threshold, and the first attribute information of the corresponding target object of any two candidate frames is
The probability difference of preset attribute information is away from not less than third predetermined threshold value;
IoU be greater than the first preset threshold, and the second attribute information of the corresponding target object of any two candidate frames it
Difference is not more than the 4th preset threshold;
IoU be greater than the second preset threshold, and the second attribute information of the corresponding target object of any two candidate frames it
Difference is greater than the 4th preset threshold;
IoU is greater than the 5th preset threshold, and the 5th preset threshold is that the first preset threshold is respectively corresponded with any two candidate frames
Target object the first attribute information difference between weighted information;
IoU is greater than the 6th preset threshold, and the 6th preset threshold is that the first preset threshold is respectively corresponded with any two candidate frames
Target object the first attribute information difference and the corresponding target object of any two candidate frames the second attribute information
For the weighted information between the probability difference of preset attribute information;
IoU is greater than the 7th preset threshold, and the 7th preset threshold is that the first preset threshold is respectively corresponded with any two candidate frames
Third attribute information difference between weighted information.
Specifically, preset rules include at least one of the following:
IoU is greater than the first preset threshold and the gender information of the corresponding target object of any two candidate frames is identical;
IoU is greater than the second preset threshold and the gender information of the corresponding target object of any two candidate frames is different, the
One preset threshold is less than the second preset threshold;
IoU is greater than the first preset threshold, and the gender of the corresponding target object of any two candidate frames is that the same sex is other
Probability difference is away from no more than third predetermined threshold value;
IoU is greater than the second preset threshold, and the gender of the corresponding target object of any two candidate frames is that the same sex is other
Probability difference is away from greater than third predetermined threshold value;
IoU is greater than the first preset threshold, and the age gap of the corresponding target object of any two candidate frames is not more than
4th preset threshold;
IoU is greater than the second preset threshold, and the age gap of the corresponding target object of any two candidate frames is greater than the
Four preset thresholds;
IoU is greater than the 5th preset threshold, the 5th preset threshold=first the+the first parameter value of preset threshold × any two candidates
The age gap of the corresponding target object of frame;
IoU is greater than the 6th preset threshold, the 6th preset threshold=first the+the first parameter value of preset threshold × any two candidates
The+the second parameter value of age gap of the corresponding target object of frame × any two candidate frames corresponding target object property
Not Wei the other probability difference of the same sex away from;
IoU is greater than the 7th preset threshold, the 7th preset threshold=first the+the second parameter value of preset threshold × any two candidates
Difference between the radian value of the corresponding direction of frame.
For example, if candidate frame is the candidate frame for face information, when carrying out non-maxima suppression algorithm, two candidates
The condition that frame merges is that IoU is greater than 0.3 and the corresponding gender attribute of two candidate frames is identical;Carrying out non-maxima suppression
When algorithm, the condition that two candidate frames merge is IoU greater than 0.3 and the corresponding gender attribute of two candidate frames is identical, or
Person IoU is greater than 0.95 and the corresponding gender attribute of two candidate frames is not identical;When carrying out non-maxima suppression algorithm, two
The condition that a candidate frame merges is that IoU is greater than 0.3 and their corresponding attribute informations are no more than for the gap of the probability of women
0.4;When carrying out non-maxima suppression algorithm, the condition that two candidate frames merge is: their IoU is greater than 0.3 and they are
The gap of the probability of women be no more than 0.4 or IoU be greater than 0.95 and they for the gap of probability of women be at least 0.4;?
When carrying out non-maxima suppression algorithm, the condition that two candidate frames merge is that greater than 0.3, their age (belongs to their IoU
Property) gap be no more than 5 years old;When carrying out non-maxima suppression algorithm, the condition that two candidate frames merge is: their IoU is big
In 0.3 and age (attribute) gap be no more than 5 years old or IoU and be greater than 0.95 and age (attribute) gap at least 5 years old;Carry out
When non-maxima suppression algorithm, the condition that two candidate frames merge is, their IoU is greater than (0.3+0.05 × their age
Gap);When carrying out non-maxima suppression algorithm, the condition that two candidate frames merge is that their IoU is greater than (0.3+0.05
× their age gap away from+0.5 × they for the probability of women gap);When carrying out non-maxima suppression algorithm, two faces
The condition that frame merges is, their IoU be greater than (0.3+0.5 × they direction radian value difference).This rule and
The difference of front is, one can consider that age and gender are a people constant attributes in a short time, but are directed towards and hold
Continue it is changed, but because direction variation have continuity, still can assist non-maxima suppression algorithm.
For example, the candidate frame for face includes candidate frame 1 and candidate frame 2, candidate frame in any image to be processed
IoU between 1 and candidate frame 2 is 0.5, and 1 corresponding gender (attribute) of candidate frame is 28,2 corresponding gender (attribute) of candidate frame
It is that 32, then candidate frame 1 and candidate frame 2 can merge.
In another example the candidate frame for face includes candidate frame 3 and candidate frame 4, between candidate frame 3 and candidate frame 4
IoU is 0.5, and the 3 corresponding age of candidate frame (attribute) is 28, and the 4 corresponding age of candidate frame (attribute) is 32, and candidate frame 3 is corresponding
Target object be the probability of women be 0.8, the corresponding target object of candidate frame 4 is that the attribute of women is 0.6, the first default threshold
Value is 0.3, and the first parameter value is 0.05, and the second parameter value is 0.5, then first the+the first parameter value of preset threshold × any two times
Select the+the second parameter value of age gap × corresponding target object of any two candidate frames of the corresponding target object of frame
Gender is the probability difference of women away from=0.3+0.05 × 4+0.5 × 0.2=0.6, i.e. the 6th preset threshold is 0.6, and candidate frame 3
IoU between candidate frame 4 is 0.5, then candidate frame 3 and candidate frame 4 cannot merge.
For the embodiment of the present application, IoU by non-maxima suppression algorithm, and based on any two candidate frames, any two
The attribute information and preset rules of the corresponding face information of each candidate frame, determine that any two candidate frames are in candidate frame
It is no to belong to pending combined candidate frame, the accuracy merged for the candidate frame of face can be improved, with promoted for
The accuracy of the modes such as face tracking, and then user experience can be promoted.
For the embodiment of the present application, place can be merged by above-mentioned preset rules not only for the candidate frame of face
Reason, for other candidate frames, can also merge processing by corresponding rule by non-maxima suppression algorithm.
For example, if target object be information of vehicles, be directed to information of vehicles candidate frame, then by any two candidate frame it
Between IoU be greater than 0.3 and belong to the candidate frame of same category and merge processing.
The alternatively possible implementation of the embodiment of the present application, may include: before step S101 step Sa (in figure not
Show), step S101 specifically can be with step S1011 (not shown), wherein
Step Sa, multimedia messages are pre-processed, obtains image to be processed.
For the embodiment of the present application, pretreatment may include: image correction process and denoising etc. for initial data
Processing mode.
Step S1011, by image to be processed by detection network, each candidate frame difference in image to be processed is determined
Corresponding location information.
For the embodiment of the present application, detecting network can be the neural network after training.
The above-mentioned method for describing image procossing in detail, it is following to pass through an application scenarios and introduced in a manner of summing-up
The method of the image procossing, specific as shown in Figure 4:
By camera video stream by obtaining frame image to be processed after pretreatment, and frame image to be processed enters inspection
Survey grid network obtains the location information of at least one face frame, then the position of frame image to be processed and at least one face frame
Information is obtained the attribute information of the corresponding target object of any face frame, is then based on and passes through face by face character network
The location information of any face frame, merges place to face frame in the attribute information and any frame image that net with attributes obtains
Reason, the location information of the face frame after being merged.
The method for describing image procossing from the angle of method flow above, angle of following combination attached drawings from virtual module
The device of image procossing is introduced, specific as follows shown:
The embodiment of the present application provides a kind of device of image procossing, as shown in Fig. 2, the device 20 of the image procossing can be with
It include: the first determining module 21, the second determining module 22, merging treatment module 23, wherein
First determining module 21, for determining the corresponding location information of each candidate frame in image to be processed.
Wherein, candidate frame is the candidate frame for target object.
Second determining module 22, each candidate frame for being determined based on image to be processed and the first determining module 21
Corresponding location information, and pass through the attribute information that preset model determines the corresponding target object of each candidate frame respectively.
For the embodiment of the present application, the first determining module 21 and the second determining module 22 can be identical determining module,
Or different determining module.In the embodiment of the present application without limitation.
Merging treatment module 23, for passing through non-maxima suppression algorithm and being based on the corresponding position of each candidate frame
The attribute information of information and the corresponding target object of each candidate frame merges processing to each candidate frame.
A kind of possible implementation of the embodiment of the present application, for any two candidate frame, merging treatment module 23, packet
It includes: the first determination unit, the second determination unit and merge processing unit, wherein
First determination unit, for determining any two candidate frame based on the corresponding location information of each candidate frame
It hands over and compares IoU.
Second determination unit, any two times for passing through non-maxima suppression algorithm and being determined based on the first determination unit
The attribute information for selecting the corresponding target object of each candidate frame in the IoU and any two candidate frames of frame, determines any two
Whether candidate frame belongs to pending combined candidate frame.
For the embodiment of the present application, the first determination unit and the second determination unit can be identical determination unit, can also
Think different determination units.In the embodiment of the present application without limitation.
Merge processing unit, for determining that any two candidate frames belong to pending combined candidate when the second determination unit
Frame then merges processing to any two candidate frame.
The alternatively possible implementation of the embodiment of the present application, the device 20 further include: preprocessing module, wherein
Preprocessing module obtains image to be processed for pre-processing to multimedia messages.
Specifically, the first determining module 21, specifically for by detection network, determining to be processed image to be processed
The corresponding location information of each candidate frame in image.
The alternatively possible implementation of the embodiment of the present application, the attribute information of the corresponding target object of any candidate frame
It includes at least one of the following:
The colouring information of target object;The density information of target object;The shape information of target object;Belonging to target object
Classification information;
If candidate frame is the candidate frame for face information, the attribute information packet of the corresponding target object of any candidate frame
It includes at least one of following:
The age information of target object;The gender information of target object;The skin color information of target object;Target object
Towards radian information.
The alternatively possible implementation of the embodiment of the present application, when candidate frame is the candidate frame for face information,
Second determination unit is specifically used for determining IoU, any two candidate frames of any two candidate frames by non-maxima suppression algorithm
In the attribute information of the corresponding face information of each candidate frame whether meet preset rules, to determine that any two candidate frames are
It is no to belong to pending combined candidate frame.
Wherein, preset rules include at least one of the following:
IoU is greater than the first attribute information phase of the first preset threshold and the corresponding target object of any two candidate frames
Together;
IoU is not greater than the first attribute information of the second preset threshold and the corresponding target object of any two candidate frames
Together, the first preset threshold is less than the second preset threshold;
IoU is greater than the first preset threshold, and the first attribute information of the corresponding target object of any two candidate frames is
The probability difference of preset attribute information is away from no more than third predetermined threshold value;
IoU is greater than the second preset threshold, and the first attribute information of the corresponding target object of any two candidate frames is
The probability difference of preset attribute information is away from not less than third predetermined threshold value;
IoU be greater than the first preset threshold, and the second attribute information of the corresponding target object of any two candidate frames it
Difference is not more than the 4th preset threshold;
IoU be greater than the second preset threshold, and the second attribute information of the corresponding target object of any two candidate frames it
Difference is greater than the 4th preset threshold;
IoU is greater than the 5th preset threshold, and the 5th preset threshold is that the first preset threshold is respectively corresponded with any two candidate frames
Target object the first attribute information difference between weighted information;
IoU is greater than the 6th preset threshold, and the 6th preset threshold is that the first preset threshold is respectively corresponded with any two candidate frames
Target object the first attribute information difference and the corresponding target object of any two candidate frames the second attribute information
For the weighted information between the probability difference of preset attribute information;
IoU is greater than the 7th preset threshold, and the 7th preset threshold is that the first preset threshold is respectively corresponded with any two candidate frames
Third attribute information difference between weighted information.
Specifically, preset rules include at least one of the following:
IoU is greater than the first preset threshold and the gender information of the corresponding target object of any two candidate frames is identical;
IoU is greater than the second preset threshold and the gender information of the corresponding target object of any two candidate frames is different, the
One preset threshold is less than the second preset threshold;
IoU is greater than the first preset threshold, and the gender of the corresponding target object of any two candidate frames is that the same sex is other
Probability difference is away from no more than third predetermined threshold value;
IoU is greater than the second preset threshold, and the gender of the corresponding target object of any two candidate frames is that the same sex is other
Probability difference is away from greater than third predetermined threshold value;
IoU is greater than the first preset threshold, and the age gap of the corresponding target object of any two candidate frames is not more than
4th preset threshold;
IoU is greater than the second preset threshold, and the age gap of the corresponding target object of any two candidate frames is greater than the
Four preset thresholds;
IoU is greater than the 5th preset threshold, the 5th preset threshold=first the+the first parameter value of preset threshold × any two candidates
The age gap of the corresponding target object of frame;
IoU is greater than the 6th preset threshold, the 6th preset threshold=first the+the first parameter value of preset threshold × any two candidates
The+the second parameter value of age gap of the corresponding target object of frame × any two candidate frames corresponding target object property
Not Wei the other probability difference of the same sex away from;
IoU is greater than the 7th preset threshold, the 7th preset threshold=first the+the second parameter value of preset threshold × any two candidates
Difference between the radian value of the corresponding direction of frame.
The embodiment of the present application provides a kind of device of image procossing, only with non-maxima suppression method in the prior art
It is that the location information based on each candidate frame merges candidate frame and compares, image to be processed is determined in the embodiment of the present application
In the corresponding location information of each candidate frame, candidate frame is the candidate frame for target object, is then based on to be processed
Image and the corresponding location information of each candidate frame, and the corresponding mesh of each candidate frame is determined by preset model respectively
Mark object attribute information, then by non-maxima suppression algorithm and based on the corresponding location information of each candidate frame with
And the attribute information of the corresponding target object of each candidate frame, processing is merged to each candidate frame.That is the embodiment of the present application
When merging processing for the candidate frame in image to be processed, be based on the corresponding location information of each candidate frame with
And the attribute information of the corresponding target object of each candidate frame, it will be closer and belong to not in image so as to reduce
It is incorporated into the probability of a candidate frame with the candidate frame of target object, and can also reduce and will belong to the multiple of same target object
Candidate frame is merged into the case where more than one frame, and then the accuracy handled candidate frame each in image can be improved,
And then user experience can be promoted.
The side for the image procossing that the application above method embodiment provides can be performed in the device of the image procossing of the present embodiment
Method, realization principle is similar, and details are not described herein again.
Above-described embodiment describes the method for image procossing from the angle of method flow and is situated between from the angle of virtual module
Continued the device of image procossing, following combination attached drawings, describes from entity structure angle and executes the electronics of image processing method and set
It is standby, specific as follows shown:
The embodiment of the present application provides a kind of electronic equipment, as shown in figure 3, electronic equipment shown in Fig. 3 3000 includes: place
Manage device 3001 and memory 3003.Wherein, processor 3001 is connected with memory 3003, is such as connected by bus 3002.It is optional
Ground, electronic equipment 3000 can also include transceiver 3004.It should be noted that transceiver 3004 is not limited to one in practical application
A, the structure of the electronic equipment 3000 does not constitute the restriction to the embodiment of the present application.
Processor 3001 can be CPU, general processor, DSP, ASIC, FPGA or other programmable logic device, crystalline substance
Body pipe logical device, hardware component or any combination thereof.It, which may be implemented or executes, combines described by present disclosure
Various illustrative logic blocks, module and circuit.Processor 3001 is also possible to realize the combination of computing function, such as wraps
It is combined containing one or more microprocessors, DSP and the combination of microprocessor etc..
Bus 3002 may include an access, and information is transmitted between said modules.Bus 3002 can be pci bus or
Eisa bus etc..Bus 3002 can be divided into address bus, data/address bus, control bus etc..Only to be used in Fig. 3 convenient for indicating
One thick line indicates, it is not intended that an only bus or a type of bus.
Memory 3003 can be ROM or can store the other kinds of static storage device of static information and instruction, RAM
Or the other kinds of dynamic memory of information and instruction can be stored, it is also possible to EEPROM, CD-ROM or other CDs
Storage, optical disc storage (including compression optical disc, laser disc, optical disc, Digital Versatile Disc, Blu-ray Disc etc.), magnetic disk storage medium
Or other magnetic storage apparatus or can be used in carry or store have instruction or data structure form desired program generation
Code and can by any other medium of computer access, but not limited to this.
Memory 3003 is used to store the application code for executing application scheme, and is held by processor 3001 to control
Row.Processor 3001 is for executing the application code stored in memory 3003, to realize aforementioned either method embodiment
Shown in content.
The embodiment of the present application provides a kind of electronic equipment, the electronic equipment in the embodiment of the present application include: memory and
Processor;At least one program is stored in the memory, when for being executed by the processor, compared with prior art
It can be achieved: determining that the corresponding location information of each candidate frame, candidate frame are in image to be processed in the embodiment of the present application
For the candidate frame of target object, it is then based on image and the corresponding location information of each candidate frame to be processed, and
It determines the attribute information of the corresponding target object of each candidate frame respectively by preset model, is then calculated by non-maxima suppression
Method and attribute information based on the corresponding location information of each candidate frame and the corresponding target object of each candidate frame, it is right
Each candidate frame merges processing.It is base when i.e. the application merges processing for the candidate frame in image to be processed
In the corresponding location information of each candidate frame and the attribute information of the corresponding target object of each candidate frame, thus
Can reduce will be closer and belong to the candidate frame of different target object in image is incorporated into the probability of a candidate frame, and
The case where multiple candidate frames for belonging to same target object are merged into more than one frame can be reduced, and then can be improved to figure
The accuracy that each candidate frame is handled as in, and then user experience can be promoted.
The method that the image procossing that the application above method embodiment provides can be performed in the electronic equipment of the present embodiment, in fact
Existing principle is similar, and details are not described herein again.
The embodiment of the present application provides a kind of computer readable storage medium, is stored on the computer readable storage medium
Computer program allows computer to execute corresponding contents in preceding method embodiment when run on a computer.With
The prior art is compared, and the corresponding location information of each candidate frame in image to be processed is determined in the embodiment of the present application, is waited
Selecting frame is the candidate frame for target object, is then based on image to be processed and the corresponding position letter of each candidate frame
Breath, and pass through the attribute information that preset model determines the corresponding target object of each candidate frame respectively, then pass through non-maximum
Restrainable algorithms and attribute based on the corresponding location information of each candidate frame and the corresponding target object of each candidate frame
Information merges processing to each candidate frame.I.e. the application merges processing for the candidate frame in image to be processed
When, it is to be believed based on the attribute of the corresponding location information of each candidate frame and the corresponding target object of each candidate frame
Breath is incorporated into the general of a candidate frame so as to reduce to be closer in image and belong to the candidate frame of different target object
Rate, and the case where multiple candidate frames for belonging to same target object are merged into more than one frame can also be reduced, and then can be with
The accuracy handled candidate frame each in image is improved, and then user experience can be promoted.
Computer readable storage medium provided by the embodiments of the present application is suitable for above method embodiment, implements in the application
It is repeated no more in example.
It should be understood that although each step in the flow chart of attached drawing is successively shown according to the instruction of arrow,
These steps are not that the inevitable sequence according to arrow instruction successively executes.Unless expressly stating otherwise herein, these steps
Execution there is no stringent sequences to limit, can execute in the other order.Moreover, at least one in the flow chart of attached drawing
Part steps may include that perhaps these sub-steps of multiple stages or stage are not necessarily in synchronization to multiple sub-steps
Completion is executed, but can be executed at different times, execution sequence, which is also not necessarily, successively to be carried out, but can be with other
At least part of the sub-step or stage of step or other steps executes in turn or alternately.
The above is only some embodiments of the invention, it is noted that for the ordinary skill people of the art
For member, various improvements and modifications may be made without departing from the principle of the present invention, these improvements and modifications are also answered
It is considered as protection scope of the present invention.
Claims (10)
1. a kind of image processing method characterized by comprising
Determine that the corresponding location information of each candidate frame in image to be processed, the candidate frame are for target object
Candidate frame;
Based on the image to be processed and the corresponding location information of each candidate frame, and pass through preset model point
The attribute information of the corresponding target object of each candidate frame is not determined;
Pass through non-maxima suppression algorithm and is based on each corresponding location information of candidate frame and each time
The attribute information for selecting the corresponding target object of frame merges processing to each candidate frame.
2. passing through non-maxima suppression the method according to claim 1, wherein being directed to any two candidate frame
Algorithm and based on each corresponding location information of candidate frame and the corresponding target object of each candidate frame
Attribute information merges processing to each candidate frame, comprising:
The friendship of any two candidate frame is determined based on the corresponding location information of each candidate frame and compares IoU;
It is by non-maxima suppression algorithm and each in the IoU based on any two candidate frames and any two described candidate frames
The attribute information of the corresponding target object of candidate frame, determines whether any two described candidate frames belong to pending combined time
Select frame;
If belonging to pending combined candidate frame, processing is merged to any two candidate frame.
3. the method according to claim 1, wherein determining that each candidate frame respectively corresponds in image to be processed
Location information, before further include:
Multimedia messages are pre-processed, image to be processed is obtained;
The corresponding location information of each candidate frame in determination image to be processed, comprising:
By the image to be processed by detection network, the corresponding position of each candidate frame in image to be processed is determined
Information.
4. method according to claim 1-3, which is characterized in that the category of the corresponding target object of any candidate frame
Property information includes at least one of the following:
The colouring information of target object;
The density information of target object;
The shape information of target object;
Target object generic information.
5. method according to claim 1-3, which is characterized in that if the candidate frame is for face information
Candidate frame, then the attribute information of the corresponding target object of any candidate frame includes at least one of the following:
The age information of target object;
The gender information of target object;
The skin color information of target object;
Target object towards radian information.
6. according to the method described in claim 5, it is characterized in that, if the candidate frame be for face information candidate frame,
Then
Each time in IoU and any two candidate frames by non-maxima suppression algorithm, and based on any two candidate frames
The attribute information of the corresponding target object of frame is selected, determines whether any two described candidate frames belong to pending combined candidate
Frame, comprising:
By non-maxima suppression algorithm, the IoU of any two candidate frames is determined, each candidate frame point in any two candidate frames
Whether the attribute information of not corresponding face information meets preset rules, with determine any two described candidate frames whether belong to into
The combined candidate frame of row;
The preset rules include at least one of the following:
The IoU is greater than the first attribute information phase of the first preset threshold and the corresponding target object of any two candidate frames
Together;
The IoU is not greater than the first attribute information of the second preset threshold and the corresponding target object of any two candidate frames
Together, the first preset threshold is less than the second preset threshold;
The IoU is greater than the first preset threshold, and the first attribute information of the corresponding target object of any two candidate frames is
The probability difference of preset attribute information is away from no more than third predetermined threshold value;
The IoU is greater than the second preset threshold, and the first attribute information of the corresponding target object of any two candidate frames is
The probability difference of preset attribute information is away from not less than third predetermined threshold value;
The IoU be greater than the first preset threshold, and the second attribute information of the corresponding target object of any two candidate frames it
Difference is not more than the 4th preset threshold;
The IoU be greater than the second preset threshold, and the second attribute information of the corresponding target object of any two candidate frames it
Difference is greater than the 4th preset threshold;
The IoU is greater than the 5th preset threshold, and the 5th preset threshold is first preset threshold and any two candidate frames
Weighted information between first attribute information difference of corresponding target object;
The IoU is greater than the 6th preset threshold, and the 6th preset threshold is first preset threshold and any two candidate frames
First attribute information difference of corresponding target object and the corresponding target object of any two described candidate frames
Second attribute information is the weighted information between the probability difference of preset attribute information;
The IoU is greater than the 7th preset threshold, and the 7th preset threshold is first preset threshold and any two candidate frames
Weighted information between corresponding third attribute information difference.
7. according to the method described in claim 6, it is characterized in that, the preset rules include at least one of the following:
The IoU is greater than the first preset threshold and the gender information of the corresponding target object of any two candidate frames is identical;
The IoU is greater than the second preset threshold and the gender information of the corresponding target object of any two candidate frames is different, the
One preset threshold is less than the second preset threshold;
The IoU is greater than the first preset threshold, and the gender of the corresponding target object of any two candidate frames is that the same sex is other
Probability difference is away from no more than third predetermined threshold value;
The IoU is greater than the second preset threshold, and the gender of the corresponding target object of any two candidate frames is that the same sex is other
Probability difference is away from greater than third predetermined threshold value;
The IoU is greater than the first preset threshold, and the age gap of the corresponding target object of any two candidate frames is not more than
4th preset threshold;
The IoU is greater than the second preset threshold, and the age gap of the corresponding target object of any two candidate frames is greater than the
Four preset thresholds;
The IoU be greater than the 5th preset threshold, the 5th preset threshold=first the+the first parameter value of preset threshold × any two
The age gap of the corresponding target object of candidate frame;
The IoU be greater than the 6th preset threshold, the 6th preset threshold=first the+the first parameter value of preset threshold × any two
The+the second parameter value of age gap of the corresponding target object of the candidate frame × corresponding target object of any two candidate frames
Gender be the other probability difference of the same sex away from;
The IoU be greater than the 7th preset threshold, the 7th preset threshold=first the+the second parameter value of preset threshold × any two
Difference between the radian value of the corresponding target object direction of candidate frame.
8. a kind of device of image procossing characterized by comprising
First determining module, for determining the corresponding location information of each candidate frame, the candidate in image to be processed
Frame is the candidate frame for target object;
Second determining module, each candidate frame for being determined based on the image to be processed and first determining module
Corresponding location information, and determine that the attribute of the corresponding target object of each candidate frame is believed respectively by preset model
Breath;
Merging treatment module, for passing through non-maxima suppression algorithm and being based on the corresponding position letter of each candidate frame
The attribute information of breath and the corresponding target object of each candidate frame, merges processing to each candidate frame.
9. a kind of electronic equipment, characterized in that it comprises:
One or more processors;
Memory;
One or more application program, wherein one or more of application programs are stored in the memory and are configured
To be executed by one or more of processors, one or more of programs are configured to: being executed according to claim 1~7
The method of described in any item image procossings.
10. a kind of computer readable storage medium, which is characterized in that the storage medium is stored at least one instruction, at least
One Duan Chengxu, code set or instruction set, at least one instruction, an at least Duan Chengxu, the code set or instruction set
Loaded by the processor and executed the method to realize the image procossing as described in claim 1 to 7 is any.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201910313968.2A CN110033424A (en) | 2019-04-18 | 2019-04-18 | Method, apparatus, electronic equipment and the computer readable storage medium of image procossing |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201910313968.2A CN110033424A (en) | 2019-04-18 | 2019-04-18 | Method, apparatus, electronic equipment and the computer readable storage medium of image procossing |
Publications (1)
Publication Number | Publication Date |
---|---|
CN110033424A true CN110033424A (en) | 2019-07-19 |
Family
ID=67239141
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201910313968.2A Pending CN110033424A (en) | 2019-04-18 | 2019-04-18 | Method, apparatus, electronic equipment and the computer readable storage medium of image procossing |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN110033424A (en) |
Cited By (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111310824A (en) * | 2020-02-13 | 2020-06-19 | 中科智云科技有限公司 | Multi-angle dense target detection inhibition optimization method and equipment |
CN111967595A (en) * | 2020-08-17 | 2020-11-20 | 成都数之联科技有限公司 | Candidate frame marking method and system, model training method and target detection method |
CN112348778A (en) * | 2020-10-21 | 2021-02-09 | 深圳市优必选科技股份有限公司 | Object identification method and device, terminal equipment and storage medium |
WO2021155661A1 (en) * | 2020-02-05 | 2021-08-12 | 华为技术有限公司 | Image processing method and related device |
CN113761245A (en) * | 2021-05-11 | 2021-12-07 | 腾讯科技(深圳)有限公司 | Image recognition method and device, electronic equipment and computer readable storage medium |
WO2022012158A1 (en) * | 2020-07-17 | 2022-01-20 | 华为技术有限公司 | Target determination method and target determination device |
Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN106651955A (en) * | 2016-10-10 | 2017-05-10 | 北京小米移动软件有限公司 | Method and device for positioning object in picture |
CN107871102A (en) * | 2016-09-23 | 2018-04-03 | 北京眼神科技有限公司 | A kind of method for detecting human face and device |
CN108009544A (en) * | 2017-12-13 | 2018-05-08 | 北京小米移动软件有限公司 | Object detection method and device |
US20180307897A1 (en) * | 2016-05-28 | 2018-10-25 | Samsung Electronics Co., Ltd. | System and method for a unified architecture multi-task deep learning machine for object recognition |
CN108875537A (en) * | 2018-02-28 | 2018-11-23 | 北京旷视科技有限公司 | Method for checking object, device and system and storage medium |
CN109472264A (en) * | 2018-11-09 | 2019-03-15 | 北京字节跳动网络技术有限公司 | Method and apparatus for generating object detection model |
-
2019
- 2019-04-18 CN CN201910313968.2A patent/CN110033424A/en active Pending
Patent Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20180307897A1 (en) * | 2016-05-28 | 2018-10-25 | Samsung Electronics Co., Ltd. | System and method for a unified architecture multi-task deep learning machine for object recognition |
CN107871102A (en) * | 2016-09-23 | 2018-04-03 | 北京眼神科技有限公司 | A kind of method for detecting human face and device |
CN106651955A (en) * | 2016-10-10 | 2017-05-10 | 北京小米移动软件有限公司 | Method and device for positioning object in picture |
CN108009544A (en) * | 2017-12-13 | 2018-05-08 | 北京小米移动软件有限公司 | Object detection method and device |
CN108875537A (en) * | 2018-02-28 | 2018-11-23 | 北京旷视科技有限公司 | Method for checking object, device and system and storage medium |
CN109472264A (en) * | 2018-11-09 | 2019-03-15 | 北京字节跳动网络技术有限公司 | Method and apparatus for generating object detection model |
Cited By (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2021155661A1 (en) * | 2020-02-05 | 2021-08-12 | 华为技术有限公司 | Image processing method and related device |
CN111310824A (en) * | 2020-02-13 | 2020-06-19 | 中科智云科技有限公司 | Multi-angle dense target detection inhibition optimization method and equipment |
WO2022012158A1 (en) * | 2020-07-17 | 2022-01-20 | 华为技术有限公司 | Target determination method and target determination device |
CN111967595A (en) * | 2020-08-17 | 2020-11-20 | 成都数之联科技有限公司 | Candidate frame marking method and system, model training method and target detection method |
CN112348778A (en) * | 2020-10-21 | 2021-02-09 | 深圳市优必选科技股份有限公司 | Object identification method and device, terminal equipment and storage medium |
CN112348778B (en) * | 2020-10-21 | 2023-10-27 | 深圳市优必选科技股份有限公司 | Object identification method, device, terminal equipment and storage medium |
CN113761245A (en) * | 2021-05-11 | 2021-12-07 | 腾讯科技(深圳)有限公司 | Image recognition method and device, electronic equipment and computer readable storage medium |
CN113761245B (en) * | 2021-05-11 | 2023-10-13 | 腾讯科技(深圳)有限公司 | Image recognition method, device, electronic equipment and computer readable storage medium |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN110033424A (en) | Method, apparatus, electronic equipment and the computer readable storage medium of image procossing | |
US10943145B2 (en) | Image processing methods and apparatus, and electronic devices | |
US10832069B2 (en) | Living body detection method, electronic device and computer readable medium | |
US20190318156A1 (en) | Face anti-counterfeiting detection methods and systems, electronic devices, programs and media | |
US20230030267A1 (en) | Method and apparatus for selecting face image, device, and storage medium | |
CN107169463A (en) | Method for detecting human face, device, computer equipment and storage medium | |
CN109961107B (en) | Training method and device for target detection model, electronic equipment and storage medium | |
CN111144215B (en) | Image processing method, device, electronic equipment and storage medium | |
CN111767906B (en) | Face detection model training method, face detection device and electronic equipment | |
CN111104925B (en) | Image processing method, image processing apparatus, storage medium, and electronic device | |
CN112598643A (en) | Depth counterfeit image detection and model training method, device, equipment and medium | |
CN115861462B (en) | Training method and device for image generation model, electronic equipment and storage medium | |
CN109583364A (en) | Image-recognizing method and equipment | |
CN111652864A (en) | Casting defect image generation method for generating countermeasure network based on conditional expression | |
CN109671055B (en) | Pulmonary nodule detection method and device | |
US20110182497A1 (en) | Cascade structure for classifying objects in an image | |
CN114463603B (en) | Training method and device for image detection model, electronic equipment and storage medium | |
CN114372931A (en) | Target object blurring method and device, storage medium and electronic equipment | |
CN114372932A (en) | Image processing method and computer program product | |
CN112906571B (en) | Living body identification method and device and electronic equipment | |
CN111126566B (en) | Abnormal furniture layout data detection method based on GAN model | |
CN116721091A (en) | Cloth flaw detection method, device and readable medium | |
KR102580281B1 (en) | Related object detection method and device | |
CN112989869B (en) | Optimization method, device, equipment and storage medium of face quality detection model | |
CN114332112A (en) | Cell image segmentation method and device, electronic equipment and storage medium |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
RJ01 | Rejection of invention patent application after publication | ||
RJ01 | Rejection of invention patent application after publication |
Application publication date: 20190719 |