CN109165644A - Object detection method and device, electronic equipment, storage medium, program product - Google Patents
Object detection method and device, electronic equipment, storage medium, program product Download PDFInfo
- Publication number
- CN109165644A CN109165644A CN201810770381.XA CN201810770381A CN109165644A CN 109165644 A CN109165644 A CN 109165644A CN 201810770381 A CN201810770381 A CN 201810770381A CN 109165644 A CN109165644 A CN 109165644A
- Authority
- CN
- China
- Prior art keywords
- interest
- area
- target
- prediction
- loss
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/20—Image preprocessing
- G06V10/25—Determination of region of interest [ROI] or a volume of interest [VOI]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/23—Clustering techniques
- G06F18/232—Non-hierarchical techniques
- G06F18/2321—Non-hierarchical techniques using statistics or function optimisation, e.g. modelling of probability density functions
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- Data Mining & Analysis (AREA)
- General Physics & Mathematics (AREA)
- Evolutionary Computation (AREA)
- Life Sciences & Earth Sciences (AREA)
- Artificial Intelligence (AREA)
- General Engineering & Computer Science (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Biophysics (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Bioinformatics & Computational Biology (AREA)
- Probability & Statistics with Applications (AREA)
- Health & Medical Sciences (AREA)
- Biomedical Technology (AREA)
- Evolutionary Biology (AREA)
- Computational Linguistics (AREA)
- General Health & Medical Sciences (AREA)
- Molecular Biology (AREA)
- Computing Systems (AREA)
- Mathematical Physics (AREA)
- Software Systems (AREA)
- Multimedia (AREA)
- Image Analysis (AREA)
Abstract
The embodiment of the present application discloses a kind of object detection method and device, electronic equipment, storage medium, program product, wherein method includes: the characteristic information based on image, obtains at least one area-of-interest;Based on the area-of-interest, corresponding at least two target area of the area-of-interest is obtained, the area-of-interest is within the target area;Based on the characteristic information of at least two target area, the corresponding object detection results of the area-of-interest are determined.More contextual informations can be fused in area-of-interest by the embodiment of the present application, improved the ability to express of area-of-interest, improved detection accuracy.
Description
Technical field
This application involves computer vision technique, especially a kind of object detection method and device, electronic equipment, storage are situated between
Matter, program product.
Background technique
With various higher and higher spaces, time, the lift-off of spectral resolution satellite launch, satellite-based remotely-sensed data is opened
Beginning is widely applied to every field, significant increase information acquisition efficiency.Of interest various of people are obtained from remotely-sensed data
Target information, all significant to multiple industries, especially military affairs, finance, security protection etc. is related to national security and economic development
Etc. key factors.
Summary of the invention
A kind of target detection technique provided by the embodiments of the present application.
According to the one aspect of the embodiment of the present application, a kind of object detection method for providing, comprising:
Characteristic information based on image obtains at least one area-of-interest;
Based on the area-of-interest, corresponding at least two target area of the area-of-interest is obtained, the sense is emerging
Interesting region is within the target area;
Based on the characteristic information of at least two target area, the corresponding target detection knot of the area-of-interest is determined
Fruit.
Optionally, described to be based on the area-of-interest, obtain at least two target areas, comprising:
Enhanced processing at least twice is carried out to the area-of-interest, obtains at least two target areas.
Optionally, the characteristic information based at least two target area determines that the area-of-interest is corresponding
Object detection results, comprising:
Feature connection processing is carried out to the characteristic information of at least two target area, obtains connection features;
Based on the connection features, the corresponding object detection results of the area-of-interest are determined.
Optionally, described to be based on the connection features, it determines the corresponding object detection results of the area-of-interest, wraps
It includes:
Based on the connection features, the first testing result is obtained;
The characteristic information of characteristic information and described image based on the area-of-interest, obtains the second testing result;
Based on first testing result and second testing result, the corresponding target inspection of the area-of-interest is determined
Survey result.
Optionally, described to be based on first testing result and second testing result, determine the area-of-interest
Corresponding object detection results, comprising:
First testing result and second testing result are averaging processing, the area-of-interest pair is obtained
The object detection results answered.
Optionally, the characteristic information based on image, before obtaining at least one area-of-interest, further includes:
Feature extraction is carried out to described image, obtains the characteristic information of described image.
Optionally, the method utilizes target detection neural fusion;
In the characteristic information based on image, before obtaining at least one area-of-interest, the method also includes:
Based on the sample image with markup information, the training target detection neural network.
Optionally, the target detection neural network includes feature extraction network, region of interesting extraction network, the first inspection
Survey neural network and the second detection neural network;
It is described based on the sample image with markup information, the training target detection neural network, comprising:
Feature extraction processing is carried out to the sample image using the feature extraction network, obtains sample characteristics;
The sample characteristics are handled using the region of interesting extraction network, it is emerging to obtain at least one prediction sense
The characteristic information of each prediction area-of-interest corresponding at least two prediction target area in interesting region;
Mesh is predicted to each prediction area-of-interest corresponding at least two using the first detection neural network
The characteristic information in mark region is handled, and the first prediction result of each prediction area-of-interest is obtained;
First prediction result based on each prediction area-of-interest at least one described prediction area-of-interest
With the markup information of the sample image, each prediction area-of-interest at least one described prediction area-of-interest is obtained
First-loss;
Based on the first-loss of each prediction area-of-interest at least one described prediction area-of-interest, described in adjustment
The network parameter of feature extraction network, the region of interesting extraction network and the first detection neural network;
Based on it is described at least one prediction area-of-interest in it is each prediction area-of-interest first-loss and it is described extremely
A few prediction area-of-interest obtains the second loss;
Based on second loss, the network ginseng of the feature extraction network and the second detection neural network is adjusted
Number.
Optionally, first damage based on each prediction area-of-interest at least one described prediction area-of-interest
At least one prediction area-of-interest, obtains the second loss described in becoming estranged, comprising:
Based on the first-loss, screening obtains at least one error sense from least one described prediction area-of-interest
Interest region;
Using the second detection neural network to the characteristic information of at least one error area-of-interest and described
Sample characteristics are handled, and the second prediction of each error area-of-interest at least one described error area-of-interest is obtained
As a result;
The second prediction result and institute based on each error area-of-interest at least one described error area-of-interest
The corresponding markup information of sample image is stated, the second loss is obtained.
Optionally, described first based on each prediction area-of-interest at least one described prediction area-of-interest is pre-
The markup information of result and the sample image is surveyed, it is interested to obtain each prediction at least one described prediction area-of-interest
The first-loss in region, comprising:
The first prediction result and institute based on each prediction area-of-interest at least one described prediction area-of-interest
The markup information of sample image is stated, the index normalization loss and the loss of thunder lattice of each prediction area-of-interest are obtained;
Based on it is described prediction area-of-interest the index normalization loss and the thunder lattice loss the sum of, determine described in
Predict the first-loss of area-of-interest.
Optionally, described first based on each prediction area-of-interest at least one described prediction area-of-interest is pre-
The markup information of result and the sample image is surveyed, it is interested to obtain each prediction at least one described prediction area-of-interest
After the first-loss in region, further includes:
Based on the first prediction result of each prediction area-of-interest at least one described prediction area-of-interest, obtain
The center loss of each prediction area-of-interest at least one described prediction area-of-interest;
The first-loss based on each prediction area-of-interest at least one described prediction area-of-interest, adjustment
The network parameter of the feature extraction network, the region of interesting extraction network and the first detection neural network, comprising:
The first-loss and institute based on each prediction area-of-interest at least one described prediction area-of-interest
Center loss is stated, the feature extraction network, the region of interesting extraction network and the first detection neural network are adjusted
Network parameter.
Optionally, described the based on each prediction area-of-interest at least one described prediction area-of-interest
One prediction result obtains the center loss of each prediction area-of-interest at least one described prediction area-of-interest, comprising:
First prediction result of each prediction area-of-interest at least one described prediction area-of-interest is gathered
Class obtains cluster centre;
First prediction result based on each prediction area-of-interest at least one described prediction area-of-interest
At a distance from the cluster centre, the center loss of each prediction area-of-interest is obtained.
Optionally, described screened from least one described prediction area-of-interest based on the first-loss is obtained at least
One error area-of-interest, comprising:
The prediction sense for being more than default loss threshold value for first-loss described at least one described prediction area-of-interest is emerging
Interesting region is determined as the error area-of-interest.
According to the one aspect of the embodiment of the present application, a kind of object detecting device for providing, comprising:
Region of interesting extraction unit obtains at least one area-of-interest for the characteristic information based on image;
It is corresponding at least to obtain the area-of-interest for being based on the area-of-interest for target area obtaining unit
Two target areas, the area-of-interest are within the target area;
Result detection unit determines the region of interest for the characteristic information based at least two target area
The corresponding object detection results in domain.
Optionally, the region of interesting extraction unit, for being carried out at amplification at least twice to the area-of-interest
Reason obtains at least two target areas.
Optionally, the result detection unit includes:
Link block carries out feature connection processing for the characteristic information at least two target area, is connected
Connect feature;
Module of target detection determines the corresponding target detection knot of the area-of-interest for being based on the connection features
Fruit.
Optionally, the module of target detection obtains the first testing result for being based on the connection features;Based on institute
The characteristic information of area-of-interest and the characteristic information of described image are stated, the second testing result is obtained;Based on first detection
As a result with second testing result, the corresponding object detection results of the area-of-interest are determined.
Optionally, the module of target detection is being based on first testing result and second testing result, determines
When the corresponding object detection results of the area-of-interest, for first testing result and second testing result into
Row average treatment obtains the corresponding object detection results of the area-of-interest.
Optionally, further includes:
Feature extraction unit obtains the characteristic information of described image for carrying out feature extraction to described image.
Optionally, described device combining target detects neural fusion target detection;
Described device further include:
Network training unit, for based on the sample image with markup information, the training target detection neural network.
Optionally, the target detection neural network includes feature extraction network, region of interesting extraction network, the first inspection
Survey neural network and the second detection neural network;
The network training unit, comprising:
Sample process module carries out feature extraction processing to the sample image using the feature extraction network, obtains
Sample characteristics;The sample characteristics are handled using the region of interesting extraction network, obtain at least one prediction sense
The characteristic information of each prediction area-of-interest corresponding at least two prediction target area in interest region;Utilize described first
Neural network is detected to carry out the characteristic information of each prediction area-of-interest corresponding at least two prediction target area
Processing obtains the first prediction result of each prediction area-of-interest;
First-loss module, for based on each prediction area-of-interest at least one described prediction area-of-interest
The markup information of first prediction result and the sample image obtains each prediction at least one prediction area-of-interest
The first-loss of area-of-interest;
First-loss training module, for based on each prediction region of interest at least one described prediction area-of-interest
The first-loss in domain adjusts the feature extraction network, the region of interesting extraction network and the first detection nerve net
The network parameter of network;
Second loss module, for based on each prediction area-of-interest at least one described prediction area-of-interest
First-loss and at least one described prediction area-of-interest, obtain the second loss;
Second loss training module, for adjusting the feature extraction network and described second based on second loss
Detect the network parameter of neural network.
Optionally, the second loss module is specifically used for feeling based on the first-loss from least one described prediction
Screening obtains at least one error area-of-interest in interest region;Using the second detection neural network at least one institute
The characteristic information and the sample characteristics for stating error area-of-interest are handled, and it is emerging to obtain at least one described described error sense
Second prediction result of each error area-of-interest in interesting region;Based at least one described described error area-of-interest
The second prediction result and the corresponding markup information of the sample image of each error area-of-interest obtain the second loss.
Optionally, the first-loss module is specifically used for based on each at least one described prediction area-of-interest
It predicts the first prediction result of area-of-interest and the markup information of the sample image, obtains each prediction region of interest
The index normalization loss and the loss of thunder lattice in domain;It is lost based on the index normalization for predicting area-of-interest and described
The sum of thunder lattice loss determines the first-loss of the prediction area-of-interest.
Optionally, the network training unit, further includes:
Module is lost at center, for based on each prediction area-of-interest at least one described prediction area-of-interest
First prediction result obtains the center loss of each prediction area-of-interest at least one described prediction area-of-interest;
The first-loss training module is specifically used for based on each prediction at least one described prediction area-of-interest
The first-loss of area-of-interest and center loss, adjust the feature extraction network, the area-of-interest mentions
Take the network parameter of network and the first detection neural network.
Optionally, module is lost at the center, is specifically used for each pre- at least one described prediction area-of-interest
The first prediction result for surveying area-of-interest is clustered, and cluster centre is obtained;Based at least one described prediction region of interest
It is emerging to obtain each prediction sense at a distance from the cluster centre for the first prediction result of each prediction area-of-interest in domain
It loses at the center in interesting region.
Optionally, the second loss module is being based on the first-loss from least one described prediction area-of-interest
When middle screening obtains at least one error area-of-interest, for first-loss at least one prediction area-of-interest by described in
Prediction area-of-interest more than default loss threshold value is determined as the error area-of-interest.
According to the other side of the embodiment of the present application, a kind of electronic equipment provided, including processor, the processor
Including object detecting device described in any one as above.
According to the other side of the embodiment of the present application, a kind of electronic equipment that provides, comprising: memory, for storing
Executable instruction;
And processor, it is as above any one to complete that the executable instruction is executed for communicating with the memory
The operation of the item object detection method.
According to the other side of the embodiment of the present application, a kind of computer readable storage medium provided, based on storing
The instruction that calculation machine can be read, described instruction are performed the operation for executing object detection method described in any one as above.
According to the other side of the embodiment of the present application, a kind of computer program product provided, including it is computer-readable
Code, when the computer-readable code is run in equipment, the processor in the equipment is executed for realizing such as taking up an official post
The instruction of one object detection method of meaning.
According to another aspect of the embodiment of the present application, another computer program product provided is calculated for storing
Machine readable instruction, described instruction is performed so that computer executes the inspection of target described in any of the above-described possible implementation
The operation of survey method.
In an optional embodiment, the computer program product is specially computer storage medium, at another
In optional embodiment, the computer program product is specially software product, such as SDK etc..
According to the embodiment of the present application also provides another object detection methods and device, electronic equipment, computer storage
Medium, computer program product, wherein the characteristic information based on image obtains at least one area-of-interest;Based on interested
Region obtains corresponding at least two target area of area-of-interest;Based on the characteristic information of at least two target areas, determine
The corresponding object detection results of area-of-interest.
A kind of object detection method and device, electronic equipment, storage medium, journey provided based on the above embodiments of the present application
Sequence product, the characteristic information based on image obtain at least one area-of-interest;Based on area-of-interest, region of interest is obtained
Corresponding at least two target area in domain;Based on the characteristic information of at least two target areas, determine that area-of-interest is corresponding
More contextual informations have been fused in area-of-interest by object detection results, improve the ability to express of area-of-interest,
Improve detection accuracy.
Below by drawings and examples, the technical solution of the application is described in further detail.
Detailed description of the invention
The attached drawing for constituting part of specification describes embodiments herein, and together with description for explaining
The principle of the application.
The application can be more clearly understood according to following detailed description referring to attached drawing, in which:
Fig. 1 is a flow diagram of the embodiment of the present application object detection method.
Fig. 2 is the structural schematic diagram that the embodiment of the present application obtains one specific example of connection features by way of more ponds.
Fig. 3 is that the embodiment of the present application training objective detects one exemplary structural schematic diagram of neural network.
Fig. 4 is that one of the embodiment of the present application object detection method applies exemplary diagram.
Fig. 5 is a structural schematic diagram of the embodiment of the present application object detecting device.
Fig. 6 is the structural representation suitable for the electronic equipment of the terminal device or server that are used to realize the embodiment of the present application
Figure.
Specific embodiment
The various exemplary embodiments of the application are described in detail now with reference to attached drawing.It should also be noted that unless in addition having
Body explanation, the unlimited system of component and the positioned opposite of step, numerical expression and the numerical value otherwise illustrated in these embodiments is originally
The range of application.
Simultaneously, it should be appreciated that for ease of description, the size of various pieces shown in attached drawing is not according to reality
Proportionate relationship draw.
Be to the description only actually of at least one exemplary embodiment below it is illustrative, never as to the application
And its application or any restrictions used.
Technology, method and apparatus known to person of ordinary skill in the relevant may be not discussed in detail, but suitable
In the case of, the technology, method and apparatus should be considered as part of specification.
It should also be noted that similar label and letter indicate similar terms in following attached drawing, therefore, once a certain Xiang Yi
It is defined in a attached drawing, then in subsequent attached drawing does not need that it is further discussed.
Since the remotely-sensed data of acquisition is usually influenced seriously by sensor and weather conditions, image quality difference is obvious, together
When target size in image of interest is relatively small or dense arrangement.It causes to deposit even if especially consistent target
In huge feature difference;And the actual size of target constrains the ability in feature extraction of algorithm, so traditional detection method
It is difficult to adapt to the target identification under various scenes, recall rate is low, and there are serious false-alarms, is not able to satisfy practical application request.
Fig. 1 is a flow diagram of the embodiment of the present application object detection method.This method can be filled by target detection
Execution, such as terminal device, server, etc. are set, the embodiment of the present application does not make the specific implementation of the object detecting device
It limits.As shown in Figure 1, the embodiment method includes:
Step 110, based on the characteristic information of image, at least one area-of-interest is obtained.
Optionally, the characteristic information of image may include the characteristic pattern of image, carry out to the characteristic pattern of image to be processed
Area-of-interest (Region of Interest, ROI) extracts, and obtains one or more area-of-interests.Area-of-interest pair
Clarification of objective region should be may include, wherein in some implementations, can by the extraction of neural fusion ROI,
The unlimited fixture body of the application obtains the structure of the neural network of area-of-interest, alternatively, can also pass through other machines study side
Method realizes the extraction of ROI, and the embodiment of the present application does not limit this.
In some embodiments, the characteristic information of image can be obtained by image progress feature extraction processing, or
Person, can also obtain the characteristic information of image from other equipment, such as the spy of image that server receiving terminal equipment is sent
Reference breath, etc., the embodiment of the present application does not limit this.
Optionally, the image in the present embodiment can be remote sensing images, since the usual coverage of remote sensing images is larger, because
This target of interest size in image is relatively small or dense arrangement, and directly carrying out target detection to remote sensing images can lead
Cause recall rate low, there are serious false-alarms, are not able to satisfy practical application request.The present embodiment, will by obtaining area-of-interest
It may include mesh target area of interest in remote sensing images to split from original image, highlight the position of target in the picture
It sets.Alternatively, the embodiment of the present application also can be applied to target other lesser situations of proportion in the picture, the application is implemented
Example does not limit this.
Step 120, it is based on area-of-interest, obtains corresponding at least two target area of area-of-interest.
Wherein, area-of-interest is within target area.
In some implementations, at least two target areas include area-of-interest and by area-of-interest into
At least one magnification region that row enhanced processing obtains.Alternatively, at least two target areas include by area-of-interest into
At least two magnification regions, etc. that row enhanced processing obtains, the embodiment of the present application does not limit this.
In one or more optional embodiments, enhanced processing at least twice is carried out to area-of-interest, is obtained at least
Two target areas.
Belong to the partial region in image due to area-of-interest, it is possible to can exist target not or not exclusively feel it is emerging
In interesting region the case where (target may a part overseas in region of interest), the case where in order to avoid omitting target, the present embodiment
The characteristic information of bigger regional scope (target area after enhanced processing) is added, to guarantee to obtain the accurate of object detection results
Property.Optionally, which can be carried out centered on the center of area-of-interest, that is to say, that the center of magnification region
It is identical as the center of area-of-interest.Enhanced processing respectively corresponds different amplification factors at least twice for this, such as: it will be interested
1.2 times and 1.5 times are amplified in region respectively, obtain two target areas.It in some implementations, can also be by area-of-interest
As a target area, it is understood that for area-of-interest is amplified 1 times of obtained target area.
The present embodiment is in order to more accurately obtain target position, using the context language of more ponds (Multi-Pool)
Adopted information fusion is basic frame with the corresponding region of area-of-interest, which is amplified different multiples, so as to extract
Partial information other than area-of-interest realizes extraction and fusion to content (content) information, such as: if when only
Observe ROI region feature can not judgment object classification when, usual observation information in larger scope can obtain more acurrate
Judgement.
Step 130, the characteristic information based at least two target areas determines the corresponding target detection knot of area-of-interest
Fruit.
By combining the characteristic information of at least two target areas, more contextual informations are obtained, are conducive to mention inspection
Survey the accuracy of result.
Wherein, the characteristic information of target area can be divided from the characteristic information of image and obtain.
Based on the above embodiments of the present application provide a kind of object detection method, the characteristic information based on image, obtain to
A few area-of-interest;Based on area-of-interest, corresponding at least two target area of area-of-interest is obtained;Based at least
The characteristic information of two target areas determines the corresponding object detection results of area-of-interest, more contextual informations is merged
It has arrived in area-of-interest, has improved the ability to express of area-of-interest, improve detection accuracy.
In one or more optional embodiments, step 130 more accurate object detection results in order to obtain are optional
Ground carries out feature connection processing to the characteristic information of at least two target areas, obtains connection features;Based on connection features, really
Determine the corresponding object detection results of area-of-interest.It wherein, in some implementations, can be to the feature of multiple target areas
Information carries out fusion treatment, or carry out channel superposition, etc., connection features are obtained, the embodiment of the present application is to connection processing
Specific implementation is not construed as limiting.
It in some implementations, can also include: the spy at least two target areas before obtaining connection features
Reference breath (such as: feature vector or characteristic pattern) it is handled, make the identical (example of size of each characteristic information obtained after processing
Such as: the dimension of feature vector is identical or the size of characteristic pattern is identical);Correspondingly, identical at least two target area of connection size
The characteristic information in domain obtains connection features.
Fig. 2 is the schematic diagram that the embodiment of the present application obtains one specific example of connection features by way of more ponds.Such as Fig. 2
Shown, the smallest frame indicates area-of-interest in left-side images, amplifies different multiples based on area-of-interest and obtains 3 respectively
Different size of target area, then down-sampling operation is carried out to 3 different size of target areas, it obtains and area-of-interest phase
With the region of size;Or down-sampling operation is carried out to target area, interpolation (such as: bilinear interpolation) is carried out to area-of-interest
Operation, to obtain the identical target area of size and area-of-interest;Based on the identical target area of size and area-of-interest
Determine objective result.The present embodiment can be applied to the target detection of remote sensing image, since the target size in remote sensing image is smaller,
The feature only extracted in many cases to target region can not well be described target, often make
At very high error, using the feature for carrying out different scale after more ponds (Multi pool) to target region of interest (ROI)
It extracts and merges, target peripheral part environmental information can be integrated into target signature so together, realized and context language
The combination of adopted information, reduces error, improves target detection precision.
Optionally, connection features are based on, determine the corresponding object detection results of area-of-interest, comprising:
Based on connection features, the first testing result is obtained;
The characteristic information of characteristic information and image based on area-of-interest, obtains the second testing result;
Based on the first testing result and the second testing result, the corresponding object detection results of area-of-interest are determined.
It, can be by two neural networks respectively to connection features and interested in order to obtain more accurate object detection results
The characteristic information in region and the characteristic information of image are handled, and the first testing result and the second testing result are obtained, specifically,
Connection features can be handled by the first detection neural network, obtain the first testing result, pass through the second detection nerve
Network handles the characteristic information of area-of-interest and the characteristic information of image, obtains the second testing result, it is then possible to
Comprehensive first testing result and the second testing result obtain the object detection results of area-of-interest.For example, the first detection nerve
Network is region convolutional neural networks (Region Convolutional Neural Network, RCNN), the second detection nerve
Network is the full convolutional network (Region-based Fully Convolutional Network, RFCN) based on region, is led to
It crosses RCNN and RFCN and carries out target detection respectively, obtain the first testing result and the second testing result, and integrate two networks
The object detection results that output result obtains are more acurrate.
Optionally, obtain object detection results mode may include: to the first testing result and the second testing result into
Row average treatment obtains the corresponding object detection results of area-of-interest.
Wherein, average treatment can include but is not limited to: computational geometry is average, calculate arithmetic mean, weighting is averaging, this
Apply for the mode of unlimited fixture body average treatment;The first testing result is combined by the object detection results that average treatment obtains
With the second testing result, false alarm rate is reduced while guaranteeing that height recalls precision, promotes detection accuracy.
It should be understood that can be applied to some or all area-of-interests of image, example to the description of area-of-interest above
Such as, above-mentioned process flow can be carried out to each area-of-interest of image, the embodiment of the present application does not limit this.
In one or more optional embodiments, prior to step 110, can also include:
Feature extraction is carried out to image, obtains the characteristic information of image.
Feature extraction is carried out to image, can be and realized by convolutional neural networks or at least one convolutional layer etc., this Shen
Please unlimited fixture body obtain the mode of characteristics of image, but in order to handle characteristics of image faster, optional structure is simpler, processing
Faster neural network carries out feature extraction, realizes the lightweight network of whole network, promotes target detection neural network and increases
Image-capable and processing speed.
In one or more optional embodiments, the present embodiment method utilizes target detection neural fusion, accordingly
It ground can also be using with markup information before carrying out target detection to image to be processed using target detection neural network
Sample image training objective detect neural network, at this point, using the markup information of sample image as supervise.Alternatively, can also be with
Neural network is detected using unsupervised mode training objective, the embodiment of the present application does not limit this.
In some optional implementations, which includes feature extraction network, area-of-interest
Extract network, the first detection neural network and the second detection neural network.Correspondingly, sample image, training characteristics can be based on
Extract network, region of interesting extraction network, the first detection neural network and the second detection neural network.Alternatively, the target is examined
Surveying neural network also may include other networks, and the embodiment of the present application does not limit this.
In the embodiment of the present application, the markup information of sample image can be concentrated based on sample image and passes through the target
Detection neural network carries out the object detection results that target detection obtains to the sample image, adjusts target detection neural network
Network parameter.For example, feature extraction network, region of interesting extraction network in adjustable target detection neural network,
The network parameter of at least one network in first detection neural network and the second detection neural network.
Neural network needs to be trained before the use, so that it has more specific aim to current task, such as: for dividing
Generic task, the sample image feature set based on known classification results are trained target detection neural network, the net after training
Network can obtain more accurately predicting classification results.
In some implementations, target detection neural network to the processing mode of sample image can with to be processed
The processing mode of image is identical.
In some instances, the feature extraction network that can use in target detection neural network carries out spy to sample image
Extraction process is levied, sample characteristics are obtained;
Sample characteristics are inputted into the region of interesting extraction network in target detection neural network, obtain at least one prediction
Area-of-interest and the characteristic information of each prediction area-of-interest corresponding at least two prediction target area;
Using the first detection neural network to each prediction area-of-interest corresponding at least two prediction target area
Characteristic information is handled, and the first prediction result of each prediction area-of-interest is obtained;
The first prediction result and sample graph based on each prediction area-of-interest at least one prediction area-of-interest
The markup information of picture obtains the first-loss of each prediction area-of-interest;
Based on the first-loss of each prediction area-of-interest at least one prediction area-of-interest, feature extraction is adjusted
The network parameter of network, region of interesting extraction network and the first detection neural network.
In some embodiments, it is also based on the network parameter of first-loss adjustment the second detection neural network.Example
Such as, it can be predicted based on the first-loss of each prediction area-of-interest at least one prediction area-of-interest at least one
Area-of-interest obtains the second loss;Based on the second loss, the ginseng of feature extraction network and the second detection neural network is adjusted
Number.
The first detection neural network and the second detection neural network are individually trained in the training process, sample image warp
It crosses feature extraction network and the first detection neural network can be obtained first-loss, realized based on the first-loss to feature extraction
The training of network and the first detection neural network;And the input of the second detection neural network is to combine prediction sense emerging with first-loss
What interesting region determined, feature extraction network and the second detection neural network are trained by the second loss, the mesh after training
Mark detection neural network first detects neural network and the second detection neural network can export more accurate first detection knot respectively
Fruit and the second testing result, to obtain more accurate object detection results.
Optionally, the second loss is obtained based on first-loss and at least one prediction area-of-interest, comprising:
Based on first-loss, screening obtains at least one error area-of-interest from least one prediction area-of-interest;
Optionally, at least one is predicted that the prediction in area-of-interest in first-loss more than default penalty values is interested
Region is determined as error area-of-interest.Alternatively, before at least one being predicted to, first-loss is maximum in area-of-interest
One or more prediction area-of-interest is determined as error area-of-interest.
Judge whether the corresponding first-loss of prediction area-of-interest is larger by preset value, when first-loss is larger
When, illustrate that the prediction area-of-interest is detected by first between the prediction result and annotation results that Processing with Neural Network obtains
Difference is excessive, at this point, handling except through the first detection neural network the prediction area-of-interest, is also added into second
Neural network is detected, is handled, is avoided only by the error area-of-interest that the second detection neural network obtains screening
Slow problem was restrained caused by being handled by the first detection neural network prediction area-of-interest, accelerates training speed
Degree.
It is carried out using characteristic information and sample characteristics of the second detection neural network at least one error area-of-interest
Processing, obtains the second prediction result of at least one error area-of-interest;
Based on the second prediction result of at least one error area-of-interest and the markup information of sample image, second is obtained
Loss.
Optionally, the present embodiment by online hard example excavate screening obtained from least one prediction area-of-interest to
A few error area-of-interest.
It is according to defeated that online hard example, which excavates the core concept of OHEM (online hard example mining) algorithm,
The loss for entering sample is screened, and is filtered out difficult sample (hard example), and expression is affected to classification and detection
Then these samples that screening obtains are applied the training in stochastic gradient descent by sample.Being in actual operation will be original
One area-of-interest network (ROI Network) is extended for two area-of-interest networks (ROI Network), the two
ROI Network shared parameter.Wherein the ROI Network of front one is only preceding to operation, is mainly used for calculating loss;Below
One ROI Network includes that forward and backward operation is calculated and lost simultaneously using difficult sample (hard example) as input
Return gradient.
The present embodiment obtains the true institute's thoughts of forecasting inaccuracy from prediction area-of-interest by OHEM, based on first-loss
Interest region obtains the second loss as error area-of-interest input the second detection neural network, with the second loss instruction
Practice the second detection neural network and feature extraction network.
Fig. 3 is that the embodiment of the present application training objective detects one exemplary structural schematic diagram of neural network.As shown in figure 3,
Using remote sensing image Small object as sample image feature, remote sensing image Small object is detected, network structure specifically includes that feature
It extracts network (conv feature), region of interesting extraction network (ROI), Duo Chihua (Multi Pool) network, be based on area
Neural network (RCNN) is detected in domain, online hard example excavates (OHEM) network and the full convolutional network (RFCN) based on region;
RCNN and RFCN (wherein R refers to region (region)) is the classical architecture of two kinds of target detections, respectively corresponds above-mentioned implementation
The first detection neural network and the second detection neural network in example;First by RCNN network to Preliminary detection to region of interest
Domain carries out preliminary target category classification and boundary and returns (bounding-box regression);It is then based on OHEM network
Descending arrangement is carried out by classification, regression combination loss to preliminary classification result, selects the mistake that there is detection mistake in more maximum probability
Poor area-of-interest;Finally using RFCN network in the more maximum probability selected exist detection mistake error area-of-interest into
The further category classification of row and boundary return, and reduce the probability of detection mistake, strengthen the study to fuzzy difficult sample simultaneously
Secondary detection is carried out to sample, is obviously improved target detection precision.
Optionally, based at least one prediction area-of-interest in it is each prediction area-of-interest the first prediction result and
The markup information of sample image obtains the first-loss of each prediction area-of-interest at least one prediction area-of-interest,
Include:
The first prediction result and sample graph based on each prediction area-of-interest at least one prediction area-of-interest
The markup information of picture obtains the index normalization loss and the loss of thunder lattice of each prediction area-of-interest;
The sum of index normalization loss and the loss of thunder lattice based on prediction area-of-interest, determine prediction area-of-interest
First-loss.
Optionally, index normalization loss (softmax loss) is obtained respectively based on the first prediction result and annotation results
(reg loss) is lost with thunder lattice;Based on the sum of softmax loss and reg loss, first-loss is determined;In this implementation for
Each prediction area-of-interest obtains an a softmax loss and reg loss, can get each prediction region of interest
The first-loss in domain adjusts feature extraction network, region of interesting extraction network and the first inspection by each first-loss respectively
Survey the network parameter of neural network.
Softmax loss and reg loss is obtained respectively based on the first prediction result and label target result to belong to now
Have technology, the present embodiment by the way that trained speed can be accelerated using the sum of softmax loss and reg loss as first-loss,
And it is more preferable to the training effect of feature extraction network, region of interesting extraction network and the first detection neural network.
Optionally, based at least one prediction area-of-interest in it is each prediction area-of-interest the first prediction result and
The markup information of sample image, obtain at least one prediction area-of-interest in it is each prediction area-of-interest first-loss it
Afterwards, further includes:
Based on the first prediction result of each prediction area-of-interest at least one prediction area-of-interest, obtain at least
The center loss of each prediction area-of-interest in one prediction area-of-interest;
Specifically, the first prediction result of each prediction area-of-interest at least one prediction area-of-interest is carried out
Cluster obtains cluster centre;
In the first prediction result and cluster based on each prediction area-of-interest at least one prediction area-of-interest
The distance of the heart obtains the center loss of each prediction area-of-interest.
At least one prediction area-of-interest respectively correspond at least first prediction result, by by least one first
Prediction result is clustered, it is known that difference between the obtained prediction result of difference prediction area-of-interests (it is determined by distance,
Such as: COS distance, Euclidean distance etc.), using the difference as in multitask supervised learning cooperate first-loss it is increased in
The heart loses (center loss) makes different classes of target guarantee the same of class inherited in assorting process to realize supervision
When reduce difference in class, improve target detection precision.
Optionally, the first-loss based on each prediction area-of-interest at least one prediction area-of-interest, adjustment
The network parameter of feature extraction network, region of interesting extraction network and the first detection neural network, comprising:
First-loss and center loss based on each prediction area-of-interest at least one prediction area-of-interest, are adjusted
The parameter of whole feature extraction network, region of interesting extraction network and the first detection neural network.
Optionally, the sum lost based on first-loss and center, adjusts feature extraction network, region of interesting extraction network
With the parameter of the first detection neural network.
It can accelerate the convergence rate of network by the loss of increase center, optionally, center is lost, it is pre- constantly to reduce
The center loss for surveying area-of-interest is target, that is, the first prediction result of same class categories is made to have similar characteristics,
Realize the reduction of difference in class.By losing (center loss) and trained network with first-loss and center, can be improved
The accuracy rate of feature extraction network, region of interesting extraction network and the first detection neural network to classification.
In one or more optional embodiments, screened from least one prediction area-of-interest based on first-loss
Obtain at least one error area-of-interest, comprising:
At least one is predicted that first-loss is more than that the prediction area-of-interest of default loss threshold value is true in area-of-interest
It is set to error area-of-interest.
The present embodiment can excavate OHEM (online hard example mining) algorithm by online hard example and realize,
The prediction area-of-interest (difficult sample (hard example)) that there is detection mistake in more maximum probability is selected by OHEM, and
It determines which prediction area-of-interest has detection mistake in more maximum probability, in this implementation is determined by first-loss,
Such as: the corresponding prediction area-of-interest of the bigger explanation of first-loss more likely there are detection mistake, can be interested by the prediction
Region is determined as error area-of-interest;It is finally emerging to the sense that there is detection mistake in the more maximum probability selected using RFCN network
Interesting region is detected, and target detection precision is obviously improved.
Fig. 4 is that one of the embodiment of the present application object detection method applies exemplary diagram.As shown in figure 4, to a remote sensing shadow
Picture, by the object detection method based on target detection neural network described in any of the above-described embodiment of the application detect to obtain to
A few detection block position, realization detect the position of the aircraft in the remote sensing image, and each frame indicates an inspection in figure
Survey result.
Those of ordinary skill in the art will appreciate that: realize that all or part of the steps of above method embodiment can pass through
The relevant hardware of program instruction is completed, and program above-mentioned can be stored in a computer readable storage medium, the program
When being executed, step including the steps of the foregoing method embodiments is executed;And storage medium above-mentioned includes: ROM, RAM, magnetic disk or light
The various media that can store program code such as disk.
Fig. 5 is a structural schematic diagram of the embodiment of the present application object detecting device.The device of the embodiment can be used for reality
The existing above-mentioned each method embodiment of the application.As shown in figure 5, the device of the embodiment includes:
Region of interesting extraction unit 51 obtains at least one area-of-interest for the characteristic information based on image.
Optionally, the characteristic information of image may include the characteristic pattern of image, carry out to the characteristic pattern of image to be processed
Area-of-interest (Region of Interest, ROI) extracts, and obtains one or more area-of-interests.Area-of-interest pair
Clarification of objective region should be may include, wherein in some implementations, can by the extraction of neural fusion ROI,
The unlimited fixture body of the application obtains the structure of the neural network of area-of-interest, alternatively, can also pass through other machines study side
Method realizes the extraction of ROI, and the embodiment of the present application does not limit this.
In some embodiments, the characteristic information of image can be obtained by image progress feature extraction processing, or
Person, can also obtain the characteristic information of image from other equipment, such as the spy of image that server receiving terminal equipment is sent
Reference breath, etc., the embodiment of the present application does not limit this.
Optionally, the image in the present embodiment can be remote sensing images, since the usual coverage of remote sensing images is larger, because
This target of interest size in image is relatively small or dense arrangement, and directly carrying out target detection to remote sensing images can lead
Cause recall rate low, there are serious false-alarms, are not able to satisfy practical application request.The present embodiment, will by obtaining area-of-interest
It may include mesh target area of interest in remote sensing images to split from original image, highlight the position of target in the picture
It sets.Alternatively, the embodiment of the present application also can be applied to target other lesser situations of proportion in the picture, the application is implemented
Example does not limit this.
Target area obtaining unit 52 obtains corresponding at least two mesh of area-of-interest for being based on area-of-interest
Mark region.
Wherein, area-of-interest is within target area.
In some implementations, at least two target areas include area-of-interest and by area-of-interest into
At least one magnification region that row enhanced processing obtains.Alternatively, at least two target areas include by area-of-interest into
At least two magnification regions, etc. that row enhanced processing obtains, the embodiment of the present application does not limit this.
In one or more optional embodiments, enhanced processing at least twice is carried out to area-of-interest, is obtained at least
Two target areas.
Optionally, which can be carried out centered on the center of area-of-interest, that is to say, that magnification region
Center is identical as the center of area-of-interest.Enhanced processing respectively corresponds different amplification factors at least twice for this, such as: it will feel
1.2 times and 1.5 times are amplified in interest region respectively, obtain two target areas.It in some implementations, can also will be interested
Region is as a target area, it is understood that for area-of-interest is amplified 1 times of obtained target area.
The present embodiment is in order to more accurately obtain target position, using the context language of more ponds (Multi-Pool)
Adopted information fusion is basic frame with the corresponding region of area-of-interest, which is amplified different multiples, so as to extract
Partial information other than area-of-interest realizes extraction and fusion to content (content) information, such as: if when only
Observe ROI region feature can not judgment object classification when, usual observation information in larger scope can obtain more acurrate
Judgement.
Result detection unit 53 determines that area-of-interest is corresponding for the characteristic information based at least two target areas
Object detection results.
By combining the characteristic information of at least two target areas, more contextual informations are obtained, are conducive to mention inspection
Survey the accuracy of result.
Based on a kind of object detecting device that the above embodiments of the present application provide, more contextual informations sense has been fused to
In interest region, the ability to express of area-of-interest is improved, detection accuracy is improved.
In one or more optional embodiments, in order to obtain as a result more accurate object detection results are optionally examined
Survey unit 53, comprising: link block carries out feature connection processing for the characteristic information at least two target areas, obtains
Connection features;Module of target detection determines the corresponding object detection results of area-of-interest for being based on connection features.Wherein,
In some implementations, fusion treatment can be carried out to the characteristic information of multiple target areas, or carries out channel superposition, etc.
Deng obtaining connection features, the embodiment of the present application is not construed as limiting the specific implementation of connection processing.
Optionally, module of target detection obtains the first testing result for being based on connection features;Based on area-of-interest
Characteristic information and image characteristic information, obtain the second testing result;Based on the first testing result and the second testing result, really
Determine the corresponding object detection results of area-of-interest.
Optionally, module of target detection is being based on the first testing result and the second testing result, determines area-of-interest pair
When the object detection results answered, for the first testing result and the second testing result to be averaging processing, region of interest is obtained
The corresponding object detection results in domain.
In one or more optional embodiments, further includes:
Feature extraction unit obtains the characteristic information of image for carrying out feature extraction to image.
Feature extraction is carried out to image, can be and realized by convolutional neural networks or at least one convolutional layer etc., this Shen
Please unlimited fixture body obtain the mode of characteristics of image, but in order to handle characteristics of image faster, optional structure is simpler, processing
Faster neural network carries out feature extraction, realizes the lightweight network of whole network, promotes target detection neural network and increases
Image-capable and processing speed.
In one or more optional embodiments, device combining target detects neural fusion target detection;
The present embodiment device further include:
Network training unit, for based on the sample image with markup information, training objective to detect neural network.This
When, using the markup information of sample image as supervision.Alternatively, nerve net can also be detected using unsupervised mode training objective
Network, the embodiment of the present application do not limit this.
In some optional implementations, target detection neural network includes that feature extraction network, area-of-interest mention
Take network, the first detection neural network and the second detection neural network;
Network training unit, comprising:
Sample process module carries out feature extraction processing to sample image using feature extraction network, obtains sample characteristics;
Sample characteristics are handled using region of interesting extraction network, obtain each prediction at least one prediction area-of-interest
The characteristic information of area-of-interest corresponding at least two prediction target area;Using the first detection neural network to each prediction
The characteristic information of area-of-interest corresponding at least two prediction target area is handled, and each prediction area-of-interest is obtained
The first prediction result;
First-loss module, for first based on each prediction area-of-interest at least one prediction area-of-interest
The markup information of prediction result and sample image obtains each prediction area-of-interest at least one prediction area-of-interest
First-loss;
First-loss training module, for based on each prediction area-of-interest at least one prediction area-of-interest
First-loss, the network parameter of adjustment feature extraction network, region of interesting extraction network and the first detection neural network.
In some embodiments, network training unit can also include: the second loss module, for being based at least one
It predicts the first-loss of each prediction area-of-interest and at least one prediction area-of-interest in area-of-interest, obtains second
Loss;
Second loss training module, for adjusting feature extraction network and the second detection neural network based on the second loss
Network parameter.
The first detection neural network and the second detection neural network are individually trained in the training process, sample image warp
It crosses feature extraction network and the first detection neural network can be obtained first-loss, realized based on the first-loss to feature extraction
The training of network and the first detection neural network;And the input of the second detection neural network is to combine prediction sense emerging with first-loss
What interesting region determined, feature extraction network and the second detection neural network are trained by the second loss, the mesh after training
Mark detection neural network first detects neural network and the second detection neural network can export more accurate first detection knot respectively
Fruit and the second testing result, to obtain more accurate object detection results.
Optionally, the second loss module, specifically for being sieved from least one prediction area-of-interest based on first-loss
Choosing obtains at least one error area-of-interest;Using the second detection neural network to the spy of at least one error area-of-interest
Reference breath and sample characteristics are handled, and second of each error area-of-interest at least one error area-of-interest is obtained
Prediction result;The second prediction result and sample based on error area-of-interest each in error area-of-interest described at least one
The corresponding markup information of this image obtains the second loss.
Optionally, first-loss module is specifically used for emerging based on prediction sense each at least one prediction area-of-interest
First prediction result in interesting region and the markup information of sample image, the index for obtaining each prediction area-of-interest normalize damage
Thunder lattice of becoming estranged loss;The sum of index normalization loss and the loss of thunder lattice based on prediction area-of-interest, determine that prediction is interested
The first-loss in region.
Optionally, network training unit, further includes:
Module is lost at center, for first based on each prediction area-of-interest at least one prediction area-of-interest
Prediction result obtains the center loss of each prediction area-of-interest at least one prediction area-of-interest;
First-loss training module is specifically used for based on each prediction region of interest at least one prediction area-of-interest
The first-loss in domain and center loss, adjustment feature extraction network, region of interesting extraction network and the first detection neural network
Network parameter.
Optionally, module is lost at center, is specifically used for interested in each prediction at least one prediction area-of-interest
First prediction result in region is clustered, and cluster centre is obtained;Based on each prediction at least one prediction area-of-interest
First prediction result of area-of-interest obtains the center loss of each prediction area-of-interest at a distance from cluster centre.
Optionally, second loss module based on first-loss from least one prediction area-of-interest in screening obtain to
When a few error area-of-interest, at least one to be predicted that first-loss is more than default loss threshold value in area-of-interest
Prediction area-of-interest be determined as error area-of-interest.
According to the other side of the embodiment of the present application, a kind of electronic equipment provided, including processor, the processor
Including object detecting device described in any one embodiment as above.
According to the other side of the embodiment of the present application, a kind of electronic equipment that provides, comprising: memory, for storing
Executable instruction;
And processor, it is as above any one to complete that the executable instruction is executed for communicating with the memory
The operation of object detection method described in embodiment.
According to the other side of the embodiment of the present application, a kind of computer readable storage medium provided, based on storing
The instruction that calculation machine can be read, described instruction are performed the operation for executing object detection method described in as above any one embodiment.
According to the other side of the embodiment of the present application, a kind of computer program product provided, including it is computer-readable
Code, when the computer-readable code is run in equipment, the processor in the equipment is executed for realizing such as taking up an official post
It anticipates the instruction of object detection method described in an embodiment.
According to another aspect of the embodiment of the present application, another computer program product provided is calculated for storing
Machine readable instruction, described instruction is performed so that computer executes the inspection of target described in any of the above-described possible implementation
The operation of survey method.
In one or more optional embodiments, the embodiment of the present application also provides a kind of productions of computer program program
Product, for storing computer-readable instruction, described instruction is performed so that computer executes any of the above-described possible realization side
The operation of object detection method described in formula.
The computer program product can be realized especially by hardware, software or its mode combined.In an alternative embodiment
In son, the computer program product is embodied as computer storage medium, in another optional example, the computer
Program product is embodied as software product, such as software development kit (Software Development Kit, SDK) etc..
According to the embodiment of the present application also provides object detection methods and device, electronic equipment, computer storage medium, meter
Calculation machine program product, wherein the characteristic information based on image obtains at least one area-of-interest;Based on area-of-interest, obtain
Obtain corresponding at least two target area of area-of-interest;Based on the characteristic information of at least two target areas, determine interested
The corresponding object detection results in region.
In some embodiments, target detection instruction can be specially call instruction, and first device can pass through calling
Mode indicate second device performance objective detect, accordingly, in response to call instruction is received, second device can be executed
State the step and/or process in any embodiment in object detection method.
It should be understood that the terms such as " first " in the embodiment of the present application, " second " are used for the purpose of distinguishing, and be not construed as
Restriction to the embodiment of the present application.
It should also be understood that in this application, " multiple " can refer to two or more, "at least one" can refer to one,
Two or more.
It should also be understood that clearly being limited or no preceding for the either component, data or the structure that are referred in the application
In the case where opposite enlightenment given hereinlater, one or more may be generally understood to.
It should also be understood that the application highlights the difference between each embodiment to the description of each embodiment,
Same or similar place can be referred to mutually, for sake of simplicity, no longer repeating one by one.
The embodiment of the present application also provides a kind of electronic equipment, such as can be mobile terminal, personal computer (PC), put down
Plate computer, server etc..Below with reference to Fig. 6, it illustrates the terminal device or the services that are suitable for being used to realize the embodiment of the present application
The structural schematic diagram of the electronic equipment 600 of device: as shown in fig. 6, electronic equipment 600 includes one or more processors, communication unit
For example Deng, one or more of processors: one or more central processing unit (CPU) 601, and/or one or more figures
As processor (GPU) 613 etc., processor can according to the executable instruction being stored in read-only memory (ROM) 602 or from
Executable instruction that storage section 608 is loaded into random access storage device (RAM) 603 and execute various movements appropriate and place
Reason.Communication unit 612 may include but be not limited to network interface card, and the network interface card may include but be not limited to IB (Infiniband) network interface card.
Processor can with communicate in read-only memory 602 and/or random access storage device 603 to execute executable instruction,
It is connected by bus 604 with communication unit 612 and is communicated through communication unit 612 with other target devices, to completes the application implementation
The corresponding operation of any one method that example provides, for example, the characteristic information based on image, obtains at least one area-of-interest;
Based on area-of-interest, obtain corresponding at least two target area of area-of-interest, area-of-interest be in target area it
It is interior;Based on the characteristic information of at least two target areas, the corresponding object detection results of area-of-interest are determined.
In addition, in RAM 603, various programs and data needed for being also stored with device operation.CPU601,ROM602
And RAM603 is connected with each other by bus 604.In the case where there is RAM603, ROM602 is optional module.RAM603 storage
Executable instruction, or executable instruction is written into ROM602 at runtime, executable instruction executes central processing unit 601
The corresponding operation of above-mentioned communication means.Input/output (I/O) interface 605 is also connected to bus 604.Communication unit 612 can integrate
Setting, may be set to be with multiple submodule (such as multiple IB network interface cards), and in bus link.
I/O interface 605 is connected to lower component: the importation 606 including keyboard, mouse etc.;It is penetrated including such as cathode
The output par, c 607 of spool (CRT), liquid crystal display (LCD) etc. and loudspeaker etc.;Storage section 608 including hard disk etc.;
And the communications portion 609 of the network interface card including LAN card, modem etc..Communications portion 609 via such as because
The network of spy's net executes communication process.Driver 610 is also connected to I/O interface 605 as needed.Detachable media 611, such as
Disk, CD, magneto-optic disk, semiconductor memory etc. are mounted on as needed on driver 610, in order to read from thereon
Computer program be mounted into storage section 608 as needed.
It should be noted that framework as shown in FIG. 6 is only a kind of optional implementation, it, can root during concrete practice
The component count amount and type of above-mentioned Fig. 6 are selected, are deleted, increased or replaced according to actual needs;It is set in different function component
It sets, separately positioned or integrally disposed and other implementations, such as the separable setting of GPU613 and CPU601 or can also be used
GPU613 is integrated on CPU601, the separable setting of communication unit, can also be integrally disposed on CPU601 or GPU613, etc..
These interchangeable embodiments each fall within protection scope disclosed in the present application.
Particularly, according to an embodiment of the present application, it may be implemented as computer above with reference to the process of flow chart description
Software program.For example, embodiments herein includes a kind of computer program product comprising be tangibly embodied in machine readable
Computer program on medium, computer program include the program code for method shown in execution flow chart, program code
It may include the corresponding instruction of corresponding execution method and step provided by the embodiments of the present application, for example, the characteristic information based on image, obtains
To at least one area-of-interest;Based on area-of-interest, corresponding at least two target area of area-of-interest is obtained, is felt emerging
Interesting region is within target area;Based on the characteristic information of at least two target areas, the corresponding mesh of area-of-interest is determined
Mark testing result.In such embodiments, which can be downloaded and be pacified from network by communications portion 609
Dress, and/or be mounted from detachable media 611.When the computer program is executed by central processing unit (CPU) 601, execute
The operation for the above-mentioned function of being limited in the present processes.
The present processes and device may be achieved in many ways.For example, can by software, hardware, firmware or
Software, hardware, firmware any combination realize the present processes and device.The said sequence of the step of for the method
Merely to be illustrated, the step of the present processes, is not limited to sequence described in detail above, special unless otherwise
It does not mentionlet alone bright.In addition, in some embodiments, also the application can be embodied as to record program in the recording medium, these programs
Including for realizing according to the machine readable instructions of the present processes.Thus, the application also covers storage for executing basis
The recording medium of the program of the present processes.
The description of the present application is given for the purpose of illustration and description, and is not exhaustively or by the application
It is limited to disclosed form.Many modifications and variations are obvious for the ordinary skill in the art.It selects and retouches
Embodiment is stated and be the principle and practical application in order to more preferably illustrate the application, and those skilled in the art is enable to manage
Solution the application is to design various embodiments suitable for specific applications with various modifications.
Claims (10)
1. a kind of object detection method characterized by comprising
Characteristic information based on image obtains at least one area-of-interest;
Based on the area-of-interest, corresponding at least two target area of the area-of-interest, the region of interest are obtained
Domain is within the target area;
Based on the characteristic information of at least two target area, the corresponding object detection results of the area-of-interest are determined.
2. obtaining at least two the method according to claim 1, wherein described be based on the area-of-interest
Target area, comprising:
Enhanced processing at least twice is carried out to the area-of-interest, obtains at least two target areas.
3. method according to claim 1 or 2, which is characterized in that the spy based at least two target area
Reference breath, determines the corresponding object detection results of the area-of-interest, comprising:
Feature connection processing is carried out to the characteristic information of at least two target area, obtains connection features;
Based on the connection features, the corresponding object detection results of the area-of-interest are determined.
4. according to the method described in claim 3, it is characterized in that, it is described be based on the connection features, determine described interested
The corresponding object detection results in region, comprising:
Based on the connection features, the first testing result is obtained;
The characteristic information of characteristic information and described image based on the area-of-interest, obtains the second testing result;
Based on first testing result and second testing result, the corresponding target detection knot of the area-of-interest is determined
Fruit.
5. according to the method described in claim 4, it is characterized in that, described examined based on first testing result with described second
It surveys as a result, determining the corresponding object detection results of the area-of-interest, comprising:
First testing result and second testing result are averaging processing, it is corresponding to obtain the area-of-interest
Object detection results.
6. a kind of object detecting device characterized by comprising
Region of interesting extraction unit obtains at least one area-of-interest for the characteristic information based on image;
Target area obtaining unit obtains the area-of-interest corresponding at least two for being based on the area-of-interest
Target area, the area-of-interest are within the target area;
Result detection unit determines the area-of-interest pair for the characteristic information based at least two target area
The object detection results answered.
7. a kind of electronic equipment, which is characterized in that including processor, the processor includes target inspection as claimed in claim 6
Survey device.
8. a kind of electronic equipment characterized by comprising memory, for storing executable instruction;
And processor, for being communicated with the memory to execute the executable instruction to complete claim 1 to 5 times
The operation of one object detection method of meaning.
9. a kind of computer readable storage medium, for storing computer-readable instruction, which is characterized in that described instruction quilt
Perform claim requires the operation of object detection method described in 1 to 5 any one when execution.
10. a kind of computer program product, including computer-readable code, which is characterized in that when the computer-readable code
When running in equipment, the processor in the equipment is executed for realizing target detection described in claim 1 to 5 any one
The instruction of method.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201810770381.XA CN109165644A (en) | 2018-07-13 | 2018-07-13 | Object detection method and device, electronic equipment, storage medium, program product |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201810770381.XA CN109165644A (en) | 2018-07-13 | 2018-07-13 | Object detection method and device, electronic equipment, storage medium, program product |
Publications (1)
Publication Number | Publication Date |
---|---|
CN109165644A true CN109165644A (en) | 2019-01-08 |
Family
ID=64897861
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201810770381.XA Pending CN109165644A (en) | 2018-07-13 | 2018-07-13 | Object detection method and device, electronic equipment, storage medium, program product |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN109165644A (en) |
Cited By (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110458087A (en) * | 2019-08-07 | 2019-11-15 | 湖南省华芯医疗器械有限公司 | A kind of patient problems position mask method based on deep learning |
CN110738123A (en) * | 2019-09-19 | 2020-01-31 | 创新奇智(北京)科技有限公司 | Method and device for identifying densely displayed commodities |
CN110837789A (en) * | 2019-10-31 | 2020-02-25 | 北京奇艺世纪科技有限公司 | Method and device for detecting object, electronic equipment and medium |
CN111950551A (en) * | 2020-08-14 | 2020-11-17 | 长春理工大学 | Target detection method based on convolutional neural network |
CN114004963A (en) * | 2021-12-31 | 2022-02-01 | 深圳比特微电子科技有限公司 | Target class identification method and device and readable storage medium |
CN116385952A (en) * | 2023-06-01 | 2023-07-04 | 华雁智能科技(集团)股份有限公司 | Distribution network line small target defect detection method, device, equipment and storage medium |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN103345738A (en) * | 2013-06-04 | 2013-10-09 | 北京百度网讯科技有限公司 | Object detection method and device based on area of interest |
CN104067314A (en) * | 2014-05-23 | 2014-09-24 | 中国科学院自动化研究所 | Human-shaped image segmentation method |
US20170206431A1 (en) * | 2016-01-20 | 2017-07-20 | Microsoft Technology Licensing, Llc | Object detection and classification in images |
CN107610113A (en) * | 2017-09-13 | 2018-01-19 | 北京邮电大学 | The detection method and device of Small object based on deep learning in a kind of image |
CN108229524A (en) * | 2017-05-25 | 2018-06-29 | 北京航空航天大学 | A kind of chimney and condensing tower detection method based on remote sensing images |
-
2018
- 2018-07-13 CN CN201810770381.XA patent/CN109165644A/en active Pending
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN103345738A (en) * | 2013-06-04 | 2013-10-09 | 北京百度网讯科技有限公司 | Object detection method and device based on area of interest |
CN104067314A (en) * | 2014-05-23 | 2014-09-24 | 中国科学院自动化研究所 | Human-shaped image segmentation method |
US20170206431A1 (en) * | 2016-01-20 | 2017-07-20 | Microsoft Technology Licensing, Llc | Object detection and classification in images |
CN108229524A (en) * | 2017-05-25 | 2018-06-29 | 北京航空航天大学 | A kind of chimney and condensing tower detection method based on remote sensing images |
CN107610113A (en) * | 2017-09-13 | 2018-01-19 | 北京邮电大学 | The detection method and device of Small object based on deep learning in a kind of image |
Non-Patent Citations (4)
Title |
---|
BOWEN CAI ET AL: "Airport Detection Using End-to-End Convolutional Neural Network with Hard Example Mining", 《REMOTE SENSING》 * |
SHIGUANG WANG ET AL: "PCN: Par t and Context Information for Pedestrian Detection with CNNs", 《ARXIV:1804.04483V1》 * |
公安部第三研究所: "多摄像机协同关注目标检测跟踪技术", 《南京:东南大学出版社》 * |
曾逸琪,关胜晓: "一种基于隔离损失函数的人脸表情识别方法", 《信息技术与网络安全 》 * |
Cited By (11)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110458087A (en) * | 2019-08-07 | 2019-11-15 | 湖南省华芯医疗器械有限公司 | A kind of patient problems position mask method based on deep learning |
CN110458087B (en) * | 2019-08-07 | 2022-03-11 | 湖南省华芯医疗器械有限公司 | Patient problem part labeling method based on deep learning |
CN110738123A (en) * | 2019-09-19 | 2020-01-31 | 创新奇智(北京)科技有限公司 | Method and device for identifying densely displayed commodities |
CN110738123B (en) * | 2019-09-19 | 2020-10-23 | 创新奇智(北京)科技有限公司 | Method and device for identifying densely displayed commodities |
CN110837789A (en) * | 2019-10-31 | 2020-02-25 | 北京奇艺世纪科技有限公司 | Method and device for detecting object, electronic equipment and medium |
CN111950551A (en) * | 2020-08-14 | 2020-11-17 | 长春理工大学 | Target detection method based on convolutional neural network |
CN111950551B (en) * | 2020-08-14 | 2024-03-08 | 长春理工大学 | Target detection method based on convolutional neural network |
CN114004963A (en) * | 2021-12-31 | 2022-02-01 | 深圳比特微电子科技有限公司 | Target class identification method and device and readable storage medium |
CN114004963B (en) * | 2021-12-31 | 2022-03-29 | 深圳比特微电子科技有限公司 | Target class identification method and device and readable storage medium |
CN116385952A (en) * | 2023-06-01 | 2023-07-04 | 华雁智能科技(集团)股份有限公司 | Distribution network line small target defect detection method, device, equipment and storage medium |
CN116385952B (en) * | 2023-06-01 | 2023-09-01 | 华雁智能科技(集团)股份有限公司 | Distribution network line small target defect detection method, device, equipment and storage medium |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN109165644A (en) | Object detection method and device, electronic equipment, storage medium, program product | |
CN111738124B (en) | Remote sensing image cloud detection method based on Gabor transformation and attention | |
CN108921159A (en) | Method and apparatus for detecting the wear condition of safety cap | |
CN110287932B (en) | Road blocking information extraction method based on deep learning image semantic segmentation | |
CN108229418B (en) | Human body key point detection method and apparatus, electronic device, storage medium, and program | |
CN108830221A (en) | The target object segmentation of image and training method and device, equipment, medium, product | |
CN110287849A (en) | A kind of lightweight depth network image object detection method suitable for raspberry pie | |
CN109492638A (en) | Method for text detection, device and electronic equipment | |
CN109154978A (en) | System and method for detecting plant disease | |
CN109784386A (en) | A method of it is detected with semantic segmentation helpers | |
CN106910204B (en) | A kind of method and system to the automatic Tracking Recognition of sea ship | |
CN110147753A (en) | Method and device for detecting small objects in image | |
CN108304816A (en) | Personal identification method, device, storage medium and electronic equipment | |
CN112699832B (en) | Target detection method, device, equipment and storage medium | |
CN113269737B (en) | Fundus retina artery and vein vessel diameter calculation method and system | |
CN108830842A (en) | A kind of medical image processing method based on Corner Detection | |
CN109409210A (en) | A kind of method for detecting human face and system based on SSD frame | |
CN108898096A (en) | A kind of quick accurate extracting method of the information towards high score image | |
CN110533100A (en) | A method of CME detection and tracking is carried out based on machine learning | |
CN108229289A (en) | Target retrieval method, apparatus and electronic equipment | |
CN111524113A (en) | Lifting chain abnormity identification method, system, equipment and medium | |
CN115546187A (en) | Agricultural pest and disease detection method and device based on YOLO v5 | |
CN114155397A (en) | Small sample image classification method and system | |
CN113989519B (en) | Long-tail target detection method and system | |
CN114445268A (en) | Garment style migration method and system based on deep learning |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
RJ01 | Rejection of invention patent application after publication | ||
RJ01 | Rejection of invention patent application after publication |
Application publication date: 20190108 |