CN108121986A - Object detection method and device, computer installation and computer readable storage medium - Google Patents

Object detection method and device, computer installation and computer readable storage medium Download PDF

Info

Publication number
CN108121986A
CN108121986A CN201711484723.3A CN201711484723A CN108121986A CN 108121986 A CN108121986 A CN 108121986A CN 201711484723 A CN201711484723 A CN 201711484723A CN 108121986 A CN108121986 A CN 108121986A
Authority
CN
China
Prior art keywords
region
target
neural networks
network
convolutional neural
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201711484723.3A
Other languages
Chinese (zh)
Other versions
CN108121986B (en
Inventor
牟永强
刘荣杰
裴超
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Shenzhen Intellifusion Technologies Co Ltd
Original Assignee
Shenzhen Intellifusion Technologies Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Shenzhen Intellifusion Technologies Co Ltd filed Critical Shenzhen Intellifusion Technologies Co Ltd
Priority to CN201711484723.3A priority Critical patent/CN108121986B/en
Publication of CN108121986A publication Critical patent/CN108121986A/en
Application granted granted Critical
Publication of CN108121986B publication Critical patent/CN108121986B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/20Image preprocessing
    • G06V10/255Detecting or recognising potential candidate objects based on visual cues, e.g. shapes
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/46Descriptors for shape, contour or point-related descriptors, e.g. scale invariant feature transform [SIFT] or bags of words [BoW]; Salient regional features
    • G06V10/462Salient features, e.g. scale invariant feature transforms [SIFT]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V2201/00Indexing scheme relating to image or video recognition or understanding
    • G06V2201/07Target detection

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • General Health & Medical Sciences (AREA)
  • General Engineering & Computer Science (AREA)
  • Biophysics (AREA)
  • Computational Linguistics (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Computation (AREA)
  • Artificial Intelligence (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • Biomedical Technology (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Health & Medical Sciences (AREA)
  • Multimedia (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Image Analysis (AREA)

Abstract

A kind of object detection method, the described method includes:Training sample set is obtained, the training sample set includes multiple target images for being labeled with target location and target angle type;It is trained using the training sample set pair acceleration region convolutional neural networks model, obtains trained acceleration region convolutional neural networks model;Obtain image to be detected;Target detection is carried out to described image to be detected using the trained acceleration region convolutional neural networks model, obtains the target area of described image to be detected and the target angle type of the target area.The present invention also provides a kind of object detecting device, computer installation and readable storage medium storing program for executing.The present invention can realize the target detection of quick high detection rate.

Description

Object detection method and device, computer installation and computer readable storage medium
Technical field
The present invention relates to technical field of image processing, and in particular to a kind of object detection method and device, computer installation And computer readable storage medium.
Background technology
Existing target detection technique includes the target detection of the complex characteristic based on simple pixel characteristic or hand-designed. Using simple pixel characteristic, such as representative HAAR, pixel value difference etc., although computational efficiency is high, real-time is good, The robustness of the factors such as the background variation for complicated variety is poor, is short of in accuracy of detection.And based on hand-designed Complex characteristic, such as HOG in DPM etc., although feature representation is more preferable, robustness is stronger, because GPU cannot be used Accelerate, calculated on CPU complicated, it is difficult to reach the requirement of real-time.
Existing target detection technique further includes the target detection based on convolutional neural networks.Based on convolutional neural networks Although object detection method improves the precision of detection, but the thing followed is the significantly promotion of calculation amount.Although GPU is calculated The computational problem of extraction convolution feature is solved, but candidate region extraction still expends for quite a long time.It is further, since entire Scheme is first to extract candidate region, then the frame flow classified, and leads to not realize and detect end to end, applies It is relatively cumbersome.
Further, since shooting angle is different, bigger variation can occur on the image for the appearance of target object, existing Target detection technique does not consider the problems of shooting angle, and the recall rate for causing target is relatively low.
The content of the invention
In view of the foregoing, it is necessary to propose a kind of object detection method and device, computer installation and computer-readable Storage medium can realize the target detection of quick high detection rate.
The first aspect of the application provides a kind of object detection method, the described method includes:
Training sample set is obtained, the training sample set includes multiple mesh for being labeled with target location and target angle type Logo image;
It is trained using the training sample set pair acceleration region convolutional neural networks model, obtains trained acceleration Region convolutional neural networks model, the acceleration region convolutional neural networks model include region and suggest network and fast area volume Product neutral net, the region suggest that network and the fast area convolutional neural networks share convolutional layer, and the convolutional layer carries The training sample is taken to concentrate the characteristic pattern of each target image, the region suggests network according to obtaining the characteristic pattern The target angle type of candidate region and the candidate region in each target image, the fast area convolutional Neural net Network is screened and adjusted to the candidate region according to the characteristic pattern, obtain each target image target area and The target angle type of the target area;
Obtain image to be detected;
Target is carried out to described image to be detected using in the trained acceleration region convolutional neural networks model Detection, obtains the target area of described image to be detected and the target angle type of the target area.
It is described to use the training sample set pair acceleration region convolutional neural networks mould in alternatively possible realization method Type be trained including:
(1) region described in Imagenet model initializations is used to suggest network, using described in training sample set training Suggest network in region;
(2) region in (1) after training is used to suggest that network generates the candidate region of each target image, utilizes institute It states candidate region and trains the fast area convolutional neural networks;
(3) the fast area convolutional neural networks in (2) after training is used to initialize the region and suggest network, use institute Stating training sample set trains the region to suggest network;
(4) region in (3) after training is used to suggest fast area convolutional neural networks described in netinit, and is kept The convolutional layer is fixed, and the fast area convolutional neural networks are trained using the training sample set.
It is described to use the training sample set pair acceleration region convolutional neural networks mould in alternatively possible realization method Type be trained including:
Network and the fast area convolutional neural networks, which are trained, is suggested to the region using back-propagation algorithm, The network parameter that network and the fast area convolutional neural networks are suggested in the region is adjusted in training process, makes loss function It minimizes, wherein the loss function includes target classification loss, angle Classification Loss and returns loss.
In alternatively possible realization method, the acceleration region convolutional neural networks model uses ZF frames, the area Suggest that network and the fast area convolutional neural networks share 5 convolutional layers in domain.
Negative sample difficulty example is added in alternatively possible realization method, in the training of the fast area convolutional network to excavate Method.
The second aspect of the application provides a kind of object detecting device, and described device includes:
First acquisition unit, for obtaining training sample set, the training sample set is labeled with target location including multiple With the target image of target angle type;
Training unit for being trained using the training sample set pair acceleration region convolutional neural networks model, is obtained To trained acceleration region convolutional neural networks model, the acceleration region convolutional neural networks model includes region and suggests net Network and fast area convolutional neural networks, the region suggest that network and the fast area convolutional neural networks share convolution Layer, the convolutional layer extract the characteristic pattern that the training sample concentrates each target image, and the region suggests network according to institute The target angle type of candidate region and the candidate region that characteristic pattern is obtained in each target image is stated, it is described fast Fast region convolutional neural networks are screened and adjusted to the candidate region according to the characteristic pattern, obtain each target The target area of image and the target angle type of the target area;
Second acquisition unit, for obtaining image to be detected;
Detection unit, for using in trained acceleration region convolutional neural networks model to described image to be detected Target detection is carried out, obtains the target area of described image to be detected and the target angle type of the target area.
In alternatively possible realization method, the training unit is specifically used for:
(1) region described in Imagenet model initializations is used to suggest network, using described in training sample set training Suggest network in region;
(2) region in (1) after training is used to suggest that network generates the candidate region of each target image, utilizes institute It states candidate region and trains the fast area convolutional neural networks;
(3) the fast area convolutional neural networks in (2) after training is used to initialize the region and suggest network, use institute Stating training sample set trains the region to suggest network;
(4) region in (3) after training is used to suggest fast area convolutional neural networks described in netinit, and is kept The convolutional layer is fixed, and the fast area convolutional neural networks are trained using the training sample set.
In alternatively possible realization method, the training unit is specifically used for:
Network and the fast area convolutional neural networks, which are trained, is suggested to the region using back-propagation algorithm, The network parameter that network and the fast area convolutional neural networks are suggested in the region is adjusted in training process, makes loss function It minimizes, wherein the loss function includes target classification loss, angle Classification Loss and returns loss.
The third aspect of the application provides a kind of computer installation, and the computer installation includes processor, the processing Device is used to realize the object detection method when performing the computer program stored in memory.
The fourth aspect of the application provides a kind of computer readable storage medium, is stored thereon with computer program, described The object detection method is realized when computer program is executed by processor.
The present invention obtains training sample set, and the training sample set is labeled with target location and target angle class including multiple The target image of type;It is trained, is trained using the training sample set pair acceleration region convolutional neural networks model Acceleration region convolutional neural networks model, the acceleration region convolutional neural networks model includes region and suggests network and quick Region convolutional neural networks, the region suggest that network and the fast area convolutional neural networks share convolutional layer, the volume Lamination extracts the characteristic pattern that the training sample concentrates each target image, and the region suggests that network is obtained according to the characteristic pattern Take the target angle type of the candidate region and the candidate region in each target image, the fast area convolution Neutral net is screened and adjusted to the candidate region according to the characteristic pattern, obtains the target of each target image Region and the target angle type of the target area;Obtain image to be detected;Utilize the trained acceleration region convolution Target detection is carried out to described image to be detected in neural network model, obtains the target area and institute of described image to be detected State the target angle type of target area.
The existing target detection based on convolutional neural networks using selective search algorithm generate candidate region, take compared with It is more, and extracted region and target detection are separated.Present invention introduce region in acceleration region convolutional neural networks model is built Network is discussed, candidate region is extracted using depth convolutional neural networks.After network training, by sharing convolution network parameter Method, the characteristic pattern that image is obtained by convolutional layer can be applied to extracted region and target detection simultaneously, that is, share The result of calculation of convolutional network so as to the speed of significantly lifting region extraction, accelerates the speed of entire testing process, realizes end To the detection scheme at end.Also, the present invention considers the problem of shooting angle difference causes recall rate to reduce, using being labeled with mesh The target image of mark angular type is trained acceleration region convolutional neural networks model, improves the recall rate of target.Cause This, the present invention can realize the target detection of quick high detection rate.
Description of the drawings
Fig. 1 is the flow chart for the object detection method that the embodiment of the present invention one provides.
Fig. 2 is the schematic diagram that network is suggested in region.
Fig. 3 is the structure chart of object detecting device provided by Embodiment 2 of the present invention.
Fig. 4 is the schematic diagram for the computer installation that the embodiment of the present invention three provides.
Specific embodiment
It is to better understand the objects, features and advantages of the present invention, below in conjunction with the accompanying drawings and specific real Applying example, the present invention will be described in detail.It should be noted that in the case where there is no conflict, embodiments herein and embodiment In feature can be mutually combined.
Elaborate many details in the following description to facilitate a thorough understanding of the present invention, described embodiment only Only it is part of the embodiment of the present invention, instead of all the embodiments.Based on the embodiments of the present invention, ordinary skill Personnel's all other embodiments obtained without making creative work, belong to the scope of protection of the invention.
Unless otherwise defined, all of technologies and scientific terms used here by the article is with belonging to technical field of the invention The normally understood meaning of technical staff is identical.Term used in the description of the invention herein is intended merely to description tool The purpose of the embodiment of body, it is not intended that in the limitation present invention.
Preferably, object detection method of the invention is applied in one or more computer installation.The computer Device be it is a kind of can be according to the instruction for being previously set or storing, the automatic equipment for carrying out numerical computations and/or information processing, Hardware includes but not limited to microprocessor, application-specific integrated circuit (Application Specific Integrated Circuit, ASIC), programmable gate array (Field-Programmable Gate Array, FPGA), digital processing unit (Digital Signal Processor, DSP), embedded device etc..
The computer installation can be that the calculating such as desktop PC, notebook, palm PC and cloud server are set It is standby.The computer installation can with user by modes such as keyboard, mouse, remote controler, touch tablet or voice-operated devices into pedestrian Machine interacts.
Embodiment one
Fig. 1 is the flow chart for the object detection method that the embodiment of the present invention one provides.The object detection method is applied to Computer installation.The object detection method can detect the position of goal-selling in image (such as vehicle, ship), and can To detect the angular type of goal-selling in image (such as front, side, back side).
As shown in Figure 1, the object detection method specifically includes following steps:
101:Obtain training sample set.
The training sample set includes multiple target images for being labeled with target location and target angle type.The target Image is the image for including goal-selling (such as ship, vehicle etc.).The target image can include one or more default Target.The target location represents the position of goal-selling in the target image.The target angle type represents goal-selling Shooting angle (such as front, the back side, side).
In one embodiment, the training sample set includes about 10000 target images.The target location can be with It is labeled as [x, y, w, h], x, y represents the top left co-ordinate of target area, and w represents the width of target area, and h represents target area Height.The target angle type includes positive angular type, flank angle type and back angle type.It is for example, described When object detection method is used to be detected ship, if target image is the direct picture of ship, the target angle marked Type is positive angular type;If target image is the side image of ship, the target angle type marked is flank angle Type;If target image is the back side image of ship, the target angle type marked is back angle type.
102:Use training sample set pair acceleration region convolutional neural networks model (the Faster Region-based Convolution Neural Network, Faster R-CNN) it is trained, obtain trained acceleration region convolutional Neural Network model.
The acceleration region convolutional neural networks model include region suggest network (Region Proposal Network, ) and fast area convolutional neural networks (Fast Region-based Convolution Neural Network, Fast RPN R-CNN).It needs to suggest region network and fast convolution network carry out alternately training.
Network is suggested in the region and the fast area convolutional neural networks have shared convolutional layer, and the convolutional layer is used In the characteristic pattern of extraction image.Suggest that network generates candidate region and the candidate regions of image according to the characteristic pattern in the region The target angle type in domain, and the target angle type of the candidate region of generation and candidate region is inputted into the fast area Convolutional neural networks.The fast area convolutional neural networks are screened and adjusted to the candidate region according to the characteristic pattern It is whole, obtain the target area of image and the target angle type of target area.
Specifically, in training, the convolutional layer extracts the characteristic pattern that the training sample concentrates each target image, institute It states region and suggests that network obtains candidate region and the candidate region in each target image according to the characteristic pattern Target angle type, the fast area convolutional neural networks according to the characteristic pattern to the candidate region carry out screening and Adjustment, obtains the target area of each target image and the target angle type of the target area.
In a preferred embodiment, the acceleration region convolutional neural networks model uses ZF frames, and the region is suggested Network and the fast area convolutional neural networks share 5 convolutional layers.
In one embodiment, the target image that training sample is concentrated can be the image of arbitrary dimension, in input institute Target image is scaled to the image of unified size (such as 1000*600) before stating convolutional layer.In one embodiment, institute The length and width for stating the characteristic pattern of convolutional layer extraction reduce 16 times compared with input picture, and the depth of characteristic pattern is 256.
In one embodiment, being trained using training sample set pair acceleration region convolutional neural networks model can be with Including:
(1) region described in Imagenet model initializations is used to suggest network, using described in training sample set training Suggest network in region.
(2) region in (1) after training is used to suggest that network generation training sample concentrates the candidate regions of each target image The fast area convolutional neural networks are trained in domain using the candidate region.At this point, network and fast area volume are suggested in region Product neutral net shares convolutional layer not yet.
(3) the fast area convolutional neural networks in (2) after training is used to initialize the region and suggest network, use instruction Practicing sample set trains the region to suggest network.
(4) region in (3) after training is used to suggest fast area convolutional neural networks described in netinit, and is kept The convolutional layer is fixed, and the fast area convolutional neural networks are trained using training sample set.At this point, region suggest network and Fast area convolutional neural networks share identical convolutional layer, constitute a unified network.
Fig. 2 is the schematic diagram that network is suggested in region.
After image is by shared convolutional layer, the characteristic pattern of image is obtained.With default size (such as 3*3) on characteristic pattern Sliding window slided according to default step-length (such as step-length be 1), the every position of sliding window corresponds to a central point.When When sliding window slides into a position, to the default scale (such as 3 kinds of scales 128,256,512) of the central point of position application and Default length-width ratio (such as 3 kinds of length-width ratios 1:1、1:2、2:1) anchor frame obtains default quantity (such as 9) candidate region.Pass through Each sliding window is mapped to the feature vector of a low-dimensional by one convolutional layer (convolutional layer is cascaded with shared convolutional layer) In (such as feature vector of 256-d or 512-d).This feature vector is exported to three full articulamentums at the same level, one is mesh Mark classification layer, one is angle classification layer, and one is that border returns layer.The target classification of target classification layer output candidate region obtains Point, it is target (i.e. prospect) or background to be used to indicate candidate region.Candidate region belongs to prospect or background, depending on candidate The registration in region and the target area (region determined by the target location marked) of mark, registration are more than some threshold value Then it is positioned as prospect, registration is then positioned as background less than threshold value.The angle classification score of angle classification layer output candidate region, It is used to indicate the target angle type of candidate region.Border returns the position of the candidate region after layer output fine tuning, for waiting The border of favored area is finely adjusted.
Region suggests that the candidate region that network is chosen is more, if can have been screened according to the target classification score of candidate region The candidate region of dry highest scoring is input to fast area convolutional neural networks, to accelerate the speed of training and detection.
In order to train region suggest network, give each candidate region distribute a label, the label include positive label and Negative label, positive label can distribute to two class candidate regions:(1) with some real goal (Ground Truth, GT) bounding box There is the candidate region that highest IoU (the ratio between Intersection over Union, intersection union) is overlapped;(2) with arbitrary GT sides Boundary's frame has the candidate region of the IoU overlappings more than 0.7.For a GT bounding box, positive label may be distributed to multiple candidate regions Domain.Negative label distribute to be below with the IoU ratios of all GT bounding boxes 0.3 candidate region.Non- just non-negative candidate region There is no any effect to training objective.
Region suggests that the training of network is trained using back-propagation algorithm, and adjustment region suggests network in training process Network parameter, minimize loss function.Suggest the prediction confidence of the candidate region of neural network forecast in loss function indicating area Degree and the difference of true confidence level.In the present embodiment, loss function includes target classification loss, angle Classification Loss and recurrence Lose three parts.
The loss function of image can be defined as:
Wherein, i is the index of candidate region in a training batch (mini-batch).
It is the target classification loss of candidate region.NclsFor the size of training batch, such as 256.piIt is i-th A candidate region is the prediction probability of target.It is GT labels, if candidate region is just (label distributed is positive label, is claimed For positive candidate region),For 1;If candidate region is negative (label distributed is negative label, referred to as negative candidate region),For 0。It may be calculated
It is the angle Classification Loss of candidate region,Meaning be referred to
It is the recurrence loss of candidate region.λ is balance weight, can be taken as 10.NregFor candidate region Quantity.It may be calculatedtiA coordinate vector, i.e. ti=(tx,ty,tw, th), represent that 4 of candidate region parameterize coordinates (such as the coordinate in the candidate region upper left corner and width, height).Be with The coordinate vector of the corresponding GT bounding boxes in positive candidate region, i.e.,(such as real goal region upper left The coordinate and width at angle, height).R is the loss function (smooth with robustnessL1), it is defined as:
Above-described embodiment considers the problem of shooting angle difference causes recall rate to reduce, in acceleration region convolutional Neural net Using the loss function for including angle Classification Loss in the training of network model, candidate regions are calculated according to the target angle type of prediction The angle Classification Loss in domain improves the recall rate of target.
The above-mentioned training method for suggesting network for region.The training method of fast area convolutional network is referred to region and builds The training method of network is discussed, details are not described herein again.
In the present embodiment, negative sample difficulty example is added in the training of fast area convolutional network and excavates (Hard Negative Mining, HNM) method.For being wrongly classified as the negative sample of positive sample by fast area convolutional network (i.e. Difficult example), the information record of these negative samples is got off, during next repetitive exercise, these negative samples are inputted again It is concentrated to training sample, and increases the weight of its loss, enhanced its influence to grader, can so ensure ceaselessly pin Classify to the negative sample being more difficult to so that from the easier to the more advanced, the sample distribution covered is also more various for the feature that grader is acquired Property.
103:Obtain image to be detected.
Image to be detected is to include the image of goal-selling (such as ship).Goal-selling is the detection in image to be detected Object.For example, when carrying out ship detection to image to be detected, goal-selling is the ship in image to be detected.
Image to be detected can be the image received from external equipment, such as the ship figure of the camera shooting near harbour Picture receives the ship image from the camera.
Alternatively, image to be detected can be the image of the computer installation shooting, such as computer installation shooting Ship image.
Alternatively, image to be detected can also be the image read from the memory of the computer installation, such as from institute State the ship image read in the memory of computer installation.
104:Image to be detected is detected using trained acceleration region convolutional neural networks model, is obtained to be checked The target area of altimetric image and the target angle type of the target area.
Specifically, the region suggests that the convolutional layer extraction that network and the fast area convolutional neural networks are shared is to be checked The characteristic pattern of altimetric image.Suggest that network obtains candidate region and institute in image to be detected according to the characteristic pattern in the region State the target angle type of candidate region.The fast area convolutional neural networks are according to the characteristic pattern to the candidate region It is screened and is adjusted, obtain the target area of image to be detected and the target angle type of the target area.
The object detection method of embodiment one obtains training sample set, and the training sample set is labeled with target including multiple Position and the target image of target angle type;It is carried out using the training sample set pair acceleration region convolutional neural networks model Training, obtains trained acceleration region convolutional neural networks model, and the acceleration region convolutional neural networks model includes area Network and fast area convolutional neural networks are suggested in domain, and the region suggests that network and the fast area convolutional neural networks are total to Convolutional layer is enjoyed, the convolutional layer extracts the characteristic pattern that the training sample concentrates each target image, and network is suggested in the region The target angle type of the candidate region and the candidate region in each target image is obtained according to the characteristic pattern, The fast area convolutional neural networks are screened and adjusted to the candidate region according to the characteristic pattern, are obtained described each The target area of a target image and the target angle type of the target area;Obtain image to be detected;Utilize the training Target detection is carried out to described image to be detected in good acceleration region convolutional neural networks model, obtains the mapping to be checked The target area of picture and the target angle type of the target area.
The existing target detection based on convolutional neural networks using selective search algorithm generate candidate region, take compared with It is more, and extracted region and target detection are separated.The object detection method of embodiment one is in acceleration region convolutional neural networks Introduce region suggests network in model, and candidate region is extracted using depth convolutional neural networks.After network training, pass through The method of shared convolution network parameter, the characteristic pattern that image is obtained by convolutional layer can be applied to extracted region and target simultaneously Detection, that is, the result of calculation of shared convolutional network, so as to the speed of significantly lifting region extraction, accelerate entire detection stream The speed of journey realizes detection scheme end to end.Also, the object detection method of embodiment one considers shooting angle difference and draws The problem of recall rate reduces is played, using being labeled with the target image of target angle type to acceleration region convolutional neural networks model It is trained, improves the recall rate of target.Therefore, the object detection method of embodiment one can realize quick high detection rate Target detection.
Embodiment two
Fig. 3 is the structure chart of object detecting device provided by Embodiment 2 of the present invention.As shown in figure 3, the target detection Device 10 can include:First acquisition unit 301, training unit 302, second acquisition unit 303, detection unit 304.
First acquisition unit 301, for obtaining training sample set.
The training sample set includes multiple target images for being labeled with target location and target angle type.The target Image is the image for including goal-selling (such as ship, vehicle etc.).The target image can include one or more default Target.The target location represents the position of goal-selling in the target image.The target angle type represents goal-selling Shooting angle (such as front, the back side, side).
In one embodiment, the training sample set includes about 10000 target images.The target location can be with It is labeled as [x, y, w, h], x, y represents the top left co-ordinate of target area, and w represents the width of target area, and h represents target area Height.The target angle type includes positive angular type, flank angle type and back angle type.It is for example, described When object detection method is used to be detected ship, if target image is the direct picture of ship, the target angle marked Type is positive angular type;If target image is the side image of ship, the target angle type marked is flank angle Type;If target image is the back side image of ship, the target angle type marked is back angle type.
Training unit 302, for using the training sample set pair acceleration region convolutional neural networks model (Faster Region-based Convolution Neural Network, Faster R-CNN) it is trained, obtain trained add Fast region convolutional neural networks model.
The acceleration region convolutional neural networks model include region suggest network (Region Proposal Network, ) and fast area convolutional neural networks (Fast Region-based Convolution Neural Network, Fast RPN R-CNN).It needs to suggest region network and fast convolution network carry out alternately training.
Network is suggested in the region and the fast area convolutional neural networks have shared convolutional layer, and the convolutional layer is used In the characteristic pattern of extraction image.Suggest that network generates candidate region and the candidate regions of image according to the characteristic pattern in the region The target angle type in domain, and the target angle type of the candidate region of generation and candidate region is inputted into the fast area Convolutional neural networks.The fast area convolutional neural networks are screened and adjusted to the candidate region according to the characteristic pattern It is whole, obtain the target area of image and the target angle type of target area.
Specifically, in training, the convolutional layer extracts the characteristic pattern that the training sample concentrates each target image, institute It states region and suggests that network obtains candidate region and the candidate region in each target image according to the characteristic pattern Target angle type, the fast area convolutional neural networks according to the characteristic pattern to the candidate region carry out screening and Adjustment, obtains the target area of each target image and the target angle type of the target area.
In a preferred embodiment, the acceleration region convolutional neural networks model uses ZF frames, and the region is suggested Network and the fast area convolutional neural networks share 5 convolutional layers.
In one embodiment, the target image that training sample is concentrated can be the image of arbitrary dimension, in input institute Target image is scaled to the image of unified size (such as 1000*600) before stating convolutional layer.In one embodiment, institute The length and width for stating the characteristic pattern of convolutional layer extraction reduce 16 times compared with input picture, and the depth of characteristic pattern is 256.
In one embodiment, being trained using training sample set pair acceleration region convolutional neural networks model can be with Including:
(1) region described in Imagenet model initializations is used to suggest network, using described in training sample set training Suggest network in region.
(2) region in (1) after training is used to suggest that network generation training sample concentrates the candidate regions of each target image The fast area convolutional neural networks are trained in domain using the candidate region.At this point, network and fast area volume are suggested in region Product neutral net shares convolutional layer not yet.
(3) the fast area convolutional neural networks in (2) after training is used to initialize the region and suggest network, use instruction Practicing sample set trains the region to suggest network.
(4) region in (3) after training is used to suggest fast area convolutional neural networks described in netinit, and is kept The convolutional layer is fixed, and the fast area convolutional neural networks are trained using training sample set.At this point, region suggest network and Fast area convolutional neural networks share identical convolutional layer, constitute a unified network.
Fig. 2 is the schematic diagram that network is suggested in region.
After image is by shared convolutional layer, the characteristic pattern of image is obtained.With default size (such as 3*3) on characteristic pattern Sliding window slided according to default step-length (such as step-length be 1), the every position of sliding window corresponds to a central point.When When sliding window slides into a position, to the default scale (such as 3 kinds of scales 128,256,512) of the central point of position application and Default length-width ratio (such as 3 kinds of length-width ratios 1:1、1:2、2:1) anchor frame obtains default quantity (such as 9) candidate region.Pass through Each sliding window is mapped to the feature vector of a low-dimensional by one convolutional layer (convolutional layer is cascaded with shared convolutional layer) In (such as feature vector of 256-d or 512-d).This feature vector is exported to three full articulamentums at the same level, one is mesh Mark classification layer, one is angle classification layer, and one is that border returns layer.The target classification of target classification layer output candidate region obtains Point, it is target (i.e. prospect) or background to be used to indicate candidate region.Candidate region belongs to prospect or background, depending on candidate The registration in region and the target area (region determined by the target location marked) of mark, registration are more than some threshold value Then it is positioned as prospect, registration is then positioned as background less than threshold value.The angle classification score of angle classification layer output candidate region, It is used to indicate the target angle type of candidate region.Border returns the position of the candidate region after layer output fine tuning, for waiting The border of favored area is finely adjusted.
Region suggests that the candidate region that network is chosen is more, if can have been screened according to the target classification score of candidate region The candidate region of dry highest scoring is input to fast area convolutional neural networks, to accelerate the speed of training and detection.
In order to train region suggest network, give each candidate region distribute a label, the label include positive label and Negative label, positive label can distribute to two class candidate regions:(1) with some real goal (Ground Truth, GT) bounding box There is the candidate region that highest IoU (the ratio between Intersection over Union, intersection union) is overlapped;(2) with arbitrary GT sides Boundary's frame has the candidate region of the IoU overlappings more than 0.7.For a GT bounding box, positive label may be distributed to multiple candidate regions Domain.Negative label distribute to be below with the IoU ratios of all GT bounding boxes 0.3 candidate region.Non- just non-negative candidate region There is no any effect to training objective.
Region suggests that the training of network is trained using back-propagation algorithm, and adjustment region suggests network in training process Network parameter, minimize loss function.Suggest the prediction confidence of the candidate region of neural network forecast in loss function indicating area Degree and the difference of true confidence level.In the present embodiment, loss function includes target classification loss, angle Classification Loss and recurrence Lose three parts.
The loss function of image can be defined as:
Wherein, i is the index of candidate region in a training batch (mini-batch).
It is the target classification loss of candidate region.NclsFor the size of training batch, such as 256.piIt is i-th A candidate region is the prediction probability of target.It is GT labels, if candidate region is just (label distributed is positive label, is claimed For positive candidate region),For 1;If candidate region is negative (label distributed is negative label, referred to as negative candidate region),For 0。It may be calculated
It is the angle Classification Loss of candidate region,Meaning be referred to
It is the recurrence loss of candidate region.λ is balance weight, can be taken as 10.NregFor candidate region Quantity.It may be calculatedtiA coordinate vector, i.e. ti=(tx,ty,tw, th), represent that 4 of candidate region parameterize coordinates (such as the coordinate in the candidate region upper left corner and width, height).Be with The coordinate vector of the corresponding GT bounding boxes in positive candidate region, i.e.,(such as real goal region upper left The coordinate and width at angle, height).R is the loss function (smooth with robustnessL1), it is defined as:
Above-described embodiment considers the problem of shooting angle difference causes recall rate to reduce, in acceleration region convolutional Neural net Using the loss function for including angle Classification Loss in the training of network model, candidate regions are calculated according to the target angle type of prediction The angle Classification Loss in domain improves the recall rate of target.
The above-mentioned training method for suggesting network for region.The training method of fast area convolutional network is referred to region and builds The training method of network is discussed, details are not described herein again.
In the present embodiment, negative sample difficulty example is added in the training of fast area convolutional network and excavates (Hard Negative Mining, HNM) method.For being wrongly classified as the negative sample of positive sample by fast area convolutional network (i.e. Difficult example), the information record of these negative samples is got off, during next repetitive exercise, these negative samples are inputted again It is concentrated to training sample, and increases the weight of its loss, enhanced its influence to grader, can so ensure ceaselessly pin Classify to the negative sample being more difficult to so that from the easier to the more advanced, the sample distribution covered is also more various for the feature that grader is acquired Property.
Second acquisition unit 303, for obtaining image to be detected.
Image to be detected is to include the image of goal-selling (such as ship).Goal-selling is the detection in image to be detected Object.For example, when carrying out ship detection to image to be detected, goal-selling is the ship in image to be detected.
Image to be detected can be the image received from external equipment, such as the ship figure of the camera shooting near harbour Picture receives the ship image from the camera.
Alternatively, image to be detected can be the image of the computer installation shooting, such as computer installation shooting Ship image.
Alternatively, image to be detected can also be the image read from the memory of the computer installation, such as from institute State the ship image read in the memory of computer installation.
Detection unit 304, for being carried out using trained acceleration region convolutional neural networks model to image to be detected Detection, obtains the target area of image to be detected and the target angle type of the target area.
Specifically, the region suggests that the convolutional layer extraction that network and the fast area convolutional neural networks are shared is to be checked The characteristic pattern of altimetric image.Suggest that network obtains candidate region and institute in image to be detected according to the characteristic pattern in the region State the target angle type of candidate region.The fast area convolutional neural networks are according to the characteristic pattern to the candidate region It is screened and is adjusted, obtain the target area of image to be detected and the target angle type of the target area.
Embodiment two obtains training sample set, and the training sample set is labeled with target location and target angle including multiple The target image of type;It is trained, is trained using the training sample set pair acceleration region convolutional neural networks model Good acceleration region convolutional neural networks model, the acceleration region convolutional neural networks model is including region suggestion network and soon Fast region convolutional neural networks, the region suggests that network and the fast area convolutional neural networks share convolutional layer, described Convolutional layer extracts the characteristic pattern that the training sample concentrates each target image, and the region suggests network according to the characteristic pattern Obtain the target angle type of the candidate region and the candidate region in each target image, the fast area volume Product neutral net is screened and adjusted to the candidate region according to the characteristic pattern, obtains the mesh of each target image Mark the target angle type of region and the target area;Obtain image to be detected;It is rolled up using the trained acceleration region Product neural network model in described image to be detected carry out target detection, obtain described image to be detected target area and The target angle type of the target area.
The existing target detection based on convolutional neural networks using selective search algorithm generate candidate region, take compared with It is more, and extracted region and target detection are separated.The introduce region in acceleration region convolutional neural networks model of embodiment two It is recommended that network, candidate region is extracted using depth convolutional neural networks.After network training, joined by shared convolutional network Several methods, the characteristic pattern that image is obtained by convolutional layer can be applied to extracted region and target detection simultaneously, that is, common The result of calculation of convolutional network is enjoyed, so as to the speed of significantly lifting region extraction, accelerates the speed of entire testing process, realizes Detection scheme end to end.Also, embodiment two considers the problem of shooting angle difference causes recall rate to reduce, and uses mark The target image for having target angle type is trained acceleration region convolutional neural networks model, improves the detection of target Rate.Therefore, embodiment two can realize the target detection of quick high detection rate.
Embodiment three
Fig. 4 is the schematic diagram for the computer installation that the embodiment of the present invention three provides.The computer installation 1 includes memory 20th, processor 30 and the computer program 40 that can be run in the memory 20 and on the processor 30, example are stored in Such as object detection program.The processor 30 is realized when performing the computer program 40 in above-mentioned object detection method embodiment The step of, such as step 101~104 shown in FIG. 1.Alternatively, the processor 30 is realized when performing the computer program 40 The function of each module/unit in above device embodiment, such as the unit 301~304 in Fig. 3.
Illustratively, the computer program 40 can be divided into one or more module/units, it is one or Multiple module/units are stored in the memory 20, and are performed by the processor 30, to complete the present invention.Described one A or multiple module/units can be the series of computation machine program instruction section that can complete specific function, which is used for Implementation procedure of the computer program 40 in the computer installation 1 is described.For example, the computer program 40 can be by First acquisition unit 301, training unit 302, second acquisition unit 303, the detection unit 304 being divided into Fig. 3, each unit tool Body function is referring to embodiment two.
The computer installation 1 can be that the calculating such as desktop PC, notebook, palm PC and cloud server are set It is standby.It will be understood by those skilled in the art that the schematic diagram 4 is only the example of computer installation 1, do not form to computer The restriction of device 1 can include either combining some components or different components, example than illustrating more or fewer components Such as described computer installation 1 can also include input-output equipment, network access equipment, bus.
Alleged processor 30 can be central processing unit (Central Processing Unit, CPU), can also be Other general processors, digital signal processor (Digital Signal Processor, DSP), application-specific integrated circuit (Application Specific Integrated Circuit, ASIC), ready-made programmable gate array (Field- Programmable Gate Array, FPGA) either other programmable logic device, discrete gate or transistor logic, Discrete hardware components etc..General processor can be microprocessor or the processor 30 can also be any conventional processor Deng the processor 30 is the control centre of the computer installation 1, utilizes various interfaces and connection entire computer dress Put 1 various pieces.
The memory 20 can be used for storing the computer program 40 and/or module/unit, and the processor 30 passes through The computer program and/or module/unit and calling that operation or execution are stored in the memory 20 are stored in memory Data in 20 realize the various functions of the computer installation 1.The memory 20 can mainly include storing program area and deposit Store up data field, wherein, storing program area can storage program area, the application program needed at least one function (for example broadcast by sound Playing function, image player function etc.) etc.;Storage data field can be stored uses created data (ratio according to computer installation 1 Such as voice data, phone directory) etc..In addition, memory 20 can include high-speed random access memory, can also include non-easy The property lost memory, such as hard disk, memory, plug-in type hard disk, intelligent memory card (Smart Media Card, SMC), secure digital (Secure Digital, SD) block, flash card (Flash Card), at least one disk memory, flush memory device or other Volatile solid-state part.
If the integrated module/unit of the computer installation 1 is realized in the form of SFU software functional unit and as independently Production marketing or in use, can be stored in a computer read/write memory medium.Based on such understanding, the present invention It realizes all or part of flow in above-described embodiment method, relevant hardware can also be instructed by computer program come complete Into the computer program can be stored in a computer readable storage medium, which is being executed by processor When, it can be achieved that the step of above-mentioned each embodiment of the method.Wherein, the computer program includes computer program code, described Computer program code can be source code form, object identification code form, executable file or some intermediate forms etc..The meter Calculation machine readable medium can include:Can carry the computer program code any entity or device, recording medium, USB flash disk, Mobile hard disk, magnetic disc, CD, computer storage, read-only memory (ROM, Read-Only Memory), random access memory Device (RAM, Random Access Memory), electric carrier signal, telecommunication signal and software distribution medium etc..Need what is illustrated It is that the content that the computer-readable medium includes can be fitted according to legislation in jurisdiction and the requirement of patent practice When increase and decrease, such as in some jurisdictions, according to legislation and patent practice, computer-readable medium, which does not include electric carrier wave, to be believed Number and telecommunication signal.
In several embodiments provided by the present invention, it should be understood that disclosed computer installation and method, it can be with It realizes by another way.For example, computer installation embodiment described above is only schematical, for example, described The division of unit is only a kind of division of logic function, can there is other dividing mode in actual implementation.
In addition, each functional unit in each embodiment of the present invention can be integrated in same treatment unit, it can also That unit is individually physically present, can also two or more units be integrated in same unit.Above-mentioned integrated list The form that hardware had both may be employed in member is realized, can also be realized in the form of hardware adds software function module.
It is obvious to a person skilled in the art that the invention is not restricted to the details of above-mentioned exemplary embodiment, Er Qie In the case of without departing substantially from spirit or essential attributes of the invention, the present invention can be realized in other specific forms.Therefore, no matter From the point of view of which point, the present embodiments are to be considered as illustrative and not restrictive, and the scope of the present invention is by appended power Profit requirement rather than above description limit, it is intended that all by what is fallen within the meaning and scope of the equivalent requirements of the claims Variation includes within the present invention.Any reference numeral in claim should not be considered as to the involved claim of limitation.This Outside, it is clear that one word of " comprising " is not excluded for other units or step, and odd number is not excluded for plural number.It is stated in computer installation claim Multiple units or computer installation can also be realized by same unit or computer installation by software or hardware.The One, the second grade words are used to indicate names, and are not represented any particular order.
Finally it should be noted that the above embodiments are merely illustrative of the technical solutions of the present invention and it is unrestricted, although reference The present invention is described in detail in preferred embodiment, it will be understood by those of ordinary skill in the art that, it can be to the present invention's Technical solution is modified or equivalent substitution, without departing from the spirit and scope of technical solution of the present invention.

Claims (10)

1. a kind of object detection method, which is characterized in that the described method includes:
Training sample set is obtained, the training sample set includes multiple target figures for being labeled with target location and target angle type Picture;
It is trained using the training sample set pair acceleration region convolutional neural networks model, obtains trained acceleration region Convolutional neural networks model, the acceleration region convolutional neural networks model include region and suggest network and fast area convolution god Through network, network is suggested in the region and the fast area convolutional neural networks share convolutional layer, and the convolutional layer extracts institute The characteristic pattern that training sample concentrates each target image is stated, it is described each that the region suggests that network is obtained according to the characteristic pattern The target angle type of candidate region and the candidate region in target image, the fast area convolutional neural networks root The candidate region is screened and adjusted according to the characteristic pattern, obtains the target area of each target image and described The target angle type of target area;
Obtain image to be detected;
Target detection is carried out to described image to be detected using in the trained acceleration region convolutional neural networks model, Obtain the target area of described image to be detected and the target angle type of the target area.
2. the method as described in claim 1, which is characterized in that described to use training sample set pair acceleration region convolution god Through network model be trained including:
(1) region described in Imagenet model initializations is used to suggest network, the region is trained using the training sample set It is recommended that network;
(2) region in (1) after training is used to suggest that network generates the candidate region of each target image, utilizes the time Favored area trains the fast area convolutional neural networks;
(3) the fast area convolutional neural networks in (2) after training is used to initialize the region and suggest network, use the instruction Practicing sample set trains the region to suggest network;
(4) region in (3) after training is used to suggest fast area convolutional neural networks described in netinit, and described in holding Convolutional layer is fixed, and the fast area convolutional neural networks are trained using the training sample set.
3. the method as described in claim 1, which is characterized in that described to use training sample set pair acceleration region convolution god Through network model be trained including:
Network and the fast area convolutional neural networks, which are trained, is suggested to the region using back-propagation algorithm, training The network parameter that network and the fast area convolutional neural networks are suggested in the region is adjusted in the process, makes loss function minimum Change, wherein the loss function includes target classification loss, angle Classification Loss and returns loss.
4. method as claimed any one in claims 1 to 3, which is characterized in that the acceleration region convolutional neural networks mould Type uses ZF frames, and the region suggests that network and the fast area convolutional neural networks share 5 convolutional layers.
5. method as claimed any one in claims 1 to 3, which is characterized in that the training of the fast area convolutional network Middle addition negative sample difficulty example method for digging.
6. a kind of object detecting device, which is characterized in that described device includes:
First acquisition unit, for obtaining training sample set, the training sample set is labeled with target location and mesh including multiple Mark the target image of angular type;
Training unit for being trained using the training sample set pair acceleration region convolutional neural networks model, is instructed The acceleration region convolutional neural networks model perfected, the acceleration region convolutional neural networks model include region suggest network and Fast area convolutional neural networks, the region suggest that network and the fast area convolutional neural networks share convolutional layer, institute It states convolutional layer and extracts the characteristic pattern that the training sample concentrates each target image, the region suggests network according to the feature Figure obtains the target angle type of the candidate region and the candidate region in each target image, the fast area Convolutional neural networks are screened and adjusted to the candidate region according to the characteristic pattern, obtain each target image Target area and the target angle type of the target area;
Second acquisition unit, for obtaining image to be detected;
Detection unit, for being carried out using in trained acceleration region convolutional neural networks model to described image to be detected Target detection obtains the target area of described image to be detected and the target angle type of the target area.
7. device as claimed in claim 6, which is characterized in that the training unit is specifically used for:
(1) region described in Imagenet model initializations is used to suggest network, the region is trained using the training sample set It is recommended that network;
(2) region in (1) after training is used to suggest that network generates the candidate region of each target image, utilizes the time Favored area trains the fast area convolutional neural networks;
(3) the fast area convolutional neural networks in (2) after training is used to initialize the region and suggest network, use the instruction Practicing sample set trains the region to suggest network;
(4) region in (3) after training is used to suggest fast area convolutional neural networks described in netinit, and described in holding Convolutional layer is fixed, and the fast area convolutional neural networks are trained using the training sample set.
8. device as claimed in claim 6, which is characterized in that the training unit is specifically used for:
Network and the fast area convolutional neural networks, which are trained, is suggested to the region using back-propagation algorithm, training The network parameter that network and the fast area convolutional neural networks are suggested in the region is adjusted in the process, makes loss function minimum Change, wherein the loss function includes target classification loss, angle Classification Loss and returns loss.
9. a kind of computer installation, it is characterised in that:The computer installation includes processor, and the processor is deposited for performing The object detection method as any one of claim 1-5 is realized during the computer program stored in reservoir.
10. a kind of computer readable storage medium, is stored thereon with computer program, it is characterised in that:The computer program The object detection method as any one of claim 1-5 is realized when being executed by processor.
CN201711484723.3A 2017-12-29 2017-12-29 Object detection method and device, computer device and computer readable storage medium Active CN108121986B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201711484723.3A CN108121986B (en) 2017-12-29 2017-12-29 Object detection method and device, computer device and computer readable storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201711484723.3A CN108121986B (en) 2017-12-29 2017-12-29 Object detection method and device, computer device and computer readable storage medium

Publications (2)

Publication Number Publication Date
CN108121986A true CN108121986A (en) 2018-06-05
CN108121986B CN108121986B (en) 2019-12-17

Family

ID=62230688

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201711484723.3A Active CN108121986B (en) 2017-12-29 2017-12-29 Object detection method and device, computer device and computer readable storage medium

Country Status (1)

Country Link
CN (1) CN108121986B (en)

Cited By (33)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108921099A (en) * 2018-07-03 2018-11-30 常州大学 Moving ship object detection method in a kind of navigation channel based on deep learning
CN109063740A (en) * 2018-07-05 2018-12-21 高镜尧 The detection model of ultrasonic image common-denominator target constructs and detection method, device
CN109101966A (en) * 2018-06-08 2018-12-28 中国科学院宁波材料技术与工程研究所 Workpiece identification positioning and posture estimation system and method based on deep learning
CN109359683A (en) * 2018-10-15 2019-02-19 百度在线网络技术(北京)有限公司 Object detection method, device, terminal and computer readable storage medium
CN109376619A (en) * 2018-09-30 2019-02-22 中国人民解放军陆军军医大学 A kind of cell detection method
CN109583445A (en) * 2018-11-26 2019-04-05 平安科技(深圳)有限公司 Character image correction processing method, device, equipment and storage medium
CN109583396A (en) * 2018-12-05 2019-04-05 广东亿迅科技有限公司 A kind of region prevention method, system and terminal based on CNN two stages human testing
CN109784385A (en) * 2018-12-29 2019-05-21 广州海昇计算机科技有限公司 A kind of commodity automatic identifying method, system, device and storage medium
CN109886998A (en) * 2019-01-23 2019-06-14 平安科技(深圳)有限公司 Multi-object tracking method, device, computer installation and computer storage medium
CN109934088A (en) * 2019-01-10 2019-06-25 海南大学 Sea ship discrimination method based on deep learning
CN110288082A (en) * 2019-06-05 2019-09-27 北京字节跳动网络技术有限公司 Convolutional neural networks model training method, device and computer readable storage medium
CN110428357A (en) * 2019-08-09 2019-11-08 厦门美图之家科技有限公司 The detection method of watermark, device, electronic equipment and storage medium in image
CN110443244A (en) * 2019-08-12 2019-11-12 深圳市捷顺科技实业股份有限公司 A kind of method and relevant apparatus of graphics process
CN110599456A (en) * 2019-08-13 2019-12-20 杭州智团信息技术有限公司 Method for extracting specific region of medical image
CN110619255A (en) * 2018-06-19 2019-12-27 杭州海康威视数字技术股份有限公司 Target detection method and device
CN110717905A (en) * 2019-09-30 2020-01-21 上海联影智能医疗科技有限公司 Brain image detection method, computer device, and storage medium
CN110796154A (en) * 2018-08-03 2020-02-14 华为技术有限公司 Method, device and equipment for training object detection model
CN110807431A (en) * 2019-11-06 2020-02-18 上海眼控科技股份有限公司 Object positioning method and device, electronic equipment and storage medium
CN110837760A (en) * 2018-08-17 2020-02-25 北京四维图新科技股份有限公司 Target detection method, training method and device for target detection
CN110991531A (en) * 2019-12-02 2020-04-10 中电科特种飞机系统工程有限公司 Training sample library construction method, device and medium based on air-to-ground small and slow target
CN111144398A (en) * 2018-11-02 2020-05-12 银河水滴科技(北京)有限公司 Target detection method, target detection device, computer equipment and storage medium
CN111310775A (en) * 2018-12-11 2020-06-19 Tcl集团股份有限公司 Data training method and device, terminal equipment and computer readable storage medium
CN111462094A (en) * 2020-04-03 2020-07-28 联觉(深圳)科技有限公司 PCBA component detection method and device and computer readable storage medium
WO2020151329A1 (en) * 2019-01-23 2020-07-30 平安科技(深圳)有限公司 Target detection based identification box determining method and device and terminal equipment
CN112001375A (en) * 2020-10-29 2020-11-27 成都睿沿科技有限公司 Flame detection method and device, electronic equipment and storage medium
CN112256906A (en) * 2020-10-23 2021-01-22 安徽启新明智科技有限公司 Method, device and storage medium for marking annotation on display screen
CN112464785A (en) * 2020-11-25 2021-03-09 浙江大华技术股份有限公司 Target detection method and device, computer equipment and storage medium
CN112513935A (en) * 2018-08-10 2021-03-16 奥林巴斯株式会社 Image processing method and image processing apparatus
CN112949614A (en) * 2021-04-29 2021-06-11 成都市威虎科技有限公司 Face detection method and device for automatically allocating candidate areas and electronic equipment
CN113239975A (en) * 2021-04-21 2021-08-10 洛阳青鸟网络科技有限公司 Target detection method and device based on neural network
CN113333321A (en) * 2021-05-11 2021-09-03 北京若贝特智能机器人科技有限公司 Automatic identification and classification conveying method, system and device and storage medium
CN113496223A (en) * 2020-03-19 2021-10-12 顺丰科技有限公司 Method and device for establishing text region detection model
WO2022001501A1 (en) * 2020-06-29 2022-01-06 华为技术有限公司 Data annotation method and apparatus, and computer device and storage medium

Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101271515A (en) * 2007-03-21 2008-09-24 株式会社理光 Image detection device capable of recognizing multi-angle objective
US20150117760A1 (en) * 2013-10-30 2015-04-30 Nec Laboratories America, Inc. Regionlets with Shift Invariant Neural Patterns for Object Detection
WO2015078185A1 (en) * 2013-11-29 2015-06-04 华为技术有限公司 Convolutional neural network and target object detection method based on same
CN106250812A (en) * 2016-07-15 2016-12-21 汤平 A kind of model recognizing method based on quick R CNN deep neural network
CN106647758A (en) * 2016-12-27 2017-05-10 深圳市盛世智能装备有限公司 Target object detection method and device and automatic guiding vehicle following method
CN106845430A (en) * 2017-02-06 2017-06-13 东华大学 Pedestrian detection and tracking based on acceleration region convolutional neural networks
CN106874894A (en) * 2017-03-28 2017-06-20 电子科技大学 A kind of human body target detection method based on the full convolutional neural networks in region
CN104299012B (en) * 2014-10-28 2017-06-30 银河水滴科技(北京)有限公司 A kind of gait recognition method based on deep learning
CN106919978A (en) * 2017-01-18 2017-07-04 西南交通大学 A kind of high ferro contact net support meanss parts recognition detection method

Patent Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101271515A (en) * 2007-03-21 2008-09-24 株式会社理光 Image detection device capable of recognizing multi-angle objective
US20150117760A1 (en) * 2013-10-30 2015-04-30 Nec Laboratories America, Inc. Regionlets with Shift Invariant Neural Patterns for Object Detection
WO2015078185A1 (en) * 2013-11-29 2015-06-04 华为技术有限公司 Convolutional neural network and target object detection method based on same
CN104299012B (en) * 2014-10-28 2017-06-30 银河水滴科技(北京)有限公司 A kind of gait recognition method based on deep learning
CN106250812A (en) * 2016-07-15 2016-12-21 汤平 A kind of model recognizing method based on quick R CNN deep neural network
CN106647758A (en) * 2016-12-27 2017-05-10 深圳市盛世智能装备有限公司 Target object detection method and device and automatic guiding vehicle following method
CN106919978A (en) * 2017-01-18 2017-07-04 西南交通大学 A kind of high ferro contact net support meanss parts recognition detection method
CN106845430A (en) * 2017-02-06 2017-06-13 东华大学 Pedestrian detection and tracking based on acceleration region convolutional neural networks
CN106874894A (en) * 2017-03-28 2017-06-20 电子科技大学 A kind of human body target detection method based on the full convolutional neural networks in region

Non-Patent Citations (4)

* Cited by examiner, † Cited by third party
Title
CHAOYUE CHEN ET AL: "Learning oriented region-based convolutional neural networks for building detection in satellite remote sensing images", 《REMOTE SENSING AND SPATIAL INFORMATION SCIENCES》 *
CHENCHEN ZHU ET AL: "CMS-RCNN:contextual multi-scale region-based CNN for unconstrained face detection", 《DEEP LEARNING FOR BIOMETRICS》 *
杨名等: "甚高速区域卷积神经网络的船舶视频检测方法", 《北京邮电大学学报》 *
胡鹏等: "基于区域卷积神经网络的多目标跟踪算法", 《西南科技大学学报》 *

Cited By (46)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109101966A (en) * 2018-06-08 2018-12-28 中国科学院宁波材料技术与工程研究所 Workpiece identification positioning and posture estimation system and method based on deep learning
CN109101966B (en) * 2018-06-08 2022-03-08 中国科学院宁波材料技术与工程研究所 Workpiece recognition positioning and attitude estimation system and method based on deep learning
CN110619255A (en) * 2018-06-19 2019-12-27 杭州海康威视数字技术股份有限公司 Target detection method and device
CN110619255B (en) * 2018-06-19 2022-08-26 杭州海康威视数字技术股份有限公司 Target detection method and device
CN108921099A (en) * 2018-07-03 2018-11-30 常州大学 Moving ship object detection method in a kind of navigation channel based on deep learning
CN109063740A (en) * 2018-07-05 2018-12-21 高镜尧 The detection model of ultrasonic image common-denominator target constructs and detection method, device
CN110796154B (en) * 2018-08-03 2023-03-24 华为云计算技术有限公司 Method, device and equipment for training object detection model
CN110796154A (en) * 2018-08-03 2020-02-14 华为技术有限公司 Method, device and equipment for training object detection model
US11605211B2 (en) 2018-08-03 2023-03-14 Huawei Cloud Computing Technologies Co., Ltd. Object detection model training method and apparatus, and device
CN112513935A (en) * 2018-08-10 2021-03-16 奥林巴斯株式会社 Image processing method and image processing apparatus
CN110837760B (en) * 2018-08-17 2022-10-14 北京四维图新科技股份有限公司 Target detection method, training method and device for target detection
CN110837760A (en) * 2018-08-17 2020-02-25 北京四维图新科技股份有限公司 Target detection method, training method and device for target detection
CN109376619A (en) * 2018-09-30 2019-02-22 中国人民解放军陆军军医大学 A kind of cell detection method
CN109376619B (en) * 2018-09-30 2021-10-15 中国人民解放军陆军军医大学 Cell detection method
CN109359683A (en) * 2018-10-15 2019-02-19 百度在线网络技术(北京)有限公司 Object detection method, device, terminal and computer readable storage medium
CN111144398A (en) * 2018-11-02 2020-05-12 银河水滴科技(北京)有限公司 Target detection method, target detection device, computer equipment and storage medium
CN109583445B (en) * 2018-11-26 2024-08-02 平安科技(深圳)有限公司 Text image correction processing method, device, equipment and storage medium
CN109583445A (en) * 2018-11-26 2019-04-05 平安科技(深圳)有限公司 Character image correction processing method, device, equipment and storage medium
CN109583396A (en) * 2018-12-05 2019-04-05 广东亿迅科技有限公司 A kind of region prevention method, system and terminal based on CNN two stages human testing
CN111310775B (en) * 2018-12-11 2023-08-25 Tcl科技集团股份有限公司 Data training method, device, terminal equipment and computer readable storage medium
CN111310775A (en) * 2018-12-11 2020-06-19 Tcl集团股份有限公司 Data training method and device, terminal equipment and computer readable storage medium
CN109784385A (en) * 2018-12-29 2019-05-21 广州海昇计算机科技有限公司 A kind of commodity automatic identifying method, system, device and storage medium
CN109934088A (en) * 2019-01-10 2019-06-25 海南大学 Sea ship discrimination method based on deep learning
CN109886998A (en) * 2019-01-23 2019-06-14 平安科技(深圳)有限公司 Multi-object tracking method, device, computer installation and computer storage medium
WO2020151166A1 (en) * 2019-01-23 2020-07-30 平安科技(深圳)有限公司 Multi-target tracking method and device, computer device and readable storage medium
WO2020151329A1 (en) * 2019-01-23 2020-07-30 平安科技(深圳)有限公司 Target detection based identification box determining method and device and terminal equipment
CN110288082A (en) * 2019-06-05 2019-09-27 北京字节跳动网络技术有限公司 Convolutional neural networks model training method, device and computer readable storage medium
CN110428357A (en) * 2019-08-09 2019-11-08 厦门美图之家科技有限公司 The detection method of watermark, device, electronic equipment and storage medium in image
CN110443244A (en) * 2019-08-12 2019-11-12 深圳市捷顺科技实业股份有限公司 A kind of method and relevant apparatus of graphics process
CN110443244B (en) * 2019-08-12 2023-12-05 深圳市捷顺科技实业股份有限公司 Graphics processing method and related device
CN110599456A (en) * 2019-08-13 2019-12-20 杭州智团信息技术有限公司 Method for extracting specific region of medical image
CN110717905B (en) * 2019-09-30 2022-07-05 上海联影智能医疗科技有限公司 Brain image detection method, computer device, and storage medium
CN110717905A (en) * 2019-09-30 2020-01-21 上海联影智能医疗科技有限公司 Brain image detection method, computer device, and storage medium
CN110807431A (en) * 2019-11-06 2020-02-18 上海眼控科技股份有限公司 Object positioning method and device, electronic equipment and storage medium
CN110991531A (en) * 2019-12-02 2020-04-10 中电科特种飞机系统工程有限公司 Training sample library construction method, device and medium based on air-to-ground small and slow target
CN113496223A (en) * 2020-03-19 2021-10-12 顺丰科技有限公司 Method and device for establishing text region detection model
CN111462094A (en) * 2020-04-03 2020-07-28 联觉(深圳)科技有限公司 PCBA component detection method and device and computer readable storage medium
WO2022001501A1 (en) * 2020-06-29 2022-01-06 华为技术有限公司 Data annotation method and apparatus, and computer device and storage medium
CN112256906A (en) * 2020-10-23 2021-01-22 安徽启新明智科技有限公司 Method, device and storage medium for marking annotation on display screen
CN112001375A (en) * 2020-10-29 2020-11-27 成都睿沿科技有限公司 Flame detection method and device, electronic equipment and storage medium
CN112464785A (en) * 2020-11-25 2021-03-09 浙江大华技术股份有限公司 Target detection method and device, computer equipment and storage medium
CN112464785B (en) * 2020-11-25 2024-08-09 浙江大华技术股份有限公司 Target detection method, device, computer equipment and storage medium
CN113239975B (en) * 2021-04-21 2022-12-20 国网甘肃省电力公司白银供电公司 Target detection method and device based on neural network
CN113239975A (en) * 2021-04-21 2021-08-10 洛阳青鸟网络科技有限公司 Target detection method and device based on neural network
CN112949614A (en) * 2021-04-29 2021-06-11 成都市威虎科技有限公司 Face detection method and device for automatically allocating candidate areas and electronic equipment
CN113333321A (en) * 2021-05-11 2021-09-03 北京若贝特智能机器人科技有限公司 Automatic identification and classification conveying method, system and device and storage medium

Also Published As

Publication number Publication date
CN108121986B (en) 2019-12-17

Similar Documents

Publication Publication Date Title
CN108121986A (en) Object detection method and device, computer installation and computer readable storage medium
CN109886998A (en) Multi-object tracking method, device, computer installation and computer storage medium
CN104424634B (en) Object tracking method and device
CN108470354A (en) Video target tracking method, device and realization device
Shen et al. Exemplar-based human action pose correction and tagging
US11574430B2 (en) Method and system for creating animal type avatar using human face
CN110472534A (en) 3D object detection method, device, equipment and storage medium based on RGB-D data
CN111185008B (en) Method and apparatus for controlling virtual character in game
CN112699832B (en) Target detection method, device, equipment and storage medium
WO2022062238A1 (en) Football detection method and apparatus, and computer-readable storage medium and robot
CN107341548A (en) A kind of data processing method, device and electronic equipment
CN106874906A (en) A kind of binarization method of picture, device and terminal
CN109902541A (en) A kind of method and system of image recognition
CN109614990A (en) A kind of object detecting device
CN107871143A (en) Image-recognizing method and device, computer installation and computer-readable recording medium
CN111915657A (en) Point cloud registration method and device, electronic equipment and storage medium
CN114972947B (en) Depth scene text detection method and device based on fuzzy semantic modeling
CN110287857A (en) A kind of training method of characteristic point detection model
CN108229658A (en) The implementation method and device of object detector based on finite sample
CN107895021B (en) image recognition method and device, computer device and computer readable storage medium
CN108256400A (en) The method for detecting human face of SSD based on deep learning
CN112837322A (en) Image segmentation method and device, equipment and storage medium
CN103295026B (en) Based on the image classification method of space partial polymerization description vectors
CN112749576B (en) Image recognition method and device, computing equipment and computer storage medium
CN109712146A (en) A kind of EM multi-threshold image segmentation method and device based on histogram

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant