CN109784145A - Object detection method and storage medium based on depth map - Google Patents

Object detection method and storage medium based on depth map Download PDF

Info

Publication number
CN109784145A
CN109784145A CN201811480757.XA CN201811480757A CN109784145A CN 109784145 A CN109784145 A CN 109784145A CN 201811480757 A CN201811480757 A CN 201811480757A CN 109784145 A CN109784145 A CN 109784145A
Authority
CN
China
Prior art keywords
candidate frame
depth
frame
size
pixel
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201811480757.XA
Other languages
Chinese (zh)
Other versions
CN109784145B (en
Inventor
彭博文
王行
李骊
周晓军
盛赞
李朔
杨淼
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing HJIMI Technology Co Ltd
Original Assignee
Beijing HJIMI Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing HJIMI Technology Co Ltd filed Critical Beijing HJIMI Technology Co Ltd
Priority to CN201811480757.XA priority Critical patent/CN109784145B/en
Publication of CN109784145A publication Critical patent/CN109784145A/en
Application granted granted Critical
Publication of CN109784145B publication Critical patent/CN109784145B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Abstract

A kind of specific objective detection method and storage medium based on depth image, the method are the size of true candidate frame to be defined by the size of object to be measured object, and calculate candidate frame window traversal interval;Depth image is traversed, the central point pixel coordinate of candidate frame is obtained;The depth value of the central point of each candidate frame is obtained, screening obtains effective candidate frame;The actual needs frame for calculating remaining effective candidate frame is long;Filtering threshold is set, the point that depth in effective candidate frame and the depth of central point have big difference is filtered, it is subsequent further to carry out deep learning pretreatment and deep learning.The present invention can increase the step-length of image traversal as far as possible, the candidate frame of partial invalidity can be filtered out again, the side length of candidate frame is calculated according to the full-size(d) of object according to depth to center value, it can prevent from needing to generate multiple dimensioned candidate frame in same position to save a large amount of calculation amount, provide good convenience for fast target detection.

Description

Object detection method and storage medium based on depth map
Technical field
The present invention relates to field of image detection, specifically, being related to a kind of object detection method based on depth map, Ke Yi The initial number of candidate frame is greatly reduced on depth image, so that model calculation times, which are greatly lowered, provides detection efficiency.
Background technique
Image classification, detection and segmentation are three big tasks of computer vision field.And as most basic part, currently Main target detection image object is all RGB image, and the development of structured light technique and TOF technology, depth image gradually at For new data source.
The rapid development of depth learning technology in recent years, specific objective detection speed and precision has significantly in image It improves.But it far can not also reach the real-time detection effect of video image.But for conventional target detection scheme, base There is higher precision and better adaptability in the target detection of deep learning.Common image object detection algorithm uses Several schemes below:
Currently based on the algorithm of target detection of the mainstream of deep learning model, it is segmented into two major classes: (1) two-stage Detection algorithm, the problem of will test are divided into two stages, first generation candidate region (region proposals), then right Candidate region is classified (generally also needing to position refine), and the Typical Representative of this kind of algorithm is based on region proposal R-CNN system algorithm, such as R-CNN, Fast R-CNN, Faster R-CNN etc.;(2) one-stage detection algorithm does not need The region proposal stage, directly generate object class probability and position coordinate value, than more typical algorithm such as YOLO and SSD。
The above method is completely dependent on random candidate frame and extracts, and has the following deficiencies:
(1) candidate frame randomness is very big, does not select targetedly target object;
(2) size of candidate frame needs to carry out multiple dimensioned selection, considerably increases candidate frame quantity;
(3) there may be largely overlapped, candidate frame quantity is caused to increase between candidate frame;
(4) initial convolved image is excessive, and operand is big.
Therefore, the quantity of candidate frame how is reduced in depth image identification, and improved in candidate frame and frame in target Between registration be quickly detected object so that reducing operand in subsequent convolution algorithm, become existing image and know The technical issues of other technology urgent need to resolve.
Summary of the invention
It is an object of the invention to propose a kind of specific objective detection method based on depth map, target object is utilized Full-size(d) information can greatly reduce the initial number of candidate frame on depth image, so that model calculating be greatly lowered Number provides detection efficiency.
To achieve this purpose, the present invention adopts the following technical scheme:
A kind of specific objective detection method based on depth image, includes the following steps:
Traversal step size computation step S110: the size of true candidate frame, and benefit are defined by the size L of object to be measured object Candidate frame window traversal interval Stride is calculated with formula (1)G, unit is pixel
StrideG=0.5*L*fxy/ D formula (1)
Wherein, L is the size of object to be measured object, fxyFor depth transducer master away from unit is that pixel is unit, and D is The maximum distance for the target object for needing to detect;
Image traversal step S120: according to the size of true candidate frame and candidate frame window traversal interval StrideG, Depth image is traversed, and obtains the central point pixel coordinate of all candidate frames;
Effective candidate frame screening step S130: obtaining the depth value of the central point of each candidate frame, it is detected with needs The maximum distance D of target object is compared, and the candidate frame less than maximum distance D is effective candidate frame, otherwise to wait in vain Select frame;
The actual needs frame of effective candidate frame is long to calculate step S140: utilizing the actual depth of the central point of each candidate frame Value d calculates the long L of actual needs frame of remaining effective candidate frame by formula (3)pixel
Lpixel=L*fxy/ d formula (3);
Filtration step S150: according to calculated effective candidate frame the long L of actual needs framepixelFiltering threshold is set, The point that depth in effective candidate frame and the depth of central point have big difference is filtered.In this way, will effectively in candidate frame depth with The depth of central point is more than that the point of filtering threshold filters out, and can filter out the prospect and background dot in candidate frame.
Optionally, in step s 110, the size of true candidate frame be 1.25-1.75 times object to be measured object it is true Size.
Optionally, in step s 110, the full-size(d) for the object to be measured object that the size of true candidate frame is 1.5 times.
Optionally, the filtering threshold is set in the actual needs long L of framepixel1/3~2/3 among.
Optionally, filtering threshold is the actual needs long L of framepixelHalf.
It optionally, further include deep learning pre-treatment step S160: depth image being sampled under fixed resolution, so Pixel value is normalized afterwards.
Optionally, the deep learning after deep learning pre-treatment step S160 is the CNN mould of one multi output of training Type is inputted as single channel image, is exported as new center point coordinate and two classification results.
The present invention further discloses a kind of storage medium, which can be used for storing the executable finger of computer It enables, it is characterised in that: the computer executable instructions execute the above-mentioned target based on depth map when being executed by processor Detection method.
The target scale information of depth map is introduced the step size settings that target detection candidate frame extracts by the present invention, and can use up can The step-length of image traversal can be increased;The filtering of depth to center value can filter out the candidate frame of partial invalidity again simultaneously;Later The side length for calculating candidate frame according to the full-size(d) of object according to depth to center value, can prevent in same position needs Multiple dimensioned candidate frame is generated to save a large amount of calculation amount, provides good convenience for fast target detection.
Detailed description of the invention
Fig. 1 is the flow chart of the specific objective detection side based on depth map of specific embodiment according to the present invention;
Fig. 2 is the schematic diagram of the central point of each candidate frame after traversing depth map in a specific embodiment of the present invention.
Specific embodiment
The present invention is described in further detail with reference to the accompanying drawings and examples.It is understood that this place is retouched The specific embodiment stated is used only for explaining the present invention rather than limiting the invention.It also should be noted that in order to just Only the parts related to the present invention are shown in description, attached drawing rather than entire infrastructure.
A kind of specific objective detection method based on depth image of the present invention, the object implemented are depth image, for The depth image is first according to the size of target object, the defined maximum distance meter for needing to detect target in depth image Target maximum window size is calculated as window interval and traverses step-length;Then according to window interval traversing graph on depth image As generating candidate frame, then judge whether candidate frame is effective according to the depth of the central point of the candidate frame;Then it calculates and effectively waits It selects the actual needs frame of frame long, and using the long filtration fraction prospect of the actual needs frame and background dot, and then is the depth of next step Degree study reduces interference when reducing the quantity and candidate frame study of candidate frame.
Further, referring to Fig. 1, the flow chart of the specific objective detection side based on depth map is shown, including walk as follows It is rapid:
Traversal step size computation step S110: the size of true candidate frame, and benefit are defined by the size L of object to be measured object Candidate frame window traversal interval Stride is calculated with formula (1)G, unit is pixel
StrideG=0.5*L*fxy/ D formula (1)
Wherein, L is the size of object to be measured object, fxyFor depth transducer master away from unit is that pixel is unit, and D is The maximum distance for the target object for needing to detect.
Optionally, the size of true candidate frame can be the full-size(d) of 1.25-1.75 times of object to be measured object, into one Step is preferred, and the size of true candidate frame can be the full-size(d) of 1.5 times of object to be measured object.
Such as assume that the size of object to be measured object is 170mm, then the size of space of candidate frame can be defined as 255mm.
In step s 110, the target scale information of depth map is introduced into the step size settings that target detection candidate frame extracts, The step-length of image traversal can be increased as far as possible.
Image traversal step S120: according to the size of true candidate frame and candidate frame window traversal interval StrideG, Depth image is traversed, and obtains the central point pixel coordinate of all candidate frames.
Referring to fig. 2, the schematic diagram of the central point of each candidate frame after traversing depth map is shown.
Effective candidate frame screening step S130: obtaining the depth value of the central point of each candidate frame, it is detected with needs The maximum distance D of target object is compared, and the candidate frame less than maximum distance D is effective candidate frame, otherwise to wait in vain Select frame.
That is, carrying out the judgement of effective candidate frame using formula (2).
Whether it is effective candidate frame that P is represented, and d indicates the actual depth value of the central point of candidate frame, when depth value is small greater than 0 Candidate frame is effective when D, other situation candidate frames are invalid.
Therefore, the candidate frame of partial invalidity can be filtered out by step S130.
The actual needs frame of effective candidate frame is long to calculate step S140: utilizing the actual depth of the central point of each candidate frame Value d calculates the long L of actual needs frame of remaining effective candidate frame by formula (3)pixel
Lpixel=L*fxy/ d formula (3)
In this step, the side length for calculating candidate frame according to the full-size(d) of object according to depth to center value, can To prevent from needing to generate multiple dimensioned candidate frame in same position to save a large amount of calculation amount.
Filtration step S150: according to calculated effective candidate frame the long L of actual needs framepixelFiltering threshold is set, The point that depth in effective candidate frame and the depth of central point have big difference is filtered.In this way, will effectively in candidate frame depth with The depth of central point is more than that the point of filtering threshold filters out, and can filter out the prospect and background dot in candidate frame.
It in an alternative embodiment, can be in the actual needs long L of framepixel1/3~2/3 setting filtering threshold.
Further alternative, filtering threshold is the actual needs long L of framepixelHalf.
Therefore, by step S110 to S150 complete using depth map full-size(d) target detection candidate frame extract and The processing of part foreground and background point, so that can quickly find target in the deep learning of next step, and reduces phase Close interference.
Deep learning pre-treatment step S160: depth image is sampled under fixed resolution, then carries out pixel value Normalization.
In an alternative embodiment, fixed point can be sampled using bilinearity difference or closest difference arithmetic Under resolution, such as 64*64,128*128, pixel value is then based on cube side length and normalizes to 0~1.
Subsequent deep learning can be, the CNN model of one multi output of training, input as single channel image, export and be New center point coordinate (returning device) and two classification results (classifier).
The present invention further discloses a kind of storage medium, which can be used for storing the executable finger of computer It enables, it is characterised in that:
The computer executable instructions execute the above-mentioned target detection side based on depth map when being executed by processor Method.
Therefore, the target scale information of depth map is introduced the step size settings that target detection candidate frame extracts by the present invention, can To increase the step-length of image traversal as far as possible;The filtering of depth to center value can filter out the candidate of partial invalidity again simultaneously Frame;The side length for calculating candidate frame according to the full-size(d) of object according to depth to center value later, can prevent same Position needs to generate multiple dimensioned candidate frame to save a large amount of calculation amount.Since candidate frame quantity is few, between candidate frame Distance is that half of target size can be positioned so target may be not at the central area of candidate frame by the recurrence device of CNN To the center of object, and classifier may determine that whether target is the target for needing to detect, and detect for fast target Provide good convenience.
The above content is a further detailed description of the present invention in conjunction with specific preferred embodiments, and it cannot be said that A specific embodiment of the invention is only limitted to this, for those of ordinary skill in the art to which the present invention belongs, is not taking off Under the premise of from present inventive concept, several simple deduction or replace can also be made, all shall be regarded as belonging to the present invention by institute Claims of submission determine protection scope.

Claims (8)

1. a kind of specific objective detection method based on depth image, includes the following steps:
Traversal step size computation step S110: the size of true candidate frame is defined by the size L of object to be measured object, and utilizes public affairs Candidate frame window traversal interval Stride is calculated in formula (1)G, unit is pixel
StrideG=0.5*L*fxy/ D formula (1)
Wherein, L is the size of object to be measured object, fxyIt is unit that master for depth transducer, which is pixel away from, unit, and D is to need The maximum distance of the target object of detection;
Image traversal step S120: according to the size of true candidate frame and candidate frame window traversal interval StrideG, to depth Image is traversed, and obtains the central point pixel coordinate of all candidate frames;
Effective candidate frame screening step S130: obtaining the depth value of the central point of each candidate frame, the target that it is detected with needs The maximum distance D of object is compared, otherwise it is invalid candidate frame that the candidate frame less than maximum distance D, which is effective candidate frame,;
The actual needs frame of effective candidate frame is long to calculate step S140: the actual depth value d of the central point of each candidate frame is utilized, The long L of actual needs frame of remaining effective candidate frame is calculated by formula (3)pixel
Lpixel=L*fxy/ d formula (3);
Filtration step S150: according to calculated effective candidate frame the long L of actual needs framepixelFiltering threshold is set, will be had The point that depth and the depth of central point have big difference in effect candidate frame filters.In this way, will depth and center effectively in candidate frame The depth of point is more than that the point of filtering threshold filters out, and can filter out the prospect and background dot in candidate frame.
2. specific objective detection method according to claim 1, it is characterised in that:
In step s 110, the full-size(d) for the object to be measured object that the size of true candidate frame is 1.25-1.75 times.
3. specific objective detection method according to claim 2, it is characterised in that:
In step s 110, the full-size(d) for the object to be measured object that the size of true candidate frame is 1.5 times.
4. specific objective detection method according to claim 1, it is characterised in that:
The filtering threshold is set in the actual needs long L of framepixel1/3~2/3 among.
5. specific objective detection method according to claim 4, it is characterised in that:
Filtering threshold is the actual needs long L of framepixelHalf.
6. specific objective detection method according to claim 1, it is characterised in that:
Further include that deep learning pre-treatment step S160: depth image is sampled under fixed resolution, then by pixel value into Row normalization.
7. specific objective detection method according to claim 6, it is characterised in that:
Deep learning after deep learning pre-treatment step S160 is the CNN model of one multi output of training, is inputted as list Channel image exports as new center point coordinate and two classification results.
8. a kind of storage medium, which can be used for storing computer executable instructions, it is characterised in that:
Computer executable instructions perform claim when being executed by processor requires described in any one of 1-7 based on depth Spend the object detection method of figure.
CN201811480757.XA 2018-12-05 2018-12-05 Target detection method based on depth map and storage medium Active CN109784145B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201811480757.XA CN109784145B (en) 2018-12-05 2018-12-05 Target detection method based on depth map and storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201811480757.XA CN109784145B (en) 2018-12-05 2018-12-05 Target detection method based on depth map and storage medium

Publications (2)

Publication Number Publication Date
CN109784145A true CN109784145A (en) 2019-05-21
CN109784145B CN109784145B (en) 2021-03-16

Family

ID=66496635

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201811480757.XA Active CN109784145B (en) 2018-12-05 2018-12-05 Target detection method based on depth map and storage medium

Country Status (1)

Country Link
CN (1) CN109784145B (en)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111127514A (en) * 2019-12-13 2020-05-08 华南智能机器人创新研究院 Target tracking method and device by robot
CN111738995A (en) * 2020-06-10 2020-10-02 苏宁云计算有限公司 RGBD image-based target detection method and device and computer equipment
CN112070759A (en) * 2020-09-16 2020-12-11 浙江光珀智能科技有限公司 Forklift pallet detection and positioning method and system

Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8111254B2 (en) * 2005-04-11 2012-02-07 Samsung Electronics Co., Ltd. Depth image-based representation method for 3D object, modeling method and apparatus, and rendering method and apparatus using the same
CN105930793A (en) * 2016-04-19 2016-09-07 中山大学 Human body detection method based on SAE characteristic visual learning
CN106778835A (en) * 2016-11-29 2017-05-31 武汉大学 The airport target by using remote sensing image recognition methods of fusion scene information and depth characteristic
US20170220876A1 (en) * 2017-04-20 2017-08-03 GM Global Technology Operations LLC Systems and methods for visual classification with region proposals
CN107301377A (en) * 2017-05-26 2017-10-27 浙江大学 A kind of face based on depth camera and pedestrian's sensory perceptual system
CN107742113A (en) * 2017-11-08 2018-02-27 电子科技大学 One kind is based on the posterior SAR image complex target detection method of destination number
CN108597003A (en) * 2018-04-20 2018-09-28 腾讯科技(深圳)有限公司 A kind of article cover generation method, device, processing server and storage medium
CN108682039A (en) * 2018-04-28 2018-10-19 国网山西省电力公司电力科学研究院 A kind of binocular stereo vision measurement method

Patent Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8111254B2 (en) * 2005-04-11 2012-02-07 Samsung Electronics Co., Ltd. Depth image-based representation method for 3D object, modeling method and apparatus, and rendering method and apparatus using the same
CN105930793A (en) * 2016-04-19 2016-09-07 中山大学 Human body detection method based on SAE characteristic visual learning
CN106778835A (en) * 2016-11-29 2017-05-31 武汉大学 The airport target by using remote sensing image recognition methods of fusion scene information and depth characteristic
US20170220876A1 (en) * 2017-04-20 2017-08-03 GM Global Technology Operations LLC Systems and methods for visual classification with region proposals
CN107301377A (en) * 2017-05-26 2017-10-27 浙江大学 A kind of face based on depth camera and pedestrian's sensory perceptual system
CN107742113A (en) * 2017-11-08 2018-02-27 电子科技大学 One kind is based on the posterior SAR image complex target detection method of destination number
CN108597003A (en) * 2018-04-20 2018-09-28 腾讯科技(深圳)有限公司 A kind of article cover generation method, device, processing server and storage medium
CN108682039A (en) * 2018-04-28 2018-10-19 国网山西省电力公司电力科学研究院 A kind of binocular stereo vision measurement method

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
RAMANPREET SINGH PAHWA等: "Locating 3D Object Proposals: A Depth-Based Online Approach", 《IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY》 *
王美华: "基于神经网络的特定目标检测与识别", 《中国优秀硕士学位论文全文数据库信息科技辑》 *
邓寒冰等: "基于DRGB的运动中肉牛形体部位识别", 《农业工程学报》 *

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111127514A (en) * 2019-12-13 2020-05-08 华南智能机器人创新研究院 Target tracking method and device by robot
CN111127514B (en) * 2019-12-13 2024-03-22 华南智能机器人创新研究院 Method and device for tracking target by robot
CN111738995A (en) * 2020-06-10 2020-10-02 苏宁云计算有限公司 RGBD image-based target detection method and device and computer equipment
WO2021249351A1 (en) * 2020-06-10 2021-12-16 苏宁易购集团股份有限公司 Target detection method, apparatus and computer device based on rgbd image
CN112070759A (en) * 2020-09-16 2020-12-11 浙江光珀智能科技有限公司 Forklift pallet detection and positioning method and system
CN112070759B (en) * 2020-09-16 2023-10-24 浙江光珀智能科技有限公司 Fork truck tray detection and positioning method and system

Also Published As

Publication number Publication date
CN109784145B (en) 2021-03-16

Similar Documents

Publication Publication Date Title
CN110533084B (en) Multi-scale target detection method based on self-attention mechanism
CN103810503B (en) Depth study based method for detecting salient regions in natural image
CN108647694B (en) Context-aware and adaptive response-based related filtering target tracking method
WO2018103608A1 (en) Text detection method, device and storage medium
CN109784145A (en) Object detection method and storage medium based on depth map
CN109741318A (en) The real-time detection method of single phase multiple dimensioned specific objective based on effective receptive field
CN103714345B (en) A kind of method and system of binocular stereo vision detection finger fingertip locus
CN104463892B (en) Based on level set and the pinpoint Colony hybridization dividing methods of GVF Snake
CN113240691A (en) Medical image segmentation method based on U-shaped network
CN110245600B (en) Unmanned aerial vehicle road detection method for self-adaptive initial quick stroke width
CN105493141A (en) Unstructured road boundary detection
CN109087337B (en) Long-time target tracking method and system based on hierarchical convolution characteristics
CN110222712B (en) Multi-special-item target detection algorithm based on deep learning
CN108876818A (en) A kind of method for tracking target based on like physical property and correlation filtering
CN106600965B (en) Traffic flow morning and evening peak period automatic identifying method based on sharpness
CN107610177A (en) A kind of method and apparatus that characteristic point is determined in synchronous superposition
CN106371614A (en) Gesture recognition optimizing method and device
CN103324361B (en) The method and system of location, touch point
CN105930793A (en) Human body detection method based on SAE characteristic visual learning
CN106023184A (en) Depth significance detection method based on anisotropy center-surround difference
CN106683125A (en) RGB-D image registration method based on 2D/3D mode switching
CN110930393A (en) Chip material pipe counting method, device and system based on machine vision
CN116129327A (en) Infrared vehicle detection method based on improved YOLOv7 algorithm
Yuming et al. Traffic signal light detection and recognition based on canny operator
Chen et al. Bi-deformation-UNet: recombination of differential channels for printed surface defect detection

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant