CN109214241A - Pedestrian detection method based on deep learning algorithm - Google Patents

Pedestrian detection method based on deep learning algorithm Download PDF

Info

Publication number
CN109214241A
CN109214241A CN201710531902.1A CN201710531902A CN109214241A CN 109214241 A CN109214241 A CN 109214241A CN 201710531902 A CN201710531902 A CN 201710531902A CN 109214241 A CN109214241 A CN 109214241A
Authority
CN
China
Prior art keywords
candidate frame
loss
scale
analysis
size
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201710531902.1A
Other languages
Chinese (zh)
Inventor
葛水英
杨东明
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
National Science Library Chinese Academy Of Sciences
Original Assignee
National Science Library Chinese Academy Of Sciences
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by National Science Library Chinese Academy Of Sciences filed Critical National Science Library Chinese Academy Of Sciences
Priority to CN201710531902.1A priority Critical patent/CN109214241A/en
Publication of CN109214241A publication Critical patent/CN109214241A/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/103Static body considered as a whole, e.g. static pedestrian or occupant recognition
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/46Descriptors for shape, contour or point-related descriptors, e.g. scale invariant feature transform [SIFT] or bags of words [BoW]; Salient regional features
    • G06V10/462Salient features, e.g. scale invariant feature transforms [SIFT]

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Artificial Intelligence (AREA)
  • Multimedia (AREA)
  • Evolutionary Computation (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • General Engineering & Computer Science (AREA)
  • Health & Medical Sciences (AREA)
  • General Health & Medical Sciences (AREA)
  • Evolutionary Biology (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • Computational Linguistics (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Human Computer Interaction (AREA)
  • Image Analysis (AREA)

Abstract

The invention discloses a kind of pedestrian detection methods based on deep learning algorithm, wherein the method comprising the steps of 1: carrying out feature extraction to input picture using basic network, the additional Analysis On Multi-scale Features layer part in network is utilized to carry out the extraction of Analysis On Multi-scale Features figure;Step 2: each location point generates fixed candidate frame according to different scale and length-width ratio in characteristic pattern;Step 3: the classification of 3*3 convolution is carried out on characteristic pattern, it includes 1 confidence score of pedestrian's object and 4 positional shift values for each candidate frame return;Step 4: loss being calculated according to trained objective function, obtains final training pattern.Gained model carries out pedestrian detection, after obtaining the confidence score and location information of candidate frame, uses non-maximum restraining operation to select final candidate frame and exports as detection.The present invention effectively improves the performance and real-time of pedestrian detection, can reach the 60 frames above speed (GPU of Titan X) per second, 75% or so Average Accuracy in the detection.

Description

Pedestrian detection method based on deep learning algorithm
Technical field
The present embodiments relate to computer image processing technology fields, more particularly, to image detection and target identification side Method.
Background technique
Pedestrian detection is an important branch of field of target recognition, is research hotspot in recent years.Pedestrian detection technology It is basis and the guarantee of the researchs such as pedestrian tracking, behavioural analysis, gait analysis, pedestrian's identification, in such as auxiliary driving, intelligence Possess broad application prospect in the fields such as energy monitoring, advanced man-machine interface.Conventional pedestrian's detection method is using engineer's Feature carries out pedestrian detection, these features are integrated into distorted pattern and block in model later, however these methods are limited In the feature of the engineer of reduced levels, the problems such as pedestrian is blocked can not be solved.Since 2012, convolutional neural networks (Convolutional neural networks, CNNs are used for target detection and introduce pedestrian detection, departing from manually setting The constraint of feature is counted, however due to the complicated network structure and computationally intensive, these methods all have poor real-time.
The present invention proposes a kind of pedestrian detection network model based on deep learning, and detection accuracy and detection speed are relatively existing Method, which has, to be substantially improved and preferably balances.Meanwhile, it is capable to guarantee detection when the lower image of input resolution ratio Precision.The end-to-end design of model entirety is so that trained and detection process becomes simply, can settle at one go.
Summary of the invention
The main purpose of the embodiment of the present invention is to provide a kind of pedestrian detection method based on deep learning algorithm.
To achieve the goals above, according to an aspect of the invention, there is provided following technical scheme:
A kind of pedestrian detection method based on deep learning algorithm, this method include at least:
Step 1: input picture being normalized into unified size as input, utilizes the facilities network in designed depth network Network part carries out the extraction of foundation characteristic figure to input picture.The additional Analysis On Multi-scale Features layer in designed depth network is utilized later Part carries out the extraction of Analysis On Multi-scale Features figure to above-mentioned gained foundation characteristic figure;
Step 2: each location point (feature map location) in Analysis On Multi-scale Features figure described in step 1 is not according to With size scale and a series of corresponding different fixation candidate frames (default box) of generation of length-breadth ratio;
Step 3: the classification of 3*3 convolution, each candidate in each location point are carried out on the Analysis On Multi-scale Features figure described in step 1 It includes 1 confidence score (score) of pedestrian's object and 4 positional shift values (offsets) for frame return.It is every when training Jaccard overlap between a candidate frame and any one ground truth is greater than certain threshold value, is just used as positive sample, Others sort according to confidence level, and that chooses 3 times of positive sample quantity is used as negative sample;
Step 4: loss, and back-propagation and iteration being calculated according to trained objective function, obtain final training pattern.Institute It obtains model and carries out pedestrian detection, wherein picture is sequentially completed step 1~3, obtains the confidence score and location information of candidate frame Afterwards, use non-maximum restraining operation to select final candidate frame to export as detection.
Further, the step 1 specifically includes:
Input picture is normalized into unified size, puts into designed pedestrian detection depth network.The network includes Basic network part and additional features layer.It is taken using Poo15 layers of VGG16 network and framework before the basic network part It builds, comprising 5 convolutional layers and 5 MaxPooling operations, which carries out the extraction of foundation characteristic figure to input picture.Institute Stating additional features layer includes 3 convolutional layers and one pool layers of average.It is related to convolution operation and is followed successively by Conv6 (3*3* 1024), Conv7 (1*1*1024), Conv8_1 (1*1*256), Conv8_2 (3*3*512-s2), Conv9_1 (1*1*128), Conv9_2 (3*3*256-s2), Conv10_1 (1*1*128), Conv10_2 (3*3*256-s2), Avg Pooling: Global, the subnetwork carry out the extraction of Analysis On Multi-scale Features figure to above-mentioned gained foundation characteristic figure.Network overall construction drawing is such as Shown in Fig. 1.
Further, gained different scale characteristic pattern specifically includes in the step 1:
19*19*1024 dimensional feature figure obtained by Conv7, the resulting 10*10*512 dimensional feature figure of Conv8_2, Conv9_2 institute Obtain 5*5*256 dimensional feature figure, 3*3*256 dimensional feature figure obtained by Conv10_2, Avg Pooling:Global gained 1*1*256 dimension Characteristic pattern, totally 5 characteristic patterns, characteristic pattern size variation obtained by these convolutional layers is bigger, for detecting under different scale Object.
Further, each location point in the step 2 in Analysis On Multi-scale Features figure is according to different size scale and length and width A series of corresponding different fixation candidate frames of generation of ratio specifically include:
The corresponding candidate frame for generating 6 kinds of size scales and length-breadth ratio of each location point in Analysis On Multi-scale Features figure.It is described The length-breadth ratio of candidate frame is respectively that default boundary frame, transverse and longitudinal ratio are respectivelyCandidate frame, such as Fig. 2 institute Show.
In 5 characteristic patterns, the candidate frame size in each characteristic pattern calculates as follows:
For above-mentioned default frame, size are as follows:
Wherein, m indicates characteristic pattern number, herein m=5.sminValue 0.2, smaxValue 0.95, it is meant that the ruler of lowermost layer Degree is 0.2, and top scale is 0.95.Then, each candidate frame size are as follows:
The center of each candidate frame is arranged are as follows:Wherein | fk| it is the size of k-th of characteristic pattern, meanwhile, i, j ∈ [0, | fk|]。
Further, the step 3 specifically includes:
The classification of 3*3 convolution is carried out on the Analysis On Multi-scale Features figure described in step 1, each candidate frame in each location point returns It includes 1 confidence score of pedestrian's object and 4 positional shift values.It is defeated altogether in 6 candidate frames on i.e. each location point (1+4) * 6 values out.For the characteristic pattern of a m*n size, i.e., it can generate (1+4) * 6*m*n output result.
In the training stage, the ground truth of input picture and each object is given.For each candidate frame and appoint As soon as the jaccard overlap for the ground truth that anticipates is greater than 0.5, it is used as positive sample, therefore allow a ground Truth corresponds to multiple candidate frames.Other frames sort according to confidence level, and that chooses 3 times of positive sample quantity is used as negative sample.
Further, the step 4 specifically includes:
Trained loss function (objective loss function) loses (confidence loss) by confidence level Loss (confidence loss) two parts weighted sum is returned with position to constitute.Confidence level is using Softmax Loss, position It puts back into and returns, be using Smooth L1 loss.Calculation formula is as follows:
Wherein: N is the candidate frame number to match with ground truth box.LconfIt is defeated using Softmax Loss Enter for the confidence level c of every one kind.LlocUsing Smooth L1 Loss.α is weight term, is set as 1.
The formula of the confidence level loss is as follows:
The formula that the position returns loss is as follows:
Detailed description of the invention
Attached drawing is as a part of the invention, and for providing further understanding of the invention, of the invention is schematic Examples and descriptions thereof are used to explain the present invention, but does not constitute an undue limitation on the present invention.Obviously, the accompanying drawings in the following description Only some embodiments to those skilled in the art without creative efforts, can be with Other accompanying drawings can also be obtained according to these attached drawings.In the accompanying drawings:
Fig. 1 is depth network structure of the invention.
Fig. 2 is default boundary frame, candidate frame length-breadth ratio is respectivelyCandidate frame.
Fig. 3 is network training data sample.
Fig. 4 is using Fig. 3 sample as the output par, c result of the Partial Feature figure of case.
Fig. 5 is that candidate frame matching result is shown by case of a part of Fig. 3 sample.
Fig. 6 is that pedestrian detection exports result case.
These attached drawings and verbal description are not intended to the conception range limiting the invention in any way, but by reference to Specific embodiment is that those skilled in the art illustrate idea of the invention.
Specific embodiment
The technical issues of with reference to the accompanying drawing and specific embodiment is solved to the embodiment of the present invention, used technical side Case and the technical effect of realization carry out clear, complete description.Obviously, described embodiment is only one of the application Divide embodiment, is not whole embodiments.Based on the embodiment in the application, those of ordinary skill in the art are not paying creation Property labour under the premise of, all other equivalent or obvious variant the embodiment obtained is fallen within the scope of protection of the present invention. The embodiment of the present invention can be embodied according to the multitude of different ways being defined and covered by claim.
Pedestrian detection method based on deep learning algorithm includes 4 steps.
Step 1: input picture being normalized into unified size, puts into designed pedestrian detection depth network.When training, For using INRIA pedestrian's data set as training set, the original image of input and the image sample after addition label are respectively such as Fig. 3 (a) and shown in Fig. 3 (b).Training carries out 60K times altogether, is updated using SGD, learning rate is initialized as 0.001, momentum Being set as 0.9, weight decay is 0.0005.
The network includes basic network part and additional features layer.The basic network part uses VGG16 network Poo15 layers and framework before are built, and comprising 5 convolutional layers and 5 MaxPooling operations, the subnetwork is to input picture Carry out the extraction of foundation characteristic figure.The additional features layer includes 3 convolutional layers and one pool layers of average.It is related to convolution Operation is followed successively by Conv6 (3*3*1024), Conv7 (1*1*1024), Conv8_1 (1*1*256), Conv8_2 (3*3*512- S2), Conv9_1 (1*1*128), Conv9_2 (3*3*256-s2), Conv10_1 (1*1*128), Conv10_2 (3*3*256- S2), Avg Pooling:Global, the subnetwork carry out the extraction of Analysis On Multi-scale Features figure to above-mentioned gained foundation characteristic figure.
Further, the different scale characteristic pattern specifically includes:
19*19*1024 dimensional feature figure obtained by Conv7, the resulting 10*10*512 dimensional feature figure of Conv8_2, Conv9_2 institute Obtain 5*5*256 dimensional feature figure, 3*3*256 dimensional feature figure obtained by Conv10_2, Avg Pooling:Global gained 1*1*256 dimension Characteristic pattern, totally 5 characteristic patterns, characteristic pattern size variation obtained by these convolutional layers is bigger, for detecting under different scale Object.Fig. 4 is illustrated using Fig. 3 sample as the output result of the Partial Feature figure of case.
Step 2: each location point in Analysis On Multi-scale Features figure described in step 1 is according to different size scale and length-breadth ratio pair A series of different fixation candidate frames should be generated.Candidate frame specifically includes the corresponding generation of each location point in Analysis On Multi-scale Features figure The candidate frame of 6 kinds of size scales and length-breadth ratio.The length-breadth ratio of the candidate frame is respectively default boundary frame, transverse and longitudinal than difference ForCandidate frame.
Candidate frame size in each characteristic pattern calculates as follows:
For above-mentioned default frame, size are as follows:
Wherein, m indicates characteristic pattern number, herein m=5.sminValue 0.2, smaxValue 0.95, it is meant that the ruler of lowermost layer Degree is 0.2, and top scale is 0.95.Then, each candidate frame size are as follows:
The center of each candidate frame is arranged are as follows:Wherein | fk| it is the size of k-th of characteristic pattern, meanwhile, i, j ∈ [0, | fk|]。
Candidate frame matching result such as Fig. 5 is shown by case of a part of Fig. 3 sample.
Step 3: the classification of 3*3 convolution, each candidate in each location point are carried out on the Analysis On Multi-scale Features figure described in step 1 It includes 1 confidence score of pedestrian's object and 4 positional shift values for frame return.6 candidate frames on i.e. each location point In output (1+4) * 6 value altogether.For the characteristic pattern of a m*n size, i.e., it can generate (1+4) * 6*m*n output result.
In the training stage, the ground truth of input picture and each object is given.For each candidate frame and appoint As soon as the jaccard overlap for the ground truth that anticipates is greater than 0.5, it is used as positive sample, therefore allow a ground Truth corresponds to multiple candidate frames.Other frames sort according to confidence level, and that chooses 3 times of positive sample quantity is used as negative sample.
Step 4: loss, and back-propagation and iteration being calculated according to trained objective function, obtain final training pattern.Instruction Experienced loss function returns loss two parts weighted sum by confidence level loss and position and constitutes.Confidence level is using Softmax Loss, it is using Smooth L1 loss that position, which returns then,.Calculation formula is as follows:
Wherein: N is the candidate frame number to match with ground truth box.LconfIt is defeated using Softmax Loss Enter for the confidence level c of every one kind.LlocUsing Smooth L1 Loss.α is weight term, is set as 1.
The formula of the confidence level loss is as follows:
The formula that the position returns loss is as follows:
Gained model carries out pedestrian detection, and wherein picture is sequentially completed step 1~3, obtains the confidence score of candidate frame After location information, uses non-maximum restraining operation to select final candidate frame and exported as detection.Output result case is for example schemed 6。
The present invention effectively improves the performance and real-time of pedestrian detection, can reach in the detection compared with existing algorithm The 60 frames above speed (GPU of Titan X) per second, 75% or so Average Accuracy, well very than current state-of-the-art method It is more.The design of this paper simultaneously can guarantee the precision of detection when the lower image of input resolution ratio.It is whole to set end to end Meter, so that training and detection can settle at one go.
Each step of the invention can be realized with general computing device, for example, they can concentrate on it is single On computing device, such as: personal computer, server computer, handheld device or portable device, laptop device or more Processor device can also be distributed over a network of multiple computing devices, they can be to be different from sequence herein Shown or described step is executed, perhaps they are fabricated to each integrated circuit modules or will be more in them A module or step are fabricated to single integrated circuit module to realize.Therefore, the present invention is not limited to any specific hardware and soft Part or its combination.
Programmable logic device can be used to realize in method provided by the invention, and it is soft also to may be embodied as computer program Part or program module (it include routines performing specific tasks or implementing specific abstract data types, programs, objects, component or Data structure etc.), such as embodiment according to the present invention can be a kind of computer program product, run the computer program Product executes computer for demonstrated method.The computer program product includes computer readable storage medium, should It include computer program logic or code section on medium, for realizing the method.The computer readable storage medium can To be the built-in medium being mounted in a computer or the removable medium (example that can be disassembled from basic computer Such as: using the storage equipment of hot plug technology).The built-in medium includes but is not limited to rewritable nonvolatile memory, Such as: RAM, ROM, flash memory and hard disk.The removable medium includes but is not limited to: and optical storage media (such as: CD- ROM and DVD), magnetic-optical storage medium (such as: MO), magnetic storage medium (such as: tape or mobile hard disk), can with built-in Rewrite the media (such as: storage card) of nonvolatile memory and the media (such as: ROM box) with built-in ROM.
Present invention is not limited to the embodiments described above, and without departing substantially from substantive content of the present invention, this field is common Any deformation, improvement or the replacement that technical staff is contemplated that each fall within the scope of the present invention.
Although having been shown above, being described and pointed out basic novel feature of the invention suitable for various embodiments Detailed description, it will be understood that do not depart from the invention is intended in the case where, those skilled in the art can be to system Form and details carry out various omissions, substitutions and changes.

Claims (6)

1. a kind of pedestrian detection method based on deep learning algorithm, which is characterized in that this method includes at least:
Step 1: input picture being normalized into unified size as input, utilizes the basic network portion in designed depth network Divide and the extraction of foundation characteristic figure is carried out to input picture.The additional Analysis On Multi-scale Features layer part in designed depth network is utilized later The extraction of Analysis On Multi-scale Features figure is carried out to above-mentioned gained foundation characteristic figure;
Step 2: each location point (feature map location) in Analysis On Multi-scale Features figure described in step 1 is according to different big Small scale and a series of corresponding different fixation candidate frames (default box) of generation of length-breadth ratio;
Step 3: the classification of 3*3 convolution is carried out on the Analysis On Multi-scale Features figure described in step 1, each candidate frame in each location point returns Return it includes 1 confidence score (score) of pedestrian's object and 4 positional shift values (offsets).Each candidate when training Jaccard overlap between frame and any one ground truth is greater than certain threshold value, is just used as positive sample, other It sorts according to confidence level, that chooses 3 times of positive sample quantity is used as negative sample;
Step 4: loss, and back-propagation and iteration being calculated according to trained objective function, obtain final training pattern.Gained mould Type carries out pedestrian detection, and wherein picture is sequentially completed step 1~3, after obtaining the confidence score and location information of candidate frame, makes It uses non-maximum restraining operation to select final candidate frame to export as detection.
2. the method according to claim 1, wherein the step 1 specifically includes:
Input picture is normalized into unified size, puts into designed pedestrian detection depth network.The network includes basis Network portion and additional features layer.The basic network part is built using Pool5 layers of VGG16 network and framework before, is wrapped Containing 5 convolutional layers and 5 MaxPooling operations, the extraction of foundation characteristic figure is carried out to input picture.The additional features layer packet Include 3 convolutional layers and one pool layers of average.It is related to convolution operation and is followed successively by Conv6 (3*3*1024), Conv7 (1*1* 1024), Conv8_1 (1*1*256), Conv8_2 (3*3*512-s2), Conv9_1 (1*1*128), Conv9_2 (3*3*256- S2), Conv10_1 (1*1*128), Conv10_2 (3*3*256-s2), Avg Pooling:Global, to above-mentioned gained basis The extraction of characteristic pattern progress Analysis On Multi-scale Features figure.
3. according to the method described in claim 2, it is characterized in that, gained different scale characteristic pattern specifically includes: obtained by Conv7 19*19*1024 dimensional feature figure, the resulting 10*10*512 dimensional feature figure of Conv8_2, Conv9_2 gained 5*5*256 dimensional feature figure, 3*3*256 dimensional feature figure obtained by Conv10_2,1*1*256 dimensional feature figure obtained by Avg Pooling:Global, totally 5 features Figure, these convolutional layers gained characteristic pattern size variation is bigger, for detecting the object under different scale.
4. the method according to claim 1, wherein the step 2 specifically includes:
The corresponding candidate frame for generating 6 kinds of size scales and length-breadth ratio of each location point in Analysis On Multi-scale Features figure.
The length-breadth ratio of the candidate frame is respectively that default boundary frame, transverse and longitudinal ratio are respectivelyCandidate frame.Often Candidate frame size in one characteristic pattern calculates as follows:
For above-mentioned default frame, size are as follows:
Wherein, m indicates characteristic pattern number, herein m=5.sminValue 0.2, smaxValue 0.95, it is meant that the scale of lowermost layer is 0.2, top scale is 0.95.Then, each candidate frame size are as follows:
The center of each candidate frame is arranged are as follows:Wherein | fk| it is the size of k-th of characteristic pattern, meanwhile, i, j ∈ [0, | fk|]。
5. the method according to claim 1, wherein the step 3 specifically includes:
The classification of 3*3 convolution is carried out on the Analysis On Multi-scale Features figure described in step 1, each candidate frame in each location point returns to its packet 1 confidence score of the object containing pedestrian and 4 positional shift values.(1+ is exported in 6 candidate frames on i.e. each location point altogether 4) * 6 values.For the characteristic pattern of a m*n size, i.e., it can generate (1+4) * 6*m*n output result.
In the training stage, the ground truth of input picture and each object is given.For each candidate frame with it is any one The jaccard overlap of a ground truth is greater than 0.5, is just used as positive sample, therefore allow a ground truth Corresponding multiple candidate frames.Other frames sort according to confidence level, and that chooses 3 times of positive sample quantity is used as negative sample.
6. the method according to claim 1, wherein the step 4 specifically includes:
Trained loss function (objective loss function) is by confidence level loss (confidence loss) and position It puts back into and loss (confidence loss) two parts weighted sum is returned to constitute.Confidence level is using Softmax Loss, and position is returned Gui Ze is using Smooth L1 loss.Calculation formula is as follows:
Wherein: N is the candidate frame number to match with ground truth box.LconfUsing Softmax Loss, it is every for inputting A kind of confidence level c.LlocUsing Smooth L1 Loss.α is weight term, is set as 1.
The formula of the confidence level loss is as follows:
The formula that the position returns loss is as follows:
CN201710531902.1A 2017-07-03 2017-07-03 Pedestrian detection method based on deep learning algorithm Pending CN109214241A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201710531902.1A CN109214241A (en) 2017-07-03 2017-07-03 Pedestrian detection method based on deep learning algorithm

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201710531902.1A CN109214241A (en) 2017-07-03 2017-07-03 Pedestrian detection method based on deep learning algorithm

Publications (1)

Publication Number Publication Date
CN109214241A true CN109214241A (en) 2019-01-15

Family

ID=64992283

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201710531902.1A Pending CN109214241A (en) 2017-07-03 2017-07-03 Pedestrian detection method based on deep learning algorithm

Country Status (1)

Country Link
CN (1) CN109214241A (en)

Cited By (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109871829A (en) * 2019-03-15 2019-06-11 北京行易道科技有限公司 A kind of detection model training method and device based on deep learning
CN110176944A (en) * 2019-04-25 2019-08-27 中国科学院上海微系统与信息技术研究所 A kind of intelligent means for anti-jamming and method based on deep learning
CN110348390A (en) * 2019-07-12 2019-10-18 创新奇智(重庆)科技有限公司 A kind of training method, computer-readable medium and the system of fire defector model
CN111414861A (en) * 2020-03-20 2020-07-14 赛特斯信息科技股份有限公司 Method for realizing detection processing of pedestrians and non-motor vehicles based on deep learning
CN111967326A (en) * 2020-07-16 2020-11-20 北京交通大学 Gait recognition method based on lightweight multi-scale feature extraction
CN112163499A (en) * 2020-09-23 2021-01-01 电子科技大学 Small target pedestrian detection method based on fusion features
CN112560649A (en) * 2020-12-09 2021-03-26 广州云从鼎望科技有限公司 Behavior action detection method, system, equipment and medium
CN112733671A (en) * 2020-12-31 2021-04-30 新大陆数字技术股份有限公司 Pedestrian detection method, device and readable storage medium

Cited By (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109871829A (en) * 2019-03-15 2019-06-11 北京行易道科技有限公司 A kind of detection model training method and device based on deep learning
CN110176944A (en) * 2019-04-25 2019-08-27 中国科学院上海微系统与信息技术研究所 A kind of intelligent means for anti-jamming and method based on deep learning
CN110348390A (en) * 2019-07-12 2019-10-18 创新奇智(重庆)科技有限公司 A kind of training method, computer-readable medium and the system of fire defector model
CN111414861A (en) * 2020-03-20 2020-07-14 赛特斯信息科技股份有限公司 Method for realizing detection processing of pedestrians and non-motor vehicles based on deep learning
CN111967326A (en) * 2020-07-16 2020-11-20 北京交通大学 Gait recognition method based on lightweight multi-scale feature extraction
CN112163499A (en) * 2020-09-23 2021-01-01 电子科技大学 Small target pedestrian detection method based on fusion features
CN112560649A (en) * 2020-12-09 2021-03-26 广州云从鼎望科技有限公司 Behavior action detection method, system, equipment and medium
CN112733671A (en) * 2020-12-31 2021-04-30 新大陆数字技术股份有限公司 Pedestrian detection method, device and readable storage medium

Similar Documents

Publication Publication Date Title
CN109214241A (en) Pedestrian detection method based on deep learning algorithm
CN108830188A (en) Vehicle checking method based on deep learning
Costea et al. Creating roadmaps in aerial images with generative adversarial networks and smoothing-based optimization
Wu et al. Milcut: A sweeping line multiple instance learning paradigm for interactive image segmentation
Xie et al. DeRPN: Taking a further step toward more general object detection
CN106845430A (en) Pedestrian detection and tracking based on acceleration region convolutional neural networks
Xia et al. An evaluation of deep learning in loop closure detection for visual SLAM
Yang et al. ConDinet++: Full-scale fusion network based on conditional dilated convolution to extract roads from remote sensing images
CN114841257B (en) Small sample target detection method based on self-supervision comparison constraint
Montoya-Zegarra et al. Mind the gap: modeling local and global context in (road) networks
CN105825502B (en) A kind of Weakly supervised method for analyzing image of the dictionary study based on conspicuousness guidance
CN108664838A (en) Based on the monitoring scene pedestrian detection method end to end for improving RPN depth networks
Chen et al. Guide local feature matching by overlap estimation
CN112287906B (en) Template matching tracking method and system based on depth feature fusion
CN104637045A (en) Image pixel labeling method based on super pixel level features
Zhang et al. DWin-HRFormer: A high-resolution transformer model with directional windows for semantic segmentation of urban construction land
Lu et al. Dual-stream class-adaptive network for semi-supervised hyperspectral image classification
Schmidt et al. Object detection of aerial image using mask-region convolutional neural network (mask R-CNN)
CN109558883B (en) Blade feature extraction method and device
Hu et al. Crowd R-CNN: An object detection model utilizing crowdsourced labels
Janani et al. Feature fusion for efficient object classification using deep and shallow learning
CN113538278B (en) Depth map completion method based on deformable convolution
Xu et al. Compressed YOLOv5 for oriented object detection with integrated network slimming and knowledge distillation
CN114240844A (en) Unsupervised key point positioning and target detection method in medical image
CN112733925A (en) Method and system for constructing light image classification network based on FPCC-GAN

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
WD01 Invention patent application deemed withdrawn after publication
WD01 Invention patent application deemed withdrawn after publication

Application publication date: 20190115