CN105740891A - Target detection method based on multilevel characteristic extraction and context model - Google Patents

Target detection method based on multilevel characteristic extraction and context model Download PDF

Info

Publication number
CN105740891A
CN105740891A CN201610056601.3A CN201610056601A CN105740891A CN 105740891 A CN105740891 A CN 105740891A CN 201610056601 A CN201610056601 A CN 201610056601A CN 105740891 A CN105740891 A CN 105740891A
Authority
CN
China
Prior art keywords
picture
spatial position
window
person
target detection
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201610056601.3A
Other languages
Chinese (zh)
Other versions
CN105740891B (en
Inventor
刘波
马增妍
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Shenzhen Xiaofeng Technology Co ltd
Original Assignee
Beijing University of Technology
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing University of Technology filed Critical Beijing University of Technology
Priority to CN201610056601.3A priority Critical patent/CN105740891B/en
Publication of CN105740891A publication Critical patent/CN105740891A/en
Application granted granted Critical
Publication of CN105740891B publication Critical patent/CN105740891B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/217Validation; Performance evaluation; Active pattern learning techniques
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • G06F18/2411Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches based on the proximity to a decision surface, e.g. support vector machines

Landscapes

  • Engineering & Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • Theoretical Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Artificial Intelligence (AREA)
  • Evolutionary Biology (AREA)
  • Evolutionary Computation (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Image Analysis (AREA)

Abstract

The invention discloses a target detection method based on multilevel characteristic extraction and a context model. The model constructed by the invention mainly carries out statistics on a spatial position relationship between images in a real picture so as to improve target detection accuracy. Both the images of the same category and the images of different categories have certain spatial position relationships. Firstly, one picture is selected and searched to generate a great quantity of region proposals, then, all region proposals of each picture are subjected to feature extraction by a seven-layer convolutional neural network, and finally, a support vector machine is adopted for classification. The invention provides a new method for finding an optimal object detection position, mainly solves the technical problem of providing a new context model, replaces an original non-maximum suppression method, and is used for obtaining better target detection accuracy.

Description

Target detection based on multi-level feature extraction and context model
Technical Field
The invention belongs to the field of computer machine learning, and particularly relates to an algorithm for positioning a target position in a picture by applying a machine learning algorithm to target detection in image processing.
Background
Target detection is an algorithm that locates the position of a target in an image, which combines the segmentation and recognition of the target into one. Summarizing the average precision of various algorithms trained on standard visual inspection pascalloc datasets between 2010-2012, it was not difficult to find that the algorithms progressed slowly, almost all being integrated systems or a little improvement over existing algorithms. Until 2012, AlexKrizhevsky applied convolutional neural networks (convolutional neural networks) to image classification and made a great progress, and on the basis of AlexKrizhevsky, rossgirschick proposed that the convolutional neural networks (convolutional neural networks) were applied to the paschaloc data set to extract the features of the image, and a linear support vector machine (supportvector machine) was used to classify the image, so as to determine the category to which the image belongs, i.e., to realize target detection.
The concrete content of RossGirshick is that firstly, a picture is selected and searched to generate a large number of regionproposals, then all the regionproposals of each picture are subjected to feature extraction, a 7-layer convolutional neural network is adopted, and finally, a support vector machine is used for classification.
The traditional target detection algorithm generally adopts a sliding window method to solve the problem of positioning, but because RossGirshick adopts a 7-layer convolutional neural network, the size of an image obtained after 5-layer convolution is very small, the image is not suitable for a sliding window, and the sliding window consumes long time. Therefore, RossGirshick generated a large number of regionproposals using the method of selective search. The graph is divided into a plurality of regions by a segmentation method, and pixels are combined by using a minimum spanning tree algorithm to form region responses. The method comprises the steps of firstly, segmenting a picture, representing the Image (Image) into a Graph structure (Graph), regarding each pixel in the Image as each vertex in the Graph, forming an edge by the relationship between the pixels, obtaining a weight value of the edge as a gray level difference value between the pixels, combining the pixels by using a minimum spanning tree algorithm, and forming regions. The second part combines the regions, firstly calculates the color similarity, the texture similarity, the size similarity and the matching similarity between all the regions, and then combines the 4 similarities to obtain the final similarity. And merging the most similar 2 regions according to the similarity, and recalculating the similarity of the newly generated region and other remaining regions. And repeating the process until the whole picture is aggregated into a large region, randomly scoring each region, sorting the regions according to scores, and selecting the first K subsets.
And taking a 7-layer convolutional neural network proposed by AlexKrizhevsky as a framework, and performing feature extraction on all regionproposals of each picture. The first 5 layers in the 7-layer convolutional neural network are volume base layers, and the last 2 layers are full-connection layers. Because the input of the convolutional neural network is 227 × 227 in a fixed format, and the sizes of the regionproposals obtained after the selection search are not fixed, the length and the width of each regionproposal are firstly adjusted and then input into the network. And finally, classifying the result output from the network by using a linear support vector machine so as to detect the target. In order to eliminate redundant frames and find the best position for object detection, a non-maximum suppression (non-maximum suppression) method is generally used.
Object of the Invention
The invention provides a new method for finding the optimal object detection position. The main technical problem to be solved is to provide a new context model to replace the original non-maximum suppression (non-maximum suppression) method, so as to obtain better target detection accuracy.
The model constructed by the invention mainly counts the spatial position relation between the images in the real picture, thereby improving the accuracy of target detection. Whether the images are of the same or different classes, there will be some specific spatial relationship. For example, the spatial position relationship between two images, such as a person and a bicycle, may only be that the person is on the bicycle (above) or the person is beside the bicycle (next-to), and the spatial position relationship that the bicycle is on the person (above) rarely occurs. The invention has the flow as shown in figure 1, and the main steps are as follows:
s1 construction context model
Firstly, the methodA context model is constructed for capturing the relationships between the target detectors. A picture is explicitly represented by a series of windows with coincidence, the position of the ith window is represented by its center and its length and width, written as IiWhere N denotes a picture with N windows, xiRepresenting the picture features extracted from the ith window, X ═ X for the entire picturei: i is 1, … … N, K represents the number of image classes (the method uses a pashellvoc data set, so K is 20), yi∈ {0, … …, K } represents the label of the ith window, 0 represents the background, and Y ═ YiI 1, … … N }. Defining the fraction between X and Y as:
S ( X , Y ) = Σ i , j w y i , y j T d i , j + Σ i w y i T x i - - - ( 1 )
wherein wyi,yjDenotes yiClass and yjWeight between classes, wyiLocal templates representing classes i, dijThe spatial position relationship between the window i and the window j is shown, and the position relationship is divided into an upper surface (above), a lower surface (below), an overlapping (overlapping) and a side surfaceEdge (next-to), near (near) and far (far), thus, dijIs a sparse one-dimensional vector, and only the corresponding items satisfying the spatial position relationship among each other will be assigned as 1. The spatial position relationship between people is always next-to, not above, so above and other positions have their corresponding values 0 and next-to 1.
Since the calculation of maxS (X, Y) is a non-deterministic polynomial NP (non-deterministic polynomial) hard, the method employs greedy algorithm-like to solve the NPhard problem. The algorithm comprises the following steps:
(1) initializing a vector Y of each window into a background class;
(2) the greedy selection is not a single window of the background class, and the value of S (X, Y) is increased to the maximum extent;
(3) when any one window is selected, the value of S (X, Y) is not increased but is decreased, and the operation is stopped;
i represents a series of instantiated window-classes I ═ pairs (I, c), let y (I) represent the associated label vectors, y when all pairs are in set IiC, otherwise yi0; the value of S (X, Y) is changed by adding a window-class pair (I, c) to set I, as shown below:
(i,c)=S(X,Y(I{(i,C)}))-S(X,Y(I))
initializing I { }, S { }, 0, (I, c) ═ wT cxiIteration of
1)(i*,c*)=argmax(i,j)I(i,c)
2)I=I{(i*,c*)}
3)S=S+(i*,c*)
4)(i,c)=(i,c)+wT c*,cdi*,i+wT c,c*di,i*
The end condition is (i, c) 0 or all windows are instantiated.
S2 convex training with tangent plane optimization
To describe the learning algorithm of the present invention, equation (1) needs to be written as follows:
equivalent to S (X, Y) ═ WTΨ(X,Y)
Wherein,
namely: Y * = arg m a x Y W T Ψ ( X , Y )
the purpose of convex training is to assume that a series of training pictures X is giveniAnd label YiIt is desirable to obtain an optimal model of W, so that a new picture X is giveniA tag vector Y ═ Y can be generatedi. Thus, the result of the convex training is to get an optimal model of W, such that Y and YiThe difference value of the W-shaped optimal model is as small as possible, the process of obtaining the W-shaped optimal model through convex training is a process of solving the extreme value of the following mathematical formula,
arg m i n w , ξ i ≥ 0 w T w + C Σ i ξ i - - - ( 3 )
s.t.,HiwTΔΨ(Xi,Yi,Hi)≥l(Yi,Hi)-ξi
wherein, Δ Ψ (X)i,Yi,Hi)=Ψ(Xi,Yi)-Ψ(Yi,Hi),
l ( Y i , H i ) = Σ i = 1 N l ( y i , h i )
HiIs a label that is calculated by itself,
for ease of optimization, the constraint problem of equation (3) is equivalent to the unconstrained problem of equation (4):
w * = arg m i n w L ( w ) - - - ( 4 )
wherein L ( w ) = 1 2 || w || 2 + C R ( w )
R ( w ) = Σ i = 1 N m a x H ( 0 , l ( Y i , H ) - w T Δ Ψ ( X i , Y i , H ) )
And (4) performing tangent plane optimization on the formula (4) to obtain an optimal model of W.
S3 Overall implementation flow
Firstly, obtaining the regionproposals of the picture through a search algorithm, secondly, adopting a trained 7-layer convolutional neural network CNNS of RossGirshick to extract the characteristics of the picture, changing the final output of the 7-layer CNNS network structure of RossGirshick into 21(20 VOC classes and 1 background class) because the training set and the test set of the invention both adopt PASCALLVOC data sets (20 classes), and finally, carrying out picture classification by using a linear Support Vector Machine (SVM), and in the classification process, in order to find the position of the best object detection, using a trained context model to carry out more accurate positioning, thereby finally obtaining better target detection accuracy.
Drawings
FIG. 1 is a flow chart of the method of the present invention.
Fig. 2 is a spatial positional relationship explanatory diagram.
Detailed Description
As shown in fig. 1-2, experiments were performed according to the above method. Experiments compare results of whether the contextual model target detection is adopted, and the experiments adopt PASCALLVOC data sets with 20 classes. Since the method only stores 7 spatial position relations (same type or non-same type) between categories, if one of the 7 spatial position relations is satisfied, the target detection accuracy of the corresponding category should be higher than that of a method which is not trained by a context model, otherwise, if the spatial position relations are not determined, the context model which is possibly learned plays a role of reverse, and due to non-maximum suppression (non-maximum suppression), the detection result may be disturbed, so that the corresponding target detection accuracy is reduced.
Table 1: comparison of Experimental results
Table 1: experimental results comparison Class No context model (unit%) Context model (unit%)
aero 66.9 70.7
bike 23.7 21.2
bird 58.3 53.7
boat 37.4 39.8
bottle 55.4 50.1
bus 73.3 35.8
car 58.7 34.8
cat 56.5 59.5
chair 9.7 9.6
cow 45.5 53
table 29.5 15.9
dog 49.3 43.6
horse 40.1 34
mbike 57.8 52.8
person 53.9 57.4
plant 33.8 13.3
sheep 60.7 36.9
soft 22.7 23.2
train 47.1 55.9
tv 41.3 41.9

Claims (2)

1. The target detection method based on multi-level feature extraction and context model is characterized in that:
the model constructed by the method mainly counts the spatial position relation between images in a real picture, so that the accuracy of target detection can be improved; whether the images are of the same type or different types, the images have certain spatial position relations; the two images of the person and the bicycle have the spatial position relationship that the person is on the bicycle (above) or the person is beside the bicycle (next-to), and the spatial position relationship that the bicycle is on the person (above) rarely occurs; the spatial position relationship of a person and a person is generally the spatial position relationship of the person beside the person (next-to) and few persons on the person (above); the main steps of the method are as follows,
s1 construction context model
Firstly, constructing a context model for capturing the relation between target detectors; a picture is explicitly represented by a series of windows with coincidence, the position of the ith window is represented by its center and its length and width, written as IiWhere N denotes a picture with N windows, xiRepresenting the picture features extracted from the ith window, X ═ X for the entire picturei: i is 1, … … N, K represents the number of image classes (the method uses a pashellvoc data set, so K is 20), yi∈ {0, … …, K } represents the label of the ith window, 0 represents the background, and Y ═ YiI is 1, … … N }; defining the fraction between X and Y as:
S ( X , Y ) = Σ i , j w y i , y j T d i , j + Σ i w y i T x i - - - ( 1 )
wherein wyi,yjDenotes yiClass and yjWeight between classes, wyiLocal templates representing classes i, dijThe spatial position relationship between window i and window j is shown, and the position relationship is divided into, upper (above), lower (below), overlapping (overlapping), next-to, near (near), and far (far), so dijIs a sparse one-dimensional vector, and only the corresponding items meeting the spatial position relationship among each other can be assigned as 1; the spatial position relation between people is always next-to, but not above, so above and other corresponding items in positions are assigned as 0, and next-to is assigned as 1;
since the calculation maxS (X, Y) is a non-deterministic polynomial NP (non-deterministic polynomial) hard, the method employs greedy algorithm-like to solve the NPhard problem;
i represents a series of instantiated window-classes I ═ pairs (I, c), let y (I) represent the associated label vectors, y when all pairs are in set IiC, otherwise yi0; the value of S (X, Y) is changed by adding a window-class pair (I, c) to set I, as shown below:
(i,c)=S(X,Y(I{(i,C)}))-S(X,Y(I))
initializing I { }, S { }, 0, (I, c) ═ wT cxiIteration of
1)(i*,c*)=argmax(i,j)I(i,c)
2)I=I{(i*,c*)}
3)S=S+(i*,c*)
4) ( i , c ) = ( i , c ) + w T c * , c d i * , i + w T c , c * d i , i *
The end condition is (i, c) 0 or all windows are instantiated;
s2 convex training with tangent plane optimization
To describe the learning algorithm of the present method, equation (1) needs to be written as follows:
equivalent to S (X, Y) ═ WTΨ(X,Y)
Wherein, W = w s w a
namely: Y * = arg m a x Y W T Ψ ( X , Y )
the purpose of convex training is to assume that a series of training pictures X is giveniAnd label YiIt is desirable to obtain an optimal model of W, so that a new picture X is giveniA tag vector Y ═ Y can be generatedi(ii) a Thus, the result of the convex training is to get an optimal model of W, such that Y and YiThe difference value of the W-shaped optimal model is as small as possible, the process of obtaining the W-shaped optimal model through convex training is a process of solving the extreme value of the following mathematical formula,
arg m i n w , ξ i ≥ 0 w T w + C Σ i ξ i - - - ( 3 )
s . t . ∀ i , H i w T Δ Ψ ( X i , Y i , H i ) ≥ l ( Y i , H i ) - ξ i
wherein, Δ Ψ (X)i,Yi,Hi)=Ψ(Xi,Yi)-Ψ(Yi,Hi),
l ( Y i , H i ) = Σ i = 1 N l ( y i , h i )
HiIs a label that is calculated by itself,
l ( y i , h i ) = 1 : y i ≠ b g ^ h i ≠ y i 1 : h i ≠ b g ^ ~ ∃ j 0 : o t h e r w i s e
for ease of optimization, the constraint problem of equation (3) is equivalent to the unconstrained problem of equation (4):
w * = arg m i n w L ( w ) - - - ( 4 )
wherein L ( w ) = 1 2 | | w | | 2 + C R ( w )
R ( w ) = Σ i = 1 N m a x H ( 0 , l ( Y i , H ) - w T Δ Ψ ( X i , Y i , H ) )
Performing tangent plane optimization on the formula (4) to obtain an optimal model of W;
s3 Overall implementation flow
Firstly, obtaining the regionproposals of the picture through a search algorithm, secondly, adopting a trained 7-layer convolutional neural network CNNS of RossGirshick to extract the characteristics of the picture, changing the final output of the 7-layer CNNS network structure of RossGirshick into 21(20 VOC classes and 1 background class) because the training set and the test set of the method both adopt PASCALLVOC data sets (20 classes), and finally, carrying out picture classification by using a linear support vector machine SVM, and in the classification process, in order to find the position of the best object detection, using a trained context model to carry out more accurate positioning so as to finally obtain better target detection accuracy.
2. The method for detecting a target based on multi-level feature extraction and context model as claimed in claim 1, wherein:
the steps of the algorithm are as follows,
(1) initializing a vector Y of each window into a background class;
(2) the greedy selection is not a single window of the background class, and the value of S (X, Y) is increased to the maximum extent;
(3) when any one of the windows is selected, the value of S (X, Y) does not increase but decreases, and stops.
CN201610056601.3A 2016-01-27 2016-01-27 Target detection based on multi level feature selection and context model Active CN105740891B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201610056601.3A CN105740891B (en) 2016-01-27 2016-01-27 Target detection based on multi level feature selection and context model

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201610056601.3A CN105740891B (en) 2016-01-27 2016-01-27 Target detection based on multi level feature selection and context model

Publications (2)

Publication Number Publication Date
CN105740891A true CN105740891A (en) 2016-07-06
CN105740891B CN105740891B (en) 2019-10-08

Family

ID=56247276

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201610056601.3A Active CN105740891B (en) 2016-01-27 2016-01-27 Target detection based on multi level feature selection and context model

Country Status (1)

Country Link
CN (1) CN105740891B (en)

Cited By (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106372597A (en) * 2016-08-31 2017-02-01 李涛 CNN traffic detection method based on adaptive context information
CN106408618A (en) * 2016-08-31 2017-02-15 上海交通大学 Image deconstruction method based on machine learning
CN106446933A (en) * 2016-08-31 2017-02-22 河南广播电视大学 Multi-target detection method based on context information
CN106991296A (en) * 2017-04-01 2017-07-28 大连理工大学 Ensemble classifier method based on the greedy feature selecting of randomization
CN107239827A (en) * 2017-06-18 2017-10-10 北京理工大学 A kind of spatial information learning method based on artificial neural network
CN108229519A (en) * 2017-02-17 2018-06-29 北京市商汤科技开发有限公司 The method, apparatus and system of image classification
CN108830903A (en) * 2018-04-28 2018-11-16 杨晓春 A kind of steel billet method for detecting position based on CNN
CN108846047A (en) * 2018-05-30 2018-11-20 百卓网络科技有限公司 A kind of picture retrieval method and system based on convolution feature
CN110298402A (en) * 2019-07-01 2019-10-01 国网内蒙古东部电力有限公司 A kind of small target deteection performance optimization method
CN111553228A (en) * 2020-04-21 2020-08-18 佳都新太科技股份有限公司 Method, device, equipment and storage medium for detecting personal bag relationship
US10963676B2 (en) 2016-12-23 2021-03-30 Samsung Electronics Co., Ltd. Image processing method and apparatus

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20070297675A1 (en) * 2006-06-26 2007-12-27 Shih-Jong J. Lee Method of directed feature development for image pattern recognition
CN102110227A (en) * 2010-11-24 2011-06-29 清华大学 Compound method for classifying multiresolution remote sensing images based on context
CN102495865A (en) * 2011-11-28 2012-06-13 南京大学 Image annotation method combined with image internal space relation and visual symbiosis relation
CN103514456A (en) * 2013-06-30 2014-01-15 安科智慧城市技术(中国)有限公司 Image classification method and device based on compressed sensing multi-core learning
CN104778466A (en) * 2015-04-16 2015-07-15 北京航空航天大学 Detection method combining various context clues for image focus region

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20070297675A1 (en) * 2006-06-26 2007-12-27 Shih-Jong J. Lee Method of directed feature development for image pattern recognition
CN102110227A (en) * 2010-11-24 2011-06-29 清华大学 Compound method for classifying multiresolution remote sensing images based on context
CN102495865A (en) * 2011-11-28 2012-06-13 南京大学 Image annotation method combined with image internal space relation and visual symbiosis relation
CN103514456A (en) * 2013-06-30 2014-01-15 安科智慧城市技术(中国)有限公司 Image classification method and device based on compressed sensing multi-core learning
CN104778466A (en) * 2015-04-16 2015-07-15 北京航空航天大学 Detection method combining various context clues for image focus region

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
刘扬 等: "高分辨率遥感影像目标分类与识别研究进展", 《地球信息科学学报》 *

Cited By (17)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106446933B (en) * 2016-08-31 2019-08-02 河南广播电视大学 Multi-target detection method based on contextual information
CN106408618A (en) * 2016-08-31 2017-02-15 上海交通大学 Image deconstruction method based on machine learning
CN106446933A (en) * 2016-08-31 2017-02-22 河南广播电视大学 Multi-target detection method based on context information
CN106372597A (en) * 2016-08-31 2017-02-01 李涛 CNN traffic detection method based on adaptive context information
CN106372597B (en) * 2016-08-31 2019-09-13 郑州禅图智能科技有限公司 CNN Vehicle Detection method based on adaptive contextual information
CN106408618B (en) * 2016-08-31 2019-05-07 上海交通大学 A kind of image destructing method based on machine learning
US10963676B2 (en) 2016-12-23 2021-03-30 Samsung Electronics Co., Ltd. Image processing method and apparatus
CN108229519B (en) * 2017-02-17 2020-09-04 北京市商汤科技开发有限公司 Image classification method, device and system
CN108229519A (en) * 2017-02-17 2018-06-29 北京市商汤科技开发有限公司 The method, apparatus and system of image classification
CN106991296B (en) * 2017-04-01 2019-12-27 大连理工大学 Integrated classification method based on randomized greedy feature selection
CN106991296A (en) * 2017-04-01 2017-07-28 大连理工大学 Ensemble classifier method based on the greedy feature selecting of randomization
CN107239827B (en) * 2017-06-18 2020-06-09 北京理工大学 Spatial information learning method based on artificial neural network
CN107239827A (en) * 2017-06-18 2017-10-10 北京理工大学 A kind of spatial information learning method based on artificial neural network
CN108830903A (en) * 2018-04-28 2018-11-16 杨晓春 A kind of steel billet method for detecting position based on CNN
CN108846047A (en) * 2018-05-30 2018-11-20 百卓网络科技有限公司 A kind of picture retrieval method and system based on convolution feature
CN110298402A (en) * 2019-07-01 2019-10-01 国网内蒙古东部电力有限公司 A kind of small target deteection performance optimization method
CN111553228A (en) * 2020-04-21 2020-08-18 佳都新太科技股份有限公司 Method, device, equipment and storage medium for detecting personal bag relationship

Also Published As

Publication number Publication date
CN105740891B (en) 2019-10-08

Similar Documents

Publication Publication Date Title
CN105740891B (en) Target detection based on multi level feature selection and context model
US11429818B2 (en) Method, system and device for multi-label object detection based on an object detection network
CN108427920B (en) Edge-sea defense target detection method based on deep learning
US11574152B2 (en) Recognition system for security check and control method thereof
CN105740909B (en) Text recognition method under a kind of natural scene based on spatial alternation
CN110334765B (en) Remote sensing image classification method based on attention mechanism multi-scale deep learning
US10867167B2 (en) Collaborative deep network model method for pedestrian detection
CN111160440B (en) Deep learning-based safety helmet wearing detection method and device
Thai et al. Image classification using support vector machine and artificial neural network
CN108388896A (en) A kind of licence plate recognition method based on dynamic time sequence convolutional neural networks
CN107423760A (en) Based on pre-segmentation and the deep learning object detection method returned
CN104992167A (en) Convolution neural network based face detection method and apparatus
CN105740892A (en) High-accuracy human body multi-position identification method based on convolutional neural network
CN106022363B (en) A kind of Chinese text recognition methods suitable under natural scene
CN107767416B (en) Method for identifying pedestrian orientation in low-resolution image
CN104599275A (en) Understanding method of non-parametric RGB-D scene based on probabilistic graphical model
CN105184298A (en) Image classification method through fast and locality-constrained low-rank coding process
CN107784288A (en) A kind of iteration positioning formula method for detecting human face based on deep neural network
CN110543906B (en) Automatic skin recognition method based on Mask R-CNN model
CN103186776B (en) Based on the human body detecting method of multiple features and depth information
CN112861970B (en) Fine-grained image classification method based on feature fusion
CN109858327B (en) Character segmentation method based on deep learning
CN113221956B (en) Target identification method and device based on improved multi-scale depth model
CN114360067A (en) Dynamic gesture recognition method based on deep learning
CN116796248A (en) Forest health environment assessment system and method thereof

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant
TR01 Transfer of patent right
TR01 Transfer of patent right

Effective date of registration: 20211122

Address after: 518052 Room 201, building A, 1 front Bay Road, Shenzhen Qianhai cooperation zone, Shenzhen, Guangdong

Patentee after: Shenzhen Xiaofeng Technology Co.,Ltd.

Address before: 100124 No. 100 Chaoyang District Ping Tian Park, Beijing

Patentee before: Beijing University of Technology