CN105740891A - Target detection method based on multilevel characteristic extraction and context model - Google Patents
Target detection method based on multilevel characteristic extraction and context model Download PDFInfo
- Publication number
- CN105740891A CN105740891A CN201610056601.3A CN201610056601A CN105740891A CN 105740891 A CN105740891 A CN 105740891A CN 201610056601 A CN201610056601 A CN 201610056601A CN 105740891 A CN105740891 A CN 105740891A
- Authority
- CN
- China
- Prior art keywords
- picture
- spatial position
- window
- person
- target detection
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 238000001514 detection method Methods 0.000 title claims abstract description 24
- 238000000605 extraction Methods 0.000 title claims abstract description 8
- 238000000034 method Methods 0.000 claims abstract description 29
- 238000013527 convolutional neural network Methods 0.000 claims abstract description 12
- 238000012706 support-vector machine Methods 0.000 claims abstract description 8
- 238000012549 training Methods 0.000 claims description 12
- 239000013598 vector Substances 0.000 claims description 8
- 238000005457 optimization Methods 0.000 claims description 6
- 238000004364 calculation method Methods 0.000 claims description 2
- 238000010276 construction Methods 0.000 claims description 2
- 238000010845 search algorithm Methods 0.000 claims description 2
- 238000012360 testing method Methods 0.000 claims description 2
- 230000007423 decrease Effects 0.000 claims 1
- 230000001629 suppression Effects 0.000 abstract description 7
- 238000002474 experimental method Methods 0.000 description 3
- 238000010801 machine learning Methods 0.000 description 2
- 230000011218 segmentation Effects 0.000 description 2
- LPVJWXJBZPCDLM-HUWNQDJBSA-N 7-boat Chemical compound O([C@H]1\C(C)=C/[C@]23O[C@@]2(C([C@H](C)[C@@H](OC(C)=O)[C@H]2[C@H](C2(C)C)[C@@H]1OC(C)=O)=O)C[C@H]([C@H]3OC(C)=O)C)C(=O)C1=CC=CC=C1 LPVJWXJBZPCDLM-HUWNQDJBSA-N 0.000 description 1
- 241000282326 Felis catus Species 0.000 description 1
- 241001494479 Pecora Species 0.000 description 1
- 230000003247 decreasing effect Effects 0.000 description 1
- 238000010586 diagram Methods 0.000 description 1
- 238000012545 processing Methods 0.000 description 1
- 230000004044 response Effects 0.000 description 1
- 238000011179 visual inspection Methods 0.000 description 1
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/21—Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
- G06F18/217—Validation; Performance evaluation; Active pattern learning techniques
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/24—Classification techniques
- G06F18/241—Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
- G06F18/2411—Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches based on the proximity to a decision surface, e.g. support vector machines
Landscapes
- Engineering & Computer Science (AREA)
- Data Mining & Analysis (AREA)
- Theoretical Computer Science (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Bioinformatics & Computational Biology (AREA)
- Artificial Intelligence (AREA)
- Evolutionary Biology (AREA)
- Evolutionary Computation (AREA)
- Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Life Sciences & Earth Sciences (AREA)
- Image Analysis (AREA)
Abstract
The invention discloses a target detection method based on multilevel characteristic extraction and a context model. The model constructed by the invention mainly carries out statistics on a spatial position relationship between images in a real picture so as to improve target detection accuracy. Both the images of the same category and the images of different categories have certain spatial position relationships. Firstly, one picture is selected and searched to generate a great quantity of region proposals, then, all region proposals of each picture are subjected to feature extraction by a seven-layer convolutional neural network, and finally, a support vector machine is adopted for classification. The invention provides a new method for finding an optimal object detection position, mainly solves the technical problem of providing a new context model, replaces an original non-maximum suppression method, and is used for obtaining better target detection accuracy.
Description
Technical Field
The invention belongs to the field of computer machine learning, and particularly relates to an algorithm for positioning a target position in a picture by applying a machine learning algorithm to target detection in image processing.
Background
Target detection is an algorithm that locates the position of a target in an image, which combines the segmentation and recognition of the target into one. Summarizing the average precision of various algorithms trained on standard visual inspection pascalloc datasets between 2010-2012, it was not difficult to find that the algorithms progressed slowly, almost all being integrated systems or a little improvement over existing algorithms. Until 2012, AlexKrizhevsky applied convolutional neural networks (convolutional neural networks) to image classification and made a great progress, and on the basis of AlexKrizhevsky, rossgirschick proposed that the convolutional neural networks (convolutional neural networks) were applied to the paschaloc data set to extract the features of the image, and a linear support vector machine (supportvector machine) was used to classify the image, so as to determine the category to which the image belongs, i.e., to realize target detection.
The concrete content of RossGirshick is that firstly, a picture is selected and searched to generate a large number of regionproposals, then all the regionproposals of each picture are subjected to feature extraction, a 7-layer convolutional neural network is adopted, and finally, a support vector machine is used for classification.
The traditional target detection algorithm generally adopts a sliding window method to solve the problem of positioning, but because RossGirshick adopts a 7-layer convolutional neural network, the size of an image obtained after 5-layer convolution is very small, the image is not suitable for a sliding window, and the sliding window consumes long time. Therefore, RossGirshick generated a large number of regionproposals using the method of selective search. The graph is divided into a plurality of regions by a segmentation method, and pixels are combined by using a minimum spanning tree algorithm to form region responses. The method comprises the steps of firstly, segmenting a picture, representing the Image (Image) into a Graph structure (Graph), regarding each pixel in the Image as each vertex in the Graph, forming an edge by the relationship between the pixels, obtaining a weight value of the edge as a gray level difference value between the pixels, combining the pixels by using a minimum spanning tree algorithm, and forming regions. The second part combines the regions, firstly calculates the color similarity, the texture similarity, the size similarity and the matching similarity between all the regions, and then combines the 4 similarities to obtain the final similarity. And merging the most similar 2 regions according to the similarity, and recalculating the similarity of the newly generated region and other remaining regions. And repeating the process until the whole picture is aggregated into a large region, randomly scoring each region, sorting the regions according to scores, and selecting the first K subsets.
And taking a 7-layer convolutional neural network proposed by AlexKrizhevsky as a framework, and performing feature extraction on all regionproposals of each picture. The first 5 layers in the 7-layer convolutional neural network are volume base layers, and the last 2 layers are full-connection layers. Because the input of the convolutional neural network is 227 × 227 in a fixed format, and the sizes of the regionproposals obtained after the selection search are not fixed, the length and the width of each regionproposal are firstly adjusted and then input into the network. And finally, classifying the result output from the network by using a linear support vector machine so as to detect the target. In order to eliminate redundant frames and find the best position for object detection, a non-maximum suppression (non-maximum suppression) method is generally used.
Object of the Invention
The invention provides a new method for finding the optimal object detection position. The main technical problem to be solved is to provide a new context model to replace the original non-maximum suppression (non-maximum suppression) method, so as to obtain better target detection accuracy.
The model constructed by the invention mainly counts the spatial position relation between the images in the real picture, thereby improving the accuracy of target detection. Whether the images are of the same or different classes, there will be some specific spatial relationship. For example, the spatial position relationship between two images, such as a person and a bicycle, may only be that the person is on the bicycle (above) or the person is beside the bicycle (next-to), and the spatial position relationship that the bicycle is on the person (above) rarely occurs. The invention has the flow as shown in figure 1, and the main steps are as follows:
s1 construction context model
Firstly, the methodA context model is constructed for capturing the relationships between the target detectors. A picture is explicitly represented by a series of windows with coincidence, the position of the ith window is represented by its center and its length and width, written as IiWhere N denotes a picture with N windows, xiRepresenting the picture features extracted from the ith window, X ═ X for the entire picturei: i is 1, … … N, K represents the number of image classes (the method uses a pashellvoc data set, so K is 20), yi∈ {0, … …, K } represents the label of the ith window, 0 represents the background, and Y ═ YiI 1, … … N }. Defining the fraction between X and Y as:
wherein wyi,yjDenotes yiClass and yjWeight between classes, wyiLocal templates representing classes i, dijThe spatial position relationship between the window i and the window j is shown, and the position relationship is divided into an upper surface (above), a lower surface (below), an overlapping (overlapping) and a side surfaceEdge (next-to), near (near) and far (far), thus, dijIs a sparse one-dimensional vector, and only the corresponding items satisfying the spatial position relationship among each other will be assigned as 1. The spatial position relationship between people is always next-to, not above, so above and other positions have their corresponding values 0 and next-to 1.
Since the calculation of maxS (X, Y) is a non-deterministic polynomial NP (non-deterministic polynomial) hard, the method employs greedy algorithm-like to solve the NPhard problem. The algorithm comprises the following steps:
(1) initializing a vector Y of each window into a background class;
(2) the greedy selection is not a single window of the background class, and the value of S (X, Y) is increased to the maximum extent;
(3) when any one window is selected, the value of S (X, Y) is not increased but is decreased, and the operation is stopped;
i represents a series of instantiated window-classes I ═ pairs (I, c), let y (I) represent the associated label vectors, y when all pairs are in set IiC, otherwise yi0; the value of S (X, Y) is changed by adding a window-class pair (I, c) to set I, as shown below:
(i,c)=S(X,Y(I{(i,C)}))-S(X,Y(I))
initializing I { }, S { }, 0, (I, c) ═ wT cxiIteration of
1)(i*,c*)=argmax(i,j)I(i,c)
2)I=I{(i*,c*)}
3)S=S+(i*,c*)
4)(i,c)=(i,c)+wT c*,cdi*,i+wT c,c*di,i*
The end condition is (i, c) 0 or all windows are instantiated.
S2 convex training with tangent plane optimization
To describe the learning algorithm of the present invention, equation (1) needs to be written as follows:
equivalent to S (X, Y) ═ WTΨ(X,Y)
Wherein,
namely:
the purpose of convex training is to assume that a series of training pictures X is giveniAnd label YiIt is desirable to obtain an optimal model of W, so that a new picture X is giveniA tag vector Y ═ Y can be generatedi. Thus, the result of the convex training is to get an optimal model of W, such that Y and YiThe difference value of the W-shaped optimal model is as small as possible, the process of obtaining the W-shaped optimal model through convex training is a process of solving the extreme value of the following mathematical formula,
s.t.,HiwTΔΨ(Xi,Yi,Hi)≥l(Yi,Hi)-ξi
wherein, Δ Ψ (X)i,Yi,Hi)=Ψ(Xi,Yi)-Ψ(Yi,Hi),
HiIs a label that is calculated by itself,
for ease of optimization, the constraint problem of equation (3) is equivalent to the unconstrained problem of equation (4):
wherein
And (4) performing tangent plane optimization on the formula (4) to obtain an optimal model of W.
S3 Overall implementation flow
Firstly, obtaining the regionproposals of the picture through a search algorithm, secondly, adopting a trained 7-layer convolutional neural network CNNS of RossGirshick to extract the characteristics of the picture, changing the final output of the 7-layer CNNS network structure of RossGirshick into 21(20 VOC classes and 1 background class) because the training set and the test set of the invention both adopt PASCALLVOC data sets (20 classes), and finally, carrying out picture classification by using a linear Support Vector Machine (SVM), and in the classification process, in order to find the position of the best object detection, using a trained context model to carry out more accurate positioning, thereby finally obtaining better target detection accuracy.
Drawings
FIG. 1 is a flow chart of the method of the present invention.
Fig. 2 is a spatial positional relationship explanatory diagram.
Detailed Description
As shown in fig. 1-2, experiments were performed according to the above method. Experiments compare results of whether the contextual model target detection is adopted, and the experiments adopt PASCALLVOC data sets with 20 classes. Since the method only stores 7 spatial position relations (same type or non-same type) between categories, if one of the 7 spatial position relations is satisfied, the target detection accuracy of the corresponding category should be higher than that of a method which is not trained by a context model, otherwise, if the spatial position relations are not determined, the context model which is possibly learned plays a role of reverse, and due to non-maximum suppression (non-maximum suppression), the detection result may be disturbed, so that the corresponding target detection accuracy is reduced.
Table 1: comparison of Experimental results
Table 1: experimental results comparison Class | No context model (unit%) | Context model (unit%) |
aero | 66.9 | 70.7 |
bike | 23.7 | 21.2 |
bird | 58.3 | 53.7 |
boat | 37.4 | 39.8 |
bottle | 55.4 | 50.1 |
bus | 73.3 | 35.8 |
car | 58.7 | 34.8 |
cat | 56.5 | 59.5 |
chair | 9.7 | 9.6 |
cow | 45.5 | 53 |
table | 29.5 | 15.9 |
dog | 49.3 | 43.6 |
horse | 40.1 | 34 |
mbike | 57.8 | 52.8 |
person | 53.9 | 57.4 |
plant | 33.8 | 13.3 |
sheep | 60.7 | 36.9 |
soft | 22.7 | 23.2 |
train | 47.1 | 55.9 |
tv | 41.3 | 41.9 |
Claims (2)
1. The target detection method based on multi-level feature extraction and context model is characterized in that:
the model constructed by the method mainly counts the spatial position relation between images in a real picture, so that the accuracy of target detection can be improved; whether the images are of the same type or different types, the images have certain spatial position relations; the two images of the person and the bicycle have the spatial position relationship that the person is on the bicycle (above) or the person is beside the bicycle (next-to), and the spatial position relationship that the bicycle is on the person (above) rarely occurs; the spatial position relationship of a person and a person is generally the spatial position relationship of the person beside the person (next-to) and few persons on the person (above); the main steps of the method are as follows,
s1 construction context model
Firstly, constructing a context model for capturing the relation between target detectors; a picture is explicitly represented by a series of windows with coincidence, the position of the ith window is represented by its center and its length and width, written as IiWhere N denotes a picture with N windows, xiRepresenting the picture features extracted from the ith window, X ═ X for the entire picturei: i is 1, … … N, K represents the number of image classes (the method uses a pashellvoc data set, so K is 20), yi∈ {0, … …, K } represents the label of the ith window, 0 represents the background, and Y ═ YiI is 1, … … N }; defining the fraction between X and Y as:
wherein wyi,yjDenotes yiClass and yjWeight between classes, wyiLocal templates representing classes i, dijThe spatial position relationship between window i and window j is shown, and the position relationship is divided into, upper (above), lower (below), overlapping (overlapping), next-to, near (near), and far (far), so dijIs a sparse one-dimensional vector, and only the corresponding items meeting the spatial position relationship among each other can be assigned as 1; the spatial position relation between people is always next-to, but not above, so above and other corresponding items in positions are assigned as 0, and next-to is assigned as 1;
since the calculation maxS (X, Y) is a non-deterministic polynomial NP (non-deterministic polynomial) hard, the method employs greedy algorithm-like to solve the NPhard problem;
i represents a series of instantiated window-classes I ═ pairs (I, c), let y (I) represent the associated label vectors, y when all pairs are in set IiC, otherwise yi0; the value of S (X, Y) is changed by adding a window-class pair (I, c) to set I, as shown below:
(i,c)=S(X,Y(I{(i,C)}))-S(X,Y(I))
initializing I { }, S { }, 0, (I, c) ═ wT cxiIteration of
1)(i*,c*)=argmax(i,j)I(i,c)
2)I=I{(i*,c*)}
3)S=S+(i*,c*)
4)
The end condition is (i, c) 0 or all windows are instantiated;
s2 convex training with tangent plane optimization
To describe the learning algorithm of the present method, equation (1) needs to be written as follows:
equivalent to S (X, Y) ═ WTΨ(X,Y)
Wherein,
namely:
the purpose of convex training is to assume that a series of training pictures X is giveniAnd label YiIt is desirable to obtain an optimal model of W, so that a new picture X is giveniA tag vector Y ═ Y can be generatedi(ii) a Thus, the result of the convex training is to get an optimal model of W, such that Y and YiThe difference value of the W-shaped optimal model is as small as possible, the process of obtaining the W-shaped optimal model through convex training is a process of solving the extreme value of the following mathematical formula,
wherein, Δ Ψ (X)i,Yi,Hi)=Ψ(Xi,Yi)-Ψ(Yi,Hi),
HiIs a label that is calculated by itself,
for ease of optimization, the constraint problem of equation (3) is equivalent to the unconstrained problem of equation (4):
wherein
Performing tangent plane optimization on the formula (4) to obtain an optimal model of W;
s3 Overall implementation flow
Firstly, obtaining the regionproposals of the picture through a search algorithm, secondly, adopting a trained 7-layer convolutional neural network CNNS of RossGirshick to extract the characteristics of the picture, changing the final output of the 7-layer CNNS network structure of RossGirshick into 21(20 VOC classes and 1 background class) because the training set and the test set of the method both adopt PASCALLVOC data sets (20 classes), and finally, carrying out picture classification by using a linear support vector machine SVM, and in the classification process, in order to find the position of the best object detection, using a trained context model to carry out more accurate positioning so as to finally obtain better target detection accuracy.
2. The method for detecting a target based on multi-level feature extraction and context model as claimed in claim 1, wherein:
the steps of the algorithm are as follows,
(1) initializing a vector Y of each window into a background class;
(2) the greedy selection is not a single window of the background class, and the value of S (X, Y) is increased to the maximum extent;
(3) when any one of the windows is selected, the value of S (X, Y) does not increase but decreases, and stops.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201610056601.3A CN105740891B (en) | 2016-01-27 | 2016-01-27 | Target detection based on multi level feature selection and context model |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201610056601.3A CN105740891B (en) | 2016-01-27 | 2016-01-27 | Target detection based on multi level feature selection and context model |
Publications (2)
Publication Number | Publication Date |
---|---|
CN105740891A true CN105740891A (en) | 2016-07-06 |
CN105740891B CN105740891B (en) | 2019-10-08 |
Family
ID=56247276
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201610056601.3A Active CN105740891B (en) | 2016-01-27 | 2016-01-27 | Target detection based on multi level feature selection and context model |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN105740891B (en) |
Cited By (11)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN106372597A (en) * | 2016-08-31 | 2017-02-01 | 李涛 | CNN traffic detection method based on adaptive context information |
CN106408618A (en) * | 2016-08-31 | 2017-02-15 | 上海交通大学 | Image deconstruction method based on machine learning |
CN106446933A (en) * | 2016-08-31 | 2017-02-22 | 河南广播电视大学 | Multi-target detection method based on context information |
CN106991296A (en) * | 2017-04-01 | 2017-07-28 | 大连理工大学 | Ensemble classifier method based on the greedy feature selecting of randomization |
CN107239827A (en) * | 2017-06-18 | 2017-10-10 | 北京理工大学 | A kind of spatial information learning method based on artificial neural network |
CN108229519A (en) * | 2017-02-17 | 2018-06-29 | 北京市商汤科技开发有限公司 | The method, apparatus and system of image classification |
CN108830903A (en) * | 2018-04-28 | 2018-11-16 | 杨晓春 | A kind of steel billet method for detecting position based on CNN |
CN108846047A (en) * | 2018-05-30 | 2018-11-20 | 百卓网络科技有限公司 | A kind of picture retrieval method and system based on convolution feature |
CN110298402A (en) * | 2019-07-01 | 2019-10-01 | 国网内蒙古东部电力有限公司 | A kind of small target deteection performance optimization method |
CN111553228A (en) * | 2020-04-21 | 2020-08-18 | 佳都新太科技股份有限公司 | Method, device, equipment and storage medium for detecting personal bag relationship |
US10963676B2 (en) | 2016-12-23 | 2021-03-30 | Samsung Electronics Co., Ltd. | Image processing method and apparatus |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20070297675A1 (en) * | 2006-06-26 | 2007-12-27 | Shih-Jong J. Lee | Method of directed feature development for image pattern recognition |
CN102110227A (en) * | 2010-11-24 | 2011-06-29 | 清华大学 | Compound method for classifying multiresolution remote sensing images based on context |
CN102495865A (en) * | 2011-11-28 | 2012-06-13 | 南京大学 | Image annotation method combined with image internal space relation and visual symbiosis relation |
CN103514456A (en) * | 2013-06-30 | 2014-01-15 | 安科智慧城市技术(中国)有限公司 | Image classification method and device based on compressed sensing multi-core learning |
CN104778466A (en) * | 2015-04-16 | 2015-07-15 | 北京航空航天大学 | Detection method combining various context clues for image focus region |
-
2016
- 2016-01-27 CN CN201610056601.3A patent/CN105740891B/en active Active
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20070297675A1 (en) * | 2006-06-26 | 2007-12-27 | Shih-Jong J. Lee | Method of directed feature development for image pattern recognition |
CN102110227A (en) * | 2010-11-24 | 2011-06-29 | 清华大学 | Compound method for classifying multiresolution remote sensing images based on context |
CN102495865A (en) * | 2011-11-28 | 2012-06-13 | 南京大学 | Image annotation method combined with image internal space relation and visual symbiosis relation |
CN103514456A (en) * | 2013-06-30 | 2014-01-15 | 安科智慧城市技术(中国)有限公司 | Image classification method and device based on compressed sensing multi-core learning |
CN104778466A (en) * | 2015-04-16 | 2015-07-15 | 北京航空航天大学 | Detection method combining various context clues for image focus region |
Non-Patent Citations (1)
Title |
---|
刘扬 等: "高分辨率遥感影像目标分类与识别研究进展", 《地球信息科学学报》 * |
Cited By (17)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN106446933B (en) * | 2016-08-31 | 2019-08-02 | 河南广播电视大学 | Multi-target detection method based on contextual information |
CN106408618A (en) * | 2016-08-31 | 2017-02-15 | 上海交通大学 | Image deconstruction method based on machine learning |
CN106446933A (en) * | 2016-08-31 | 2017-02-22 | 河南广播电视大学 | Multi-target detection method based on context information |
CN106372597A (en) * | 2016-08-31 | 2017-02-01 | 李涛 | CNN traffic detection method based on adaptive context information |
CN106372597B (en) * | 2016-08-31 | 2019-09-13 | 郑州禅图智能科技有限公司 | CNN Vehicle Detection method based on adaptive contextual information |
CN106408618B (en) * | 2016-08-31 | 2019-05-07 | 上海交通大学 | A kind of image destructing method based on machine learning |
US10963676B2 (en) | 2016-12-23 | 2021-03-30 | Samsung Electronics Co., Ltd. | Image processing method and apparatus |
CN108229519B (en) * | 2017-02-17 | 2020-09-04 | 北京市商汤科技开发有限公司 | Image classification method, device and system |
CN108229519A (en) * | 2017-02-17 | 2018-06-29 | 北京市商汤科技开发有限公司 | The method, apparatus and system of image classification |
CN106991296B (en) * | 2017-04-01 | 2019-12-27 | 大连理工大学 | Integrated classification method based on randomized greedy feature selection |
CN106991296A (en) * | 2017-04-01 | 2017-07-28 | 大连理工大学 | Ensemble classifier method based on the greedy feature selecting of randomization |
CN107239827B (en) * | 2017-06-18 | 2020-06-09 | 北京理工大学 | Spatial information learning method based on artificial neural network |
CN107239827A (en) * | 2017-06-18 | 2017-10-10 | 北京理工大学 | A kind of spatial information learning method based on artificial neural network |
CN108830903A (en) * | 2018-04-28 | 2018-11-16 | 杨晓春 | A kind of steel billet method for detecting position based on CNN |
CN108846047A (en) * | 2018-05-30 | 2018-11-20 | 百卓网络科技有限公司 | A kind of picture retrieval method and system based on convolution feature |
CN110298402A (en) * | 2019-07-01 | 2019-10-01 | 国网内蒙古东部电力有限公司 | A kind of small target deteection performance optimization method |
CN111553228A (en) * | 2020-04-21 | 2020-08-18 | 佳都新太科技股份有限公司 | Method, device, equipment and storage medium for detecting personal bag relationship |
Also Published As
Publication number | Publication date |
---|---|
CN105740891B (en) | 2019-10-08 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN105740891B (en) | Target detection based on multi level feature selection and context model | |
US11429818B2 (en) | Method, system and device for multi-label object detection based on an object detection network | |
CN108427920B (en) | Edge-sea defense target detection method based on deep learning | |
US11574152B2 (en) | Recognition system for security check and control method thereof | |
CN105740909B (en) | Text recognition method under a kind of natural scene based on spatial alternation | |
CN110334765B (en) | Remote sensing image classification method based on attention mechanism multi-scale deep learning | |
US10867167B2 (en) | Collaborative deep network model method for pedestrian detection | |
CN111160440B (en) | Deep learning-based safety helmet wearing detection method and device | |
Thai et al. | Image classification using support vector machine and artificial neural network | |
CN108388896A (en) | A kind of licence plate recognition method based on dynamic time sequence convolutional neural networks | |
CN107423760A (en) | Based on pre-segmentation and the deep learning object detection method returned | |
CN104992167A (en) | Convolution neural network based face detection method and apparatus | |
CN105740892A (en) | High-accuracy human body multi-position identification method based on convolutional neural network | |
CN106022363B (en) | A kind of Chinese text recognition methods suitable under natural scene | |
CN107767416B (en) | Method for identifying pedestrian orientation in low-resolution image | |
CN104599275A (en) | Understanding method of non-parametric RGB-D scene based on probabilistic graphical model | |
CN105184298A (en) | Image classification method through fast and locality-constrained low-rank coding process | |
CN107784288A (en) | A kind of iteration positioning formula method for detecting human face based on deep neural network | |
CN110543906B (en) | Automatic skin recognition method based on Mask R-CNN model | |
CN103186776B (en) | Based on the human body detecting method of multiple features and depth information | |
CN112861970B (en) | Fine-grained image classification method based on feature fusion | |
CN109858327B (en) | Character segmentation method based on deep learning | |
CN113221956B (en) | Target identification method and device based on improved multi-scale depth model | |
CN114360067A (en) | Dynamic gesture recognition method based on deep learning | |
CN116796248A (en) | Forest health environment assessment system and method thereof |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
C06 | Publication | ||
PB01 | Publication | ||
C10 | Entry into substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant | ||
TR01 | Transfer of patent right | ||
TR01 | Transfer of patent right |
Effective date of registration: 20211122 Address after: 518052 Room 201, building A, 1 front Bay Road, Shenzhen Qianhai cooperation zone, Shenzhen, Guangdong Patentee after: Shenzhen Xiaofeng Technology Co.,Ltd. Address before: 100124 No. 100 Chaoyang District Ping Tian Park, Beijing Patentee before: Beijing University of Technology |