CN103177264B - The image classification method that view-based access control model dictionary Global Topological is expressed - Google Patents

The image classification method that view-based access control model dictionary Global Topological is expressed Download PDF

Info

Publication number
CN103177264B
CN103177264B CN201310081556.3A CN201310081556A CN103177264B CN 103177264 B CN103177264 B CN 103177264B CN 201310081556 A CN201310081556 A CN 201310081556A CN 103177264 B CN103177264 B CN 103177264B
Authority
CN
China
Prior art keywords
image
vision word
word
global topological
sift feature
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201310081556.3A
Other languages
Chinese (zh)
Other versions
CN103177264A (en
Inventor
黄凯奇
谭铁牛
王冲
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Institute of Automation of Chinese Academy of Science
Original Assignee
Institute of Automation of Chinese Academy of Science
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Institute of Automation of Chinese Academy of Science filed Critical Institute of Automation of Chinese Academy of Science
Priority to CN201310081556.3A priority Critical patent/CN103177264B/en
Publication of CN103177264A publication Critical patent/CN103177264A/en
Application granted granted Critical
Publication of CN103177264B publication Critical patent/CN103177264B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Abstract

The invention discloses the image classification method that a kind of view-based access control model dictionary Global Topological is expressed, including training and two processes of identification, specifically include step: the target image having marked classification is carried out feature extraction, the feature extracted is carried out Global Topological coding on visual dictionary, coding result is trained and models;The image of unknown classification is carried out feature extraction, the feature extracted is carried out Global Topological coding on visual dictionary, be input to train the model obtained by coding result, it is thus achieved that the classification of target image.Expressing the manifold for image due to Global Topological to express and have invariance, therefore the present invention uses the Global Topological of view-based access control model dictionary to express the precision improving image recognition, and this technology has great importance for the understanding of dynamic image.The present invention is expressed by the Global Topological of study visual dictionary, can identify the classification of image accurately, and this technology can be widely applied to safety verification, the field such as web search and digital entertainment.

Description

The image classification method that view-based access control model dictionary Global Topological is expressed
Technical field
The present invention relates to area of pattern recognition and computer vision field, open up particularly to a kind of view-based access control model dictionary overall situation Flutter the image classification method of expression.
Background technology
In computer vision, image classification is basic studying a question.In the past few decades, researchers taste Examination goes the physiology according to human cognitive according to studying classification problem.And as one of them important foundation, researchers Think that the fundamental that image is classified is the potential manifold expression of image.This manifold is expressed and is come from neuron coding, And researchers propose utilize interneuronal topological relation to obtain certain topological invariance, and this invariance for figure The manifold of picture is expressed and is served important function.Therefore, vision topology is to connect neuron coding and the bridge of manifold expression, and The feature coding that this relationship mechanism has been widely applied in image classification is expressed with manifold.Numerous studies in the last few years Show, describe vision topology and can preferably reconstruct local feature so that the manifold of image is expressed more smooth, to reach to promote The purpose of image classification accuracy.Therefore, the vision topology in the classification of research image is necessary.
In recent years, increasing researcher considered topological structure in image is classified, and especially compiled at local feature Code part.Topological structure is used for feature coding and comes from topology expression neutral net, and researchers in a network the earliest Utilize the topological relation between neuron to come more accurately neuron is encoded.Expressed by this topology and inspired, many Local feature is encoded by the topological relation that researcher starts to be devoted to utilize the vision word in visual dictionary model.Mesh Front topological structure mainly considers melting of the topological relation of two vision word, such as two closest vision word Close, the distance of two vision word and angular relationship etc., and this relation of topological structure two-by-two achieves certain effect.
But, the theory expressed according to manifold, a limitation of this topological structure two-by-two is that it can not good table Levy the manifold invariance of feature space.Accordingly, it is considered to one can allow manifold express more smooth higher order topology structure and just show Must be highly desirable to, such as, consider the global Topological Structure between multiple vision word.Another one limitation is due to vision Single contamination forms exponential increase along with the increase of vision word number, it is considered to the topological structure of this high-order is very challenging 's.Therefore, in the present invention, it will thus provide the image classification method that a kind of view-based access control model dictionary Global Topological is expressed overcomes above Two limitation, to reach better image recognition effect.
Summary of the invention
In order to solve the problem that prior art exists, the present invention provides the figure that a kind of view-based access control model dictionary Global Topological is expressed As sorting technique.
The image classification method that a kind of view-based access control model dictionary Global Topological that the present invention proposes is expressed, comprises the following steps:
Step 1, gathers several training images, and several training images carry out local sampling respectively, and in the local obtained Extract Scale invariant features transform (SIFT) feature on sampling block, thus obtain the SIFT feature set of training image;
Step 2, carries out cluster and generates multiple cluster centres, and with described cluster centre be the SIFT feature set obtained Vision word composition visual dictionaryWherein, C represents visual dictionary, its vision tieed up by M D Word ciComposition,Represent the subspace that M point in D dimension space forms;
Step 3, carries out Global Topological coding to each SIFT feature of every width training image;
Step 4, all SIFT feature to every width training imageGlobal Topological coding V1, V2..., VNCarry out maximum aggregation operator, generate the image expression F of this training image;
Step 5, sends into the image expression of all training images in grader and is trained, generate training pattern;
Step 6, similar to described step 1, every image to be identified is carried out local sampling, and at the local sampling obtained Extract Scale invariant features transform SIFT feature on block, obtain the SIFT feature set of every image to be identified;
Step 7, based on the visual dictionary being made up of vision word obtained in described step 2, utilizes described step 3 to often Each SIFT feature of image to be identified carries out Global Topological coding;
Step 8, similar to described step 4, the Global Topological of all SIFT feature of every image to be identified is coded into The maximum aggregation operator of row, generates the image expression of image to be identified;
Step 9, the image expression of the image to be identified described step 8 obtained sends into the training mould that described step 5 generates Type is tested, thus obtains the other recognition result of target class in image to be identified.
The method according to the invention, in complex situations, the topological structure of image still can be by robust for target image Ground obtains, thus carries out the robust control policy of image.In intelligent visual surveillance system, this technology can be used for identification monitoring system field The classification of target in scape so that monitoring system can well identify the behavior that target object is current.
Accompanying drawing explanation
Fig. 1 is the image classification method flow chart that view-based access control model dictionary Global Topological of the present invention is expressed.
Detailed description of the invention
For making the object, technical solutions and advantages of the present invention clearer, below in conjunction with specific embodiment, and reference Accompanying drawing, the present invention is described in more detail.
Fig. 1 is the image classification method flow chart that view-based access control model dictionary Global Topological of the present invention is expressed, as it is shown in figure 1, this The image classification method that the view-based access control model dictionary Global Topological that invention is proposed is expressed comprises the following steps:
Step 1, gathers several training images, and several training images carry out local sampling, local sampling block pixel respectively Size can be such as 16*16, and on the local sampling block obtained extract Scale invariant features transform SIFT feature, thus Obtain the SIFT feature set of training image;
Step 2, carries out cluster and generates multiple cluster centres, and with described cluster centre be the SIFT feature set obtained Vision word composition visual dictionaryWherein, C represents visual dictionary, its vision tieed up by M D Word ciComposition,Represent the subspace that M point in D dimension space forms;
In this step, described cluster can use clustering algorithm commonly used in the prior art, such as k-means clustering algorithm.
Step 3, carries out Global Topological coding to each SIFT feature of every width training image;
Described step 3 farther includes three sub-steps:
Step 3.1, calculates the dependency between vision word in described visual dictionary;
In an embodiment of the present invention, utilize a kind of independent increment formula algorithm based on distance and angle obtain vision word it Between dependency, before described independent increment formula algorithm based on distance and angle is introduced, first define several parameter:
At visual dictionaryIn, for vision word ciIf, cjWith ciIt is relevant, Then define them in feature space, form a cone:
c i &RightArrow; c j = { y | | | y - c i | | 2 2 &le; &mu; i , &Delta; ( c i , y , c j ) < &theta; } - - - ( 1 )
Wherein, ci→cjRepresent ciAnd cjCone in the feature space formed between two vision word, y is appointing in cone Anticipate a characteristic point, y} represents this cone,Represent y and ciBetween Euclidean distance, μiIt is ciAfter clustering with k-means Belong to ciCharacteristic point between maximum Euclidean distance, θ be control cone angle;Δ(ci, y, cj) it is vector y-ciWith vector cj- ciBetween angle, be defined as follows:
&Delta; ( c i , y , c j ) = arccos &lang; y - c i , c j - c i &rang; | | y - c i | | 2 &CenterDot; | | c j - c i | | 2 - - - ( 2 )
Wherein, < y-ci, cj-ci> represent vector y-ciWith vector cj-ciBetween inner product, " " represents between two vectors Dot product, | | | |2Represent two norms of vector.
Next the angle between the vector that the cone defined according to formula (1) and formula (2) define, it is possible to based on distance With the independent increment algorithm of angle obtains the dependency between vision word, the base of described independent increment algorithm based on distance and angle This thinking is that each relevant vision word occupies an independent cone region.Specifically, described based on away from walk-off angle The independent increment algorithm of degree includes:
Firstly, for a certain vision word ci, all of vision word arrives c according to itiEuclidean distance carry out from the near to the remote Sequence, wherein, using nearest vision word as initial relevant vision word ci1
Secondly, except vision word ciOther outer vision word check according to distance and angle criterion one by one, if A certain vision word cjBe satisfied by described distance and angle criterion through inspection, then this vision word is marked as multi view list Word cij, and add set S toi, finally give vision word ciThe set S of multi view wordi
Wherein, described distance criterion is defined as:
||ci-cij||2< τ | | ci-ci1||2 (3)
Wherein, τ is used for controlling ciEuclidean distance to other vision word, say, that distance ciThe most remote vision word Do not account for;
Described angle rule definition is:
Δ(ci, cij, ck) > θ, &ForAll; { c k } &Element; S i - - - ( 4 )
Wherein, SiIt it is vision word ciThe set of multi view word, and ckRepresentative has been added to gather SiIn phase Close vision word.
Finally, to all of vision word traversal processing, i.e. can get the related words set S of visual dictionary C:
Wherein, D represents the dimension of each vision word, SiIt it is vision word ciThe set of related words, QiIt is SiIn Vision word number.
Step 3.2, utilizes dependency between the vision word obtained in described step 3.1 to each SIFT feature K the vision word of neighbour sets up Global Topological model;
In this step, it is assumed that the SIFT feature collection obtained is extracted for every width training image and is combined intoWherein, D is intrinsic dimensionality, and N is Characteristic Number.Then for each feature xj, its overall situation is opened up Flutterring model is:
Δ (C, xj, S) and=[Δ (c1, xj, S1) ..., Δ (cM, xj, SM)] (7)
&ForAll; &Delta; ( c i , x j , S i ) = [ &Delta; ( c i , x j , c i 1 ) , . . . , &Delta; ( c i , x j , c iQ i ) ] &Element; R 1 &times; Q i - - - ( 8 )
Wherein, Δ (ci, xj, Si) it is vector xj-ciWith set SiIn each multi view word cijWith ciConstitute to Amount cij-ciBetween angle:
&Delta; ( c i , x j , S i ) = [ &Delta; ( c i , x j , c i 1 ) , . . . , &Delta; ( c i , x j , c iQ i ) ] ,
&Delta; ( c i , x j , c ij ) = arccos &lang; x j - x i , c ij - c i &rang; | | x j - c i | | 2 &CenterDot; | | c ij - c i | | 2 , <xj-ci, cij-ci> represent vector xj-ciAnd vector cij-ciBetween inner product, " " represents the dot product between two vectors, | | | |2Represent two norms of vector.
Step 3.3, the Global Topological model utilizing described step 3.2 to obtain carries out Global Topological volume to each SIFT feature Code.
In this step, utilize following formula to each SIFT feature xiCarry out Global Topological coding:
arg min V i | | x i - SV i T | | 2 2 + [ &lambda;T ( V i ) + &alpha; | | V i &CenterDot; &Delta; ( C , x i , S ) | | 2 2 ] - - - ( 9 )
Wherein, λ is penalty term coefficient, T (Vi) it is to ViAny form of penalty term, α is the punishment of Global Topological model , " " represents the dot product between two vectors, ViIt is xiCoding on S, is defined as follows:
Wherein, (vk1)iIt is this local feature xiIn vision word ckFirst multi view word ck1On response, depend on Secondary analogize.
Then, formula (9) is optimized and solves, i.e. can get SIFT feature xiGlobal Topological coding Vi
Step 4, all SIFT feature to every width training imageGlobal Topological coding V1, V2..., VNCarry out maximum aggregation operator, generate the image expression F of this training image;
In this step, carry out maximum aggregation operator according to the following formula, the image expression F of generation training image:
F=maxcolumn[V1 T, V2 T..., VN T]T (12)
Wherein, maxcolumnRepresent the every string for matrix and only retain its maximum.
Step 5, sends into the image expression of all training images in grader and is trained, generate training pattern;
In this step, described grader can use grader commonly used in the prior art, such as support vector machine.
Step 6, similar to described step 1, every image to be identified is carried out local sampling, and at the local sampling obtained Extract Scale invariant features transform SIFT feature on block, obtain the SIFT feature set of every image to be identified;
Step 7, based on the visual dictionary being made up of vision word obtained in described step 2, utilizes described step 3 (ratio Such as formula (7)-formula (11)), each SIFT feature of every image to be identified is carried out Global Topological coding;
Step 8, similar to described step 4, utilize the image expression that formula (12) defines, the institute to every image to be identified The Global Topological coding having SIFT feature carries out maximum aggregation operator, generates the image expression of image to be identified;
Step 9, the image expression of the image to be identified described step 8 obtained sends into the training mould that described step 5 generates Type is tested, thus obtains the other recognition result of target class in image to be identified.
From the foregoing, it will be observed that the present invention propose view-based access control model dictionary Global Topological express image classification method comprise training and Identifying two processes, next as a example by the vehicle detecting system in some monitoring scene, enforcement to the method is said Bright.Described vehicle detecting system may determine that in monitoring scene whether contain vehicle.
First having to collect substantial amounts of vehicle image (1000) and non-vehicle image (1000), these images are used for training Vehicle identification model.Training step S1 is:
Step S11: 1000 vehicle images (positive sample) and 1000 non-vehicle images (negative sample) are carried out SIFT spy Levy extraction, generate 2000 groups of SIFT feature, calculate containing 2000 SIFT feature with average often group, altogether extract 4000000 (2000 × 2000) individual SIFT feature;
Step S12: 4000000 SIFT feature carry out k-menas cluster operation, generates 1 and comprises 1000 visions The visual dictionary of word;
Step S13: in actual application scenarios, 3 vision word of the arest neighbors taking each SIFT local feature are entered Row Global Topological encodes, and code length is 3000 dimensions;
Step S14: the Global Topological coding of 2000 SIFT local features of every training image is carried out maximum gathering Operation, obtains the image expression of image, a length of 3000 dimensions;
Step S15: the image expression of all training images is sent in grader and is trained, obtain training pattern;
At cognitive phase, camera signals is accessed computer by capture card and carrys out collecting test picture, specifically identify step Rapid S2 is:
Step S21: input a test image, it is carried out SIFT feature and extracts operation, generate 1 group of SIFT feature, contain There are 2000 SIFT feature.
Step S22: in actual application scenarios, 3 vision word of the arest neighbors taking each SIFT local feature are entered Row Global Topological encodes, and code length is 3000 dimensions;
Step S23: the Global Topological coding of 2000 SIFT local features of every test image is carried out maximum gathering Operation, obtains the image expression of image, a length of 3000 dimensions;
Step S24: the image expression of the image to be identified step S23 obtained sends into the training pattern that step S15 generates Test, thus obtain the other recognition result of target class in image to be identified.
To sum up, the present invention proposes the Image Classfication Technology that a kind of effective view-based access control model dictionary Global Topological is expressed.This Invention is easily achieved, stable performance, it is possible to increase the intelligent monitor system understandability to monitoring scene, is intelligence of future generation simultaneously Key technology in energy video monitoring system.
Particular embodiments described above, has been carried out the purpose of the present invention, technical scheme and beneficial effect the most in detail Describe in detail bright, be it should be understood that the specific embodiment that the foregoing is only the present invention, be not limited to the present invention, all Within the spirit and principles in the present invention, any modification, equivalent substitution and improvement etc. done, should be included in the guarantor of the present invention Within the scope of protecting.

Claims (8)

1. the image classification method that a view-based access control model dictionary Global Topological is expressed, it is characterised in that the method includes following step Rapid:
Step 1, gathers several training images, and several training images carry out local sampling respectively, and at the local sampling obtained Extract Scale invariant features transform (SIFT) feature on block, thus obtain the SIFT feature set of training image;
Step 2, carries out cluster and generates multiple cluster centres the SIFT feature set obtained, and with described cluster centre as vision Word composition visual dictionaryWherein, C represents visual dictionary, its vision word tieed up by M D ciComposition,Represent the subspace that M point in D dimension space forms;
Step 3, carries out Global Topological coding to each SIFT feature of every width training image;
Step 4, all SIFT feature to every width training imageGlobal Topological coding V1, V2..., VNCarry out maximum aggregation operator, generate the image expression F of this training image;
Step 5, sends into the image expression of all training images in grader and is trained, generate training pattern;
Step 6, similar to described step 1, every image to be identified is carried out local sampling, and on the local sampling block obtained Extract Scale invariant features transform SIFT feature, obtain the SIFT feature set of every image to be identified;
Step 7, based on the visual dictionary being made up of vision word obtained in described step 2, utilizes described step 3 to treat every width Identify that each SIFT feature of image carries out Global Topological coding;
Step 8, similar to described step 4, the Global Topological coding of all SIFT feature of every image to be identified is carried out Big aggregation operator, generates the image expression of image to be identified;
Step 9, the training pattern that the image expression described step 5 of feeding of the image to be identified described step 8 obtained generates is entered Row test, thus obtain the other recognition result of target class in image to be identified;
Described step 3 farther includes three sub-steps:
Step 3.1, calculates the dependency between vision word in described visual dictionary;
Step 3.2, utilizes the arest neighbors to each SIFT feature of the dependency between the vision word obtained in described step 3.1 K vision word set up Global Topological model;
Step 3.3, the Global Topological model utilizing described step 3.2 to obtain carries out Global Topological coding to each SIFT feature;
In described step 3.2, it is assumed that the SIFT feature collection obtained is extracted for every width training image and is combined intoWherein, D is intrinsic dimensionality, and N is Characteristic Number, then for each feature xj, its Global Topological Model representation is:
Δ (C, xj, S) and=[Δ (c1, xj, S1) ..., Δ (cM, xj, SM)],
&ForAll; &Delta; ( c i , x j , S i ) = &lsqb; &Delta; ( c i , x j , c i 1 ) , ... , &Delta; ( c i , x j , c iQ i ) &rsqb; &Element; R 1 &times; Q i ,
Wherein, Δ (ci, xj, Si) it is vector xj-ciWith set SiIn each related words cijWith ciThe vectorial c constitutedij-ciIt Between angle:
&Delta; ( c i , x j , S i ) = [ &Delta; ( c i , x j , c i 1 ) , . . . , &Delta; ( c i , x j , c iQ i ) ] ,
<xj-ci, cij-ci> represent vector xj-ciWith vector cij-ci Between inner product, " " represents the dot product between two vectors, | | | |2Represent two norms of vector.
Method the most according to claim 1, it is characterised in that in described step 2, uses k-means clustering algorithm to carry out Cluster.
Method the most according to claim 1, it is characterised in that utilize independent increment formula algorithm based on distance and angle to obtain Dependency between vision word.
Method the most according to claim 3, it is characterised in that described independent increment algorithm based on distance and angle includes:
Firstly, for a certain vision word ci, all of vision word arrives c according to itiEuclidean distance arrange from the near to the remote Sequence, wherein, using nearest vision word as initial relevant vision word ci1
Secondly, except vision word ciOther outer vision word check according to distance and angle criterion one by one, regard if a certain Feel word cjBe satisfied by described distance and angle criterion through inspection, then this vision word is marked as multi view word cij, And add set S toi, finally give vision word ciThe set S of multi view wordi
Finally, to all of vision word traversal processing, i.e. can get the related words set S of visual dictionary C.
Method the most according to claim 4, it is characterised in that wherein, described distance criterion is defined as:
||ci-cij||2< τ | | ci-ci1| | 2,
Wherein, τ is used for controlling ciEuclidean distance to other vision word;
Described angle rule definition is:
Δ(ci, cij, ck) > θ,
Wherein, SiIt it is vision word ciThe set of multi view word, and ckRepresentative has been added to gather SiIn vision list Word.
Method the most according to claim 4, it is characterised in that the related words set S of described visual dictionary C can represent For:
Wherein, D represents the dimension of each vision word, SiIt it is vision word ciThe set of multi view word, QiIt is SiIn Vision word number.
Method the most according to claim 1, it is characterised in that in described step 3.3, utilizes following formula to each SIFT feature xiCarry out Global Topological coding:
argmin V i | | x i - SV i T | | 2 2 + &lsqb; &lambda; T ( V i ) + &alpha; | | V i &CenterDot; &Delta; ( C , x i , S ) | | 2 2 &rsqb; ,
Wherein, λ is penalty term coefficient, T (Vi) it is to ViAny form of penalty term, α is the penalty term of Global Topological model, " " represents the dot product between two vectors, ViIt is xiGlobal Topological coding on S:
Wherein, (vk1)iIt it is each SIFT feature xiIn vision word ckFirst multi view word ck1On response, successively Analogize.
Method the most according to claim 1, it is characterised in that in described step 4, carries out grasping maximum gathering according to the following formula Make, the image expression F of generation training image:
F=maxcolumn[V1 T, V2 T..., VN T]T
Wherein, maxcolumnRepresent the every string for matrix and only retain its maximum.
CN201310081556.3A 2013-03-14 2013-03-14 The image classification method that view-based access control model dictionary Global Topological is expressed Active CN103177264B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201310081556.3A CN103177264B (en) 2013-03-14 2013-03-14 The image classification method that view-based access control model dictionary Global Topological is expressed

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201310081556.3A CN103177264B (en) 2013-03-14 2013-03-14 The image classification method that view-based access control model dictionary Global Topological is expressed

Publications (2)

Publication Number Publication Date
CN103177264A CN103177264A (en) 2013-06-26
CN103177264B true CN103177264B (en) 2016-09-14

Family

ID=48637105

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201310081556.3A Active CN103177264B (en) 2013-03-14 2013-03-14 The image classification method that view-based access control model dictionary Global Topological is expressed

Country Status (1)

Country Link
CN (1) CN103177264B (en)

Families Citing this family (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104751198B (en) 2013-12-27 2018-04-27 华为技术有限公司 The recognition methods of object in image and device
CN103984959B (en) * 2014-05-26 2017-07-21 中国科学院自动化研究所 A kind of image classification method based on data and task-driven
CN104464079B (en) * 2014-12-29 2016-10-05 北京邮电大学 Multiple Currencies face amount recognition methods based on template characteristic point and topological structure thereof
CN104598898B (en) * 2015-02-13 2018-02-06 合肥工业大学 A kind of Aerial Images system for rapidly identifying and its method for quickly identifying based on multitask topology learning
CN108319907A (en) * 2018-01-26 2018-07-24 腾讯科技(深圳)有限公司 A kind of vehicle identification method, device and storage medium

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101315663A (en) * 2008-06-25 2008-12-03 中国人民解放军国防科学技术大学 Nature scene image classification method based on area dormant semantic characteristic
CN102402621A (en) * 2011-12-27 2012-04-04 浙江大学 Image retrieval method based on image classification
CN102509110A (en) * 2011-10-24 2012-06-20 中国科学院自动化研究所 Method for classifying images by performing pairwise-constraint-based online dictionary reweighting
CN102609732A (en) * 2012-01-31 2012-07-25 中国科学院自动化研究所 Object recognition method based on generalization visual dictionary diagram

Family Cites Families (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2001338294A (en) * 2000-05-24 2001-12-07 Monolith Co Ltd Form analyzer targeting on topology
US8766982B2 (en) * 2010-01-19 2014-07-01 Disney Enterprises, Inc. Vectorization of line drawings using global topology and storing in hybrid form

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101315663A (en) * 2008-06-25 2008-12-03 中国人民解放军国防科学技术大学 Nature scene image classification method based on area dormant semantic characteristic
CN102509110A (en) * 2011-10-24 2012-06-20 中国科学院自动化研究所 Method for classifying images by performing pairwise-constraint-based online dictionary reweighting
CN102402621A (en) * 2011-12-27 2012-04-04 浙江大学 Image retrieval method based on image classification
CN102609732A (en) * 2012-01-31 2012-07-25 中国科学院自动化研究所 Object recognition method based on generalization visual dictionary diagram

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
"Global topological features of cancer proteins in the human";Pall F. Jonsson .etc;《Bioinformatics》;20060915;第22卷(第18期);全文 *
"基于全局拓扑结构的分级三角剖分图像拼接";曾丹 等;《计算机研究与发展》;20121230;第49卷(第1期);全文 *
C. Bonatti .etc."Topological classification of gradient-like diffeomorphisms on 3-manifolds".《ELSEVIER》.2004,369–391. *

Also Published As

Publication number Publication date
CN103177264A (en) 2013-06-26

Similar Documents

Publication Publication Date Title
Sun et al. RSOD: Real-time small object detection algorithm in UAV-based traffic monitoring
CN107767405B (en) Nuclear correlation filtering target tracking method fusing convolutional neural network
CN104966104B (en) A kind of video classification methods based on Three dimensional convolution neutral net
CN108090472B (en) Pedestrian re-identification method and system based on multi-channel consistency characteristics
CN103177264B (en) The image classification method that view-based access control model dictionary Global Topological is expressed
Peng et al. Rgb-t crowd counting from drone: A benchmark and mmccn network
CN107133569A (en) The many granularity mask methods of monitor video based on extensive Multi-label learning
CN103632168A (en) Classifier integration method for machine learning
CN103699874B (en) Crowd abnormal behavior identification method based on SURF (Speed-Up Robust Feature) stream and LLE (Locally Linear Embedding) sparse representation
CN103020658B (en) Recognition method for objects in two-dimensional images
CN108319672A (en) Mobile terminal malicious information filtering method and system based on cloud computing
Mussina et al. Multi-modal data fusion using deep neural network for condition monitoring of high voltage insulator
CN111460881A (en) Traffic sign countermeasure sample detection method and classification device based on neighbor discrimination
CN109523514A (en) To the batch imaging quality assessment method of Inverse Synthetic Aperture Radar ISAR
Bhuiyan et al. Video analytics using deep learning for crowd analysis: a review
Wei et al. Lightweight multimodal feature graph convolutional network for dangerous driving behavior detection
CN112990282B (en) Classification method and device for fine-granularity small sample images
Ye et al. The image data and backbone in weakly supervised fine-grained visual categorization: A revisit and further thinking
CN106056131A (en) Image feature extraction method based on LRR-LDA
Li et al. Action recognition with spatio-temporal augmented descriptor and fusion method
CN102609732B (en) Object recognition method based on generalization visual dictionary diagram
Zhao et al. Multi-angle insulator recognition method in infrared image based on parallel deep convolutional neural networks
Qu et al. A time sequence location method of long video violence based on improved C3D network
Wang et al. A lightweight and stochastic depth residual attention network for remote sensing scene classification
Masilamani et al. Art classification with pytorch using transfer learning

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
C14 Grant of patent or utility model
GR01 Patent grant