CN103049514B - A kind of equilibrium image clustering method based on hierarchical cluster - Google Patents
A kind of equilibrium image clustering method based on hierarchical cluster Download PDFInfo
- Publication number
- CN103049514B CN103049514B CN201210545637.XA CN201210545637A CN103049514B CN 103049514 B CN103049514 B CN 103049514B CN 201210545637 A CN201210545637 A CN 201210545637A CN 103049514 B CN103049514 B CN 103049514B
- Authority
- CN
- China
- Prior art keywords
- cluster
- clustering
- data
- image
- clustering cluster
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Expired - Fee Related
Links
Abstract
The invention discloses a kind of equilibrium image clustering method based on hierarchical cluster, the present invention is directed to dress ornament class commodity image high dimensional feature data, use method based on hierarchical clustering, it is thus achieved that the clustering cluster of size equalization, and the data volume that single clustering cluster comprises is less than the threshold value limited.During retrieval, after the data that will be retrieved and all cluster centres carry out distance calculating, choose nearest multiple clustering cluster, inside multiple clustering cluster, carry out data traversal, it is thus achieved that last Query Result.Relative to general indexing means based on cluster, this method avoid the problem that ergodic data amount is excessive when the data that are retrieved are in big clustering cluster, it is ensured that the performance of inquiry.Meanwhile, by the way of traversal multi-cluster bunch, Query Result has higher registration with the Query Result of SSA, improves inquiry effect.
Description
Technical field
The present invention relates to picture search technical field, particularly relate to a kind of image higher-dimension based on hierarchical cluster to
Amount quickly approximation k-neighbour's search method.
Background technology
In picture search technology based on content (Content-Based Image Retrieval, CBIR),
When user uploads a width commodity image and expects to search the commodity same or like with this figure, search engine pair
The commodity image that user uploads carries out feature extraction, and chooses and it from thumbnail feature vector data storehouse
K image closest in higher dimensional space returns as result.A large amount of index property data bases are looked into
Asking k nearest characteristics of image, most basic method is SSA method.SSA method is retrieved by calculating
Image has put the distance of image in storage with each, and the mode being then ranked up these distances obtains nearest k
Individual image.This is a kind of accurate k neighbour retrieval (k-Nearest Neighbor, kNN).But,
When in characteristics of image dimension and storehouse, amount of images is bigger, the inquiry of the method is the biggest, it is impossible to meet
Requirement of engineering.
The method of cluster is introduced in CBIR.The method using cluster, by data according to it at higher dimensional space
Distribution, assembles and becomes clustering cluster;During retrieval, first calculate the image distance with the center of all bunches that is retrieved,
Determine the clustering cluster being retrieved belonging to image, then bunch interior data are traveled through, it is thus achieved that nearest k
Image.Owing to needing the minimizing of the data volume of traversal, the mode recall precision that the method travels through relative to forward
Increase, but there is problems in that
1, query time efficiency depend on be queried belonging to image bunch size, if cluster produce bunch
Size is unbalanced, and query time can be caused to produce lack of uniformity.Belong to comprise image number when being queried image
When bigger bunch, need the image volume of traversal and inquiry is time-consuming increases.Bunch representative big owing to comprising data volume
The more characteristics of image of " common ", is queried probability that image falls wherein more than comprising the clustering cluster that data volume is few.
Therefore, if the data volume that certain clustering cluster comprises is far above meansigma methods, it will have a strong impact on commodity image and search
The average response time that index is held up.
2, in data traversal is limited at bunch, if there being k-neighbour's data to be in other bunches, then at retrieval knot
It is lost in Guo, causes inquiring about effect and reduce.
Summary of the invention
Present invention aims to the deficiencies in the prior art, it is provided that the image clustering method of a kind of optimization.
It is an object of the invention to be achieved through the following technical solutions: a kind of equilibrium figures based on hierarchical cluster
As clustering method, comprise the steps of:
(1) when setting up index, first image feature data is carried out initial clustering;
(2) each clustering cluster obtaining step (1) carries out clustering slicing operation.Concretely comprise the following steps: checking should
The image number that cluster is comprised.If the image number that this cluster centre comprises is more than upper limit N arrangedtop,
Then carry out two segregation classes at intra-cluster.If the data volume that the result of two segregation classes comprises is still above Ntop, then
This process of result iteration to two segregation classes.By data volume less than NtopClustering cluster central record to cluster in
In heart file.Afterwards all for this classification image feature datas are organized according to the cluster centre obtained.
(3) when retrieval, the characteristic to query image, calculate its all cluster centres arriving affiliated classification
Distance, and these distances are carried out ascending sort, obtain front c the clustering cluster mark that distance is minimum, c value
Specified by systematic parameter.Carry out data traversal in the inside of c clustering cluster afterwards, obtain last inquiry knot
Really.
The invention has the beneficial effects as follows, this patent, for dress ornament class commodity image high dimensional feature data, uses base
Method in hierarchical clustering, it is thus achieved that the clustering cluster of size equalization, and the data volume that single clustering cluster comprises do not surpasses
Cross the threshold value limited.During retrieval, after the data that will be retrieved and all cluster centres carry out distance calculating, choose
Nearest multiple clustering cluster, carry out data traversal, it is thus achieved that last Query Result inside multiple clustering cluster.
Relative to general indexing means based on cluster, this method avoid when the data that are retrieved are in big clustering cluster
Time the excessive problem of ergodic data amount, it is ensured that the performance of inquiry.Meanwhile, by traveling through the side of multi-cluster bunch
Formula, Query Result has higher registration with the Query Result of SSA, improves inquiry effect.
Accompanying drawing explanation
Fig. 1 is that commodity image characteristic indexes Establishing process;
Fig. 2 is that commodity image characteristic clusters cutting flow chart;
Fig. 3 is that commodity image characteristic puts flow chart in storage;
Fig. 4 is retrieval flow figure;
Fig. 5 is " edge effect " schematic diagram under two-dimensional case.
Detailed description of the invention
Below with the cluster of dress ornament class commodity image, index set up, retrieve and safeguard as a example by, detailed in conjunction with accompanying drawing
Carefully describe the present invention, the purpose of the present invention and effect will be apparent from.
As it is shown in figure 1, the present invention based on hierarchical cluster equilibrium image clustering method index set up include as
Lower step:
Step 1: commodity image is carried out image characteristics extraction, view data is converted into feature vector data.
The purpose of feature extraction is that the low structure obtaining image describes.Each feature is represented by d n dimensional vector n.
The present invention uses the global characteristics of image, the corresponding high dimensional feature vectors of the most each sub-picture.
Every one dimensional numerical of characteristic vector is all used for phenogram picture feature in terms of some, such as shape, color,
The information such as texture, structure.Image characteristic extracting method is a lot, and MPEG-7 visual feature extraction tools is one
Plant popular method.The method include color layout describe (Color Layout Descriptor, CLD),
Edge histogram descriptor (Edge Histogram Descriptor, EHD) etc..Wherein, CLD uses 8*8
12 coefficients of DCT, are suitable for the compactest and that resolution is constant color and represent.EHD uses 80
Rectangular histogram window describes the content from 16 subimages.
For the ease of data storage with calculate, it is whole that every one-dimensional characteristic value is quantified as in the range of [0,255] by we
Number.Characteristic vector after quantization, each dimension can be stored as a byte.
Step 2: the initial characteristic data obtaining step 1 carries out initial clustering, cluster centre number is set to
One less integer.The purpose that data carry out initial clustering is the distribution shape in order to probably embody data
State.The algorithm that cluster is used is k-average (K-Means).
N data object of input is divided into K-Means algorithm k cluster so that obtained is poly-
Class meets: the object similarity in same cluster is higher;And object similarity in different cluster is less.
The major parameter of K-Means algorithm includes: cluster number k and range formula d (x, y).
Owing to simply data being carried out initial clustering, smaller k value is set here.It is desirable that carry out
After initial clustering, the average amount that each clustering cluster comprises is fixing value Ns.The k of initial clustering
Data total amount N when value can be set up by indextotalAnd NsCalculate:
For range formula, general distance include manhatton distance (L1 distance), Euclidean distance (L2 away from
From), mahalanobis distance (Mahalanobis distance) etc..Range formula does not requires to calculate all of feature dimensions
Degree, can carry out distance with the dimension that discrimination in selected characteristic vector is bigger and calculate.For same classification, should
The follow-up step that range formula needs keep consistent.Different distances can be used for different classifications
Computing formula.
Step 3: the data obtaining step 2 carry out clustering cutting, such as Fig. 2.I.e. obtain step 2 is every
Individual clustering cluster, proceeds as follows:
Step 3.1: check the image number that this cluster is comprised.If the image number that this cluster centre comprises
Less than threshold value N arrangedtopThen jump to step 3.3, otherwise jump to step 3.2.
NtopSetting depend on calculating and the IO performance of server.The complexity calculated due to higher dimensional space distance
Property, we ignore the amount of calculation of distance-taxis and merger calculating when, and be conceived to make single inquiry away from
Minimum from calculation times.If the clustering cluster number of single inquiry traversal is c.It is the most saturated that we introduce clustering cluster
Degree α, is defined as data volume N that average each clustering cluster comprisesmeanWith NtopRatio:
Therefore, number N of total clustering clusterc:
Retrieval time, it would be desirable to distance calculate include two parts: choose c nearest cluster centre,
And travel through inside c cluster centre.In order to obtain c nearest cluster centre, the distance of needs
Calculation times is equal to number N of clustering clusterc.The distance that carrying out traversal inside c cluster centre needs calculates
Number of times is c × Nmean.Total distance calculation times is:
WhenTime, above formula obtains minima.Now:
Illustrating: comprise 5 for a certain classification, 000,000 image, single inquiry travels through nearest 8 and gathers
Class center, then
Step 3.2: carry out 2 cluster centre clusters in this clustering cluster.Owing to our target is to ensure that each
The data volume that clustering cluster comprises is less than Ntop.Therefore, for comprising data volume more than NtopClustering cluster, I
Carry out inside this clustering cluster 2 cluster centres k-Means cluster.The process of cluster and step 2 phase
Seemingly, difference is that k value here is set to 2 by force.Cluster obtains 2 new clustering cluster, and the two
The data volume that new clustering cluster comprises still can exceed that Ntop.For data volume more than NtopNew clustering cluster, weight
Perform the operation of step 3.2 again, until there is no the data volume that clustering cluster comprises more than Ntop;
Step 3.3: for comprising data volume less than or equal to NtopClustering cluster, by the centre coordinate of this clustering cluster
It is written in cluster centre file.The dimension of cluster centre is equal with the dimension of tag file.In order to ensure essence
Degree, every one-dimensional data of cluster centre is saved as a floating number by us.One cluster centre file may
Comprise the cluster centre data of multiple classification, and classification comprises the coordinate of multiple cluster centre point.
Step 4: as it is shown on figure 3, after obtaining the cluster centre of classification, it would be desirable to initial characteristic data
Reorganize according to the cluster centre obtained, in order to read during retrieval.Concrete method is, for
Each image feature data in primitive character file, calculates its distance to all cluster centres of this classification,
Choose wherein minimum distance, and it is special that this image feature data belongs to clustering cluster corresponding to minimum range
Solicit articles in part.
Step 5: preserve each clustering cluster data to fixed disk file.Characteristics of image file depositing on hard disk
Need to ensure that the data of same clustering cluster save as continuous print hard disc data block.
As shown in Figure 4, the retrieval of the image clustering method of a kind of improvement of the present invention comprises the steps:
Step 6: load cluster centre data to internal memory;
For inquiring about each time, the cluster centre data of classification belonging to commodity all can be used for distance and calculate,
Access for cluster centre data is the most frequently.Therefore, cluster centre data need to reside in internal memory
In.
Knowable to the analysis of step 3.3, the image number that average each clustering cluster comprises is: Ntop× α is corresponding
One cluster centre.Each cluster centre coordinate is d floating number, takies the internal memory of d*4 byte.Cause
This, EMS memory occupation altogether is:
Vcenter=Ntotal/(α×Ntop)×d×4;
Such as, for the characteristic of some classification, Ntotal=5000000, Ntop=2000, d=600,
α=0.6.Memory headroom required for so storing this classification cluster centre is about 10M byte.
Step 7: carry out image characteristics extraction to being queried image;It is identical that this operates in step 1, here
Repeat no more.
Step 8: the image feature data obtaining step 7, calculates its all cluster centres to affiliated classification
Distance, obtain front c the CID that distance is minimum after sequence;
The result of K-neighbour retrieval, in higher dimensional space, constitutes one to be queried vector surpassing as the centre of sphere
Ball.
Traditional index based on cluster and inquiry mode, each inquiry only travels through with to be queried image nearest
Data in clustering cluster.If the center being queried the image feature data clustering cluster affiliated with it is relatively near, then
Hypersphere is completely contained in the envelope of this clustering cluster.So, even if only traveling through nearest 1 cluster centre,
Can also obtain and travel through identical Query Result with full storehouse forward.Q in Fig. 50Illustrate two-dimensional space
In this situation.C0, C1, C2, C3, C4 are 5 cluster centres, q0It is to be queried characteristics of image
Vector.Under two-dimensional case, hypersphere is degenerated to one with q0Centered by circle.In Figure 5, with q0For the center of circle
Circle be completely in C0 clustering cluster envelope.
If being queried the center of the image feature data clustering cluster affiliated with it farther out, then traditional cluster
Inquiry mode can cause loss of data.Such as q in Fig. 51Shown in, it is queried characteristics of image vector and is in C1 cluster
Bunch marginal position, Query Result hypersphere intersects with tetra-clustering cluster of C1, C2, C3, C4 simultaneously.If
Only traversal C1 clustering cluster, then hypersphere is lost with the data of C2, C3, C4 intersecting area.This loss
Effect is properly termed as " edge effect ".
In the case of high-dimensional, " edge effect " is exaggerated.Imagining a space envelope is the clustering cluster of hypersphere,
Radius is r, then the volume of this hypersphere is:
V (r)=a × rd,
Wherein, α is invariant.
We be defined into the distance at clustering cluster center be within r/2 for from center " close to ", then this " nearer "
Domain Volume be:
V (r/s)=a × (r/2)d;
" nearer " Domain Volume accounts for the ratio of whole volume of hypersphere:
Exponentially reduce along with the increase of dimension d due to this ratio, therefore, in higher dimensional space, be queried
It is normality that data fall at clustering cluster edge, and " edge effect " cannot be left in the basket.In order to give for change by " edge effect "
The data lost, use the mode of multi-cluster center traversal, can significantly reduce the data volume of loss.Example
As in the example of fig. 5, traversal C1 and C2 clustering cluster simultaneously, recall ratio is substantially higher in only traveling through C1
Clustering cluster.Travel through C1, C2, C3 simultaneously and then can obtain the Query Result identical with full storehouse traversal with C4.
Traveling through multiple neighbour's clustering cluster can bring bigger distance to calculate and data IO expense simultaneously, and this problem can
With by controlling NtopMode, the mode reducing the data volume that each clustering cluster is comprised overcomes.
Traversal neighbour's clustering cluster number c can be as the controllable parameter of search engine system.When hardware performance carries
When height or system loading are relatively low, can preferably inquire about effect by increasing c acquisition.When under hardware performance
When fall or system loading are higher, can be by reducing c, sacrificial section inquiry effect is to obtain preferably inquiry
Efficiency.
Step 9: each CID obtaining step 8, reads its corresponding characteristic from disk, calculates
These distances are entered by the distance of the image feature data that step 7 obtains each data internal with this clustering cluster
Row ascending order arranges, it is thus achieved that the image identification (Image ID, IID) of k the image that distance is minimum and correspondence
Distance value;
This step comprises substantial amounts of distance and calculates and digital independent.In step 3.1, we enter amount of calculation of adjusting the distance
Go detail discussion.Here the hard disk IO expense that our more consideration digital independent cause.
If the data volume that average each clustering cluster comprises is D (Byte).The average seek time of hard disk is tf,
Data reading speed is sread.The data of each clustering cluster are continuous distribution on hard disk.So single is looked into
Ask the average data IO time be:
For an application example, c=8, D=1MB.We calculate the digital independent under the conditions of different hardware
Time-consumingly.
If data deposit in common SAS (Serial Attached SCSI) hard disk, tfRepresentative value be 3ms,
sreadRepresentative value be 150MB/s.So digital independent of single inquiry is time-consumingly: 3*8+8/150*1000=
75ms.And for for solid state hard disc (Solid State Disk, SSD), tfRepresentative value be 0.1ms, sread
Representative value be 500MB/s, single inquiry digital independent be time-consumingly 8*0.1+8/500*1000=17ms.
For structure used in the present invention, owing to using multi-cluster bunch retrieval, hard disk tracking can be increased
Number of times, now the low tracking time response of SSD can be substantially reduced the data IO time.As can be seen here, use
SSD replaces SAS hard disk, can effectively save the data IO time, reduces the average response time of engine.
Step 10: c the IID sequence obtained in merger step 9, obtain k wherein minimum IID and
Corresponding distance value, returns as result.
The image clustering method Query Result that the present invention optimizes is higher with the result registration that full storehouse travels through;Inquiry
The most shorter, more equalize;By simple mode, inquiry effect and performance can be weighed.
Claims (2)
1. an equilibrium image clustering method based on hierarchical cluster, it is characterised in that comprise the steps of:
(1) when setting up index, first image feature data is carried out initial clustering;
(2) each clustering cluster obtaining step (1) carries out clustering slicing operation;Concretely comprise the following steps: checking should
The image number that cluster is comprised;If the image number that this cluster centre comprises is more than upper limit N arrangedtop,
Then carry out two segregation classes at intra-cluster;If the data volume that the result of two segregation classes comprises is still above Ntop, then
This process of result iteration to two segregation classes;By data volume less than NtopClustering cluster central record to cluster in
In heart file;Afterwards all for this classification image feature datas are organized according to the cluster centre obtained;
(3) when retrieval, the characteristic to query image, calculate its all cluster centres arriving affiliated classification
Distance, and these distances are carried out ascending sort, obtain front c the clustering cluster mark that distance is minimum, c value
Specified by systematic parameter;Carry out data traversal in the inside of c clustering cluster afterwards, obtain last inquiry knot
Really.
Clustering method the most according to claim 1, it is characterised in that described cluster divides 2 steps to carry out,
And two steps use identical characteristics of image and range formula.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201210545637.XA CN103049514B (en) | 2012-12-14 | 2012-12-14 | A kind of equilibrium image clustering method based on hierarchical cluster |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201210545637.XA CN103049514B (en) | 2012-12-14 | 2012-12-14 | A kind of equilibrium image clustering method based on hierarchical cluster |
Publications (2)
Publication Number | Publication Date |
---|---|
CN103049514A CN103049514A (en) | 2013-04-17 |
CN103049514B true CN103049514B (en) | 2016-08-10 |
Family
ID=48062155
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201210545637.XA Expired - Fee Related CN103049514B (en) | 2012-12-14 | 2012-12-14 | A kind of equilibrium image clustering method based on hierarchical cluster |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN103049514B (en) |
Families Citing this family (14)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN103744934A (en) * | 2013-12-30 | 2014-04-23 | 南京大学 | Distributed index method based on LSH (Locality Sensitive Hashing) |
CN107636678B (en) * | 2015-06-29 | 2021-12-14 | 北京市商汤科技开发有限公司 | Method and apparatus for predicting attributes of image samples |
CN106874268A (en) * | 2015-12-10 | 2017-06-20 | 富士通株式会社 | Image search method and image retrieval apparatus |
CN106548196A (en) * | 2016-10-20 | 2017-03-29 | 中国科学院深圳先进技术研究院 | A kind of random forest sampling approach and device for non-equilibrium data |
CN106851437A (en) * | 2017-01-17 | 2017-06-13 | 南通同洲电子有限责任公司 | A kind of method for extracting video frequency abstract |
CN108171252A (en) * | 2017-11-16 | 2018-06-15 | 柳州健鱼科技有限公司 | A kind of balanced image clustering method based on hierarchical cluster |
CN108304849A (en) * | 2018-01-15 | 2018-07-20 | 浙江理工大学 | A kind of bird plumage color character extracting method |
CN110874417B (en) | 2018-09-04 | 2024-04-16 | 华为技术有限公司 | Data retrieval method and device |
CN110909197A (en) * | 2019-11-04 | 2020-03-24 | 深圳力维智联技术有限公司 | High-dimensional feature processing method and device |
CN110861089B (en) * | 2019-11-29 | 2020-11-06 | 北京理工大学 | Task balanced distribution cooperative work control method for multi-robot system |
CN113297331B (en) * | 2020-09-27 | 2022-09-09 | 阿里云计算有限公司 | Data storage method and device and data query method and device |
CN113095397A (en) * | 2021-04-03 | 2021-07-09 | 国家计算机网络与信息安全管理中心 | Image data compression method based on hierarchical clustering method |
CN113743533B (en) * | 2021-09-17 | 2023-08-01 | 重庆紫光华山智安科技有限公司 | Picture clustering method and device and storage medium |
CN113792172B (en) * | 2021-11-15 | 2022-02-11 | 西安热工研究院有限公司 | Image retrieval method, system, device and storage medium based on triangle inequality |
Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN101571855A (en) * | 2008-04-30 | 2009-11-04 | 福特安(苏州)图像管理有限公司 | Image searching and classifying method |
CN102800120A (en) * | 2012-06-15 | 2012-11-28 | 北京理工大学 | Emergency disaster situation display system and method based on multiple intelligent bodies |
Family Cites Families (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5983224A (en) * | 1997-10-31 | 1999-11-09 | Hitachi America, Ltd. | Method and apparatus for reducing the computational requirements of K-means data clustering |
WO2009060722A1 (en) * | 2007-11-06 | 2009-05-14 | National University Corporation Hokkaido University | Similar image retrieving device |
US8488873B2 (en) * | 2009-10-07 | 2013-07-16 | Apple Inc. | Method of computing global-to-local metrics for recognition |
-
2012
- 2012-12-14 CN CN201210545637.XA patent/CN103049514B/en not_active Expired - Fee Related
Patent Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN101571855A (en) * | 2008-04-30 | 2009-11-04 | 福特安(苏州)图像管理有限公司 | Image searching and classifying method |
CN102800120A (en) * | 2012-06-15 | 2012-11-28 | 北京理工大学 | Emergency disaster situation display system and method based on multiple intelligent bodies |
Also Published As
Publication number | Publication date |
---|---|
CN103049514A (en) | 2013-04-17 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN103049514B (en) | A kind of equilibrium image clustering method based on hierarchical cluster | |
Arora et al. | Hd-index: Pushing the scalability-accuracy boundary for approximate knn search in high-dimensional spaces | |
CN104199827B (en) | The high dimensional indexing method of large scale multimedia data based on local sensitivity Hash | |
Scott et al. | Entropy-balanced bitmap tree for shape-based object retrieval from large-scale satellite imagery databases | |
CN102521366B (en) | Image retrieval method integrating classification with hash partitioning and image retrieval system utilizing same | |
Schindler et al. | City-scale location recognition | |
Bailo et al. | Efficient adaptive non-maximal suppression algorithms for homogeneous spatial keypoint distribution | |
Shashank et al. | Private content based image retrieval | |
CN107220285B (en) | Space-time index construction method for massive trajectory point data | |
CN110070121B (en) | Rapid approximate K nearest neighbor method based on tree strategy and balanced K mean clustering | |
CN107256262A (en) | A kind of image search method based on object detection | |
CN104834693A (en) | Depth-search-based visual image searching method and system thereof | |
WO2013129580A1 (en) | Approximate nearest neighbor search device, approximate nearest neighbor search method, and program | |
CN103116610A (en) | Vector space big data storage method based on HBase | |
Cha et al. | The GC-tree: a high-dimensional index structure for similarity search in image databases | |
CN108388902B (en) | Composite 3D descriptor construction method combining global framework point and local SHOT characteristics | |
CN104036012A (en) | Dictionary learning method, visual word bag characteristic extracting method and retrieval system | |
CN102436491A (en) | System and method used for searching huge amount of pictures and based on BigBase | |
CN106933511B (en) | Space data storage organization method and system considering load balance and disk efficiency | |
CN112395288B (en) | R-tree index merging and updating method, device and medium based on Hilbert curve | |
CN103207889A (en) | Method for retrieving massive face images based on Hadoop | |
Liu et al. | TOP-SIFT: the selected SIFT descriptor based on dictionary learning | |
CN101266607A (en) | High dimension data index method based on maximum clearance space mappings | |
CN110059148A (en) | The accurate searching method that spatial key applied to electronic map is inquired | |
Yuan et al. | A novel index structure for large scale image descriptor search |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
C06 | Publication | ||
PB01 | Publication | ||
C10 | Entry into substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
C14 | Grant of patent or utility model | ||
GR01 | Patent grant | ||
CF01 | Termination of patent right due to non-payment of annual fee |
Granted publication date: 20160810 Termination date: 20201214 |
|
CF01 | Termination of patent right due to non-payment of annual fee |