CN107798093B - Image retrieval method - Google Patents

Image retrieval method Download PDF

Info

Publication number
CN107798093B
CN107798093B CN201711006969.XA CN201711006969A CN107798093B CN 107798093 B CN107798093 B CN 107798093B CN 201711006969 A CN201711006969 A CN 201711006969A CN 107798093 B CN107798093 B CN 107798093B
Authority
CN
China
Prior art keywords
node
content
image data
image
data
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201711006969.XA
Other languages
Chinese (zh)
Other versions
CN107798093A (en
Inventor
方引
杨洋
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Chengdu Jinzhi Zhiyuan Technology Co ltd
Original Assignee
Chengdu Jinzhi Zhiyuan Technology Co ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Chengdu Jinzhi Zhiyuan Technology Co ltd filed Critical Chengdu Jinzhi Zhiyuan Technology Co ltd
Priority to CN201711006969.XA priority Critical patent/CN107798093B/en
Publication of CN107798093A publication Critical patent/CN107798093A/en
Application granted granted Critical
Publication of CN107798093B publication Critical patent/CN107798093B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/50Information retrieval; Database structures therefor; File system structures therefor of still image data
    • G06F16/51Indexing; Data structures therefor; Storage structures
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/50Information retrieval; Database structures therefor; File system structures therefor of still image data
    • G06F16/58Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
    • G06F16/583Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/213Feature extraction, e.g. by transforming the feature space; Summarisation; Mappings, e.g. subspace methods
    • G06F18/2135Feature extraction, e.g. by transforming the feature space; Summarisation; Mappings, e.g. subspace methods based on approximation criteria, e.g. principal component analysis
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • G06F18/2411Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches based on the proximity to a decision surface, e.g. support vector machines
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/243Classification techniques relating to the number of classes
    • G06F18/24323Tree-organised classifiers
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/46Descriptors for shape, contour or point-related descriptors, e.g. scale invariant feature transform [SIFT] or bags of words [BoW]; Salient regional features
    • G06V10/462Salient features, e.g. scale invariant feature transforms [SIFT]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/48Extraction of image or video features by mapping characteristic values of the pattern into a parameter space, e.g. Hough transformation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/168Feature extraction; Face representation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/172Classification, e.g. identification
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/46Descriptors for shape, contour or point-related descriptors, e.g. scale invariant feature transform [SIFT] or bags of words [BoW]; Salient regional features
    • G06V10/467Encoded features or binary features, e.g. local binary patterns [LBP]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/513Sparse representations

Abstract

The invention provides an image retrieval method, which comprises the following steps: establishing association between the actual data block of the face image and the storage position, and establishing association between the retrieval information and the actual storage data block; the data owner of the face image retrieves in the index through request information sent by the client, and then feeds back the related retrieval result to the client. The invention provides an image retrieval method, which is beneficial to improving the accuracy of face recognition under the conditions of face shielding, low sample quantity and quality and information loss and reducing the recognition running time.

Description

Image retrieval method
Technical Field
The invention relates to cloud computing, in particular to an image retrieval method.
Background
With the continuous development of society and the continuous progress of science and technology, the research on human face information processing has become one of the current research hotspots. The research content of face recognition relates to the fields of pattern recognition, computer image processing, machine learning, artificial intelligence and the like, and the face recognition system is widely applied to the business fields of biological feature recognition, man-machine interaction, content retrieval, video monitoring, security systems and the like. Although many face recognition algorithms have achieved better recognition performance, the face recognition system still faces many challenges in practical applications, including: the problem of face identification by shading caused by illumination change, ornaments and the like; the number of samples which can be collected under the uncontrollable condition is small, and the human face information is lost due to posture change and the like. How to overcome the above problems on the basis of the existing method to further improve the accuracy of face recognition, and simultaneously reduce the operation time of recognition to improve the effectiveness is a problem to be solved urgently at present.
Disclosure of Invention
In order to solve the problems in the prior art, the invention provides an image retrieval method, which comprises the following steps:
establishing association between the actual data block of the face image and the storage position, and establishing association between the retrieval information and the actual storage data block;
the data owner of the face image retrieves in the index through request information sent by the client, and then feeds back the related retrieval result to the client.
Preferably, address segmentation retrieval is used to divide the data block storage address.
Preferably, the storage format of the face image divides the address information of the storage data block into 3 segments according to the image data visitor/image data owner/content.
Preferably, the data and the index of the image contents are stored as one unit, and a path of each content is composed of a HASH value of an image content name and an operation time.
Preferably, the process of the client node issuing the request content includes:
(1) the image data owner provides a content query request to the file system, the request is sent to a local Agent, and after the request is analyzed, whether the request is to be forwarded to the cloud end is determined according to whether the request can be cached or not;
(2) the local Agent service inquires whether the requested content exists in the local storage service, and if so, the operation goes to (12);
(3) if the requested content is not found in the local storage service, forwarding the request to a content query service;
(4) the content query service checks whether a path address similar to the request content query exists in the image data visitor node list;
(5) if the image data visitor node list has an address similar to the request content query, directly transferring to the corresponding image data owner node;
(6) if the image data visitor node list does not have the query similar to the request content query, the content query service sends the query to a lower-layer data storage and retrieval system module to query the image data owner node of the content query;
(7) the data storage retrieval system finds out the corresponding image data owner node by searching;
(8) the image data visitor node, the image data owner node or the image content node inquires corresponding index tables of the image data visitor node, the image data owner node or the image content node, whether corresponding indexes for inquiring the request content exist or not is confirmed, and if the index for inquiring the request content is not found in the image data visitor node at the moment, the next image data visitor node or the image content node is continuously inquired through the data storage and retrieval system;
(9) the image data visitor node, the image data owner node or the image content node returns the inquired index information to the content inquiry service of the node which initiates the inquiry request; at this time, if the returned result contains the image data visitor node which is requested to be inquired, the content inquiry service updates the image data visitor node list;
(10) forwarding the return information to a local Agent service;
(11) and if the returned data is empty, the local Agent service acquires the content through the cloud end, otherwise, the content is acquired from the corresponding node according to the returned result. Then storing the backup of the content in the local storage service, and issuing corresponding index information to corresponding image data visitor nodes and image data owner nodes;
(12) the content is sent to a file system.
Compared with the prior art, the invention has the following advantages:
the invention provides an image retrieval method, which is beneficial to improving the accuracy of face recognition under the conditions of face shielding, low sample quantity and quality and information loss and reducing the recognition running time.
Drawings
Fig. 1 is a flowchart of an image retrieval method according to an embodiment of the present invention.
Detailed Description
A detailed description of one or more embodiments of the invention is provided below along with accompanying figures that illustrate the principles of the invention. The invention is described in connection with such embodiments, but the invention is not limited to any embodiment. The scope of the invention is limited only by the claims and the invention encompasses numerous alternatives, modifications and equivalents. In the following description, numerous specific details are set forth in order to provide a thorough understanding of the present invention. These details are provided for the purpose of example and the invention may be practiced according to the claims without some or all of these specific details.
One aspect of the present invention provides an image retrieval method. Fig. 1 is a flowchart of an image retrieval method according to an embodiment of the present invention.
The mass data storage and retrieval system based on cloud computing comprises data nodes and index nodes. The index nodes maintain indexes of the face image data blocks, mapping relations among the data blocks and data block attributes, and the data nodes store actual face image data blocks by taking different image data owners as units. When the image data owner accesses the storage system, an independent space is obtained. Each data block is assigned a different ChunkID, and each image block and copy is stored on each data node. The index of the face image data block includes the following attributes: ChunkID, name, type, size, image data owner name, access time, and location information thereof.
The user block records ChunkID, sharing mode, image data block name corresponding relation and image data owner name of the image data owner stored in the system. The system acquires the user block by accessing the ChunkID mapped by OwnerID of the owner of the image data and gives the owner of the image data an independent user block space. The user block of the owner of the image data only allows the owner of the image data corresponding to OwnerID to have the right to access all the user blocks, assigns a ChunkID to the user blocks, and stores the user blocks in each data node of the system after the user blocks.
And the data storage and retrieval system is also internally provided with super nodes, the super nodes are provided with high-speed bandwidth and high-performance nodes, each super node maintains a routing table, and the routing table can be adjusted according to the strength of the self capacity. All supernodes in the data storage retrieval system form a storage ring. The routing table of the super node n comprises m rows, each row comprises x items, x represents the strength of the node performance, the kth row comprises x items, x is equally divided into intervals [ n +3i, n +3i +1 ], wherein k is more than 0 and less than m, and each super node dynamically adjusts the x value according to the current capability of the super node. The storage ring is used for routing query requests. Each supernode is responsible for maintaining its predecessor and successor nodes. The successor node of the supernode n is the immediate successor node of the node n on the memory ring, i.e. the first supernode in the clockwise direction from n on the memory ring, and likewise, the predecessor node of n is the immediate predecessor supernode of n on the memory ring. Each super node also maintains a data node list, each item in the routing table of the super node points to one super node, all data nodes from the super node to a subsequent node are recorded in the data node list, and the node pointed in the data node list can be used for backing up data information on the super node or transferring overflowed image data to the node pointed in the data node list when image storage data of the super node overflows. All requested information on the data storage retrieval system is routed through the memory ring. Wherein a query request of a data node is first forwarded to its successor supernode, and then routing can be performed on the storage ring, eventually reaching the destination. The query requests of the supernodes are routed directly on the storage ring.
The indexing mechanism of the face image data block comprises the establishment of an index and block retrieval. The index establishment is used for establishing association between the actual data block and the storage position, establishing association between the retrieval information and the actual storage data block, and storing information of necessary data blocks. The block retrieval comprises retrieving request information sent by an image data owner through a client in an index, and then feeding back a related retrieval result to the client. The block search involves the storage format of the image data block itself, the node composition structure and the block search mode. The invention divides the data block storage address by adopting address segmentation retrieval.
The face image storage format divides address information of a storage data block into 3 segments according to an image data visitor/image data owner/content. The database of the image data visitor records a list of all image data owners owned by the image data visitor; the database of the image data owner maintains all the contents in the image data owner server; the image content service is responsible for storing, deleting and searching image content on the node device. The data of the image content and the index are stored together and saved as one unit. The path of each content is composed of the HASH value of the image content name and the operation time.
The method of the present invention is to divide the keyword identifier into 3 parts. The three parts are respectively an image data visitor k according to the sequence of the keyword identifier from high to low1Owner k of image data2Content of image k3ID of keyword of image content query value according to k1k2k3Are linked in sequence. The absolute value thereof is: i k1|,|k2|,|k3I is calculated according to the following sequence, firstly, the HASH value of the first content query value is calculated according to the sequence of the image content query feature, and | k is taken according to the sequence from high order to low order1| position as k1Then calculating the HASH value of the second content query value to take | k2| position as k2Then HASH is calculated for the rest of the content query value and k is taken3| position as k3The value of (c). Similar queries are arranged in close proximity.
The image data block is divided into s according to its key word when stored, and its ID is divided into s1/s2/s3Arranged on the routing memory ring from high to low 3 parts, s1Is a field of picture data visitor addresses, s2Is the image data owner address field, s3Is a picture content address field. By data block address segmentation, when image data owner determines numberAnd after the position of the image data visitor or the image data owner where the block is located is determined, the specific position of the image data block is quickly determined.
The invention records the node storing the visitorID as the image data visitor node for content query, and the node storing the ownerID as the image data owner node for content query. And finally, recording the node storing the content query ID as the image content node of the image content query. The index information of the content query may be stored in all of its image data visitor nodes, image data owner nodes, and image content nodes.
Each data node maintains local storage content, an image data visitor content index table, an image data owner content index table and a content index table, and simultaneously maintains an image data visitor node list, an image data owner node list and an image content node list. Each row in the image data visitor content index table comprises a content query address, the last access time of image content, a plurality of nodes where the image content corresponding to the content query is located and the like; similarly, the content index table of the image data owner and the content index table also contain the items, but the image content node of the content query in the content index table is at the node; the list of image content nodes includes image data content. The image data visitor node list maintains the image data owner node with the highest recent access frequency of the node; the image data owner node list maintains the image content node which has the highest frequency of being accessed recently; the image content node stores content specific content.
Each node operates a local storage server by a local proxy server; the local storage server shares the storage content with other nodes in the network through the data storage and retrieval system platform. And performing content query service on a data storage and retrieval system platform at the bottom layer, providing a retrieval information request and returning a retrieval result of the data storage and retrieval system to the local Agent. All routing is provided through the underlying data storage retrieval system. The content query comprises an image data visitor service, an image data owner service and a content service, and comprises query, insertion and deletion operations. The task of the content query service is to determine whether to forward to the image data owner node or send to the underlying routing mechanism based on the information in the image data visitor node list. And inserting the index information of the corresponding inquiry image data owner node into the corresponding node according to the image data owner node and the image content node information returned by the content inquiry service.
The data storage retrieval system carries the inquired address in the inquiry operation, if the path image content node list of the request node does not have the address similar to the inquiry of the corresponding request content, the inquiry information sent to the data storage retrieval system firstly inquires an image data visitor node, after the image data visitor node is reached, if the corresponding index item is not found, the ID of the image data owner node is calculated, and the inquiry is continuously carried out in the data storage retrieval system from the image data visitor node until the result is found or the image content node is reached. If the image data visitor node list of the request node has a path address similar to the corresponding request content query, the path address is directly forwarded to the corresponding image data owner node, and the query is started from the image data owner node.
According to the block retrieval mode, the block retrieval process can be divided into two steps, wherein firstly, the node issues request content information; and secondly, the data block retrieves the data block according to the request content information and feeds back the request.
The workflow of the node for issuing the request content comprises the following steps:
(1) the image data owner provides a content query request to the file system, the request is sent to a local Agent, and after the request is analyzed, whether the request is to be forwarded to the cloud end is determined according to whether the request can be cached or not;
(2) the local Agent service inquires whether the requested content exists in the local storage service, and if so, the operation goes to (12);
(3) if the requested content is not found in the local storage service, forwarding the request to a content query service;
(4) the content query service checks whether a path address similar to the request content query exists in the image data visitor node list;
(5) if the image data visitor node list has an address similar to the request content query, directly transferring to the corresponding image data owner node;
(6) if the image data visitor node list does not have the query similar to the request content query, the content query service sends the query to a lower-layer data storage and retrieval system module to query the image data owner node of the content query;
(7) the data storage retrieval system finds out the corresponding image data owner node by searching;
(8) the image data visitor node, the image data owner node or the image content node inquires corresponding index tables of the image data visitor node, the image data owner node or the image content node, whether corresponding indexes for inquiring the request content exist or not is confirmed, and if the index for inquiring the request content is not found in the image data visitor node at the moment, the next image data visitor node or the image content node is continuously inquired through the data storage and retrieval system;
(9) the image data visitor node, the image data owner node or the image content node returns the inquired index information to the content inquiry service of the node which initiates the inquiry request; at this time, if the returned result contains the image data visitor node which requests the query, the content query service updates the image data visitor node list.
(10) Forwarding the return information to a local Agent service;
(11) and if the returned data is empty, the local Agent service acquires the content through the cloud end, otherwise, the content is acquired from the corresponding node according to the returned result. Then storing the backup of the content in the local storage service, and issuing corresponding index information to corresponding image data visitor nodes and image data owner nodes;
(12) the content is sent to a file system.
And wherein the block retrieving the requested content and feeding back the information comprises:
and after the content is acquired, the path index information of the content is issued to the corresponding image data visitor node and the image content node. When a node n sends a content query request content, the node firstly checks whether the content needs to be queried and obtained through a lower layer routing mechanism, and if the content needs to be queried, the node checks whether an image data visitor node of the requested content query exists in an image data visitor node list. If the image data visitor node does not request the image data visitor node of the content in the image data visitor node list, at the moment, the visitorID of the content is calculated according to the content query address and sent to the subjacent data storage and retrieval system to query the viewer image content node of the content, through the query, the data storage and retrieval system finds the image data visitor node of the content, then the image data visitor node checks whether the index queried by the content exists in the image data visitor content index table, the image data owner content index table and the content index table, if the index exists, the path image content node returns the address of the node where the requested content is located to the node n, and the node n obtains the content by requesting the proxy node; if the returned proxy node does not have the requested content, the request is sent to the cloud server. Otherwise, if the index does not exist, computing the ownerID and sending the ownerID to the image content node of the viewer, continuing to query from the image data visitor node by the data storage and retrieval system, finding out the image data owner node through query, checking the index table information on the image data owner node, returning a piece of proxy node information to the local node if corresponding index information exists, otherwise continuing to query the next image data owner node, sequentially circulating the storage loop until the index information of the requested content query exists on a certain image data owner node or the image content node of the requested content query is found, and finishing the query. And if the index table of the image content node inquired by the content does not have the index of the requested content, the request is sent to the cloud server. During this query, each image data owner node will return its own address information to the local node n.
If the image data visitor node list has an image data owner node requesting content inquiry, the inquiry is directly forwarded to the image data owner node, the image data owner node checks the index table, if the index exists, a proxy node is returned, otherwise, the next image data owner node is continuously inquired until the image content node is found.
When a node obtains a content, it will issue information to the image data owner node and the image content node issuing the request, informing these nodes that it has the requested content of the content query, and the image data owner node and the image content node update their index tables accordingly. In the query process, if the image data owner node requesting content query is stored in the image data owner node list of the node n, the image data owner node directly forwards the image data owner node to start the query.
In the aspect of face image feature extraction, the invention uses the difference value of the pixel point of the center of the area and the pixel point of the storage ring neighborhood to represent the texture feature value of the pixel point, takes the neighborhood of the pixel point as a texture unit of the image, quantizes the texture unit through binary value to obtain the local texture feature value, obtains the texture feature vector of the description image by counting the texture unit in the image and carrying out normalization operation, and the detailed steps of using the method to carry out feature extraction are as follows:
first, the image is binary-coded. Randomly selecting a region from the collected face image, wherein any pixel point in the region can be described by G (y, z), and the geometric center point can be hcDescribing the neighborhood pixel point h in the 3 x 3 window0To h7The binary conversion processing is performed as follows:
hd=t(h0-hc),…t(h7-hc);
wherein
Figure BDA0001444612930000101
Weighting the binary conversion result to obtain a local binary pattern value at the center of the window:
Figure BDA0001444612930000102
q is set to describe K feature types, Q ∈ (0, 1, 2, …, K-1). Dividing the collected face image into n × p blocks, and counting the occurrence times of each mode in each block, that is, counting the feature types in the sub-region of each block of face image to obtain a face image feature component U ═ U (U) composed of n × p histograms1,U2,…Un×p). Wherein the content of the first and second substances,
Figure BDA0001444612930000103
molecule Pj(Q) the number of features used to describe that the local binary pattern value in the jth sub-region is Q,
Figure BDA0001444612930000104
a binary pattern histogram for describing the jth sub-region.
A face image feature histogram is built according to the method set forth above, thereby providing a data basis for face image retrieval.
In order to reduce the noise influence in the image processing process, the invention carries out denoising based on the median filtering human eye visual characteristics. First, a noise point is determined, the size of the image R is set to m × n, and a window of 3 × 3 is adopted to slide on the image.
Defining the gray value as the central pixel value of the window, and then the value set of all the pixel points in the window is:
wi,j={g(i+k,j+r)|k,r=(1,0,-1)}
calculating the average value of pixels in the window
Figure BDA0001444612930000111
Finding out the maximum gray value and the minimum gray value of the image R, and respectively marking as Imax(m×n)、Imin(m × n). The threshold value for marking the central pixel point is Hi,j
Then when the gray value of the central pixel point meets the following conditions, the pixel point can be judged as a noise point:
if | g (i, j) -wm|>Hi,jThen the pixel point is a noise point.
If | g (I, j) | ═ Imax(m.times.n) or Imin(m × n), the pixel is a noise point.
For the above conditions, the present invention determines the threshold H according to the noise sensitivity factor λi,jThe size of (2). Defining the noise sensitivity coefficient lambda of the central pixel point g (i, j) of the window as
Figure BDA0001444612930000112
At this time, whether the pixel point is a noise point is judged, and only the calculated noise sensitivity coefficient lambda is neededi,jIf | g (i, j) -wm|>λi,jThe condition is satisfied.
After dividing image pixel points into two types of noise points and non-noise points, smoothing the image g (i, j) by using a NURBS function, wherein the image can be regarded as uniform sampling of a curved surface, is a result of discrete convolution of an original image and k-times spline and l-times spline function, and is described as follows:
Figure BDA0001444612930000113
wherein, Bk(x-i) and BlAnd (y-i) are k-order splines and l-order spline convolution templates of NURBS respectively, if g (i, j) is a noise point, a 3 multiplied by 3 filtering window is taken, filtered values are obtained, and then median filtering is carried out, so that a final value is obtained.
If i takes [0,255], the resolution function is defined as follows:
Fr(i)=N(i)/max[N(i)]
the membership function defining the target area is:
Figure BDA0001444612930000121
wherein f (i) is a monotonically increasing function, and satisfies the conditions f (a) 0 and f (b) 1. When the gray value is in the interval of [0, a ], the pixel point belongs to a background area, when the gray value is in the interval of [ b, 255], the pixel point belongs to a target area, and when the gray value is in the interval of [ a, b ], the pixel point needs to be further represented by a fuzzy function.
The features of the face image are extracted according to the method, and a human eye visual perception model is established according to the features, so that face image retrieval is realized.
The method adopts a feature operator to take the peak value of a pixel gradient direction histogram in the neighborhood of a feature point as the main direction of the feature point, and rotates a coordinate axis as the main direction of the feature point. Calculating two vector histograms Hi(x) And Hj(x) Similarity of (2):
Figure BDA0001444612930000122
wherein, | | HiI and HjAnd | | represents the length of the histogram feature vector.
And then detecting abnormal characteristic point pairs by combining the scale direction, and finally discarding the abnormal pairs by using random sampling consistency. The whole process is to fit an image transformation matrix through the sample data set. Initial sample data n ═ min { n ═ n0,max{ns,nslog2μn0}}。n0Is the number of matched feature points, n, determined according to a K-nearest neighbor algorithmsTo discard the number of matching feature points before the outlier feature point pair, μ is the tuning parameter. Of the original image (x)1,y1) With the target image (x)2,y2) The transformation relationship is as follows:
Figure BDA0001444612930000131
Figure BDA0001444612930000132
is a transformation matrix of 8 parameters, and at least four characteristic feature point pairs are required for obtaining the parameters of the matrixSolving the matrix parameters by weighted least square method
K=[k1k2k3k4k5k6k7k8]
L=-[x2y2]T
Figure BDA0001444612930000133
The transformation is then:
K=-[GTG]-1GTL
and (3) firstly setting the initial value of mu to be 1 and obtaining the initial value of K, and then continuously carrying out iterative computation on mu to finally obtain stable K. The specific algorithm is as follows:
(1) randomly extracting matching characteristic point pairs of different planes, and calculating a transformation matrix K of the point pairs;
(2) for the matching point pair (x, y) to be detected, if the condition | K · x-y | < epsilon, epsilon is a tolerance value, the point is an interior point. If the number of the inner points is larger than the set threshold value t, recalculating the matrix K by an iterative weighted least square method, updating the number of the inner points, and if the number of the inner points is smaller than t, returning to the step (1);
(3) and if the maximum number of the inner point sets is determined and is greater than t after W iterations, calculating a transformation matrix K according to the combination of the inner points.
In a further aspect, the invention provides a condition that the threshold is satisfied, and the processing is carried out through the judgment condition to prevent the misjudgment of the noise point. Assuming that coordinates (x, y) and (x, y) represent feature operators of the source image and the target image, respectively, the feature operator of each feature point pair can be obtained in the following manner.
Δx=x-sm(x*·cos(Δθm)-y*·sin(Δθm))
Δy=y-sm(x*·cos(Δθm)-y*·sin(Δθm))
Where Δ x and Δ y represent histogram representations of the extracted features. Four-item(s)m,Δθm,Δxm,Δym) Transform near representing pairs of deleted outlier featuresSimilarly, the following conditions are satisfied:
|Δx-Δxm|>Δxt;|Δy-Δym|>Δyt
according to the width of the histogram, Δ xtAnd Δ ytRepresenting the threshold values for the horizontal and vertical differences of the histogram, respectively.
After the specific points are matched through the image transformation, the decision tree is applied to the positioning of the human face characteristic points. Firstly, the positioning of the face feature points is trained by adopting shape index pixel gray scale features. Two coordinate points are randomly sampled in a local coordinate system established by two reference points, a pixel gray difference value between the two points is made, then a random offset is added to the middle point of the two reference points to generate a characteristic point, and the pixel gray difference value of the two characteristic points is used as a characteristic.
In training the decision tree, the inputs of the tree are the face image I, the shape S composed of the coordinates of the corresponding reference points, and the true shape S' of the reference points, and the output is the predicted offset Δ S of the reference points. Training of the decision tree first determines the split to non-leaf nodes in the tree. And I (p, delta x, delta y) represents the gray value of a (delta x, delta y) pixel point in a local coordinate system established by taking the p-th reference point as an origin in the obtained picture after the shape of the current reference point is subjected to similarity transformation. Setting a segmentation threshold for a current node
Figure BDA0001444612930000141
The value range is [ -255, 255 [)]. In a local coordinate system established by taking the reference point p as an origin, the difference of the gray value characteristics of the shape index pixels of two points and a threshold value are taken
Figure BDA0001444612930000143
Comparing if it is less than the threshold
Figure BDA0001444612930000142
The training samples are divided into left child nodes, otherwise, the training samples are divided into right child nodes.
Optimal characteristic function f0And an optimum thresholdValue of
Figure BDA0001444612930000144
The selection of (c) can be described by the following formula:
Figure BDA0001444612930000145
wherein Δ SLIs composed of
Figure BDA0001444612930000146
The Δ S portion of time; delta SRIs composed of
Figure BDA0001444612930000147
The Δ S portion of time; f (I) ═ I (p, Δ x)1,Δy1)-I(p,Δx2,Δy2);
Var(ΔSL) Variance value, Var (Δ S), representing the offset of the corresponding p-th reference point in the left child nodeR) Representing the variance value of the offset of the corresponding pth reference point in the right child node.
For each non-leaf node, selecting a feature function f to extract shape index features of all samples corresponding to the node, and then selecting a threshold
Figure BDA0001444612930000151
The shape index features are divided, and the training sample (I, S, S') of the current node is divided into a left child node part and a right child node part (I)L,SL,S’L) And (I)R,SR,S’R)。
Each internal node of the decision tree is trained in the manner described above, such that the decision trees trained for each keypoint are combined into a decision forest. And outputting information of samples contained in leaf nodes in the decision tree to be represented as a binary feature vector, and connecting the binary features of all the decision trees in the decision forest back and forth to form a one-dimensional feature vector. The corresponding feature mapping of a face image in the t-th decision forest is represented by the following formula:
δt={δt i}i=1,…L
wherein t represents the number of layers where the decision tree is located, and L represents the number of reference points in the face shape. Deltat iThe feature vector is formed by connecting binary features extracted from all decision trees corresponding to the ith reference point in series and is called as local binary feature. By extracting delta corresponding to each reference point in the facet iAfter characterization, all δ are addedt iAre connected in series into a final binary feature vector to represent the feature mapping relation delta of the human facet
Randomly extracting images from all face images as a training sample set, and taking the rest images as a test sample set; and respectively extracting SIFT features and DCT features from all the training images, wherein the SIFT features comprise SIFT phase features and SIFT amplitude features.
The face image vector is mapped into a high-dimensional feature space F through a nonlinear function phi, and then principal component analysis transformation is carried out in the high-dimensional feature space F. When principal component analysis transformation is carried out, a nonlinear function E meeting kernel conditions is introduced to replace inner product operation of vectors, namely E (x)i,xj)=Φ(xi)·Φ(xj). The process of principal component analysis is:
training sample face vector x with m dimensions1,x2,...,xtMapping to a high-dimensional feature space F by using a nonlinear function phi to obtain phi (x)1),Φ(x2),...,Φ(xt);
In F for phi (x)i) And (6) carrying out transformation. Solving a characteristic equation l lambdaΦα ═ K α, where K ═ E (x)i,xj))l×iSo as to obtain a feature vector as:
Figure BDA0001444612930000161
corresponding characteristic value is lambdaΦ 1,λΦ 2,…λΦ l(ii) a Taking the first m features in the feature valueThe values and corresponding eigenvectors are used to obtain an eigen matrix MΦ=(DΦ)1/2(VΦ)TWherein:
DΦ=diag(λΦ 1,λΦ 2,…λΦ m)
VΦ=(v1,v2,...,vm)
so the training samples are transformed in space F to:
Figure BDA0001444612930000162
determining a corresponding separation matrix WΦ
Mapping any test sample y to a space F to form phi (y), and extracting a feature vector of the test sample y
Figure BDA0001444612930000163
After the principal component analysis process is completed, obtaining a kernel independent feature vector and a feature subspace; performing feature fusion on the kernel independent features to obtain a one-dimensional feature vector, and finally obtaining all feature vectors of the training sample set; training an SVM model by using the obtained feature vector;
after obtaining the feature vectors of all the test sample sets by adopting the same method, projecting the feature vectors of the test sample sets to subspaces of the test sample sets respectively to obtain kernel independent feature vectors of the test sample sets;
and (4) using the kernel independent feature vector in a trained SVM model for classification test to obtain a primary recognition result of the face image.
The invention preferably further divides the training sample into overlapped blocks, respectively calculates the discrimination rate of each block, then selects the block construction template with higher discrimination rate, filters the training sample, constructs a new dictionary from the filtered training sample, and finally classifies by sparse representation.
Given an inclusionSet of n samples of C classes a ═ a1*,A2*,…An*]。AiDenotes the ith image matrix. Each training image is divided into k overlapped blocks, and the block matrix of each image is converted into a vector, namely Ai*=[ai,1,ai,2,…ai,k]. Denote the whole training dictionary set a as a ═ a1,A2,…An]Wherein A isiThe ith module vector representing all images.
For each module set AiBy using
Figure BDA0001444612930000175
Which represents the corresponding mean vector of the image,
Figure BDA0001444612930000171
ith module vector a representing all images in class cc,iC is [1, C ]]. Then module AiThe discrimination ratios of (1) are as follows:
Figure BDA0001444612930000172
and sorting the module discrimination rates from high to low, and only keeping the first h modules to construct a template T. The template is used to filter test and training sample images. Filtered training set fA ═ fA1,fa2,...,fah]In which fa isiIs the filtered image AiThe vector of x represents, and h represents the number of templates included in the template.
To further reduce the amount of computation, principal components are extracted by principal component analysis on fA, and a projection matrix P is constructed, then the dimensions of the training image and the test sample y can be further reduced to:
fpA=P'fA
fpy=P'fy
fpy may be represented as fpLinear combination of A
fpy=fpA·X
And X is a sparse matrix, and the test sample is classified into a class corresponding to the minimum reconstruction residual according to the class residual:
Figure BDA0001444612930000173
wherein
Figure BDA0001444612930000174
Is a selection function. | () | laces2Is 12And (5) norm constraint.
In summary, the invention provides an image retrieval method, which is helpful for improving the accuracy of face recognition under the conditions of face occlusion, low sample quantity and quality and information loss, and simultaneously reducing the operation time of the recognition.
It will be apparent to those skilled in the art that the modules or steps of the present invention described above may be implemented in a general purpose computing system, centralized on a single computing system, or distributed across a network of computing systems, and optionally implemented in program code that is executable by the computing system, such that the program code is stored in a storage system and executed by the computing system. Thus, the present invention is not limited to any specific combination of hardware and software.
It is to be understood that the above-described embodiments of the present invention are merely illustrative of or explaining the principles of the invention and are not to be construed as limiting the invention. Therefore, any modification, equivalent replacement, improvement and the like made without departing from the spirit and scope of the present invention should be included in the protection scope of the present invention. Further, it is intended that the appended claims cover all such variations and modifications as fall within the scope and boundaries of the appended claims or the equivalents of such scope and boundaries.

Claims (1)

1. An image retrieval method, comprising:
establishing association between the actual data block of the face image and the storage position, and establishing association between the retrieval information and the actual storage data block;
searching in the index through request information sent by a data owner of the face image through the client, and then feeding back a related search result to the client;
the method also comprises the steps of dividing the storage address of the data block by adopting address segmentation retrieval;
the storage format of the face image divides the address information of the storage data block into 3 sections according to the image data visitor/image data owner/content;
storing the data and index of the image content as a unit, wherein the path of each content is composed of the HASH value and the operation time of the image content name;
the process of the client node issuing the request content comprises the following steps:
(1) the image data owner provides a content query request to the file system, the request is sent to a local Agent, and after the request is analyzed, whether the request can be cached or not is determined to be forwarded to the cloud end;
(2) the local Agent service inquires whether the requested content exists in the local storage service, and if so, the operation goes to (12);
(3) if the requested content is not found in the local storage service, forwarding the request to a content query service;
(4) the content query service checks whether a path address similar to the request content query exists in the image data visitor node list;
(5) if the image data visitor node list has an address similar to the request content query, directly transferring to the corresponding image data owner node;
(6) if the image data visitor node list does not have the query similar to the request content query, the content query service sends the query to a lower-layer data storage and retrieval system module to query the image data owner node of the content query;
(7) the data storage retrieval system finds out the corresponding image data owner node by searching;
(8) the image data visitor node, the image data owner node or the image content node inquires corresponding index tables of the image data visitor node, confirms whether the index corresponding to the inquiry of the request content exists or not, and if the index of the inquiry of the request content is not found by the image data visitor node at the moment, continuously inquires the next image data visitor node or the image content node through the data storage and retrieval system;
(9) the image data visitor node, the image data owner node or the image content node returns the inquired index information to the content inquiry service of the node which initiates the inquiry request; at this time, if the returned result contains the image data visitor node which is requested to be inquired, the content inquiry service updates the image data visitor node list;
(10) forwarding the return information to a local Agent service;
(11) if the returned data is empty, the local Agent service acquires the content through the cloud end, otherwise, the content is acquired from the corresponding node according to the returned result; then storing the backup of the content in the local storage service, and issuing corresponding index information to corresponding image data visitor nodes and image data owner nodes;
(12) sending the content to a file system;
the data storage and retrieval system is also internally provided with super nodes, each super node maintains a routing table, and the routing table is adjusted according to the self capacity; forming a storage ring by all supernodes in the data storage and retrieval system; the routing table of the super node n comprises m rows, each row comprises x items, x represents the node performance, the kth row comprises x items, wherein k is more than 0 and less than m, and each super node dynamically adjusts the x value according to the current capability of the super node; the storage ring is used for routing the query request; each super node maintains its predecessor and successor nodes; the successor node of the supernode n is the immediate successor node of the node n on the storage ring, namely the first supernode in the clockwise direction from n on the storage ring, and the predecessor node of the supernode n is the immediate predecessor supernode of the node n on the storage ring; each super node also maintains a data node list, each item in the routing table of the super node points to one super node, all data nodes from the super node to a subsequent node are recorded in the data node list, and the node pointed in the data node list can be used for backing up data information on the super node or transferring overflowed image data to the node pointed in the data node list when image storage data of the super node overflows; all request information on the data storage and retrieval system is routed through the storage ring; the query request of the data node is firstly forwarded to a subsequent supernode of the data node, then routing is carried out on a storage ring, and finally the data node reaches a destination; the query requests of the supernodes are routed directly on the storage ring.
CN201711006969.XA 2017-10-25 2017-10-25 Image retrieval method Active CN107798093B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201711006969.XA CN107798093B (en) 2017-10-25 2017-10-25 Image retrieval method

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201711006969.XA CN107798093B (en) 2017-10-25 2017-10-25 Image retrieval method

Publications (2)

Publication Number Publication Date
CN107798093A CN107798093A (en) 2018-03-13
CN107798093B true CN107798093B (en) 2022-05-03

Family

ID=61533611

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201711006969.XA Active CN107798093B (en) 2017-10-25 2017-10-25 Image retrieval method

Country Status (1)

Country Link
CN (1) CN107798093B (en)

Families Citing this family (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110688510B (en) * 2018-06-20 2022-06-14 浙江宇视科技有限公司 Face background image acquisition method and system
CN111695419A (en) * 2020-04-30 2020-09-22 华为技术有限公司 Image data processing method and related device
CN113868440B (en) * 2020-06-30 2023-06-27 华为技术有限公司 Feature library management method, device, equipment and medium
CN113159211B (en) * 2021-04-30 2022-11-08 杭州好安供应链管理有限公司 Method, computing device and computer storage medium for similar image retrieval
CN115953584B (en) * 2023-01-30 2023-07-07 盐城工学院 End-to-end target detection method and system with learning sparsity

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102609434A (en) * 2010-12-20 2012-07-25 微软公司 Face recognition using social data
CN102982165A (en) * 2012-12-10 2013-03-20 南京大学 Large-scale human face image searching method
CN103164452A (en) * 2011-12-15 2013-06-19 深圳市腾讯计算机系统有限公司 Method, device and system of picture storage and reading
CN105159895A (en) * 2014-05-28 2015-12-16 国际商业机器公司 Method and system for storing and inquiring data
CN105243084A (en) * 2015-09-07 2016-01-13 广东欧珀移动通信有限公司 Photographed image file storage method and system and photographed image file search method and system

Family Cites Families (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7324711B2 (en) * 2004-02-26 2008-01-29 Xerox Corporation Method for automated image indexing and retrieval
US7769892B2 (en) * 2007-08-27 2010-08-03 International Business Machines Corporation System and method for handling indirect routing of information between supernodes of a multi-tiered full-graph interconnect architecture
CN101217565B (en) * 2008-01-04 2011-12-14 华中科技大学 A network organization method of classification retrieval in peer-to-peer network video sharing system
US20090198956A1 (en) * 2008-02-01 2009-08-06 Arimilli Lakshminarayana B System and Method for Data Processing Using a Low-Cost Two-Tier Full-Graph Interconnect Architecture
US20090274157A1 (en) * 2008-05-01 2009-11-05 Vaidya Aniruddha S Method and apparatus for hierarchical routing in multiprocessor mesh-based systems
CN101447937A (en) * 2009-02-27 2009-06-03 北京理工大学 Rapid data positioning method based on path division and multi-distributed-directory
CN103036964B (en) * 2012-12-04 2015-05-20 杭州顺网科技股份有限公司 Internet bar server data updating method based on peer-to-peer (P2P)
CN103544261B (en) * 2013-10-16 2016-06-22 国家计算机网络与信息安全管理中心 A kind of magnanimity structuring daily record data global index's management method and device
CN104301215A (en) * 2014-10-10 2015-01-21 北京邮电大学 Construction method of overlay network
CN104933154B (en) * 2015-06-25 2017-12-29 北京邮电大学 It is a kind of that method of the image by content quick-searching is realized based on distributed network
CN105976399A (en) * 2016-04-29 2016-09-28 北京航空航天大学 Moving object detection method based on SIFT (Scale Invariant Feature Transform) feature matching
CN106599311A (en) * 2016-12-29 2017-04-26 广州市奥威亚电子科技有限公司 Cloud computation-based internet education platform resource library image retrieval method
CN106875325B (en) * 2017-02-14 2020-05-12 四川理工学院 Searchable image encryption algorithm

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102609434A (en) * 2010-12-20 2012-07-25 微软公司 Face recognition using social data
CN103164452A (en) * 2011-12-15 2013-06-19 深圳市腾讯计算机系统有限公司 Method, device and system of picture storage and reading
CN102982165A (en) * 2012-12-10 2013-03-20 南京大学 Large-scale human face image searching method
CN105159895A (en) * 2014-05-28 2015-12-16 国际商业机器公司 Method and system for storing and inquiring data
CN105243084A (en) * 2015-09-07 2016-01-13 广东欧珀移动通信有限公司 Photographed image file storage method and system and photographed image file search method and system

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
Indexing and Retrieval of Images Using Wavelets-Segmentation Using Bisection K-Means Clustering;Vijayakalyari Natarajan el;《2011 First International Conference on Informatics and Computation Intelligence》;20120130;第3页 *
基于图像特征索引的并行检索技术研究;蔡晓东 等;《电视技术》;20150702;第2-3页 *

Also Published As

Publication number Publication date
CN107798093A (en) 2018-03-13

Similar Documents

Publication Publication Date Title
CN107798093B (en) Image retrieval method
US9424277B2 (en) Methods and apparatus for automated true object-based image analysis and retrieval
US8392484B2 (en) Shape based picture search
Hartmann et al. Recent developments in large-scale tie-point matching
CN104991959B (en) A kind of method and system of the same or similar image of information retrieval based on contents
CN107729855B (en) Mass data processing method
CN107679235B (en) Retrieval system based on cloud platform
Rashno et al. Content-based image retrieval system with most relevant features among wavelet and color features
EP1419458A2 (en) Modular intelligent multimedia analysis
WO2023221790A1 (en) Image encoder training method and apparatus, device, and medium
Yu et al. A content-based goods image recommendation system
Brust et al. Active and incremental learning with weak supervision
CN113313170A (en) Full-time global training big data platform based on artificial intelligence
US7751621B1 (en) Method and system for rapid object recall within images
CN115115855A (en) Training method, device, equipment and medium for image encoder
JP6173754B2 (en) Image search system, image search apparatus, and image search method
Srinivasa Rao et al. Content and context based image retrieval classification based on firefly-neural network
CN109190637A (en) A kind of image characteristic extracting method
CN110334290B (en) MF-Octree-based spatio-temporal data rapid retrieval method
JP5971722B2 (en) Method for determining transformation matrix of hash function, hash type approximate nearest neighbor search method using the hash function, apparatus and computer program thereof
Mounika et al. Content based video retrieval using histogram of gradients and frame fusion
CN107957865B (en) Neuron reconstruction result matching method
Shaikh et al. Contemporary integration of content based image retrieval
Lu et al. L 1-medial skeleton-based 3D point cloud model retrieval
Niemueller et al. Life-long learning perception using cloud database technology

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant