WO2017032243A1 - Image feature extraction method, apparatus, terminal device, and system - Google Patents

Image feature extraction method, apparatus, terminal device, and system Download PDF

Info

Publication number
WO2017032243A1
WO2017032243A1 PCT/CN2016/095524 CN2016095524W WO2017032243A1 WO 2017032243 A1 WO2017032243 A1 WO 2017032243A1 CN 2016095524 W CN2016095524 W CN 2016095524W WO 2017032243 A1 WO2017032243 A1 WO 2017032243A1
Authority
WO
WIPO (PCT)
Prior art keywords
image
structured
feature
training
sub
Prior art date
Application number
PCT/CN2016/095524
Other languages
French (fr)
Chinese (zh)
Inventor
刘荣
易东
张帆
张伦
楚汝峰
Original Assignee
阿里巴巴集团控股有限公司
刘荣
易东
张帆
张伦
楚汝峰
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 阿里巴巴集团控股有限公司, 刘荣, 易东, 张帆, 张伦, 楚汝峰 filed Critical 阿里巴巴集团控股有限公司
Publication of WO2017032243A1 publication Critical patent/WO2017032243A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/168Feature extraction; Face representation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/172Classification, e.g. identification

Definitions

  • the present application relates to the field of electronic technologies, and in particular, to an image feature extraction method, an image feature extraction device, an image feature extraction terminal device, and an image feature extraction system.
  • Face recognition research began in the 1990s. At the beginning, there were proposed eigen face methods for describing faces with image principal components and fisher face methods for describing face images with distinguishing features. After entering this century, based on LBP Gabor's face local feature description method and boosting-based distinguishing feature learning method have quickly become mainstream; in recent years, with the deep learning method proposed, face recognition technology has been pushed to a new Steps. At present, there are several leading edge technologies in the field of face recognition:
  • the first is the American facebook company, which introduced the deep learning method to face recognition for the first time. Using five convolutional layers and two deep-separated neural networks, it extracts 4096-dimensional visual features from the entire face image. Described, the recognition accuracy has been significantly improved.
  • the domestic face++ company also uses the deep learning method to learn a deep neural network by pyramid structure, and analyzes the entire face image, which also makes a breakthrough in face recognition technology.
  • the present application provides an image feature extraction method, an image feature extraction device, an image feature extraction terminal device, and an image feature extraction system.
  • the technical solution adopted in this application is:
  • the application provides an image feature extraction method, including:
  • the model obtained by the structured model training is used to calculate the structured feature data to obtain image feature data.
  • the constructing the multiple structured sub-images on the registered image comprises:
  • the registered image is cut according to the structured reference point position and the shape parameter of the sub-image to obtain a plurality of structured sub-images.
  • the determining the location of the structured reference point of the registered image comprises:
  • a structured reference point location of the registered image is determined based on the spatial location.
  • the mathematical algorithm for cutting the registered image according to the structured reference point position and the shape parameter of the sub-image to obtain a plurality of structured sub-images is:
  • a ij C(a,p ij (x,y),s ij )
  • a ij denotes a structured sub-image in which the structural order is in the i-th row and j-th in the vertical row
  • C is a constructing function of the structured sub-image
  • a represents the image input by the user
  • p ij represents the order in the horizontal row i, vertical jth structured reference points
  • p ij (x, y) indicates that the structured reference point p ij is at the coordinates (x, y) of the image input by the user
  • s ij represents the structuring Shape parameters of the image, including rectangular, circular, elliptical and other arbitrary planar shapes and their dimensions.
  • the feature model obtained by the multi-model training is obtained by the following method:
  • the plurality of structured sub-training images are subjected to feature model training by using a visual feature learning algorithm to extract corresponding plurality of sub-training image visual features, and a feature model is obtained.
  • the visual feature learning algorithm includes any one of the following:
  • the mathematical expression of the feature model is:
  • a ij denotes a sub-training image in which the structural order is located in the i-th row and the j-th row in the horizontal row
  • M ij is a feature model trained on the corresponding sub-training image a ij
  • q ij is a feature model parameter obtained by training
  • v ij is a sub-training image visual feature extracted by the feature model M ij for the sub-training image a ij .
  • the structurally merging the visual features of the plurality of structured sub-images to obtain the structured feature data includes:
  • the mathematical expression of the structured feature data is:
  • v ij represents the visual feature of the structured sub-image
  • k is the data of the k-th dimension
  • d is the structured feature data after the fusion.
  • the model obtained by the structured model training is obtained by:
  • the structured feature model is trained on the structured image data of the training image by using the visual feature learning algorithm, and the model obtained by the structured model training is obtained.
  • the mathematical expression of the model obtained by the structured model training is:
  • M is a model obtained by performing structured model training based on the fused training image feature data d
  • q is a model parameter obtained by training
  • v is a corresponding visual feature obtained by merging the training image feature data d by the model M.
  • the image feature extraction method further includes:
  • the comparing the image feature data with each predetermined image feature data in a predetermined image database including:
  • the output comparison results include:
  • each of the differences is greater than a predetermined similarity threshold, information having no similar image is output, otherwise, an image corresponding to predetermined image feature data having the smallest difference from the image feature data, and/or an image Information output.
  • the algorithm for calculating a difference between the image feature data and each predetermined image feature data in a predetermined image database includes any one of the following:
  • Euclidean distance calculation method Cosine distance calculation method or Joint Bayesian distance calculation method.
  • the image includes: a face image.
  • the application also provides an image feature extraction device, including:
  • An image receiving unit configured to receive an image input by a user
  • a registration unit configured to register an image input by the user to obtain a registered image
  • a sub-image construction unit configured to construct a plurality of structured sub-images on the registered image
  • a visual feature extraction unit configured to extract a visual feature of each of the structured sub-images by using a feature model obtained by multi-model training
  • a merging unit configured to structurally fuse the visual features of the plurality of structured sub-images to obtain structured feature data
  • the operation unit is configured to use the model obtained by the structural model training, and perform operation on the structured feature data to obtain image feature data.
  • the registration unit includes:
  • a reference point determining subunit for determining a structured reference point position of the registered image
  • a shape parameter determining subunit for determining a shape parameter of the sub image
  • a cutting subunit configured to cut the registered image according to the structured reference point position and the shape parameter of the sub image to obtain a plurality of structured sub-images.
  • the reference point determining subunit includes:
  • a feature reference point determining subunit configured to determine a structured reference point position of the registered image according to the image feature point
  • a spatial reference point determining subunit is configured to determine a structured reference point position of the registered image based on the spatial location.
  • the mathematical algorithm used by the cutting subunit is:
  • a ij C(a,p ij (x,y),s ij )
  • a ij denotes a structured sub-image in which the structural order is in the i-th row and j-th in the vertical row
  • C is a constructing function of the structured sub-image
  • a represents the image input by the user
  • p ij represents the order in the horizontal row i, vertical jth structured reference points
  • p ij (x, y) indicates that the structured reference point p ij is at the coordinates (x, y) of the image input by the user
  • s ij represents the structuring Shape parameters of the image, including rectangular, circular, elliptical and other arbitrary planar shapes and their dimensions.
  • the image feature extraction device further includes:
  • a multi-model training unit for obtaining a feature model by multi-model training
  • the multi-model training unit includes:
  • Training image library selection subunit for selecting a predetermined training image library
  • a training image registration sub-unit configured to register each training image in the predetermined training image library according to a unified registration method, to obtain a plurality of the registered training images
  • a sub-training image construction sub-unit configured to respectively construct a plurality of structured sub-training images for the plurality of registered training images
  • the feature model acquisition sub-unit is configured to perform feature model training on the plurality of structured sub-training images by using a visual feature learning algorithm to extract corresponding plurality of sub-training image visual features, and obtain a feature model.
  • the visual feature learning algorithm adopted by the feature model acquisition subunit includes any one of the following:
  • the fusion unit includes:
  • a reference point fusion subunit configured to structurally fuse the visual features of the plurality of structured sub-images according to the determined structured reference point position when constructing the plurality of structured sub-images, to obtain structured feature data
  • the structured feature data includes feature space relationships and feature information.
  • the image feature extraction device further includes:
  • a structured model training unit for training a model through a structured model
  • the structured model training unit includes:
  • a sub-training image fusion sub-unit configured to structurally fuse the plurality of sub-training image visual features to obtain training image structured feature data
  • the model acquisition subunit is configured to perform structural model training on the structured image data of the training image by using a visual feature learning algorithm, and obtain a model obtained by training the structured model.
  • the image feature extraction device further includes:
  • a comparison unit configured to sequentially compare the image feature data with each predetermined image feature data in a predetermined image database
  • An output unit for outputting the comparison result.
  • the comparison unit includes:
  • a difference calculation subunit configured to sequentially calculate a difference between the image feature data and each predetermined image feature data in a predetermined image database
  • the output unit includes:
  • a difference determining subunit configured to sequentially determine whether each of the difference values is greater than a predetermined difference threshold
  • An information output unit configured to: if each of the differences is greater than a predetermined similarity threshold, output information without a similar image; otherwise, an image corresponding to predetermined image feature data having a minimum difference from the image feature data , and / or image information output.
  • the algorithm for calculating, by the comparing unit, the difference between the image feature data and each predetermined image feature data in the predetermined image database includes any one of the following:
  • Euclidean distance calculation method Cosine distance calculation method or Joint Bayesian distance calculation method.
  • the application also provides an image feature extraction terminal device, including:
  • the image feature extraction method provided by the present application is stored in the memory; and can be run according to the above method after startup.
  • the present application also provides an image feature extraction system, including a client and a remote server.
  • the client captures an image and/or selects an image in the album.
  • the remote server extracts the image feature data, compares it with the image in the predetermined image database, and sends the comparison result to the client, and finally outputs the ratio by the client. For the result.
  • An image feature extraction method provided by the present application first receives an image input by a user; then, an image input by the user is registered to obtain a registered image; and then multiple structures are constructed on the registered image.
  • the feature model obtained by the multi-model training is used to extract the visual features of each of the structured sub-images; then the visual features of the plurality of structured sub-images are structurally fused to obtain the structural features.
  • Data; finally, the model obtained by the structured model training is used to calculate the structured feature data to obtain image feature data.
  • the spatial position information between the structured sub-images is preserved by constructing the structured sub-image, and thus the extracted visual features of the structured sub-image include features simultaneously.
  • the spatial relationship and feature information not only retain the descriptiveness of each visual feature, but also preserve the spatial relationship of each visual feature, so that the final image feature data is the feature vector, and the feature vector can be used.
  • the distance describes the difference between different images, and because the feature vector and the model in the method better maintain the structural characteristics of the image during the training process, the image feature data has higher accuracy and identifiability. .
  • Applying the image feature extraction method provided by the present application in image recognition, especially face recognition, has higher accuracy, thereby obtaining a better recognition effect.
  • FIG. 1 is a flowchart of an embodiment of an image feature extraction method provided by the present application.
  • FIG. 2 is a flow chart of constructing a plurality of structured sub-images in an embodiment of an image feature extraction method provided by the present application
  • FIG. 3 is a diagram showing an example of determining a structured reference point according to a spatial position relationship provided by the present application
  • FIG. 4 is an exemplary diagram of determining a structured reference point according to a face feature point provided by the present application
  • FIG. 5 is a flowchart of multi-model training in an embodiment of an image feature extraction method provided by the present application.
  • FIG. 6 is a schematic structural fusion diagram of the feature provided by the present application.
  • FIG. 7 is a schematic diagram of an embodiment of an image feature extraction device provided by the present application.
  • the present application provides an image feature extraction method, an image feature extraction device, an image feature extraction terminal device, and an image feature extraction system.
  • the embodiments of the present application are described in detail below with reference to the accompanying drawings.
  • FIG. 1 is a flowchart of an embodiment of an image feature extraction method provided by the present application.
  • the image feature extraction method includes the following steps:
  • Step S101 Receive an image input by the user.
  • the image input by the user is first received, and the user can select an image input from the electronic album of the terminal device, or take an image and input by the camera.
  • the purpose of the present application is image recognition, so that the image input by the user is prioritized as a static image, but in order to improve the general applicability of the method, in one embodiment of the present application, a dynamic image input by the user may be received.
  • pre-processing is performed to extract only a specific frame (such as the first frame) of the dynamic image as an image input by the user, and all of the above are within the protection scope of the present application.
  • the present image feature extraction method is used for face image recognition, and therefore, the image includes a face image.
  • Step S102 register the image input by the user to obtain a registered image.
  • step S101 the image input by the user has been received, and then the image input by the user needs to be registered.
  • the registration method commonly used in the prior art is to detect the image feature point first, and then perform image simulation according to the feature point. The transformation is performed, the image is normalized to a predetermined size and scale, and the registered avatar is obtained for identification and comparison.
  • the image feature extraction method is used for face image recognition, and the image is a face image.
  • the image is a face image.
  • Step S103 Construct a plurality of structured sub-images on the registered images.
  • step S103 the registered image is obtained by registering the image input by the user, and then, a plurality of structured sub-images are required to be constructed on the registered image.
  • FIG. 2 It is a flowchart of constructing a plurality of structured sub-images in an embodiment of an image feature extraction method provided by the present application, and the constructing a plurality of structured sub-images on the registered image may be performed by the following sub-steps:
  • Step S1031 Determine a structured reference point position of the registered image.
  • a plurality of structured sub-images are constructed, that is, a plurality of sub-images are segmented from the image according to a certain structure, position, and constraints.
  • the structured reference point position of the registered image is determined to determine the cutting position of the structured sub-image.
  • the structured reference point is used as a center point of the structured sub-image cutting.
  • the upper and lower left and right relations are basically kept unchanged.
  • the method for determining the structured reference point may be determined by determining a structured reference point position of the registered image according to a spatial position, or determining a structured reference point of the registered image according to the image feature point. position.
  • a set of 4 ⁇ 4 structured reference points is determined according to the spatial positional relationship, and the distance between them is completely fixed with respect to the image.
  • it determines 3 ⁇ 3 structured reference points according to the face feature points.
  • the nine structured reference points in the figure are from top to bottom and from left to right: right eye center point and two eyes. The center point, the left eye center point, the right cheek point, the nose point, the left cheek point, the right mouth corner point, the lip center point, and the left corner corner point.
  • the positional relationship of the nine structured reference points is different for different people, gestures, and expressions. A slight change occurs, but the approximate rectangular structure relationship is also satisfied.
  • the method of determining the structured reference point can be selected according to the main content of the image, and the number of structured reference points is not limited to the above 4 ⁇ 4 sum.
  • the case of 3 ⁇ 3 can be flexibly determined according to the actual situation, and will not be further described herein, and all of them are within the protection scope of the present application.
  • Step S1032 Determine a shape parameter of the sub image.
  • step S1031 the structured reference point position of the registered image has been determined, and then, the shape parameter of the sub-image needs to be determined, that is, the reference position of the structured reference point is used as a reference, and a certain ratio is
  • the size determines a sub-image area, the shape parameter including a shape of the sub-image, such as an arbitrary planar shape such as a rectangle, a circle, an ellipse, and the size of the sub-image, such as a length and a width of the rectangle, and a radius of the circle Wait.
  • the upper left is respectively Two rectangular sub-image areas of different sizes centered on the lower right two structured reference points.
  • Step S1033 Cutting the registered image according to the structured reference point position and the shape parameter of the sub-image to obtain a plurality of structured sub-images.
  • the structured reference point position and the shape parameter of the sub image have been determined by step S1031 and step S1032, and then, the configuration is required to be cut according to the structured reference point position and the shape parameter of the sub image.
  • the quasi-image is extracted to extract a plurality of structured sub-images, and the positional relationship of the structured reference points is recorded and stored as structural information.
  • the mathematical algorithm of the structured sub-image may be:
  • a ij C(a,p ij (x,y),s ij )
  • a ij denotes a structured sub-image in which the structural order is in the i-th row and j-th in the vertical row
  • C is a constructing function of the structured sub-image
  • a represents the image input by the user
  • p ij represents the order in the horizontal row i, vertical jth structured reference points
  • p ij (x, y) indicates that the structured reference point p ij is at the coordinates (x, y) of the image input by the user
  • s ij represents the structuring Shape parameters of the image, including rectangular, circular, elliptical and other arbitrary planar shapes and their dimensions.
  • Step S104 Extract a visual feature of each of the structured sub-images by using a feature model obtained by multi-model training.
  • step S103 a plurality of structured sub-images have been constructed for the registered image, and then, a feature model obtained by multi-model training is required to extract visual features of each of the structured sub-images, the feature model It is a mathematical expression obtained by multi-model training to extract visual features of an image.
  • the input is an overall or partial image
  • the output is a corresponding visual feature.
  • the visual feature is a mathematical expression based on an image that can describe the overall or local shape, texture, color, etc. of the image, and is generally expressed in the form of a vector.
  • the multi-model training is a process of estimating the parameters of the feature model, and the estimation of the feature model parameters is generally completed according to a certain criterion by a large number of images.
  • FIG. 5 is a flowchart of multi-model training in an image feature extraction method embodiment provided by the present application, and the feature model obtained by the multi-model training is through the following sub- Steps to achieve:
  • Step S1041 Select a predetermined training image library.
  • a predetermined training image library is first selected, and the predetermined training image library is a set of a plurality of training images that are consistent with the image subject content input by the user, and the preferred image of the face image is For example, if the image input by the user is a face image, the predetermined training image library is selected as a face training image library, and the face training image library can adopt a representative open face database in the industry, such as 1fw, CASIA_WebFace, etc., can also use their own face database organized according to uniform standards.
  • Step S1042 Register each training image in the predetermined training image library according to a unified registration method to obtain a plurality of registered training images.
  • step S104 a predetermined training image library has been selected.
  • the registration method is the same as the registration method described in the step S102. For details, refer to the description of the above step S102, which is not described here, and is within the protection scope of the present application.
  • Step S1043 respectively construct a plurality of structured sub-training images for the plurality of registered training images.
  • step S1042 each training image in the predetermined training image library has been registered according to a unified registration method, and a plurality of registered training images are obtained, and then, after the registration is required
  • the plurality of training images respectively construct a plurality of structured sub-training images.
  • Step S1044 Perform feature model training on the plurality of structured sub-training images by using a visual feature learning algorithm to extract corresponding plurality of sub-training image visual features, and obtain a feature model.
  • step S1043 a plurality of structured sub-training images are respectively constructed for the plurality of registered training images, and then, the feature model training is performed on the plurality of structured sub-training images by using a visual feature learning algorithm to extract Corresponding multiple sub-training image visual features and obtaining feature models.
  • multi-model training is performed on each structured sub-training image to extract the most characteristic visual features for each structured sub-training image.
  • the visual feature learning algorithm includes any one of the following: a deep learning method, a boosting algorithm, an svm algorithm, or a local feature combination learning algorithm.
  • the mathematical expression of the feature model is:
  • a ij denotes a sub-training image in which the structural order is located in the i-th row and the j-th row in the horizontal row
  • M ij is a feature model trained corresponding to the sub-training image a ij
  • q ij is a feature model parameter obtained by training
  • v Ij is a sub-training image visual feature extracted by the feature model M ij for the sub-training image a ij .
  • steps S1041 to S1044 multi-model training is completed, and the feature model and the feature model parameters are determined. Next, the plurality of structured sub-images are substituted into the feature model, and each of the structurators can be calculated. The visual characteristics of the image.
  • Step S105 Structurally merging the visual features of the plurality of structured sub-images to obtain structured feature data.
  • step S104 the feature features obtained by the multi-model training are used to extract the visual features of each of the structured sub-images, and then the visual features of the plurality of structured sub-images are structurally fused to obtain the structured features. data.
  • the structurally merging the visual features of the plurality of structured sub-images to obtain structured feature data includes:
  • the visual features of the structured sub-image are spatially structured and fused according to the structured reference point position determined in the above step S103, so that on the spatial plane It may be reflected that a visual feature of each of the structured sub-images is based on a spatial relationship of the structured reference point locations, and a feature axis of the visual features of the structured sub-images reflects feature information of each of the structured sub-images Its length represents the feature dimension.
  • FIG. 6 which is a schematic diagram of feature structure fusion provided by the present application.
  • the feature value image 602 of the feature reference point 601 is extracted by a corresponding feature model, and the feature vector 603 is structured by structural fusion.
  • the feature data 604 since the process of structured merging maintains the spatial positional relationship of the structured reference point 601 with respect to other structured reference points, the structured feature data 604 also includes feature space relationships and feature information.
  • the mathematical representation of the structured feature data is:
  • v ij represents the visual feature of the structured sub-image
  • k is the data of the k-th dimension
  • d is the structured feature data obtained after the fusion.
  • Step S106 The model obtained by training the structured model is used to perform operation on the structured feature data to obtain image feature data.
  • step S105 Structurally merging the visual features of the plurality of structured sub-images by step S105, The structured feature data is obtained.
  • the obtained model is trained by using the structured model, and the structured feature data is calculated to obtain image feature data.
  • the structured model training is a subsequent step of the multi-model training described in the above steps S1041 to S1044.
  • the structured model training is a subsequent step of the multi-model training described in the above steps S1041 to S1044.
  • the structured model training is to train the structured feature data, and the feature information is better integrated while maintaining the feature space relationship.
  • the structured model training includes:
  • the structured feature model is trained on the structured image data of the training image by using the visual feature learning algorithm, and the model obtained by the structured model training is obtained.
  • the mathematical expression of the model obtained by the structured model training is:
  • M is a model obtained by performing structured model training based on the fused training image feature data d
  • q is a model parameter obtained by training
  • v is a corresponding visual feature obtained by merging the training image feature data d by the model M.
  • the model and model parameters can be determined.
  • the final image feature data v can be calculated.
  • the process of the image feature extraction method provided by the present application is completed in steps S101 to S106.
  • the spatial position information between the structured sub-images is preserved by constructing the structured sub-image, and thus the extracted location is
  • the visual features of the structured sub-image include both the feature space relationship and the feature information.
  • the structural fusion is performed, the descriptiveness of each visual feature is preserved, and the spatial relationship of each visual feature is preserved, so that the finally obtained image feature data is obtained.
  • the feature distance between feature vectors can be used to describe the difference between different images. Because the feature vector and model in this method better preserve the structural characteristics of the image during the training process, the image features are The data is more accurate and identifiable. Applying the image feature extraction method provided by the present application in image recognition, especially face recognition, has higher accuracy, thereby obtaining a better recognition effect.
  • the image feature data of the image input by the user has been extracted, and then The image input data may be used to identify the image input by the user, and may be used to determine whether the image input by the user is similar to a certain image, or determine whether there is an image input by the user in a certain image database. A similar picture, or a picture similar to the image input by the user is selected in a certain image database.
  • the image feature extraction method further includes the steps of:
  • the comparison result may be a degree of similarity between the image input by the user and each picture in the predetermined image database, or may be a picture in a predetermined image database and its information and the like that bring the degree of similarity to a predetermined threshold.
  • the predetermined image database may be a criminal face database in a public security pursuit application, an employee face database in the attendance system, a member face database in the member management system, or a star in the star face retrieval system.
  • the comparison result may be whether the image input by the user is a fugitive, whether the image input by the user is a registered employee or a member, and whether the appearance of the attendant is consistent with the record in the attendance system. , the image input by the user is similar to the appearance of which star, and the like.
  • the degree of similarity can be characterized by the distance between the vectors. The smaller the distance, the higher the degree of similarity such as the Euclidean distance, the Cosine distance or the Joint Bayesian distance.
  • the comparing the image feature data with each predetermined image feature data in a predetermined image database including:
  • the output comparison results include:
  • the algorithm for calculating a difference between the image feature data and each predetermined image feature data in a predetermined image database includes any one of the following:
  • Euclidean distance calculation method Cosine distance calculation method or Joint Bayesian distance calculation method.
  • FIG. 7 is a schematic diagram of an embodiment of an image feature extraction apparatus provided by the present application. Since the device embodiment is basically similar to the method embodiment, the description is relatively simple, and the relevant parts can be referred to the description of the method embodiment.
  • the device embodiments described below are merely illustrative.
  • the image feature extraction device includes: an image receiving unit 701, configured to receive an image input by a user; and a registration unit 702, configured to perform an image input by the user Registration, obtaining a registered image; a sub-image construction unit 703 for constructing a plurality of structured sub-images for the registered image; a visual feature extraction unit 704 for using a multi-model training to obtain a feature model Extracting a visual feature of each of the structured sub-images; a fusing unit 705, configured to structurally fuse the visual features of the plurality of structured sub-images to obtain structured feature data; and an operation unit 706, configured to adopt a structure
  • the model obtained by the model training is operated on the structured feature data to obtain image feature data.
  • the registration unit 702 includes:
  • a reference point determining subunit for determining a structured reference point position of the registered image
  • a shape parameter determining subunit for determining a shape parameter of the sub image
  • a cutting subunit configured to cut the registered image according to the structured reference point position and the shape parameter of the sub image to obtain a plurality of structured sub-images.
  • the reference point determining subunit includes:
  • a feature reference point determining subunit configured to determine a structured reference point position of the registered image according to the image feature point
  • a spatial reference point determining subunit is configured to determine a structured reference point position of the registered image based on the spatial location.
  • the mathematical algorithm used by the cutting subunit is:
  • a ij C(a,p ij (x,y),s ij )
  • a ij denotes a structured sub-image in which the structural order is in the i-th row and j-th in the vertical row
  • C is a constructing function of the structured sub-image
  • a represents the image input by the user
  • p ij represents the order in the horizontal row i, vertical j-th structured reference points
  • p ij (x, y) indicates that the structured reference point p ij is at the coordinates (x, y) of the image input by the user
  • s ij represents the structuring Shape parameters of the image, including rectangular, circular, elliptical and other arbitrary planar shapes and their dimensions.
  • the image feature extraction device further includes: a multi-model training unit, configured to obtain a feature model by multi-model training.
  • the multi-model training unit includes:
  • Training image library selection subunit for selecting a predetermined training image library
  • a training image registration sub-unit configured to register each training image in the predetermined training image library according to a unified registration method, to obtain a plurality of the registered training images
  • a sub-training image construction sub-unit configured to respectively construct a plurality of structured sub-training images for the plurality of registered training images
  • the feature model acquisition sub-unit is configured to perform feature model training on the plurality of structured sub-training images by using a visual feature learning algorithm to extract corresponding plurality of sub-training image visual features, and obtain a feature model.
  • the visual feature learning algorithm adopted by the feature model acquisition subunit includes any one of the following:
  • the merging unit 705 includes:
  • a reference point fusion subunit configured to structurally fuse the visual features of the plurality of structured sub-images according to the determined structured reference point position when constructing the plurality of structured sub-images, to obtain structured feature data
  • the structured feature data includes feature space relationships and feature information.
  • the image feature extraction device further includes:
  • a structured model training unit for training models through structured model training.
  • the structured model training unit includes:
  • a sub-training image fusion sub-unit configured to structurally fuse the plurality of sub-training image visual features to obtain training image structured feature data
  • the model acquisition subunit is configured to perform structural model training on the structured image data of the training image by using a visual feature learning algorithm, and obtain a model obtained by training the structured model.
  • the image feature extraction device further includes:
  • a matching unit for using the image feature data with each predetermined map in a predetermined image database Performing alignments like feature data in sequence;
  • An output unit for outputting the comparison result.
  • the comparison unit includes:
  • a difference calculation subunit configured to sequentially calculate a difference between the image feature data and each predetermined image feature data in a predetermined image database
  • the output unit includes:
  • a difference determining subunit configured to sequentially determine whether each of the difference values is greater than a predetermined difference threshold
  • An information output unit configured to: if each of the differences is greater than a predetermined similarity threshold, output information without a similar image; otherwise, an image corresponding to predetermined image feature data having a minimum difference from the image feature data , and / or image information output.
  • the algorithm for calculating, by the comparing unit, the difference between the image feature data and each predetermined image feature data in the predetermined image database includes any one of the following:
  • Euclidean distance calculation method Cosine distance calculation method or Joint Bayesian distance calculation method.
  • the application also provides an image feature extraction terminal device, including:
  • the image feature extraction method provided by the present application is stored in the memory; and can be run according to the above method after startup.
  • the client is a tablet computer
  • the user takes a photo with the tablet or selects a face photo from the album
  • the tablet calls the image feature extraction method provided by the application to extract the image feature of the photo.
  • Data is compared with the image in the pre-stored star face image database to obtain a star image with the highest similarity to the photo, and the character information of the star is retrieved, and then the star image and the person information are displayed.
  • the present application also provides an image feature extraction system, including a client and a remote server.
  • the system is provided with the image feature extraction device provided by the application.
  • the client captures an image and/or selects an album.
  • the image in the image is sent to the remote server, which extracts the image features
  • the data is compared with an image in a predetermined image database, and the comparison result is sent to the client, and finally the comparison result is output by the client.
  • the client is a smart phone
  • the user takes a photo by using a smart phone or selects a face photo from the album, and then sends the photo to the remote server, and the remote server invokes the image feature extraction method provided by the application. Extracting image feature data of the photo, and comparing with the image in the pre-stored star face image database, obtaining a star image with the highest similarity to the photo, and retrieving the character information of the star, and then the star
  • the image and character information are sent to the client and finally output on the display screen of the client.
  • the image feature extraction method is used.
  • details refer to the description of the image feature extraction method embodiment, and details are not described herein.
  • a computing device includes one or more processors (CPUs), input/output interfaces, network interfaces, and memory.
  • processors CPUs
  • input/output interfaces network interfaces
  • memory volatile and non-volatile memory
  • the memory may include non-persistent memory, random access memory (RAM), and/or non-volatile memory in a computer readable medium, such as read only memory (ROM) or flash memory.
  • RAM random access memory
  • ROM read only memory
  • Memory is an example of a computer readable medium.
  • Computer readable media including both permanent and non-persistent, removable and non-removable media may be implemented by any method or technology.
  • the information can be computer readable instructions, data structures, modules of programs, or other data.
  • Examples of computer storage media include, but are not limited to, phase change memory (PRAM), static random access memory (SRAM), dynamic random access memory (DRAM), other types of random access memory (RAM), read only memory. (ROM), electrically erasable programmable read only memory (EEPROM), flash memory or other memory technology, compact disk read only memory (CD-ROM), digital versatile disk (DVD) or other optical storage, Magnetic tape cartridges, magnetic tape storage or other magnetic storage devices or any other non-transportable media can be used to store information that can be accessed by a computing device.
  • computer readable media does not include non-transitory computer readable media, such as modulated data signals and carrier waves.
  • embodiments of the present application can be provided as a method, system, or computer program product. Therefore, the present application may employ an entirely hardware embodiment, an entirely software embodiment, or a combination of software. And in the form of an embodiment of the hardware aspect. Moreover, the application can take the form of a computer program product embodied on one or more computer-usable storage media (including but not limited to disk storage, CD-ROM, optical storage, etc.) including computer usable program code.
  • computer-usable storage media including but not limited to disk storage, CD-ROM, optical storage, etc.

Abstract

The present application provides an image feature extraction method. Firstly, an image input by a user is received. Then, the image input by the user is registered to obtain a registered image. Then, a plurality of structured sub-images is constructed for the registered image. Then, visual features of each structured sub-image are extracted by means of a feature model obtained by multi-model training. Then, the visual features of the plurality of structured sub-images are structurally combined to obtain structured feature data. Finally, the structured feature data is operated by means of a model obtained by structured model training to obtain image feature data. Compared with the prior art, the present application has the advantages that obtained image feature data is a feature vector; because the feature vector and a model keep structured characteristics of an image in a training process, the image feature data has higher accuracy and recognizability, and will have higher accuracy when being applied to image recognition, particularly to face recognition, thereby obtaining a better recognition effect.

Description

图像特征提取方法、装置、终端设备及系统Image feature extraction method, device, terminal device and system 技术领域Technical field
本申请涉及电子技术领域,具体的说是一种图像特征提取方法、一种图像特征提取装置、一种图像特征提取终端设备以及一种图像特征提取系统。The present application relates to the field of electronic technologies, and in particular, to an image feature extraction method, an image feature extraction device, an image feature extraction terminal device, and an image feature extraction system.
背景技术Background technique
人脸识别研究始于上世纪90年代,一开始提出的有以图像主成分来描述人脸的eigen face方法与以区分性特征来描述人脸图像的fisher face方法;进入本世纪后,基于LBP与Gabor的人脸局部特征描述方法以及基于boosting的区分性特征学习方法迅速成为主流;近些年,随着深度学习(deep learning)方法的提出,人脸识别技术又被推上了一个新的台阶。目前人脸识别领域比较有代表性的前沿技术有以下几个:Face recognition research began in the 1990s. At the beginning, there were proposed eigen face methods for describing faces with image principal components and fisher face methods for describing face images with distinguishing features. After entering this century, based on LBP Gabor's face local feature description method and boosting-based distinguishing feature learning method have quickly become mainstream; in recent years, with the deep learning method proposed, face recognition technology has been pushed to a new Steps. At present, there are several leading edge technologies in the field of face recognition:
首先是美国的facebook公司,首度将深度学习方法引入到人脸识别,利用5个卷积层与2个全连层构建的深度神经网络,对整幅人脸图像提取4096维的视觉特征来进行描述,在识别准确性上得到了显著的提高。The first is the American facebook company, which introduced the deep learning method to face recognition for the first time. Using five convolutional layers and two deep-separated neural networks, it extracts 4096-dimensional visual features from the entire face image. Described, the recognition accuracy has been significantly improved.
国内的face++公司同样利用深度学习方法,以金字塔结构分级学习了一个较深的神经网络,对整幅人脸图像进行分析,同样在人脸识别技术上取得了突破。The domestic face++ company also uses the deep learning method to learn a deep neural network by pyramid structure, and analyzes the entire face image, which also makes a breakthrough in face recognition technology.
香港中文大学汤晓鸥教授所在的研究组,对基于深度学习的人脸识别技术进行了更加深入的研究,他们用多个人脸子图像分别训练深度神经网络,再将各子神经网络输出的特征串联起来,得到了更好的识别效果,但是,这种对各子图像提取的特征简单的串联损失了图像本身的结构特点。The research team of Prof. Tang Xiaoou from the Chinese University of Hong Kong has conducted a more in-depth study on face recognition technology based on deep learning. They used multiple face images to train deep neural networks and then connected the features of each sub-neural network. A better recognition effect is obtained, but this simple series connection of the features extracted from each sub-image loses the structural characteristics of the image itself.
发明内容Summary of the invention
鉴于上述问题,本申请提供一种图像特征提取方法、一种图像特征提取装置、一种图像特征提取终端设备以及一种图像特征提取系统。本申请采用的技术方案是:In view of the above problems, the present application provides an image feature extraction method, an image feature extraction device, an image feature extraction terminal device, and an image feature extraction system. The technical solution adopted in this application is:
本申请提供一种图像特征提取方法,包括:The application provides an image feature extraction method, including:
接收用户输入的图像;Receiving an image input by a user;
对所述用户输入的图像进行配准,获得配准后的图像;Registering the image input by the user to obtain a registered image;
对所述配准后的图像构建多个结构化子图像; Constructing a plurality of structured sub-images for the registered image;
采用多模型训练获得的特征模型提取每个所述结构化子图像的视觉特征;Extracting a visual feature of each of the structured sub-images using a feature model obtained by multi-model training;
将所述多个结构化子图像的视觉特征进行结构化融合,获得结构化特征数据;Structurally merging the visual features of the plurality of structured sub-images to obtain structured feature data;
采用结构化模型训练得到的模型,对所述结构化特征数据进行运算,获得图像特征数据。The model obtained by the structured model training is used to calculate the structured feature data to obtain image feature data.
可选的,所述对所述配准后的图像构建多个结构化子图像,包括:Optionally, the constructing the multiple structured sub-images on the registered image comprises:
确定所述配准后的图像的结构化基准点位置;Determining a structured reference point location of the registered image;
确定子图像的形状参数;Determining a shape parameter of the sub image;
根据所述结构化基准点位置及所述子图像的形状参数,切割所述配准后的图像,获得多个结构化子图像。The registered image is cut according to the structured reference point position and the shape parameter of the sub-image to obtain a plurality of structured sub-images.
可选的,所述确定所述配准后的图像的结构化基准点位置,包括:Optionally, the determining the location of the structured reference point of the registered image comprises:
根据图像特征点确定所述配准后的图像的结构化基准点位置;或者,Determining a structured reference point location of the registered image based on image feature points; or
根据空间位置确定所述配准后的图像的结构化基准点位置。A structured reference point location of the registered image is determined based on the spatial location.
可选的,所述根据所述结构化基准点位置及所述子图像的形状参数,切割所述配准后的图像,获得多个结构化子图像的数学算法为:Optionally, the mathematical algorithm for cutting the registered image according to the structured reference point position and the shape parameter of the sub-image to obtain a plurality of structured sub-images is:
aij=C(a,pij(x,y),sij)a ij =C(a,p ij (x,y),s ij )
式中aij表示结构顺序位于横排第i个、竖排第j个的结构化子图像,C为结构化子图像的构建函数,a表示用户输入的图像,pij表示顺序位于横排第i个、竖排第j个的结构化基准点,pij(x,y)表示结构化基准点pij处于所述用户输入的图像的坐标(x,y)处,sij表示结构化子图像的形状参数,包括矩形、圆形、椭圆形等任意平面形状及其尺寸。Where a ij denotes a structured sub-image in which the structural order is in the i-th row and j-th in the vertical row, C is a constructing function of the structured sub-image, a represents the image input by the user, and p ij represents the order in the horizontal row i, vertical jth structured reference points, p ij (x, y) indicates that the structured reference point p ij is at the coordinates (x, y) of the image input by the user, and s ij represents the structuring Shape parameters of the image, including rectangular, circular, elliptical and other arbitrary planar shapes and their dimensions.
可选的,所述多模型训练获得的特征模型是通过以下方法获得的:Optionally, the feature model obtained by the multi-model training is obtained by the following method:
选择预定的训练图像库;Select a predetermined training image library;
将所述预定的训练图像库中的每个训练图像按照统一的配准方法进行配准,获得配准后的多个训练图像;Registering each training image in the predetermined training image library according to a unified registration method to obtain a plurality of registered training images;
对所述配准后的多个训练图像分别构建多个结构化子训练图像;Constructing a plurality of structured sub-training images for the plurality of registered training images;
采用视觉特征学习算法对所述多个结构化子训练图像进行特征模型训练以提取相应的多个子训练图像视觉特征,并获得特征模型。The plurality of structured sub-training images are subjected to feature model training by using a visual feature learning algorithm to extract corresponding plurality of sub-training image visual features, and a feature model is obtained.
可选的,所述视觉特征学习算法包括以下任一种: Optionally, the visual feature learning algorithm includes any one of the following:
深度学习方法、boosting算法、svm算法或局部特征组合的学习算法。Learning algorithm for deep learning method, boosting algorithm, svm algorithm or local feature combination.
可选的,所述特征模型的数学表达为:Optionally, the mathematical expression of the feature model is:
vij=Mij(aij,qij)v ij =M ij (a ij ,q ij )
式中aij表示结构顺序位于横排第i个、竖排第j个的子训练图像,Mij为对应子训练图像aij上训练得到的特征模型,qij为训练得到的特征模型参数,vij为通过特征模型Mij对子训练图像aij提取的子训练图像视觉特征。Where a ij denotes a sub-training image in which the structural order is located in the i-th row and the j-th row in the horizontal row, M ij is a feature model trained on the corresponding sub-training image a ij , and q ij is a feature model parameter obtained by training. v ij is a sub-training image visual feature extracted by the feature model M ij for the sub-training image a ij .
可选的,所述将所述多个结构化子图像的视觉特征进行结构化融合,获得结构化特征数据,包括:Optionally, the structurally merging the visual features of the plurality of structured sub-images to obtain the structured feature data includes:
根据构建多个结构化子图像时的确定的结构化基准点位置,将所述多个结构化子图像的视觉特征进行结构化融合,获得结构化特征数据,所述结构化特征数据包括特征空间关系和特征信息。And structurally merging the visual features of the plurality of structured sub-images according to the determined structured reference point positions when constructing the plurality of structured sub-images to obtain structured feature data, where the structured feature data includes the feature space Relationship and feature information.
可选的,所述结构化特征数据的数学表达为:Optionally, the mathematical expression of the structured feature data is:
d(i,j,k)=vij(k)d(i,j,k)=v ij (k)
式中vij表示结构化子图像的视觉特征,k为第k维的数据,d为融合后的结构化特征数据。Where v ij represents the visual feature of the structured sub-image, k is the data of the k-th dimension, and d is the structured feature data after the fusion.
可选的,所述结构化模型训练得到的模型是通过以下方式获得的:Optionally, the model obtained by the structured model training is obtained by:
将所述多个子训练图像视觉特征进行结构化融合,获得训练图像结构化特征数据;Performing structural merging of the plurality of sub-training image visual features to obtain training image structured feature data;
采用视觉特征学习算法对所述训练图像结构化特征数据进行结构化模型训练,获得结构化模型训练得到的模型。The structured feature model is trained on the structured image data of the training image by using the visual feature learning algorithm, and the model obtained by the structured model training is obtained.
可选的,所述结构化模型训练得到的模型的数学表达为:Optionally, the mathematical expression of the model obtained by the structured model training is:
v=M(d,q)v=M(d,q)
其中M为基于融合后的训练图像特征数据d进行结构化模型训练得到的模型,q为训练得到的模型参数,v为通过模型M对训练图像特征数据d融合得到的相应视觉特征。Where M is a model obtained by performing structured model training based on the fused training image feature data d, q is a model parameter obtained by training, and v is a corresponding visual feature obtained by merging the training image feature data d by the model M.
可选的,所述图像特征提取方法还包括:Optionally, the image feature extraction method further includes:
将所述图像特征数据与预定的图像数据库中的各个预定图像特征数据依次进行比对;And sequentially comparing the image feature data with each predetermined image feature data in a predetermined image database;
输出比对结果。 Output comparison results.
可选的,所述将所述图像特征数据与预定的图像数据库中的各个预定图像特征数据依次进行比对,包括:Optionally, the comparing the image feature data with each predetermined image feature data in a predetermined image database, including:
依次计算所述图像特征数据与预定的图像数据库中的各个预定图像特征数据之间的差值;Calculating a difference between the image feature data and each predetermined image feature data in a predetermined image database;
所述输出比对结果包括:The output comparison results include:
依次判断每个所述差值是否大于预定的差值阈值;Determining, in turn, whether each of the difference values is greater than a predetermined difference threshold;
若每个所述差值都大于预定的相似度阈值,则输出没有相似图像的信息,否则,则将与所述图像特征数据差值最小的预定图像特征数据对应的图像,和/或图像的信息输出。If each of the differences is greater than a predetermined similarity threshold, information having no similar image is output, otherwise, an image corresponding to predetermined image feature data having the smallest difference from the image feature data, and/or an image Information output.
可选的,所述计算所述图像特征数据与预定的图像数据库中的各个预定图像特征数据之间的差值的算法包括以下任一种:Optionally, the algorithm for calculating a difference between the image feature data and each predetermined image feature data in a predetermined image database includes any one of the following:
欧氏距离计算方法、Cosine距离计算方法或Joint Bayesian距离计算方法。Euclidean distance calculation method, Cosine distance calculation method or Joint Bayesian distance calculation method.
可选的,所述图像包括:人脸图像。Optionally, the image includes: a face image.
本申请还提供一种图像特征提取装置,包括:The application also provides an image feature extraction device, including:
图像接收单元,用于接收用户输入的图像;An image receiving unit, configured to receive an image input by a user;
配准单元,用于对所述用户输入的图像进行配准,获得配准后的图像;a registration unit, configured to register an image input by the user to obtain a registered image;
子图像构建单元,用于对所述配准后的图像构建多个结构化子图像;a sub-image construction unit, configured to construct a plurality of structured sub-images on the registered image;
视觉特征提取单元,用于采用多模型训练获得的特征模型提取每个所述结构化子图像的视觉特征;a visual feature extraction unit, configured to extract a visual feature of each of the structured sub-images by using a feature model obtained by multi-model training;
融合单元,用于将所述多个结构化子图像的视觉特征进行结构化融合,获得结构化特征数据;a merging unit, configured to structurally fuse the visual features of the plurality of structured sub-images to obtain structured feature data;
运算单元,用于采用结构化模型训练得到的模型,对所述结构化特征数据进行运算,获得图像特征数据。The operation unit is configured to use the model obtained by the structural model training, and perform operation on the structured feature data to obtain image feature data.
可选的,所述配准单元,包括:Optionally, the registration unit includes:
基准点确定子单元,用于确定所述配准后的图像的结构化基准点位置;a reference point determining subunit for determining a structured reference point position of the registered image;
形状参数确定子单元,用于确定子图像的形状参数;a shape parameter determining subunit for determining a shape parameter of the sub image;
切割子单元,用于根据所述结构化基准点位置及所述子图像的形状参数,切割所述配准后的图像,获得多个结构化子图像。And a cutting subunit, configured to cut the registered image according to the structured reference point position and the shape parameter of the sub image to obtain a plurality of structured sub-images.
可选的,所述基准点确定子单元,包括: Optionally, the reference point determining subunit includes:
特征基准点确定子单元,用于根据图像特征点确定所述配准后的图像的结构化基准点位置;或者,a feature reference point determining subunit, configured to determine a structured reference point position of the registered image according to the image feature point; or
空间基准点确定子单元,用于根据空间位置确定所述配准后的图像的结构化基准点位置。A spatial reference point determining subunit is configured to determine a structured reference point position of the registered image based on the spatial location.
可选的,所述切割子单元采用的数学算法为:Optionally, the mathematical algorithm used by the cutting subunit is:
aij=C(a,pij(x,y),sij)a ij =C(a,p ij (x,y),s ij )
式中aij表示结构顺序位于横排第i个、竖排第j个的结构化子图像,C为结构化子图像的构建函数,a表示用户输入的图像,pij表示顺序位于横排第i个、竖排第j个的结构化基准点,pij(x,y)表示结构化基准点pij处于所述用户输入的图像的坐标(x,y)处,sij表示结构化子图像的形状参数,包括矩形、圆形、椭圆形等任意平面形状及其尺寸。Where a ij denotes a structured sub-image in which the structural order is in the i-th row and j-th in the vertical row, C is a constructing function of the structured sub-image, a represents the image input by the user, and p ij represents the order in the horizontal row i, vertical jth structured reference points, p ij (x, y) indicates that the structured reference point p ij is at the coordinates (x, y) of the image input by the user, and s ij represents the structuring Shape parameters of the image, including rectangular, circular, elliptical and other arbitrary planar shapes and their dimensions.
可选的,所述图像特征提取装置,还包括:Optionally, the image feature extraction device further includes:
多模型训练单元,用于通过多模型训练获得特征模型;a multi-model training unit for obtaining a feature model by multi-model training;
所述多模型训练单元包括:The multi-model training unit includes:
训练图像库选择子单元,用于选择预定的训练图像库;Training image library selection subunit for selecting a predetermined training image library;
训练图像配准子单元,用于将所述预定的训练图像库中的每个训练图像按照统一的配准方法进行配准,获得配准后的多个训练图像;And a training image registration sub-unit, configured to register each training image in the predetermined training image library according to a unified registration method, to obtain a plurality of the registered training images;
子训练图像构建子单元,用于对所述配准后的多个训练图像分别构建多个结构化子训练图像;a sub-training image construction sub-unit, configured to respectively construct a plurality of structured sub-training images for the plurality of registered training images;
特征模型获取子单元,用于采用视觉特征学习算法对所述多个结构化子训练图像进行特征模型训练以提取相应的多个子训练图像视觉特征,并获得特征模型。The feature model acquisition sub-unit is configured to perform feature model training on the plurality of structured sub-training images by using a visual feature learning algorithm to extract corresponding plurality of sub-training image visual features, and obtain a feature model.
可选的,所述特征模型获取子单元采用的视觉特征学习算法包括以下任一种:Optionally, the visual feature learning algorithm adopted by the feature model acquisition subunit includes any one of the following:
深度学习方法、boosting算法、svm算法或局部特征组合的学习算法。Learning algorithm for deep learning method, boosting algorithm, svm algorithm or local feature combination.
可选的,所述融合单元包括:Optionally, the fusion unit includes:
基准点融合子单元,用于根据构建多个结构化子图像时的确定的结构化基准点位置,将所述多个结构化子图像的视觉特征进行结构化融合,获得结构化特征数据,所述结构化特征数据包括特征空间关系和特征信息。 a reference point fusion subunit, configured to structurally fuse the visual features of the plurality of structured sub-images according to the determined structured reference point position when constructing the plurality of structured sub-images, to obtain structured feature data, The structured feature data includes feature space relationships and feature information.
可选的,所述图像特征提取装置,还包括:Optionally, the image feature extraction device further includes:
结构化模型训练单元,用于通过结构化模型训练获得模型;a structured model training unit for training a model through a structured model;
所述结构化模型训练单元包括:The structured model training unit includes:
子训练图像融合子单元,用于将所述多个子训练图像视觉特征进行结构化融合,获得训练图像结构化特征数据;a sub-training image fusion sub-unit, configured to structurally fuse the plurality of sub-training image visual features to obtain training image structured feature data;
模型获取子单元,用于采用视觉特征学习算法对所述训练图像结构化特征数据进行结构化模型训练,获得结构化模型训练得到的模型。The model acquisition subunit is configured to perform structural model training on the structured image data of the training image by using a visual feature learning algorithm, and obtain a model obtained by training the structured model.
可选的,所述图像特征提取装置,还包括:Optionally, the image feature extraction device further includes:
比对单元,用于将所述图像特征数据与预定的图像数据库中的各个预定图像特征数据依次进行比对;a comparison unit, configured to sequentially compare the image feature data with each predetermined image feature data in a predetermined image database;
输出单元,用于输出比对结果。An output unit for outputting the comparison result.
可选的,所述比对单元包括:Optionally, the comparison unit includes:
差值计算子单元,用于依次计算所述图像特征数据与预定的图像数据库中的各个预定图像特征数据之间的差值;a difference calculation subunit, configured to sequentially calculate a difference between the image feature data and each predetermined image feature data in a predetermined image database;
所述输出单元包括:The output unit includes:
差值判断子单元,用于依次判断每个所述差值是否大于预定的差值阈值;a difference determining subunit, configured to sequentially determine whether each of the difference values is greater than a predetermined difference threshold;
信息输出单元,用于若每个所述差值都大于预定的相似度阈值,则输出没有相似图像的信息,否则,则将与所述图像特征数据差值最小的预定图像特征数据对应的图像,和/或图像的信息输出。An information output unit, configured to: if each of the differences is greater than a predetermined similarity threshold, output information without a similar image; otherwise, an image corresponding to predetermined image feature data having a minimum difference from the image feature data , and / or image information output.
可选的,所述比对单元计算所述图像特征数据与预定的图像数据库中的各个预定图像特征数据之间的差值的算法包括以下任一种:Optionally, the algorithm for calculating, by the comparing unit, the difference between the image feature data and each predetermined image feature data in the predetermined image database includes any one of the following:
欧氏距离计算方法、Cosine距离计算方法或Joint Bayesian距离计算方法。Euclidean distance calculation method, Cosine distance calculation method or Joint Bayesian distance calculation method.
本申请还提供一种图像特征提取终端设备,包括:The application also provides an image feature extraction terminal device, including:
中央处理器;CPU;
输入输出单元;Input and output unit;
存储器;所述存储器中存储有本申请提供的图像特征提取方法;并在启动后能够根据上述方法运行。a memory; the image feature extraction method provided by the present application is stored in the memory; and can be run according to the above method after startup.
本申请还提供一种图像特征提取系统,包括客户端和远端服务器,使用本申请提供的图像特征提取装置,所述客户端拍摄图像和/或选取相册中的图像发 送到远端服务器,所述远端服务器提取出图像特征数据,并与预定的图像数据库中的图像进行比对,并将比对结果发送至所述客户端,最终由所述客户端输出比对结果。The present application also provides an image feature extraction system, including a client and a remote server. Using the image feature extraction device provided by the application, the client captures an image and/or selects an image in the album. Sending to the remote server, the remote server extracts the image feature data, compares it with the image in the predetermined image database, and sends the comparison result to the client, and finally outputs the ratio by the client. For the result.
与现有技术相比,本申请具有以下优点:Compared with the prior art, the present application has the following advantages:
本申请提供的一种图像特征提取方法,首先接收用户输入的图像;然后对所述用户输入的图像进行配准,获得配准后的图像;再对所述配准后的图像构建多个结构化子图像;接下来,采用多模型训练获得的特征模型提取每个所述结构化子图像的视觉特征;然后将所述多个结构化子图像的视觉特征进行结构化融合,获得结构化特征数据;最后采用结构化模型训练得到的模型,对所述结构化特征数据进行运算,获得图像特征数据。相较于现有技术的图像特征提取方法,本申请中,通过构建结构化子图像保留了结构化子图像之间的空间位置信息,因此提取的所述结构化子图像的视觉特征同时包括特征空间关系和特征信息,在进行结构化融合时既保留了各视觉特征的描述性,又保留了各视觉特征的空间关系,从而最终获得的图像特征数据为特征向量,可以用特征向量间的特征距离描述不同图像之间的差异,又由于本方法中特征向量与模型在训练过程中更好的保持了图像的结构化特性,因此,所述图像特征数据具有更高的准确性和可辨识性。在图像识别尤其是人脸识别中应用本申请提供的图像特征提取方法,会具有更高的准确性,从而获得更好的识别效果。An image feature extraction method provided by the present application first receives an image input by a user; then, an image input by the user is registered to obtain a registered image; and then multiple structures are constructed on the registered image. Next, the feature model obtained by the multi-model training is used to extract the visual features of each of the structured sub-images; then the visual features of the plurality of structured sub-images are structurally fused to obtain the structural features. Data; finally, the model obtained by the structured model training is used to calculate the structured feature data to obtain image feature data. Compared with the image feature extraction method of the prior art, in the present application, the spatial position information between the structured sub-images is preserved by constructing the structured sub-image, and thus the extracted visual features of the structured sub-image include features simultaneously. The spatial relationship and feature information not only retain the descriptiveness of each visual feature, but also preserve the spatial relationship of each visual feature, so that the final image feature data is the feature vector, and the feature vector can be used. The distance describes the difference between different images, and because the feature vector and the model in the method better maintain the structural characteristics of the image during the training process, the image feature data has higher accuracy and identifiability. . Applying the image feature extraction method provided by the present application in image recognition, especially face recognition, has higher accuracy, thereby obtaining a better recognition effect.
附图说明DRAWINGS
图1是本申请提供的一种图像特征提取方法实施例的流程图;1 is a flowchart of an embodiment of an image feature extraction method provided by the present application;
图2是本申请提供的一种图像特征提取方法实施例中构建多个结构化子图像的流程图;2 is a flow chart of constructing a plurality of structured sub-images in an embodiment of an image feature extraction method provided by the present application;
图3是本申请提供的根据空间位置关系确定结构化基准点的示例图;3 is a diagram showing an example of determining a structured reference point according to a spatial position relationship provided by the present application;
图4是本申请提供的根据人脸特征点确定结构化基准点的示例图;4 is an exemplary diagram of determining a structured reference point according to a face feature point provided by the present application;
图5是本申请提供的一种图像特征提取方法实施例中多模型训练的流程图;FIG. 5 is a flowchart of multi-model training in an embodiment of an image feature extraction method provided by the present application; FIG.
图6是本申请提供的特征结构化融合示意图;6 is a schematic structural fusion diagram of the feature provided by the present application;
图7是本申请提供的一种图像特征提取装置实施例的示意图。FIG. 7 is a schematic diagram of an embodiment of an image feature extraction device provided by the present application.
具体实施方式detailed description
在下面的描述中阐述了很多具体细节以便于充分理解本申请。但是本申请 能够以很多不同于在此描述的其它方式来实施,本领域技术人员可以在不违背本申请内涵的情况下做类似推广,因此本申请不受下面公开的具体实施的限制。Numerous specific details are set forth in the description below in order to provide a thorough understanding of the application. But this application It can be implemented in many other ways than those described herein, and those skilled in the art can make similar promotion without departing from the spirit of the present application, and thus the present application is not limited by the specific embodiments disclosed below.
本申请提供了一种图像特征提取方法、一种图像特征提取装置、一种图像特征提取终端设备以及一种图像特征提取系统,下面依次结合附图对本申请的实施例进行详细说明。The present application provides an image feature extraction method, an image feature extraction device, an image feature extraction terminal device, and an image feature extraction system. The embodiments of the present application are described in detail below with reference to the accompanying drawings.
请参考图1,其为本申请提供的一种图像特征提取方法实施例的流程图,所述图像特征提取方法包括如下步骤:Please refer to FIG. 1 , which is a flowchart of an embodiment of an image feature extraction method provided by the present application. The image feature extraction method includes the following steps:
步骤S101:接收用户输入的图像。Step S101: Receive an image input by the user.
本步骤中,首先接收用户输入的图像,用户可以从终端设备的电子相册中选择一幅图像输入,也可以通过摄像装置拍摄一幅图像并输入。需要说明的是,本申请的目的在于图像识别,因此优先考虑用户输入的图像为静态图像,但为了提高本方法的普遍适用性,在本申请的一个实施例中,可以接收用户输入的动态图像,但会进行预处理,只提取所述动态图像的特定帧(如第一帧)作为用户输入的图像,以上均在本申请的保护范围之内。In this step, the image input by the user is first received, and the user can select an image input from the electronic album of the terminal device, or take an image and input by the camera. It should be noted that the purpose of the present application is image recognition, so that the image input by the user is prioritized as a static image, but in order to improve the general applicability of the method, in one embodiment of the present application, a dynamic image input by the user may be received. However, pre-processing is performed to extract only a specific frame (such as the first frame) of the dynamic image as an image input by the user, and all of the above are within the protection scope of the present application.
在本申请的一个优选实施例中,本图像特征提取方法用于人脸图像识别,因此,所述图像包括人脸图像。In a preferred embodiment of the present application, the present image feature extraction method is used for face image recognition, and therefore, the image includes a face image.
步骤S102:对所述用户输入的图像进行配准,获得配准后的图像。Step S102: register the image input by the user to obtain a registered image.
通过步骤S101,已接收到用户输入的图像,接下来,需要对所述用户输入的图像进行配准,现有技术中常用的配准方法是先检测图像特征点,然后根据特征点进行图像仿射变换,将图像归一化到预定的大小与比例,获得配准后的头像,以便进行识别和比对。In step S101, the image input by the user has been received, and then the image input by the user needs to be registered. The registration method commonly used in the prior art is to detect the image feature point first, and then perform image simulation according to the feature point. The transformation is performed, the image is normalized to a predetermined size and scale, and the registered avatar is obtained for identification and comparison.
在本申请的一个优选实施例中,本图像特征提取方法用于人脸图像识别,所述图像为人脸图像,在进行配准时,首先检测人脸图像的特征点,如眼睛、嘴、鼻子的位置等,然后根据所述特征点进行图像仿射变换,归一化到预定的大小和比例,通过这种方式,将需要与所述人脸图像进行比对的图像也进行配准,使其与所述人脸图像的大小和比例一致,即可在相同的标准下进行比对,进而提高比对的准确性。In a preferred embodiment of the present application, the image feature extraction method is used for face image recognition, and the image is a face image. When performing registration, first detecting feature points of the face image, such as eyes, mouth, and nose. Position and the like, then performing image affine transformation according to the feature points, normalizing to a predetermined size and scale, and in this way, the images that need to be compared with the face image are also registered, so that Consistent with the size and proportion of the face image, the comparison can be performed under the same standard, thereby improving the accuracy of the comparison.
步骤S103:对所述配准后的图像构建多个结构化子图像。Step S103: Construct a plurality of structured sub-images on the registered images.
通过步骤S103,已通过对所述用户输入的图像进行配准,获得了配准后的图像,接下来,需要对所述配准后的图像构建多个结构化子图像,请参考图2, 其为本申请提供的一种图像特征提取方法实施例中构建多个结构化子图像的流程图,所述对所述配准后的图像构建多个结构化子图像可通过以下子步骤进行:In step S103, the registered image is obtained by registering the image input by the user, and then, a plurality of structured sub-images are required to be constructed on the registered image. Referring to FIG. 2, It is a flowchart of constructing a plurality of structured sub-images in an embodiment of an image feature extraction method provided by the present application, and the constructing a plurality of structured sub-images on the registered image may be performed by the following sub-steps:
步骤S1031:确定所述配准后的图像的结构化基准点位置。Step S1031: Determine a structured reference point position of the registered image.
构建多个结构化子图像,即按照一定的结构、位置及限制条件从图像中分割出多个子图像。首先,要确定所述配准后的图像的结构化基准点位置,以用来确定结构化子图像的切割位置。A plurality of structured sub-images are constructed, that is, a plurality of sub-images are segmented from the image according to a certain structure, position, and constraints. First, the structured reference point position of the registered image is determined to determine the cutting position of the structured sub-image.
在本申请提供的一个实施例中,将所述结构化基准点作为结构化子图像切割的中心点,为了保持图像的结构特点以及方便后续的计算,一般选择上下左右关系基本保持不变的一组大致矩形分布的基准点。In an embodiment provided by the present application, the structured reference point is used as a center point of the structured sub-image cutting. In order to maintain the structural characteristics of the image and facilitate subsequent calculations, generally, the upper and lower left and right relations are basically kept unchanged. A set of reference points that are roughly rectangularly distributed.
结构化基准点的确定方法有多种,可以是根据空间位置确定所述配准后的图像的结构化基准点位置,也可以根据图像特征点确定所述配准后的图像的结构化基准点位置。The method for determining the structured reference point may be determined by determining a structured reference point position of the registered image according to a spatial position, or determining a structured reference point of the registered image according to the image feature point. position.
仍以上述人脸图像的优选实施例为例,如图3所示,其根据空间位置关系确定了一组4×4个结构化基准点,他们之间的距离相对于图像是完全固定的。如图4所示,其根据人脸特征点确定了3×3个结构化基准点,图中9个结构化基准点自上至下、自左至右依次为:右眼中心点、两眼中心点、左眼中心点、右脸颊点、鼻尖点、左脸颊点、右嘴角点、嘴唇中心点和左嘴角点,这9个结构化基准点的位置关系对于不同的人、姿态、表情会发生稍许变化,但同样满足近似的矩形结构关系。Still taking the preferred embodiment of the above-described face image as an example, as shown in FIG. 3, a set of 4×4 structured reference points is determined according to the spatial positional relationship, and the distance between them is completely fixed with respect to the image. As shown in Fig. 4, it determines 3×3 structured reference points according to the face feature points. The nine structured reference points in the figure are from top to bottom and from left to right: right eye center point and two eyes. The center point, the left eye center point, the right cheek point, the nose point, the left cheek point, the right mouth corner point, the lip center point, and the left corner corner point. The positional relationship of the nine structured reference points is different for different people, gestures, and expressions. A slight change occurs, but the approximate rectangular structure relationship is also satisfied.
以上仅以人脸图像为例举例说明,对于不同类别的图像,在实施时可根据图像的主体内容选择确定结构化基准点的方法,同时结构化基准点的数量也不限于上述4×4和3×3的情形,可根据实际情况灵活确定,此处不再赘述,其均在本申请的保护范围之内。The above only exemplifies the face image. For different types of images, the method of determining the structured reference point can be selected according to the main content of the image, and the number of structured reference points is not limited to the above 4×4 sum. The case of 3×3 can be flexibly determined according to the actual situation, and will not be further described herein, and all of them are within the protection scope of the present application.
步骤S1032:确定子图像的形状参数。Step S1032: Determine a shape parameter of the sub image.
通过步骤S1031,已确定所述配准后的图像的结构化基准点位置,接下来,需要确定子图像的形状参数,即以所述结构化基准点位置做参考,在其周围以一定比例与大小确定一个子图像区域,所述形状参数包括所述子图像的形状,如矩形、圆形、椭圆形等任意平面形状,以及所述子图像的尺寸,如矩形的长宽、圆形的半径等。By step S1031, the structured reference point position of the registered image has been determined, and then, the shape parameter of the sub-image needs to be determined, that is, the reference position of the structured reference point is used as a reference, and a certain ratio is The size determines a sub-image area, the shape parameter including a shape of the sub-image, such as an arbitrary planar shape such as a rectangle, a circle, an ellipse, and the size of the sub-image, such as a length and a width of the rectangle, and a radius of the circle Wait.
仍以上述人脸图像的优选实施例为例,如图3所示,确定了分别以左上与 右下两个结构化基准点为中心的不同尺寸的两个矩形子图像区域。Still taking the preferred embodiment of the above-described face image as an example, as shown in FIG. 3, it is determined that the upper left is respectively Two rectangular sub-image areas of different sizes centered on the lower right two structured reference points.
步骤S1033:根据所述结构化基准点位置及所述子图像的形状参数,切割所述配准后的图像,获得多个结构化子图像。Step S1033: Cutting the registered image according to the structured reference point position and the shape parameter of the sub-image to obtain a plurality of structured sub-images.
通过步骤S1031和步骤S1032,已确定所述结构化基准点位置及所述子图像的形状参数,接下来,需要根据所述结构化基准点位置及所述子图像的形状参数,切割所述配准后的图像,从而提取出多个结构化子图像,同时将所述结构化基准点的位置关系作为结构信息进行记录保存。The structured reference point position and the shape parameter of the sub image have been determined by step S1031 and step S1032, and then, the configuration is required to be cut according to the structured reference point position and the shape parameter of the sub image. The quasi-image is extracted to extract a plurality of structured sub-images, and the positional relationship of the structured reference points is recorded and stored as structural information.
仍以上述人脸图像的优选实施例为例,所述结构化子图像的数学算法可以为:Still taking the preferred embodiment of the above-described face image as an example, the mathematical algorithm of the structured sub-image may be:
aij=C(a,pij(x,y),sij)a ij =C(a,p ij (x,y),s ij )
式中aij表示结构顺序位于横排第i个、竖排第j个的结构化子图像,C为结构化子图像的构建函数,a表示用户输入的图像,pij表示顺序位于横排第i个、竖排第j个的结构化基准点,pij(x,y)表示结构化基准点pij处于所述用户输入的图像的坐标(x,y)处,sij表示结构化子图像的形状参数,包括矩形、圆形、椭圆形等任意平面形状及其尺寸。Where a ij denotes a structured sub-image in which the structural order is in the i-th row and j-th in the vertical row, C is a constructing function of the structured sub-image, a represents the image input by the user, and p ij represents the order in the horizontal row i, vertical jth structured reference points, p ij (x, y) indicates that the structured reference point p ij is at the coordinates (x, y) of the image input by the user, and s ij represents the structuring Shape parameters of the image, including rectangular, circular, elliptical and other arbitrary planar shapes and their dimensions.
步骤S104:采用多模型训练获得的特征模型提取每个所述结构化子图像的视觉特征。Step S104: Extract a visual feature of each of the structured sub-images by using a feature model obtained by multi-model training.
通过步骤S103,已对所述配准后的图像构建多个结构化子图像,接下来,需要采用多模型训练获得的特征模型提取每个所述结构化子图像的视觉特征,所述特征模型是通过多模型训练获得的用来提取图像视觉特征的数学表达,其输入是整体或者局部图像,输出是相应的视觉特征。所述视觉特征是基于图像提炼出来的能描述图像整体或者局部形状、纹理、颜色等特点的数学表达,一般用向量的形式来表示。所述多模型训练是估计特征模型参数的过程,一般通过大批图像按照某种准则完成特征模型参数的估计。By step S103, a plurality of structured sub-images have been constructed for the registered image, and then, a feature model obtained by multi-model training is required to extract visual features of each of the structured sub-images, the feature model It is a mathematical expression obtained by multi-model training to extract visual features of an image. The input is an overall or partial image, and the output is a corresponding visual feature. The visual feature is a mathematical expression based on an image that can describe the overall or local shape, texture, color, etc. of the image, and is generally expressed in the form of a vector. The multi-model training is a process of estimating the parameters of the feature model, and the estimation of the feature model parameters is generally completed according to a certain criterion by a large number of images.
在本申请提供的一个实施例中,请参考图5,其为本申请提供的一种图像特征提取方法实施例中多模型训练的流程图,所述多模型训练获得的特征模型是通过以下子步骤实现的:In an embodiment provided by the present application, please refer to FIG. 5 , which is a flowchart of multi-model training in an image feature extraction method embodiment provided by the present application, and the feature model obtained by the multi-model training is through the following sub- Steps to achieve:
步骤S1041:选择预定的训练图像库。Step S1041: Select a predetermined training image library.
本步骤,首先选择预定的训练图像库,所述预定的训练图像库是与所述用户输入的图像主题内容一致的多个训练图像的集合,以上述人脸图像的优选实 施例为例,所述用户输入的图像为人脸图像,则选择预定的训练图像库为人脸训练图像库,所述人脸训练图像库可以采用业内具有代表性的公开人脸数据库,如1fw、CASIA_WebFace等,也可以使用自己按照统一标准整理的人脸数据库。In this step, a predetermined training image library is first selected, and the predetermined training image library is a set of a plurality of training images that are consistent with the image subject content input by the user, and the preferred image of the face image is For example, if the image input by the user is a face image, the predetermined training image library is selected as a face training image library, and the face training image library can adopt a representative open face database in the industry, such as 1fw, CASIA_WebFace, etc., can also use their own face database organized according to uniform standards.
步骤S1042:将所述预定的训练图像库中的每个训练图像按照统一的配准方法进行配准,获得配准后的多个训练图像。Step S1042: Register each training image in the predetermined training image library according to a unified registration method to obtain a plurality of registered training images.
通过步骤S1041,已选择预定的训练图像库,接下来,为了保证所述多模型训练获得的特征模型可以适用于所述用户输入的图像,需要将所述预定的训练图像库中的训练图像全部采用与步骤S102中所述的配准方法一致的配准方法进行配准,具体请参照上述步骤S102的说明,此处不再赘述,其均在本申请的保护范围之内。By step S1041, a predetermined training image library has been selected. Next, in order to ensure that the feature model obtained by the multi-model training can be applied to the image input by the user, it is necessary to all the training images in the predetermined training image library. The registration method is the same as the registration method described in the step S102. For details, refer to the description of the above step S102, which is not described here, and is within the protection scope of the present application.
步骤S1043:对所述配准后的多个训练图像分别构建多个结构化子训练图像。Step S1043: respectively construct a plurality of structured sub-training images for the plurality of registered training images.
通过步骤S1042,已将所述预定的训练图像库中的每个训练图像按照统一的配准方法进行配准,获得了配准后的多个训练图像,接下来,需要对所述配准后的多个训练图像分别构建多个结构化子训练图像。具体实施方式请参考上述步骤S103的说明,此处不再赘述,其均在本申请的保护范围之内。In step S1042, each training image in the predetermined training image library has been registered according to a unified registration method, and a plurality of registered training images are obtained, and then, after the registration is required The plurality of training images respectively construct a plurality of structured sub-training images. For details, please refer to the description of the above step S103, which is not described herein again, and is within the protection scope of the present application.
步骤S1044:采用视觉特征学习算法对所述多个结构化子训练图像进行特征模型训练以提取相应的多个子训练图像视觉特征,并获得特征模型。Step S1044: Perform feature model training on the plurality of structured sub-training images by using a visual feature learning algorithm to extract corresponding plurality of sub-training image visual features, and obtain a feature model.
通过步骤S1043,已对所述配准后的多个训练图像分别构建多个结构化子训练图像,接下来,采用视觉特征学习算法对所述多个结构化子训练图像进行特征模型训练以提取相应的多个子训练图像视觉特征,并获得特征模型。本步骤对各个结构化子训练图像分别进行多模型训练,以便对各结构化子训练图像提取最有表征性的视觉特征。By step S1043, a plurality of structured sub-training images are respectively constructed for the plurality of registered training images, and then, the feature model training is performed on the plurality of structured sub-training images by using a visual feature learning algorithm to extract Corresponding multiple sub-training image visual features and obtaining feature models. In this step, multi-model training is performed on each structured sub-training image to extract the most characteristic visual features for each structured sub-training image.
所述视觉特征学习算法包括以下任一种:深度学习方法、boosting算法、svm算法或局部特征组合的学习算法。以上均为现有技术中的成熟的学习算法,此处不再赘述,其均在本申请的保护范围之内。The visual feature learning algorithm includes any one of the following: a deep learning method, a boosting algorithm, an svm algorithm, or a local feature combination learning algorithm. The above are all mature learning algorithms in the prior art, and are not described herein again, and are all within the protection scope of the present application.
在本申请提供的一个实施例中,所述特征模型的数学表达为:In one embodiment provided by the present application, the mathematical expression of the feature model is:
vij=Mij(aij,qij)v ij =M ij (a ij ,q ij )
式中aij表示结构顺序位于横排第i个、竖排第j个的子训练图像,Mij为对应子训练图像aij训练得到的特征模型,qij为训练得到的特征模型参数,vij为通过特征模型Mij对子训练图像aij提取的子训练图像视觉特征。 Where a ij denotes a sub-training image in which the structural order is located in the i-th row and the j-th row in the horizontal row, M ij is a feature model trained corresponding to the sub-training image a ij , and q ij is a feature model parameter obtained by training, v Ij is a sub-training image visual feature extracted by the feature model M ij for the sub-training image a ij .
通过步骤S1041至S1044,完成了多模型训练,确定了特征模型及特征模型参数,接下来,将所述多个结构化子图像代入上述特征模型,即可计算得知每个所述结构化子图像的视觉特征。Through steps S1041 to S1044, multi-model training is completed, and the feature model and the feature model parameters are determined. Next, the plurality of structured sub-images are substituted into the feature model, and each of the structurators can be calculated. The visual characteristics of the image.
步骤S105:将所述多个结构化子图像的视觉特征进行结构化融合,获得结构化特征数据。Step S105: Structurally merging the visual features of the plurality of structured sub-images to obtain structured feature data.
通过步骤S104,已采用多模型训练获得的特征模型提取每个所述结构化子图像的视觉特征,接下来,将所述多个结构化子图像的视觉特征进行结构化融合,获得结构化特征数据。By step S104, the feature features obtained by the multi-model training are used to extract the visual features of each of the structured sub-images, and then the visual features of the plurality of structured sub-images are structurally fused to obtain the structured features. data.
在本申请提供的一个实施例中,所述将所述多个结构化子图像的视觉特征进行结构化融合,获得结构化特征数据,包括:In an embodiment provided by the present application, the structurally merging the visual features of the plurality of structured sub-images to obtain structured feature data includes:
根据构建多个结构化子图像时的确定的结构化基准点位置,将所述多个结构化子图像的视觉特征进行结构化融合,获得结构化特征数据,所述结构化特征数据包括特征空间关系和特征信息。And structurally merging the visual features of the plurality of structured sub-images according to the determined structured reference point positions when constructing the plurality of structured sub-images to obtain structured feature data, where the structured feature data includes the feature space Relationship and feature information.
仍以上述人脸图像的优选实施例为例,根据上述步骤S103确定的所述结构化基准点位置,对所述结构化子图像的视觉特征在空间上进行结构化融合,这样在空间平面上可以反映各所述结构化子图像的视觉特征基于所述结构化基准点位置的空间关系,而所述结构化子图像的视觉特征的特征轴则反映了各所述结构化子图像的特征信息,其长短代表了特征维度。请参考图6,其为本申请提供的特征结构化融合示意图,特征化基准点601位置的特征值图像602经过对应的特征模型抽取特征向量603,所述特征向量603经过结构化融合获得结构化特征数据604,由于结构化融合的过程保持了结构化基准点601相对于其他结构化基准点的空间位置关系,因此所述结构化特征数据604中也包含了特征空间关系和特征信息。Taking the preferred embodiment of the above-described face image as an example, the visual features of the structured sub-image are spatially structured and fused according to the structured reference point position determined in the above step S103, so that on the spatial plane It may be reflected that a visual feature of each of the structured sub-images is based on a spatial relationship of the structured reference point locations, and a feature axis of the visual features of the structured sub-images reflects feature information of each of the structured sub-images Its length represents the feature dimension. Please refer to FIG. 6 , which is a schematic diagram of feature structure fusion provided by the present application. The feature value image 602 of the feature reference point 601 is extracted by a corresponding feature model, and the feature vector 603 is structured by structural fusion. The feature data 604, since the process of structured merging maintains the spatial positional relationship of the structured reference point 601 with respect to other structured reference points, the structured feature data 604 also includes feature space relationships and feature information.
在本申请提供的一个实施例中,所述结构化特征数据的数学表达为:In one embodiment provided by the present application, the mathematical representation of the structured feature data is:
d(i,j,k)=vij(k)d(i,j,k)=v ij (k)
式中vij表示结构化子图像的视觉特征,k为第k维的数据,d为融合后获得的结构化特征数据。Where v ij represents the visual feature of the structured sub-image, k is the data of the k-th dimension, and d is the structured feature data obtained after the fusion.
步骤S106:采用结构化模型训练得到的模型,对所述结构化特征数据进行运算,获得图像特征数据。Step S106: The model obtained by training the structured model is used to perform operation on the structured feature data to obtain image feature data.
通过步骤S105,已将所述多个结构化子图像的视觉特征进行结构化融合, 获得结构化特征数据,接下来,采用结构化模型训练得到的模型,对所述结构化特征数据进行运算,获得图像特征数据。Structurally merging the visual features of the plurality of structured sub-images by step S105, The structured feature data is obtained. Next, the obtained model is trained by using the structured model, and the structured feature data is calculated to obtain image feature data.
所述结构化模型训练是上述步骤S1041至S1044描述的多模型训练的后续步骤,相关之处请参照上述步骤S1041至S1044的描述,此处不再赘述,以下对结构化模型训练进行说明。The structured model training is a subsequent step of the multi-model training described in the above steps S1041 to S1044. For related information, refer to the description of the above steps S1041 to S1044, and details are not described herein. The following describes the structured model training.
所述结构化模型训练是对结构化特征数据进行训练,在保持特征空间关系的同时,对特征信息进行更好的融合。在本申请提供的一个实施例中,所述结构化模型训练包括:The structured model training is to train the structured feature data, and the feature information is better integrated while maintaining the feature space relationship. In one embodiment provided by the present application, the structured model training includes:
将所述多个子训练图像视觉特征进行结构化融合,获得训练图像结构化特征数据;Performing structural merging of the plurality of sub-training image visual features to obtain training image structured feature data;
采用视觉特征学习算法对所述训练图像结构化特征数据进行结构化模型训练,获得结构化模型训练得到的模型。The structured feature model is trained on the structured image data of the training image by using the visual feature learning algorithm, and the model obtained by the structured model training is obtained.
在本申请提供的一个实施例中,所述结构化模型训练得到的模型的数学表达为:In an embodiment provided by the present application, the mathematical expression of the model obtained by the structured model training is:
v=M(d,q)v=M(d,q)
其中M为基于融合后的训练图像特征数据d进行结构化模型训练得到的模型,q为训练得到的模型参数,v为通过模型M对训练图像特征数据d融合得到的相应视觉特征。Where M is a model obtained by performing structured model training based on the fused training image feature data d, q is a model parameter obtained by training, and v is a corresponding visual feature obtained by merging the training image feature data d by the model M.
通过上述多模型训练,可以确定模型及模型参数,接下来,将所述结构化特征数据代入上述模型中的d,即可计算得到最终的图像特征数据v。Through the multi-model training described above, the model and model parameters can be determined. Next, by substituting the structured feature data into d in the above model, the final image feature data v can be calculated.
至此,通过步骤S101至步骤S106完成了本申请提供的图像特征提取方法实施例的流程,本申请中,通过构建结构化子图像保留了结构化子图像之间的空间位置信息,因此提取的所述结构化子图像的视觉特征同时包括特征空间关系和特征信息,在进行结构化融合时既保留了各视觉特征的描述性,又保留了各视觉特征的空间关系,从而最终获得的图像特征数据为特征向量,可以用特征向量间的特征距离描述不同图像之间的差异,又由于本方法中特征向量与模型在训练过程中更好的保持了图像的结构化特性,因此,所述图像特征数据具有更高的准确性和可辨识性。在图像识别尤其是人脸识别中应用本申请提供的图像特征提取方法,会具有更高的准确性,从而获得更好的识别效果。The process of the image feature extraction method provided by the present application is completed in steps S101 to S106. In the present application, the spatial position information between the structured sub-images is preserved by constructing the structured sub-image, and thus the extracted location is The visual features of the structured sub-image include both the feature space relationship and the feature information. When the structural fusion is performed, the descriptiveness of each visual feature is preserved, and the spatial relationship of each visual feature is preserved, so that the finally obtained image feature data is obtained. For feature vectors, the feature distance between feature vectors can be used to describe the difference between different images. Because the feature vector and model in this method better preserve the structural characteristics of the image during the training process, the image features are The data is more accurate and identifiable. Applying the image feature extraction method provided by the present application in image recognition, especially face recognition, has higher accuracy, thereby obtaining a better recognition effect.
通过以上步骤,已经提取出所述用户输入的图像的图像特征数据,接下来 可以利用所述图像特征数据对所述用户输入的图像进行识别,可用于判断所述用户输入的图像与某一图像的相似程度,或者判断某一图像数据库中是否有与所述用户输入的图像相似的图片,或者在某一图像数据库中筛选出与所述用户输入的图像相似的图片,在本申请提供的一个实施例中,所述图像特征提取方法还包括步骤:Through the above steps, the image feature data of the image input by the user has been extracted, and then The image input data may be used to identify the image input by the user, and may be used to determine whether the image input by the user is similar to a certain image, or determine whether there is an image input by the user in a certain image database. A similar picture, or a picture similar to the image input by the user is selected in a certain image database. In an embodiment provided by the present application, the image feature extraction method further includes the steps of:
将所述图像特征数据与预定的图像数据库中的各个预定图像特征数据依次进行比对;And sequentially comparing the image feature data with each predetermined image feature data in a predetermined image database;
输出比对结果。Output comparison results.
所述比对结果可以是所述用户输入的图像与预定的图像数据库中每个图片的相似程度,也可以是将相似程度达到预定阈值的预定的图像数据库中的图片及其信息等。在实际应用时,所述预定的图像数据库可以是公安追逃应用中的罪犯人脸数据库、考勤系统中的员工人脸数据库、会员管理系统中的会员人脸数据库或者明星脸检索系统中的明星人脸数据库等等,所述比对结果可以是所述用户输入的图像是否为在逃罪犯、所述用户输入的图像是否为已注册员工或会员、考勤人员的相貌是否与考勤系统中的记录一致,所述用户输入的图像与哪个明星的相貌相似等等。The comparison result may be a degree of similarity between the image input by the user and each picture in the predetermined image database, or may be a picture in a predetermined image database and its information and the like that bring the degree of similarity to a predetermined threshold. In practical application, the predetermined image database may be a criminal face database in a public security pursuit application, an employee face database in the attendance system, a member face database in the member management system, or a star in the star face retrieval system. a face database or the like, the comparison result may be whether the image input by the user is a fugitive, whether the image input by the user is a registered employee or a member, and whether the appearance of the attendant is consistent with the record in the attendance system. , the image input by the user is similar to the appearance of which star, and the like.
考虑到所述图像特征数据为向量,所述相似程度可以采用向量之间的距离来表征,距离越小,相似程度越高例如欧氏距离、Cosine距离或Joint Bayesian距离等。Considering that the image feature data is a vector, the degree of similarity can be characterized by the distance between the vectors. The smaller the distance, the higher the degree of similarity such as the Euclidean distance, the Cosine distance or the Joint Bayesian distance.
在本申请提供的一个实施例中,所述将所述图像特征数据与预定的图像数据库中的各个预定图像特征数据依次进行比对,包括:In an embodiment provided by the present application, the comparing the image feature data with each predetermined image feature data in a predetermined image database, including:
依次计算所述图像特征数据与预定的图像数据库中的各个预定图像特征数据之间的差值;Calculating a difference between the image feature data and each predetermined image feature data in a predetermined image database;
所述输出比对结果包括:The output comparison results include:
依次判断每个计算后的相似度是否大于预定的差值阈值;Determining, in turn, whether each calculated similarity is greater than a predetermined difference threshold;
若每个计算后的差值都大于预定的相似度阈值,则输出没有相似图像的信息,否则,则将与所述图像特征数据差值最小的预定图像特征数据对应的图像,和/或图像的信息输出。If each of the calculated differences is greater than a predetermined similarity threshold, information having no similar image is output, otherwise, an image corresponding to predetermined image feature data having the smallest difference from the image feature data, and/or an image is output. Information output.
其中所述计算所述图像特征数据与预定的图像数据库中的各个预定图像特征数据之间的差值的算法包括以下任一种: The algorithm for calculating a difference between the image feature data and each predetermined image feature data in a predetermined image database includes any one of the following:
欧氏距离计算方法、Cosine距离计算方法或Joint Bayesian距离计算方法。Euclidean distance calculation method, Cosine distance calculation method or Joint Bayesian distance calculation method.
以上,为本申请提供的一种图像特征提取方法实施例,与其相应的,本申请还提供了一种图像特征提取装置。请参考图7,其为本申请提供的一种图像特征提取装置实施例的示意图。由于装置实施例基本相似于方法实施例,所以描述得比较简单,相关之处参见方法实施例的部分说明即可。下述描述的装置实施例仅仅是示意性的。The above is an embodiment of an image feature extraction method provided by the present application. Correspondingly, the present application further provides an image feature extraction device. Please refer to FIG. 7, which is a schematic diagram of an embodiment of an image feature extraction apparatus provided by the present application. Since the device embodiment is basically similar to the method embodiment, the description is relatively simple, and the relevant parts can be referred to the description of the method embodiment. The device embodiments described below are merely illustrative.
本申请提供的一种图像特征提取装置实施例中,所述图像特征提取装置包括:图像接收单元701,用于接收用户输入的图像;配准单元702,用于对所述用户输入的图像进行配准,获得配准后的图像;子图像构建单元703,用于对所述配准后的图像构建多个结构化子图像;视觉特征提取单元704,用于采用多模型训练获得的特征模型提取每个所述结构化子图像的视觉特征;融合单元705,用于将所述多个结构化子图像的视觉特征进行结构化融合,获得结构化特征数据;运算单元706,用于采用结构化模型训练得到的模型,对所述结构化特征数据进行运算,获得图像特征数据。In an embodiment of the image feature extraction device provided by the present application, the image feature extraction device includes: an image receiving unit 701, configured to receive an image input by a user; and a registration unit 702, configured to perform an image input by the user Registration, obtaining a registered image; a sub-image construction unit 703 for constructing a plurality of structured sub-images for the registered image; a visual feature extraction unit 704 for using a multi-model training to obtain a feature model Extracting a visual feature of each of the structured sub-images; a fusing unit 705, configured to structurally fuse the visual features of the plurality of structured sub-images to obtain structured feature data; and an operation unit 706, configured to adopt a structure The model obtained by the model training is operated on the structured feature data to obtain image feature data.
可选的,所述配准单元702,包括:Optionally, the registration unit 702 includes:
基准点确定子单元,用于确定所述配准后的图像的结构化基准点位置;a reference point determining subunit for determining a structured reference point position of the registered image;
形状参数确定子单元,用于确定子图像的形状参数;a shape parameter determining subunit for determining a shape parameter of the sub image;
切割子单元,用于根据所述结构化基准点位置及所述子图像的形状参数,切割所述配准后的图像,获得多个结构化子图像。And a cutting subunit, configured to cut the registered image according to the structured reference point position and the shape parameter of the sub image to obtain a plurality of structured sub-images.
可选的,所述基准点确定子单元,包括:Optionally, the reference point determining subunit includes:
特征基准点确定子单元,用于根据图像特征点确定所述配准后的图像的结构化基准点位置;或者,a feature reference point determining subunit, configured to determine a structured reference point position of the registered image according to the image feature point; or
空间基准点确定子单元,用于根据空间位置确定所述配准后的图像的结构化基准点位置。A spatial reference point determining subunit is configured to determine a structured reference point position of the registered image based on the spatial location.
可选的,所述切割子单元采用的数学算法为:Optionally, the mathematical algorithm used by the cutting subunit is:
aij=C(a,pij(x,y),sij)a ij =C(a,p ij (x,y),s ij )
式中aij表示结构顺序位于横排第i个、竖排第j个的结构化子图像,C为结构化子图像的构建函数,a表示用户输入的图像,pij表示顺序位于横排第i个、竖排第j个的结构化基准点,pij(x,y)表示结构化基准点pij处于所述用户输入的 图像的坐标(x,y)处,sij表示结构化子图像的形状参数,包括矩形、圆形、椭圆形等任意平面形状及其尺寸。Where a ij denotes a structured sub-image in which the structural order is in the i-th row and j-th in the vertical row, C is a constructing function of the structured sub-image, a represents the image input by the user, and p ij represents the order in the horizontal row i, vertical j-th structured reference points, p ij (x, y) indicates that the structured reference point p ij is at the coordinates (x, y) of the image input by the user, and s ij represents the structuring Shape parameters of the image, including rectangular, circular, elliptical and other arbitrary planar shapes and their dimensions.
可选的,所述图像特征提取装置,还包括:多模型训练单元,用于通过多模型训练获得特征模型。Optionally, the image feature extraction device further includes: a multi-model training unit, configured to obtain a feature model by multi-model training.
可选的,所述多模型训练单元包括:Optionally, the multi-model training unit includes:
训练图像库选择子单元,用于选择预定的训练图像库;Training image library selection subunit for selecting a predetermined training image library;
训练图像配准子单元,用于将所述预定的训练图像库中的每个训练图像按照统一的配准方法进行配准,获得配准后的多个训练图像;And a training image registration sub-unit, configured to register each training image in the predetermined training image library according to a unified registration method, to obtain a plurality of the registered training images;
子训练图像构建子单元,用于对所述配准后的多个训练图像分别构建多个结构化子训练图像;a sub-training image construction sub-unit, configured to respectively construct a plurality of structured sub-training images for the plurality of registered training images;
特征模型获取子单元,用于采用视觉特征学习算法对所述多个结构化子训练图像进行特征模型训练以提取相应的多个子训练图像视觉特征,并获得特征模型。The feature model acquisition sub-unit is configured to perform feature model training on the plurality of structured sub-training images by using a visual feature learning algorithm to extract corresponding plurality of sub-training image visual features, and obtain a feature model.
可选的,所述特征模型获取子单元采用的视觉特征学习算法包括以下任一种:Optionally, the visual feature learning algorithm adopted by the feature model acquisition subunit includes any one of the following:
深度学习方法、boosting算法、svm算法或局部特征组合的学习算法。Learning algorithm for deep learning method, boosting algorithm, svm algorithm or local feature combination.
可选的,所述融合单元705包括:Optionally, the merging unit 705 includes:
基准点融合子单元,用于根据构建多个结构化子图像时的确定的结构化基准点位置,将所述多个结构化子图像的视觉特征进行结构化融合,获得结构化特征数据,所述结构化特征数据包括特征空间关系和特征信息。a reference point fusion subunit, configured to structurally fuse the visual features of the plurality of structured sub-images according to the determined structured reference point position when constructing the plurality of structured sub-images, to obtain structured feature data, The structured feature data includes feature space relationships and feature information.
可选的,所述图像特征提取装置还包括:Optionally, the image feature extraction device further includes:
结构化模型训练单元,用于通过结构化模型训练获得模型。A structured model training unit for training models through structured model training.
可选的,所述结构化模型训练单元包括:Optionally, the structured model training unit includes:
子训练图像融合子单元,用于将所述多个子训练图像视觉特征进行结构化融合,获得训练图像结构化特征数据;a sub-training image fusion sub-unit, configured to structurally fuse the plurality of sub-training image visual features to obtain training image structured feature data;
模型获取子单元,用于采用视觉特征学习算法对所述训练图像结构化特征数据进行结构化模型训练,获得结构化模型训练得到的模型。The model acquisition subunit is configured to perform structural model training on the structured image data of the training image by using a visual feature learning algorithm, and obtain a model obtained by training the structured model.
可选的,所述图像特征提取装置还包括:Optionally, the image feature extraction device further includes:
比对单元,用于将所述图像特征数据与预定的图像数据库中的各个预定图 像特征数据依次进行比对;a matching unit for using the image feature data with each predetermined map in a predetermined image database Performing alignments like feature data in sequence;
输出单元,用于输出比对结果。An output unit for outputting the comparison result.
可选的,所述比对单元包括:Optionally, the comparison unit includes:
差值计算子单元,用于依次计算所述图像特征数据与预定的图像数据库中的各个预定图像特征数据之间的差值;a difference calculation subunit, configured to sequentially calculate a difference between the image feature data and each predetermined image feature data in a predetermined image database;
所述输出单元包括:The output unit includes:
差值判断子单元,用于依次判断每个所述差值是否大于预定的差值阈值;a difference determining subunit, configured to sequentially determine whether each of the difference values is greater than a predetermined difference threshold;
信息输出单元,用于若每个所述差值都大于预定的相似度阈值,则输出没有相似图像的信息,否则,则将与所述图像特征数据差值最小的预定图像特征数据对应的图像,和/或图像的信息输出。An information output unit, configured to: if each of the differences is greater than a predetermined similarity threshold, output information without a similar image; otherwise, an image corresponding to predetermined image feature data having a minimum difference from the image feature data , and / or image information output.
可选的,所述比对单元计算所述图像特征数据与预定的图像数据库中的各个预定图像特征数据之间的差值的算法包括以下任一种:Optionally, the algorithm for calculating, by the comparing unit, the difference between the image feature data and each predetermined image feature data in the predetermined image database includes any one of the following:
欧氏距离计算方法、Cosine距离计算方法或Joint Bayesian距离计算方法。Euclidean distance calculation method, Cosine distance calculation method or Joint Bayesian distance calculation method.
以上,为本申请提供的一种图像特征提取装置实施例。The above is an embodiment of an image feature extraction device provided by the present application.
本申请还提供一种图像特征提取终端设备,包括:The application also provides an image feature extraction terminal device, including:
中央处理器;CPU;
输入输出单元;Input and output unit;
存储器;所述存储器中存储有本申请提供的图像特征提取方法;并在启动后能够根据上述方法运行。a memory; the image feature extraction method provided by the present application is stored in the memory; and can be run according to the above method after startup.
例如,所述客户端为一平板电脑,用户用平板电脑自拍一张照片或从相册中选择一张人脸照片,所述平板电脑即调用本申请提供的图像特征提取方法提取出照片的图像特征数据,并与预存的明星脸图像数据库中的图像进行比对,得到与所述照片相似度最高的明星图像,并调取所述明星的人物信息,然后将所述明星图像及人物信息在显示屏上输出。For example, the client is a tablet computer, and the user takes a photo with the tablet or selects a face photo from the album, and the tablet calls the image feature extraction method provided by the application to extract the image feature of the photo. Data is compared with the image in the pre-stored star face image database to obtain a star image with the highest similarity to the photo, and the character information of the star is retrieved, and then the star image and the person information are displayed. On-screen output.
由于本终端设备使用上述图像特征提取方法,相关之处请参见上述图像特征提取方法实施例的说明,此处不再赘述。For the description of the embodiment of the image feature extraction method, the description of the embodiment of the image feature extraction method is omitted here.
本申请还提供了一种图像特征提取系统,包括客户端和远端服务器,本系统部署有本申请提供的所述图像特征提取装置,在运行时,所述客户端拍摄图像和/或选取相册中的图像发送到远端服务器,所述远端服务器提取出图像特征 数据,并与预定的图像数据库中的图像进行比对,并将比对结果发送至所述客户端,最终由所述客户端输出比对结果。The present application also provides an image feature extraction system, including a client and a remote server. The system is provided with the image feature extraction device provided by the application. During operation, the client captures an image and/or selects an album. The image in the image is sent to the remote server, which extracts the image features The data is compared with an image in a predetermined image database, and the comparison result is sent to the client, and finally the comparison result is output by the client.
例如,所述客户端为一智能手机,用户用智能手机自拍一张照片或从相册中选择一张人脸照片,然后发送到远端服务器,远端服务器即调用本申请提供的图像特征提取方法提取出照片的图像特征数据,并与预存的明星脸图像数据库中的图像进行比对,得到与所述照片相似度最高的明星图像,并调取所述明星的人物信息,然后将所述明星图像及人物信息发送至所述客户端,最终在所述客户端的显示屏上输出。For example, the client is a smart phone, and the user takes a photo by using a smart phone or selects a face photo from the album, and then sends the photo to the remote server, and the remote server invokes the image feature extraction method provided by the application. Extracting image feature data of the photo, and comparing with the image in the pre-stored star face image database, obtaining a star image with the highest similarity to the photo, and retrieving the character information of the star, and then the star The image and character information are sent to the client and finally output on the display screen of the client.
由于本图像特征提取系统使用上述图像特征提取方法,相关之处请参见上述图像特征提取方法实施例的说明,此处不再赘述。For the image feature extraction system, the image feature extraction method is used. For details, refer to the description of the image feature extraction method embodiment, and details are not described herein.
本申请虽然以较佳实施例公开如上,但其并不是用来限定本申请,任何本领域技术人员在不脱离本申请的精神和范围内,都可以做出可能的变动和修改,因此本申请的保护范围应当以本申请权利要求所界定的范围为准。The present application is disclosed in the above preferred embodiments, but it is not intended to limit the present application, and any person skilled in the art can make possible changes and modifications without departing from the spirit and scope of the present application. The scope of protection should be based on the scope defined by the claims of the present application.
在一个典型的配置中,计算设备包括一个或多个处理器(CPU)、输入/输出接口、网络接口和内存。In a typical configuration, a computing device includes one or more processors (CPUs), input/output interfaces, network interfaces, and memory.
内存可能包括计算机可读介质中的非永久性存储器,随机存取存储器(RAM)和/或非易失性内存等形式,如只读存储器(ROM)或闪存(flash RAM)。内存是计算机可读介质的示例。The memory may include non-persistent memory, random access memory (RAM), and/or non-volatile memory in a computer readable medium, such as read only memory (ROM) or flash memory. Memory is an example of a computer readable medium.
1、计算机可读介质包括永久性和非永久性、可移动和非可移动媒体可以由任何方法或技术来实现信息存储。信息可以是计算机可读指令、数据结构、程序的模块或其他数据。计算机的存储介质的例子包括,但不限于相变内存(PRAM)、静态随机存取存储器(SRAM)、动态随机存取存储器(DRAM)、其他类型的随机存取存储器(RAM)、只读存储器(ROM)、电可擦除可编程只读存储器(EEPROM)、快闪记忆体或其他内存技术、只读光盘只读存储器(CD-ROM)、数字多功能光盘(DVD)或其他光学存储、磁盒式磁带,磁带磁磁盘存储或其他磁性存储设备或任何其他非传输介质,可用于存储可以被计算设备访问的信息。按照本文中的界定,计算机可读介质不包括非暂存电脑可读媒体(transitory media),如调制的数据信号和载波。1. Computer readable media including both permanent and non-persistent, removable and non-removable media may be implemented by any method or technology. The information can be computer readable instructions, data structures, modules of programs, or other data. Examples of computer storage media include, but are not limited to, phase change memory (PRAM), static random access memory (SRAM), dynamic random access memory (DRAM), other types of random access memory (RAM), read only memory. (ROM), electrically erasable programmable read only memory (EEPROM), flash memory or other memory technology, compact disk read only memory (CD-ROM), digital versatile disk (DVD) or other optical storage, Magnetic tape cartridges, magnetic tape storage or other magnetic storage devices or any other non-transportable media can be used to store information that can be accessed by a computing device. As defined herein, computer readable media does not include non-transitory computer readable media, such as modulated data signals and carrier waves.
2、本领域技术人员应明白,本申请的实施例可提供为方法、系统或计算机程序产品。因此,本申请可采用完全硬件实施例、完全软件实施例或结合软件 和硬件方面的实施例的形式。而且,本申请可采用在一个或多个其中包含有计算机可用程序代码的计算机可用存储介质(包括但不限于磁盘存储器、CD-ROM、光学存储器等)上实施的计算机程序产品的形式。 2. Those skilled in the art will appreciate that embodiments of the present application can be provided as a method, system, or computer program product. Therefore, the present application may employ an entirely hardware embodiment, an entirely software embodiment, or a combination of software. And in the form of an embodiment of the hardware aspect. Moreover, the application can take the form of a computer program product embodied on one or more computer-usable storage media (including but not limited to disk storage, CD-ROM, optical storage, etc.) including computer usable program code.

Claims (28)

  1. 一种图像特征提取方法,其特征在于,包括:An image feature extraction method, comprising:
    接收用户输入的图像;Receiving an image input by a user;
    对所述用户输入的图像进行配准,获得配准后的图像;Registering the image input by the user to obtain a registered image;
    对所述配准后的图像构建多个结构化子图像;Constructing a plurality of structured sub-images for the registered image;
    采用多模型训练获得的特征模型提取每个所述结构化子图像的视觉特征;Extracting a visual feature of each of the structured sub-images using a feature model obtained by multi-model training;
    将所述多个结构化子图像的视觉特征进行结构化融合,获得结构化特征数据;Structurally merging the visual features of the plurality of structured sub-images to obtain structured feature data;
    采用结构化模型训练得到的模型,对所述结构化特征数据进行运算,获得图像特征数据。The model obtained by the structured model training is used to calculate the structured feature data to obtain image feature data.
  2. 根据权利要求1所述的图像特征提取方法,其特征在于,所述对所述配准后的图像构建多个结构化子图像,包括:The image feature extraction method according to claim 1, wherein the constructing the plurality of structured sub-images to the registered image comprises:
    确定所述配准后的图像的结构化基准点位置;Determining a structured reference point location of the registered image;
    确定子图像的形状参数;Determining a shape parameter of the sub image;
    根据所述结构化基准点位置及所述子图像的形状参数,切割所述配准后的图像,获得多个结构化子图像。The registered image is cut according to the structured reference point position and the shape parameter of the sub-image to obtain a plurality of structured sub-images.
  3. 根据权利要求2所述的图像特征提取方法,其特征在于,所述确定所述配准后的图像的结构化基准点位置,包括:The image feature extraction method according to claim 2, wherein the determining the structured reference point position of the registered image comprises:
    根据图像特征点确定所述配准后的图像的结构化基准点位置;或者,Determining a structured reference point location of the registered image based on image feature points; or
    根据空间位置确定所述配准后的图像的结构化基准点位置。A structured reference point location of the registered image is determined based on the spatial location.
  4. 根据权利要求2所述的图像特征提取方法,其特征在于,所述根据所述结构化基准点位置及所述子图像的形状参数,切割所述配准后的图像,获得多个结构化子图像的数学算法为:The image feature extraction method according to claim 2, wherein the image is cut according to the structured reference point position and the shape parameter of the sub-image to obtain a plurality of structuring elements The mathematical algorithm of the image is:
    aij=C(a,pij(x,y),sij)a ij =C(a,p ij (x,y),s ij )
    式中aij表示结构顺序位于横排第i个、竖排第j个的结构化子图像,C为结构化子图像的构建函数,a表示用户输入的图像,pij表示顺序位于横排第i个、竖排第j个的结构化基准点,pij(x,y)表示结构化基准点pij处于所述用户输入的图像的坐标(x,y)处,sij表示结构化子图像的形状参数,包括矩形、圆形、椭圆形等任意平面形状及其尺寸。 Where a ij denotes a structured sub-image in which the structural order is in the i-th row and j-th in the vertical row, C is a constructing function of the structured sub-image, a represents the image input by the user, and p ij represents the order in the horizontal row i, vertical jth structured reference points, p ij (x, y) indicates that the structured reference point p ij is at the coordinates (x, y) of the image input by the user, and s ij represents the structuring Shape parameters of the image, including rectangular, circular, elliptical and other arbitrary planar shapes and their dimensions.
  5. 根据权利要求1所述的图像特征提取方法,其特征在于,所述多模型训练获得的特征模型是通过以下方法获得的:The image feature extraction method according to claim 1, wherein the feature model obtained by the multi-model training is obtained by the following method:
    选择预定的训练图像库;Select a predetermined training image library;
    将所述预定的训练图像库中的每个训练图像按照统一的配准方法进行配准,获得配准后的多个训练图像;Registering each training image in the predetermined training image library according to a unified registration method to obtain a plurality of registered training images;
    对所述配准后的多个训练图像分别构建多个结构化子训练图像;Constructing a plurality of structured sub-training images for the plurality of registered training images;
    采用视觉特征学习算法对所述多个结构化子训练图像进行特征模型训练以提取相应的多个子训练图像视觉特征,并获得特征模型。The plurality of structured sub-training images are subjected to feature model training by using a visual feature learning algorithm to extract corresponding plurality of sub-training image visual features, and a feature model is obtained.
  6. 根据权利要求5所述的图像特征提取方法,其特征在于,所述视觉特征学习算法包括以下任一种:The image feature extraction method according to claim 5, wherein the visual feature learning algorithm comprises any one of the following:
    深度学习方法、boosting算法、svm算法或局部特征组合的学习算法。Learning algorithm for deep learning method, boosting algorithm, svm algorithm or local feature combination.
  7. 根据权利要求5所述的图像特征提取方法,其特征在于,所述特征模型的数学表达为:The image feature extraction method according to claim 5, wherein the mathematical expression of the feature model is:
    vij=Mij(aij,qij)v ij =M ij (a ij ,q ij )
    式中aij表示结构顺序位于横排第i个、竖排第j个的子训练图像,Mij为对应子训练图像aij上训练得到的特征模型,qij为训练得到的特征模型参数,vij为通过特征模型Mij对子训练图像aij提取的子训练图像视觉特征。Where a ij denotes a sub-training image in which the structural order is located in the i-th row and the j-th row in the horizontal row, M ij is a feature model trained on the corresponding sub-training image a ij , and q ij is a feature model parameter obtained by training. v ij is a sub-training image visual feature extracted by the feature model M ij for the sub-training image a ij .
  8. 根据权利要求1所述的图像特征提取方法,其特征在于,所述将所述多个结构化子图像的视觉特征进行结构化融合,获得结构化特征数据,包括:The image feature extraction method according to claim 1, wherein the structurally merging the visual features of the plurality of structured sub-images to obtain structured feature data comprises:
    根据构建多个结构化子图像时的确定的结构化基准点位置,将所述多个结构化子图像的视觉特征进行结构化融合,获得结构化特征数据,所述结构化特征数据包括特征空间关系和特征信息。And structurally merging the visual features of the plurality of structured sub-images according to the determined structured reference point positions when constructing the plurality of structured sub-images to obtain structured feature data, where the structured feature data includes the feature space Relationship and feature information.
  9. 根据权利要求8所述的图像特征提取方法,其特征在于,所述结构化特征数据的数学表达为:The image feature extraction method according to claim 8, wherein the mathematical expression of the structured feature data is:
    d(i,j,k)=vij(k)d(i,j,k)=v ij (k)
    式中vij表示结构化子图像的视觉特征,k为第k维的数据,d为融合后的结构化特征数据。Where v ij represents the visual feature of the structured sub-image, k is the data of the k-th dimension, and d is the structured feature data after the fusion.
  10. 根据权利要求5所述的图像特征提取方法,其特征在于,所述结构化模型训练得到的模型是通过以下方式获得的: The image feature extraction method according to claim 5, wherein the model obtained by the structured model training is obtained by:
    将所述多个子训练图像视觉特征进行结构化融合,获得训练图像结构化特征数据;Performing structural merging of the plurality of sub-training image visual features to obtain training image structured feature data;
    采用视觉特征学习算法对所述训练图像结构化特征数据进行结构化模型训练,获得结构化模型训练得到的模型。The structured feature model is trained on the structured image data of the training image by using the visual feature learning algorithm, and the model obtained by the structured model training is obtained.
  11. 根据权利要求5所述的图像特征提取方法,其特征在于,所述结构化模型训练得到的模型的数学表达为:The image feature extraction method according to claim 5, wherein the mathematical expression of the model obtained by the structured model training is:
    v=M(d,q)v=M(d,q)
    其中M为基于融合后的训练图像特征数据d进行结构化模型训练得到的模型,q为训练得到的模型参数,v为通过模型M对训练图像特征数据d融合得到的相应视觉特征。Where M is a model obtained by performing structured model training based on the fused training image feature data d, q is a model parameter obtained by training, and v is a corresponding visual feature obtained by merging the training image feature data d by the model M.
  12. 根据权利要求1所述的图像特征提取方法,其特征在于,还包括:The image feature extraction method according to claim 1, further comprising:
    将所述图像特征数据与预定的图像数据库中的各个预定图像特征数据依次进行比对;And sequentially comparing the image feature data with each predetermined image feature data in a predetermined image database;
    输出比对结果。Output comparison results.
  13. 根据权利要求12所述的图像特征提取方法,其特征在于,所述将所述图像特征数据与预定的图像数据库中的各个预定图像特征数据依次进行比对,包括:The image feature extraction method according to claim 12, wherein the comparing the image feature data with each predetermined image feature data in a predetermined image database, comprises:
    依次计算所述图像特征数据与预定的图像数据库中的各个预定图像特征数据之间的差值;Calculating a difference between the image feature data and each predetermined image feature data in a predetermined image database;
    所述输出比对结果包括:The output comparison results include:
    依次判断每个所述差值是否大于预定的差值阈值;Determining, in turn, whether each of the difference values is greater than a predetermined difference threshold;
    若每个所述差值都大于预定的相似度阈值,则输出没有相似图像的信息,否则,则将与所述图像特征数据差值最小的预定图像特征数据对应的图像,和/或图像的信息输出。If each of the differences is greater than a predetermined similarity threshold, information having no similar image is output, otherwise, an image corresponding to predetermined image feature data having the smallest difference from the image feature data, and/or an image Information output.
  14. 根据权利要求13所述的图像特征提取方法,其特征在于,所述计算所述图像特征数据与预定的图像数据库中的各个预定图像特征数据之间的差值的算法包括以下任一种:The image feature extraction method according to claim 13, wherein the algorithm for calculating a difference between the image feature data and each predetermined image feature data in a predetermined image database comprises any one of the following:
    欧氏距离计算方法、Cosine距离计算方法或Joint Bayesian距离计算方法。Euclidean distance calculation method, Cosine distance calculation method or Joint Bayesian distance calculation method.
  15. 根据权利要求1至14任一项所述的图像特征提取方法,其特征在于, 所述图像包括:人脸图像。The image feature extraction method according to any one of claims 1 to 14, wherein The image includes: a face image.
  16. 一种图像特征提取装置,其特征在于,包括:An image feature extraction device, comprising:
    图像接收单元,用于接收用户输入的图像;An image receiving unit, configured to receive an image input by a user;
    配准单元,用于对所述用户输入的图像进行配准,获得配准后的图像;a registration unit, configured to register an image input by the user to obtain a registered image;
    子图像构建单元,用于对所述配准后的图像构建多个结构化子图像;a sub-image construction unit, configured to construct a plurality of structured sub-images on the registered image;
    视觉特征提取单元,用于采用多模型训练获得的特征模型提取每个所述结构化子图像的视觉特征;a visual feature extraction unit, configured to extract a visual feature of each of the structured sub-images by using a feature model obtained by multi-model training;
    融合单元,用于将所述多个结构化子图像的视觉特征进行结构化融合,获得结构化特征数据;a merging unit, configured to structurally fuse the visual features of the plurality of structured sub-images to obtain structured feature data;
    运算单元,用于采用结构化模型训练得到的模型,对所述结构化特征数据进行运算,获得图像特征数据。The operation unit is configured to use the model obtained by the structural model training, and perform operation on the structured feature data to obtain image feature data.
  17. 根据权利要求16所述的图像特征提取装置,其特征在于,所述配准单元,包括:The image feature extraction device according to claim 16, wherein the registration unit comprises:
    基准点确定子单元,用于确定所述配准后的图像的结构化基准点位置;a reference point determining subunit for determining a structured reference point position of the registered image;
    形状参数确定子单元,用于确定子图像的形状参数;a shape parameter determining subunit for determining a shape parameter of the sub image;
    切割子单元,用于根据所述结构化基准点位置及所述子图像的形状参数,切割所述配准后的图像,获得多个结构化子图像。And a cutting subunit, configured to cut the registered image according to the structured reference point position and the shape parameter of the sub image to obtain a plurality of structured sub-images.
  18. 根据权利要求17所述的图像特征提取装置,其特征在于,所述基准点确定子单元,包括:The image feature extraction device according to claim 17, wherein the reference point determination subunit comprises:
    特征基准点确定子单元,用于根据图像特征点确定所述配准后的图像的结构化基准点位置;或者,a feature reference point determining subunit, configured to determine a structured reference point position of the registered image according to the image feature point; or
    空间基准点确定子单元,用于根据空间位置确定所述配准后的图像的结构化基准点位置。A spatial reference point determining subunit is configured to determine a structured reference point position of the registered image based on the spatial location.
  19. 根据权利要求17所述的图像特征提取装置,其特征在于,所述切割子单元采用的数学算法为:The image feature extraction device according to claim 17, wherein the mathematical algorithm used by the cutting subunit is:
    aij=C(a,pij(x,y),sij)a ij =C(a,p ij (x,y),s ij )
    式中aij表示结构顺序位于横排第i个、竖排第j个的结构化子图像,C为结构化子图像的构建函数,a表示用户输入的图像,pij表示顺序位于横排第i个、竖排第j个的结构化基准点,pij(x,y)表示结构化基准点pij处于所述用户输入的 图像的坐标(x,y)处,sij表示结构化子图像的形状参数,包括矩形、圆形、椭圆形等任意平面形状及其尺寸。Where a ij denotes a structured sub-image in which the structural order is in the i-th row and j-th in the vertical row, C is a constructing function of the structured sub-image, a represents the image input by the user, and p ij represents the order in the horizontal row i, vertical j-th structured reference points, p ij (x, y) indicates that the structured reference point p ij is at the coordinates (x, y) of the image input by the user, and s ij represents the structuring Shape parameters of the image, including rectangular, circular, elliptical and other arbitrary planar shapes and their dimensions.
  20. 根据权利要求16所述的图像特征提取装置,其特征在于,还包括:The image feature extraction device according to claim 16, further comprising:
    多模型训练单元,用于通过多模型训练获得特征模型;a multi-model training unit for obtaining a feature model by multi-model training;
    所述多模型训练单元包括:The multi-model training unit includes:
    训练图像库选择子单元,用于选择预定的训练图像库;Training image library selection subunit for selecting a predetermined training image library;
    训练图像配准子单元,用于将所述预定的训练图像库中的每个训练图像按照统一的配准方法进行配准,获得配准后的多个训练图像;And a training image registration sub-unit, configured to register each training image in the predetermined training image library according to a unified registration method, to obtain a plurality of the registered training images;
    子训练图像构建子单元,用于对所述配准后的多个训练图像分别构建多个结构化子训练图像;a sub-training image construction sub-unit, configured to respectively construct a plurality of structured sub-training images for the plurality of registered training images;
    特征模型获取子单元,用于采用视觉特征学习算法对所述多个结构化子训练图像进行特征模型训练以提取相应的多个子训练图像视觉特征,并获得特征模型。The feature model acquisition sub-unit is configured to perform feature model training on the plurality of structured sub-training images by using a visual feature learning algorithm to extract corresponding plurality of sub-training image visual features, and obtain a feature model.
  21. 根据权利要求20所述的图像特征提取装置,其特征在于,所述特征模型获取子单元采用的视觉特征学习算法包括以下任一种:The image feature extraction device according to claim 20, wherein the visual feature learning algorithm adopted by the feature model acquisition subunit comprises any one of the following:
    深度学习方法、boosting算法、svm算法或局部特征组合的学习算法。Learning algorithm for deep learning method, boosting algorithm, svm algorithm or local feature combination.
  22. 根据权利要求16所述的图像特征提取装置,其特征在于,所述融合单元包括:The image feature extraction device according to claim 16, wherein the fusion unit comprises:
    基准点融合子单元,用于根据构建多个结构化子图像时的确定的结构化基准点位置,将所述多个结构化子图像的视觉特征进行结构化融合,获得结构化特征数据,所述结构化特征数据包括特征空间关系和特征信息。a reference point fusion subunit, configured to structurally fuse the visual features of the plurality of structured sub-images according to the determined structured reference point position when constructing the plurality of structured sub-images, to obtain structured feature data, The structured feature data includes feature space relationships and feature information.
  23. 根据权利要求20所述的图像特征提取装置,其特征在于,还包括:The image feature extraction device according to claim 20, further comprising:
    结构化模型训练单元,用于通过结构化模型训练获得模型;a structured model training unit for training a model through a structured model;
    所述结构化模型训练单元包括:The structured model training unit includes:
    子训练图像融合子单元,用于将所述多个子训练图像视觉特征进行结构化融合,获得训练图像结构化特征数据;a sub-training image fusion sub-unit, configured to structurally fuse the plurality of sub-training image visual features to obtain training image structured feature data;
    模型获取子单元,用于采用视觉特征学习算法对所述训练图像结构化特征数据进行结构化模型训练,获得结构化模型训练得到的模型。The model acquisition subunit is configured to perform structural model training on the structured image data of the training image by using a visual feature learning algorithm, and obtain a model obtained by training the structured model.
  24. 根据权利要求16所述的图像特征提取装置,其特征在于,还包括: The image feature extraction device according to claim 16, further comprising:
    比对单元,用于将所述图像特征数据与预定的图像数据库中的各个预定图像特征数据依次进行比对;a comparison unit, configured to sequentially compare the image feature data with each predetermined image feature data in a predetermined image database;
    输出单元,用于输出比对结果。An output unit for outputting the comparison result.
  25. 根据权利要求24所述的图像特征提取装置,其特征在于,所述比对单元包括:The image feature extraction device according to claim 24, wherein the comparison unit comprises:
    差值计算子单元,用于依次计算所述图像特征数据与预定的图像数据库中的各个预定图像特征数据之间的差值;a difference calculation subunit, configured to sequentially calculate a difference between the image feature data and each predetermined image feature data in a predetermined image database;
    所述输出单元包括:The output unit includes:
    差值判断子单元,用于依次判断每个所述差值是否大于预定的差值阈值;a difference determining subunit, configured to sequentially determine whether each of the difference values is greater than a predetermined difference threshold;
    信息输出单元,用于若每个所述差值都大于预定的相似度阈值,则输出没有相似图像的信息,否则,则将与所述图像特征数据差值最小的预定图像特征数据对应的图像,和/或图像的信息输出。An information output unit, configured to: if each of the differences is greater than a predetermined similarity threshold, output information without a similar image; otherwise, an image corresponding to predetermined image feature data having a minimum difference from the image feature data , and / or image information output.
  26. 根据权利要求25所述的图像特征提取装置,其特征在于,所述比对单元计算所述图像特征数据与预定的图像数据库中的各个预定图像特征数据之间的差值的算法包括以下任一种:The image feature extraction device according to claim 25, wherein the algorithm for calculating a difference between the image feature data and each predetermined image feature data in a predetermined image database by the comparison unit includes any of the following Kind:
    欧氏距离计算方法、Cosine距离计算方法或Joint Bayesian距离计算方法。Euclidean distance calculation method, Cosine distance calculation method or Joint Bayesian distance calculation method.
  27. 一种图像特征提取终端设备,包括:An image feature extraction terminal device includes:
    中央处理器;CPU;
    输入输出单元;Input and output unit;
    存储器;所述存储器中存储有权利要求1至权利要求15所述的图像特征提取方法;并在启动后能够根据上述方法运行。a memory; the image feature extraction method according to claim 1 to claim 15 is stored in the memory; and can be operated according to the above method after startup.
  28. 一种图像特征提取系统,包括客户端和远端服务器,其特征在于,使用权利要求16至权利要求26所述的图像特征提取装置,所述客户端拍摄图像和/或选取相册中的图像发送到远端服务器,所述远端服务器提取出图像特征数据,并与预定的图像数据库中的图像进行比对,并将比对结果发送至所述客户端,最终由所述客户端输出比对结果。 An image feature extraction system comprising a client and a remote server, characterized by using the image feature extraction device of claim 16 to claim 26, the client capturing an image and/or selecting an image transmission in an album Going to the remote server, the remote server extracts the image feature data, compares it with the image in the predetermined image database, and sends the comparison result to the client, and finally outputs the comparison by the client. result.
PCT/CN2016/095524 2015-08-26 2016-08-16 Image feature extraction method, apparatus, terminal device, and system WO2017032243A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN201510531886.7A CN106485186B (en) 2015-08-26 2015-08-26 Image feature extraction method and device, terminal equipment and system
CN201510531886.7 2015-08-26

Publications (1)

Publication Number Publication Date
WO2017032243A1 true WO2017032243A1 (en) 2017-03-02

Family

ID=58099586

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2016/095524 WO2017032243A1 (en) 2015-08-26 2016-08-16 Image feature extraction method, apparatus, terminal device, and system

Country Status (2)

Country Link
CN (1) CN106485186B (en)
WO (1) WO2017032243A1 (en)

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108875530A (en) * 2018-01-12 2018-11-23 北京旷视科技有限公司 Vivo identification method, vivo identification equipment, electronic equipment and storage medium
CN109522778A (en) * 2017-09-20 2019-03-26 顾泽苍 A kind of image-recognizing method can reach image understanding level
CN110569765A (en) * 2019-08-02 2019-12-13 北京旷视科技有限公司 picture identification method, picture comparison method and device
CN112241704A (en) * 2020-10-16 2021-01-19 百度(中国)有限公司 Method and device for judging portrait infringement, electronic equipment and storage medium
CN113256822A (en) * 2020-02-11 2021-08-13 阿里巴巴集团控股有限公司 Spatial relationship prediction, data processing method, device and storage medium
CN113408208A (en) * 2021-06-25 2021-09-17 成都欧珀通信科技有限公司 Model training method, information extraction method, related device and storage medium

Families Citing this family (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108319894A (en) * 2017-12-28 2018-07-24 杭州乔戈里科技有限公司 Fruit recognition methods based on deep learning and device
CN109064578B (en) * 2018-09-12 2020-11-03 山西巨擘天浩科技有限公司 Attendance system and method based on cloud service
CN112184843B (en) * 2020-11-09 2021-06-29 新相微电子(上海)有限公司 Redundant data removing system and method for image data compression
CN112712501B (en) * 2020-12-28 2021-10-26 江苏合泰飞梵科技有限公司 Rearview mirror assembly production method based on artificial intelligence

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101477692A (en) * 2009-02-13 2009-07-08 阿里巴巴集团控股有限公司 Method and apparatus for image characteristic extraction
WO2010150156A1 (en) * 2009-06-24 2010-12-29 Koninklijke Philips Electronics N.V. Establishing a contour of a structure based on image information
CN102654903A (en) * 2011-03-04 2012-09-05 井维兰 Face comparison method
CN102663351A (en) * 2012-03-16 2012-09-12 江南大学 Face characteristic point automation calibration method based on conditional appearance model
CN103218827A (en) * 2013-03-21 2013-07-24 上海交通大学 Contour tracing method based on shape-transmitting united division and image-matching correction
CN103886589A (en) * 2014-02-27 2014-06-25 四川农业大学 Goal-oriented automatic high-precision edge extraction method

Family Cites Families (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101599121A (en) * 2009-06-30 2009-12-09 徐勇 The authenticating colorized face images system and method
CN101770578B (en) * 2010-03-24 2011-07-27 上海交通大学 Image characteristic extraction method
CN102592136B (en) * 2011-12-21 2013-10-16 东南大学 Three-dimensional human face recognition method based on intermediate frequency information in geometry image
CN102722906B (en) * 2012-05-23 2015-01-07 温州大学 Feature-based top-down image modeling method
CN102903126B (en) * 2012-08-08 2015-11-04 公安部第三研究所 The system and method for a kind of video image texture feature extraction and structural description
CN103093237B (en) * 2013-01-15 2015-12-23 中国科学院自动化研究所 A kind of method for detecting human face of structure based model
CN104361131B (en) * 2014-12-08 2018-06-15 黑龙江大学 The method for building up of four-dimensional faceform's database

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101477692A (en) * 2009-02-13 2009-07-08 阿里巴巴集团控股有限公司 Method and apparatus for image characteristic extraction
WO2010150156A1 (en) * 2009-06-24 2010-12-29 Koninklijke Philips Electronics N.V. Establishing a contour of a structure based on image information
CN102654903A (en) * 2011-03-04 2012-09-05 井维兰 Face comparison method
CN102663351A (en) * 2012-03-16 2012-09-12 江南大学 Face characteristic point automation calibration method based on conditional appearance model
CN103218827A (en) * 2013-03-21 2013-07-24 上海交通大学 Contour tracing method based on shape-transmitting united division and image-matching correction
CN103886589A (en) * 2014-02-27 2014-06-25 四川农业大学 Goal-oriented automatic high-precision edge extraction method

Cited By (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109522778A (en) * 2017-09-20 2019-03-26 顾泽苍 A kind of image-recognizing method can reach image understanding level
CN108875530A (en) * 2018-01-12 2018-11-23 北京旷视科技有限公司 Vivo identification method, vivo identification equipment, electronic equipment and storage medium
CN110569765A (en) * 2019-08-02 2019-12-13 北京旷视科技有限公司 picture identification method, picture comparison method and device
CN110569765B (en) * 2019-08-02 2022-04-22 北京旷视科技有限公司 Picture identification method, picture comparison method and device
CN113256822A (en) * 2020-02-11 2021-08-13 阿里巴巴集团控股有限公司 Spatial relationship prediction, data processing method, device and storage medium
CN113256822B (en) * 2020-02-11 2024-02-13 阿里巴巴集团控股有限公司 Spatial relationship prediction, data processing method, device and storage medium
CN112241704A (en) * 2020-10-16 2021-01-19 百度(中国)有限公司 Method and device for judging portrait infringement, electronic equipment and storage medium
CN113408208A (en) * 2021-06-25 2021-09-17 成都欧珀通信科技有限公司 Model training method, information extraction method, related device and storage medium
CN113408208B (en) * 2021-06-25 2023-06-09 成都欧珀通信科技有限公司 Model training method, information extraction method, related device and storage medium

Also Published As

Publication number Publication date
CN106485186B (en) 2020-02-18
CN106485186A (en) 2017-03-08

Similar Documents

Publication Publication Date Title
WO2017032243A1 (en) Image feature extraction method, apparatus, terminal device, and system
CN109214343B (en) Method and device for generating face key point detection model
CN109284733B (en) Shopping guide negative behavior monitoring method based on yolo and multitask convolutional neural network
WO2019218824A1 (en) Method for acquiring motion track and device thereof, storage medium, and terminal
US9892344B1 (en) Activation layers for deep learning networks
US10616475B2 (en) Photo-taking prompting method and apparatus, an apparatus and non-volatile computer storage medium
CN106203242B (en) Similar image identification method and equipment
WO2017101267A1 (en) Method for identifying living face, terminal, server, and storage medium
TW201911130A (en) Method and device for remake image recognition
WO2017088432A1 (en) Image recognition method and device
US20190325197A1 (en) Methods and apparatuses for searching for target person, devices, and media
CN108734185B (en) Image verification method and device
US9626552B2 (en) Calculating facial image similarity
CN109614910B (en) Face recognition method and device
CN110427859A (en) A kind of method for detecting human face, device, electronic equipment and storage medium
US9171195B1 (en) Recognizing three-dimensional objects
JP2018524678A (en) Business discovery from images
Chen et al. Shape-Former: Bridging CNN and Transformer via ShapeConv for multimodal image matching
US10489636B2 (en) Lip movement capturing method and device, and storage medium
WO2020073601A1 (en) Goods recognition method, goods recognition apparatus, and storage medium
CN111814620A (en) Face image quality evaluation model establishing method, optimization method, medium and device
US9129152B2 (en) Exemplar-based feature weighting
US9633272B2 (en) Real time object scanning using a mobile phone and cloud-based visual search engine
CN109886223B (en) Face recognition method, bottom library input method and device and electronic equipment
US20230326173A1 (en) Image processing method and apparatus, and computer-readable storage medium

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 16838502

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 16838502

Country of ref document: EP

Kind code of ref document: A1