CN110147796A - Image matching method and device - Google Patents

Image matching method and device Download PDF

Info

Publication number
CN110147796A
CN110147796A CN201810144894.XA CN201810144894A CN110147796A CN 110147796 A CN110147796 A CN 110147796A CN 201810144894 A CN201810144894 A CN 201810144894A CN 110147796 A CN110147796 A CN 110147796A
Authority
CN
China
Prior art keywords
image
images
recognized
sample
convolutional layer
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201810144894.XA
Other languages
Chinese (zh)
Inventor
张鼎
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Hangzhou Hikvision Digital Technology Co Ltd
Original Assignee
Hangzhou Hikvision Digital Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Hangzhou Hikvision Digital Technology Co Ltd filed Critical Hangzhou Hikvision Digital Technology Co Ltd
Priority to CN201810144894.XA priority Critical patent/CN110147796A/en
Publication of CN110147796A publication Critical patent/CN110147796A/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/20Image preprocessing
    • G06V10/25Determination of region of interest [ROI] or a volume of interest [VOI]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/44Local feature extraction by analysis of parts of the pattern, e.g. by detecting edges, contours, loops, corners, strokes or intersections; Connectivity analysis, e.g. of connected components
    • G06V10/443Local feature extraction by analysis of parts of the pattern, e.g. by detecting edges, contours, loops, corners, strokes or intersections; Connectivity analysis, e.g. of connected components by matching or filtering

Abstract

The invention discloses a kind of image matching method and devices, belong to field of computer technology.The described method includes: obtaining images to be recognized and benchmark image;Model is extracted based on fisrt feature, in benchmark image, extracts fisrt feature image, and model is extracted based on second feature, in images to be recognized, extracts the location information of the characteristic image and each candidate object of multiple candidate objects in the images to be recognized;In the characteristic image of the multiple candidate object, the determining and highest second feature image of the fisrt feature images match degree;By location information of the corresponding candidate object of the second feature image in the images to be recognized, it is determined as the location information in the region to match in the images to be recognized with the benchmark image, and by the matching degree of the second feature image and the fisrt feature image, it is determined as the matching degree of image in the benchmark image and the region.Using the present invention, the efficiency of images match can be made to improve.

Description

Image matching method and device
Technical field
The present invention relates to field of computer technology, in particular to a kind of image matching method and device.
Background technique
Images match identifies same or similar silhouette target in two width or several pictures by certain matching algorithm Method, have a very wide range of applications in fields such as computer vision, pattern-recognitions, such as recognition of face, three-dimensional reconstruction, object Identification, motion tracking and panorama sketch creation etc..
By taking recognition of face as an example, the general step of traditional image matching method is: firstly, determining face figure to be identified Then picture and benchmark face image select a kind of Feature Descriptor, the characteristic point of two images is calculated using Feature Descriptor, Then, the characteristic point for comparing two images determines similar characteristic point between two images using arest neighbors method, by every two Similar characteristic point is determined as matching pair, finally utilizes certain geological information constraint (epipolar geom etry, affine transformation, single strain Change) removal mistake matching pair, obtain correctly match pair, finally calculate correctly matching pair matching degree, merge these Matching degree obtains final matching degree, so that it may determine the matching degree of facial image to be identified and benchmark face image, then The location information of the correct matching pair of analysis, determines the position in matched region in benchmark face image and facial image to be identified Confidence breath.
In the implementation of the present invention, inventor find the relevant technologies the prior art has at least the following problems:
Based on above-mentioned image matching method, the process of characteristic point is calculated using Feature Descriptor, calculation amount is larger, it is time-consuming compared with Long, leading to the efficiency of images match reduces.
Summary of the invention
In order to solve problems in the prior art, the embodiment of the invention provides a kind of image matching method and devices.It is described Technical solution is as follows:
According to a first aspect of the embodiments of the present invention, a kind of image matching method is provided, which comprises
Obtain images to be recognized and benchmark image;
Model is extracted based on fisrt feature, in the benchmark image, extracts fisrt feature image, and be based on second feature Model is extracted, in the images to be recognized, the characteristic image and each candidate object for extracting multiple candidate objects are described Location information in images to be recognized;
In the characteristic image of the multiple candidate object, the determining and fisrt feature images match degree highest second Characteristic image;
By location information of the corresponding candidate object of the second feature image in the images to be recognized, it is determined as institute State the location information in the region to match in images to be recognized with the benchmark image, and by the second feature image with it is described The matching degree of fisrt feature image is determined as the matching degree of image in the benchmark image and the region.
Optionally, it includes the first convolutional layer that the fisrt feature, which extracts model,;
It is described that model is extracted based on fisrt feature, in the benchmark image, extract fisrt feature image, comprising:
The benchmark image is inputted into first convolutional layer, obtains fisrt feature image.
Optionally, it includes the second convolutional layer, region of interest ROI pond layer and candidate regions that the second feature, which extracts model, Domain generation module RPN;
It is described that model is extracted based on second feature, in the images to be recognized, extract the characteristic pattern of multiple candidate objects The location information of picture and each candidate object in the images to be recognized, comprising:
The images to be recognized is inputted into second convolutional layer, obtains third feature image;
The third feature image is inputted into the RPN, obtains position of multiple candidate objects in the images to be recognized Confidence breath;
Location information and the third feature image of the multiple candidate object in the images to be recognized is defeated Enter the pond ROI layer, obtains the characteristic image of the multiple candidate object.
Optionally, before the acquisition images to be recognized and benchmark image, further includes:
Obtain multiple training samples, wherein each training sample includes sample images to be recognized, sample benchmark image, more A sample position information and sample matches degree, wherein the multiple sample position information is multiple sample candidate objects described Location information in sample images to be recognized, the sample matches degree are the sample benchmark image and sample images to be recognized phase The matching degree of image in matched region;
It is inputted using the sample images to be recognized and the sample benchmark image as training, the multiple sample bit confidence Breath and the sample matches degree as output reference value, to initial first convolutional layer, initial second convolutional layer and Initial R PN into Row training, first convolutional layer, second convolutional layer and the RPN after being trained.
Optionally, described to be inputted using the sample images to be recognized and the sample benchmark image as training, it is described more A sample position information and the sample matches degree are as output reference value, to initial first convolutional layer, initial second convolutional layer And Initial R PN is trained, first convolutional layer, second convolutional layer and the RPN after being trained, packet It includes:
It is inputted using the sample images to be recognized and the sample benchmark image as training, the multiple sample bit confidence Breath and the sample matches degree are as output reference value, based on comparison loss function, to initial first convolutional layer, initial volume Two Lamination is trained, and based on detection frame regression function Smooth L1 Loss and softmax loss, is instructed to Initial R PN Practice, first convolutional layer, second convolutional layer and the RPN after being trained.
According to a second aspect of the embodiments of the present invention, a kind of image matching apparatus is provided, described device includes:
First obtains module, for obtaining images to be recognized and benchmark image;
Extraction module in the benchmark image, extracts fisrt feature image for extracting model based on fisrt feature, And model is extracted based on second feature, in the images to be recognized, extract the characteristic image of multiple candidate objects and each Location information of the candidate object in the images to be recognized;
First determining module, for determining and the fisrt feature figure in the characteristic image of the multiple candidate object As the highest second feature image of matching degree;
Second determining module, for by the corresponding candidate object of the second feature image in the images to be recognized Location information is determined as the location information in the region to match in the images to be recognized with the benchmark image, and will be described The matching degree of second feature image and the fisrt feature image is determined as of image in the benchmark image and the region With degree.
Optionally, it includes the first convolutional layer that the fisrt feature, which extracts model,;
The extraction module is used for:
The benchmark image is inputted into first convolutional layer, obtains fisrt feature image.
Optionally, it includes the second convolutional layer, region of interest ROI pond layer and candidate regions that the second feature, which extracts model, Domain generation module RPN;
The extraction module is used for:
The images to be recognized is inputted into second convolutional layer, obtains third feature image;
The third feature image is inputted into the RPN, obtains position of multiple candidate objects in the images to be recognized Confidence breath;
Location information and the third feature image of the multiple candidate object in the images to be recognized is defeated Enter the pond ROI layer, the characteristic image and each candidate object for obtaining multiple candidate objects are in the images to be recognized Location information.
Optionally, described device further include:
Second obtains module, for obtaining multiple training samples before obtaining images to be recognized and benchmark image, In, each training sample includes sample images to be recognized, sample benchmark image, multiple sample position information and sample matches degree, Wherein, the multiple sample position information is location information of multiple sample candidate objects in the sample images to be recognized, The sample matches degree is the matching degree of the image in the region that the sample benchmark image matches with sample images to be recognized;
Training module, it is described for being inputted using the sample images to be recognized and the sample benchmark image as training Multiple sample position information and the sample matches degree are as output reference value, to initial first convolutional layer, initial second convolution Layer and Initial R PN are trained, first convolutional layer, second convolutional layer and the RPN after being trained.
Optionally, the training module is used for:
It is inputted using the sample images to be recognized and the sample benchmark image as training, the multiple sample bit confidence Breath and the sample matches degree are as output reference value, based on comparison loss function, to initial first convolutional layer, initial volume Two Lamination is trained, and based on detection frame regression function Smooth L1 Loss and softmax loss, is instructed to Initial R PN Practice, first convolutional layer, second convolutional layer and the RPN after being trained.
According to a third aspect of the embodiments of the present invention, a kind of electronic equipment is provided, the electronic equipment include processor and Memory, is stored at least one instruction, at least a Duan Chengxu, code set or instruction set in the memory, and described at least one Item instruction, an at least Duan Chengxu, the code set or instruction set are loaded by the processor and are executed to realize such as first Image matching method described in aspect.
According to a fourth aspect of the embodiments of the present invention, a kind of computer readable storage medium is provided, in the storage medium It is stored at least one instruction, at least a Duan Chengxu, code set or instruction set, described at least one instructs, is at least one section described Program, the code set or instruction set are loaded as the processor and are executed to realize the images match as described in first aspect Method.
Technical solution provided in an embodiment of the present invention has the benefit that
In the embodiment of the present invention, images to be recognized and benchmark image are obtained;Model is extracted based on fisrt feature, in the base In quasi- image, fisrt feature image is extracted, and model is extracted based on second feature, in the images to be recognized, extracted multiple The location information of the characteristic image of candidate object and each candidate object in the images to be recognized;In the multiple candidate In the characteristic image of object, the determining and highest second feature image of the fisrt feature images match degree;It is special by described second Levy location information of the corresponding candidate object of image in the images to be recognized, be determined as in the images to be recognized with it is described The location information in the region that benchmark image matches, and by the matching of the second feature image and the fisrt feature image Degree is determined as the matching degree of image in the benchmark image and the region.In this way, carrying out image using trained model Match, the calculation amount than calculating characteristic point using Feature Descriptor is small, and the time of consuming is few, therefore, can make the effect of images match Rate improves.
It should be understood that above general description and following detailed description be only it is exemplary and explanatory, not It can the limitation present invention.
Detailed description of the invention
To describe the technical solutions in the embodiments of the present invention more clearly, make required in being described below to embodiment Attached drawing is briefly described, it should be apparent that, drawings in the following description are only some embodiments of the invention, for For those of ordinary skill in the art, without creative efforts, it can also be obtained according to these attached drawings other Attached drawing.
Fig. 1 is a kind of flow chart of image matching method provided in an embodiment of the present invention;
Fig. 2 is a kind of flow diagram of image matching method provided in an embodiment of the present invention;
Fig. 3 is a kind of schematic diagram of a scenario of image matching method provided in an embodiment of the present invention;
Fig. 4 is a kind of structural schematic diagram of image matching apparatus provided in an embodiment of the present invention;
Fig. 5 is a kind of structural schematic diagram of image matching apparatus provided in an embodiment of the present invention;
Fig. 6 is a kind of structural block diagram of terminal provided in an embodiment of the present invention;
Fig. 7 is a kind of structural schematic diagram of server provided in an embodiment of the present invention.
Through the above attached drawings, it has been shown that the specific embodiment of the present invention will be hereinafter described in more detail.These attached drawings It is not intended to limit the scope of the inventive concept in any manner with verbal description, but is by referring to specific embodiments Those skilled in the art illustrate idea of the invention.
Specific embodiment
To make the object, technical solutions and advantages of the present invention clearer, below in conjunction with attached drawing to embodiment party of the present invention Formula is described in further detail.
The embodiment of the invention provides a kind of image matching method, this method can be realized by server or terminal.
Server may include the components such as processor, memory, transceiver.Processor can be CPU (Central Processing Unit, central processing unit) etc., can be used for extracting characteristic image, determine in images to be recognized with reference map As the location information in the region to match, calculating matching degree, more multiple matching degrees determine that maximum value therein etc. is handled.Storage Device can be RAM (Random Access Memory, random access memory) that Flash (flash memory) etc. can be used for storing Data needed for the data that receive, treatment process, the data generated in treatment process etc., as images to be recognized, benchmark image, Fisrt feature image, location information in images to be recognized of characteristic image, each candidate object of multiple candidate objects etc..It receives Device is sent out, can be used for carrying out data transmission with terminal or other servers, for example, images to be recognized and benchmark image etc. are received, Transceiver may include antenna, match circuit, modem etc..
Terminal may include the components such as processor, memory.Processor can be CPU (Central Processing Unit, central processing unit) etc., it can be used for extracting characteristic image, determine and match with benchmark image in images to be recognized The location information in region, calculating matching degree, more multiple matching degrees determine that maximum value therein etc. is handled.Memory, Ke Yiwei RAM (Random Access Memory, random access memory), Flash (flash memory) etc. can be used for storing the number received According to the data etc. generated in data, treatment process needed for, treatment process, such as images to be recognized, benchmark image, fisrt feature figure As the location information etc. in images to be recognized of characteristic image, each candidate object of, multiple candidate objects.Terminal can also wrap Include transceiver, screen, image pickup section, audio output part and audio input means etc..Transceiver, can be used for it is other Equipment carries out data transmission, and may include antenna, match circuit, modulatedemodulate for example, receiving images to be recognized and benchmark image etc. Adjust device etc..Screen is displayed for matching degree and location information etc..Image pickup section can be video camera etc..Audio is defeated Component can be speaker, earphone etc. out.Audio input means can be microphone etc..
As shown in Figure 1, the process flow of this method may include following step:
In a step 101, images to be recognized and benchmark image are obtained.
In an alternative embodiment, by taking recognition of face as an example, recognition of face needs to judge facial image and benchmark to be identified Whether facial image matches, if it does, then being verified, if it does not match, verifying does not pass through.
When needing to carry out images match, electronic equipment obtains images to be recognized and benchmark image, the images to be recognized and Benchmark image can be that electronic equipment is pre-stored, be also possible to temporarily to input, the present invention is it is not limited here.
Optionally, it before obtaining images to be recognized and benchmark image, needs to extract the first initial characteristics ready for use Model and the second initial characteristics extract model and are trained, and corresponding treatment process, which can be such that, obtains multiple training samples, In, each training sample includes sample images to be recognized, sample benchmark image, multiple sample position information and sample matches degree, Wherein, multiple sample position information are location information of multiple sample candidate objects in sample images to be recognized, sample matches Degree is the matching degree of the image in sample benchmark image and region;Using sample images to be recognized and sample benchmark image as training Input, multiple sample position information and sample matches degree are as output reference value, to initial first convolutional layer, initial second convolution Layer and initial candidate Area generation module RPN (Region Proposal Network, candidate region network) are trained, The first convolutional layer, the second convolutional layer and RPN after being trained.
In an alternative embodiment, in training the first convolutional layer ready for use, the second convolutional layer, ROI (Region of Interest, area-of-interest) pond layer and when RPN, firstly, electronic equipment obtains multiple training samples, each trained sample This includes sample images to be recognized, sample benchmark image, multiple sample position information and sample matches degree, wherein sample matches The value of degree can be 0 or 1, should when there is region identical with sample benchmark image in the sample images to be recognized Sample matches degree value is 1;It, should when region identical with sample benchmark image is not present in the sample images to be recognized Sample matches degree value is 0.Then, it is inputted using sample images to be recognized and sample benchmark image as training, the knot that will be obtained Fruit compares with multiple sample position information and sample matches degree, to initial first convolutional layer, initial second convolutional layer and Initial R PN is trained, the first convolutional layer, the second convolutional layer, pond layer and RPN after being trained;Wherein, the first convolution Layer is shared with the second convolutional layer weight.
Optionally, in the training process, different layers can be trained by different functions, is had reached preferably Training effect, the detailed process of above-mentioned steps can be such that sample images to be recognized and sample benchmark image is defeated as training Enter, multiple sample position information and sample matches degree are as output reference value, based on comparison loss function Contrastive Loss is trained initial first convolutional layer, initial second convolutional layer, based on detection frame regression function Smooth L1 Loss is trained Initial R PN, the first convolutional layer, the second convolutional layer and RPN after being trained.
In an alternative embodiment, sample benchmark image is input to initial first convolutional layer, sample images to be recognized is defeated Enter to initial second convolutional layer, the result that feature extraction obtains is inputted into the pond RPN and ROI layer respectively, may finally be obtained Prediction and matching degree and multiple predicted position information.
Sample matches degree in prediction and matching degree and training sample is brought into Contrastive Loss, loss is calculated Value, is trained initial first convolutional layer and initial second convolutional layer according to obtained penalty values.Contrastive The optimization aim of Loss is to keep the characteristic distance of same characteristic features closer, and the characteristic distance of different characteristic is farther, and here is The formula of Contrastive Loss:
Wherein, L indicates error, and N indicates the number of training sample, and y indicates that sample matches degree, d indicate prediction and matching degree, Margin represents preset threshold.
After obtaining penalty values by Contrastive Loss, initial first convolutional layer and initial volume Two are constantly adjusted Each weight in lamination, so that obtained penalty values constantly reduce, the certain the number of iterations of training determines this until model convergence Each weight in the first convolutional layer and the second convolutional layer carved, respectively trained first convolutional layer and the second convolutional layer In each weight.
All areas in images to be recognized are input in the softmax classifier in RPN, softmax classifier can To filter out background image from multiple candidate images, i.e., there is no the image of actual object, and the background that will be screened in image Image removal, the image for having actual object obtained after output screening, as candidate image.Then, it is based on Smooth L1 Loss and softmax loss is trained Initial R PN, the RPN after being trained.Wherein, Smooth L1 Loss's is excellent Changing target is that candidate frame is allow more accurately to navigate to object, the optimization aim of Softmax be in determining provincial characteristics whether Comprising object,
In a step 102, model is extracted based on fisrt feature, in benchmark image, extracts fisrt feature image, and be based on Second feature extracts model, and in images to be recognized, the characteristic image and each candidate object for extracting multiple candidate objects exist Location information in images to be recognized.
In an alternative embodiment, after obtaining images to be recognized and benchmark image through the above steps, electronic equipment is by benchmark Image inputs fisrt feature and extracts model, and by feature extraction, the corresponding fisrt feature image of available benchmark image (can claim Make reference characteristic image).
Meanwhile images to be recognized input second feature is extracted into model, the feature extraction of model is extracted by second feature, Location information of the characteristic image and each candidate object of available multiple candidate objects in images to be recognized, the position Confidence breath can be indicated in the form of coordinate.
Optionally, the fisrt feature in above-mentioned steps 102 extracts model, may include the first convolutional layer, therefore, above-mentioned step Rapid 102 treatment process, which may is that, inputs the first convolutional layer for benchmark image, obtains fisrt feature image.
In an alternative embodiment, as shown in Fig. 2, the fisrt feature in above-mentioned steps, which is extracted, includes at least a volume in model Lamination (as the first convolutional layer), benchmark image is input in first convolutional layer by electronic equipment, by the first convolutional layer Process of convolution, the corresponding fisrt feature image (can be referred to as reference characteristic image) of available benchmark image.
Optionally, it includes the second convolutional layer, the pond ROI layer and RPN that the second feature of above-mentioned steps 102, which extracts model,;Cause This, the treatment process of above-mentioned steps 102, which may is that, inputs the second convolutional layer for images to be recognized, obtains third feature image;It will Third feature image inputs RPN, obtains location information of multiple candidate objects in images to be recognized;Multiple candidate objects are existed Location information and third feature image in images to be recognized input the pond ROI layer, obtain the characteristic pattern of multiple candidate objects Picture.
In an alternative embodiment, as shown in Fig. 2, it may include convolutional layer that the second feature in above-mentioned steps, which is extracted in model, (i.e. the second convolutional layer), the pond ROI layer and RPN.
Images to be recognized is input in the second convolutional layer by electronic equipment, can be with by the process of convolution of the second convolutional layer Obtain third feature image (can be referred to as characteristic image to be identified).Characteristic image to be identified is input in RPN, RPN can be In characteristic image to be identified, the region where candidate object is recognized, and then obtains the rectangular area where multiple candidate objects Location information in images to be recognized, the location information can be indicated in the form of coordinate, such as can use (x, y, W, h) indicate the location information of rectangular area where each candidate object, wherein and x indicates the rectangle where candidate's object The abscissa of the top left corner apex in region, y indicate the ordinate of the top left corner apex of the rectangular area where candidate's object, w For the width of the rectangular area, h is the length of the rectangular area.
Then, location information obtained above and characteristic image to be identified are inputted into ROI (Region of Interest, area-of-interest) pond layer, after the processing of the pondization of the pond ROI layer, available multiple candidate objects, The identical characteristic image of dimension size.
Meanwhile by the direct output model of location information obtained above, multiple location information according to put in order with Putting in order for the characteristic image of multiple candidate's objects is identical, so that the subsequent characteristic image according to candidate object determines this feature The corresponding location information of image.
It should be noted that the ruler of the characteristic image of reference characteristic image obtained in above-mentioned steps and multiple candidate objects Very little size is all the same.
In step 103, in the characteristic image of multiple candidate objects, determination is highest with fisrt feature images match degree Second feature image.
In an alternative embodiment, through the above steps, obtain reference characteristic image, multiple candidate objects characteristic image with And after location information of each candidate object in images to be recognized, electronic equipment calculating benchmark characteristic image and each candidate The matching degree of the characteristic image of body obtains multiple matching degrees, this multiple matching degree is compared, and determines wherein highest matching Degree, and determine the characteristic image of the corresponding candidate object of the matching degree, this feature image, which is determined as second feature image, (can claim Make most matching characteristic image).
At step 104, the location information by the corresponding candidate object of second feature image in images to be recognized, determines Location information for the region to match in images to be recognized with benchmark image, and by second feature image and fisrt feature image Matching degree, be determined as the matching degree of image in benchmark image and region.
In an alternative embodiment, after determining most matching characteristic image through the above steps, by of most matching characteristic image Be determined as the matching degree of benchmark image and images to be recognized with degree, and according to most matching characteristic image multiple candidate objects spy Putting in order in sign image, determines the most corresponding location information of matching characteristic image, which is determined as wait know The location information in the region that other image and benchmark image match.
Benchmark image is obtained with after the matching degree of image in region, subsequent processing can be carried out, such as by the matching degree and in advance If matching degree threshold value compare, and different processing is carried out according to comparing result.For example, police can supervised with recognition of face Suspect is found in control image, in this process, obtains benchmark face image and people to be identified through the above steps The matching degree of face image is the location information in 0.87 and the corresponding region of the matching degree, can be with selecting the modes such as frame to mark the area Domain, as shown in Figure 3.And the pre-set matching degree threshold value of technical staff is 0.70, the matching degree obtained due to images match 0.87 is greater than preset matching degree threshold value 0.70, thus may determine that the people for selecting circle to go out is suspect.
In the embodiment of the present invention, images to be recognized and benchmark image are obtained;Model is extracted based on fisrt feature, in the base In quasi- image, fisrt feature image is extracted, and model is extracted based on second feature, in the images to be recognized, extracted multiple The location information of the characteristic image of candidate object and each candidate object in the images to be recognized;In the multiple candidate In the characteristic image of object, the determining and highest second feature image of the fisrt feature images match degree;It is special by described second Levy location information of the corresponding candidate object of image in the images to be recognized, be determined as in the images to be recognized with it is described The location information in the region that benchmark image matches, and by the matching of the second feature image and the fisrt feature image Degree is determined as the matching degree of image in the benchmark image and the region.In this way, carrying out image using trained model Match, the calculation amount than calculating characteristic point using Feature Descriptor is small, and the time of consuming is few, therefore, can make the effect of images match Rate improves.
Based on the same technical idea, the embodiment of the invention also provides a kind of image matching apparatus, which can be Electronic equipment in above-described embodiment, as shown in figure 4, the device includes: the first acquisition module 410, extraction module 420, first Determining module 430 and the second determining module 440.
The first acquisition module 410, is configured as obtaining images to be recognized and benchmark image;
The extraction module 420 is configured as extracting model based on fisrt feature, in the benchmark image, extracts first Characteristic image, and model is extracted based on second feature, in the images to be recognized, extract the characteristic image of multiple candidate objects And location information of each candidate object in the images to be recognized;
First determining module 430 is configured as in the characteristic image of the multiple candidate object, determining with described the The highest second feature image of one characteristic image matching degree;
Second determining module 440 is configured as the corresponding candidate object of the second feature image described wait know Location information in other image is determined as the position letter in the region to match in the images to be recognized with the benchmark image Breath, and by the matching degree of the second feature image and the fisrt feature image, it is determined as the benchmark image and the area The matching degree of image in domain.
Optionally, it includes the first convolutional layer that the fisrt feature, which extracts model,;
The extraction module 420 is configured as:
The benchmark image is inputted into first convolutional layer, obtains fisrt feature image.
Optionally, it includes the second convolutional layer, region of interest ROI pond layer and candidate regions that the second feature, which extracts model, Domain generation module RPN;
The extraction module 420 is configured as:
The images to be recognized is inputted into second convolutional layer, obtains third feature image;
The third feature image is inputted into the RPN, obtains position of multiple candidate objects in the images to be recognized Confidence breath;
Location information and the third feature image of the multiple candidate object in the images to be recognized is defeated Enter the pond ROI layer, obtains the characteristic image of the multiple candidate object.
Optionally, as shown in figure 5, described device further include:
Second obtains module 510, is configured as before obtaining images to be recognized and benchmark image, obtains multiple trained samples This, wherein each training sample includes sample images to be recognized, sample benchmark image, multiple sample position information and sample With degree, wherein the multiple sample position information is position of multiple sample candidate objects in the sample images to be recognized Information, the sample matches degree are of the image in the region that the sample benchmark image matches with sample images to be recognized With degree;
Training module 520 is configured as the sample images to be recognized and the sample benchmark image is defeated as training Enter, the multiple sample position information and the sample matches degree are as output reference value, to initial first convolutional layer, initial the Two convolutional layers, Initial R OI pond layer and Initial R PN are trained, first convolutional layer after being trained, described second Convolutional layer, the pond ROI layer and the RPN.
Optionally, the training module 520 is configured as:
It is inputted using the sample images to be recognized and the sample benchmark image as training, the multiple sample bit confidence Breath and the sample matches degree are as output reference value, based on comparison loss function, to initial first convolutional layer, initial volume Two Lamination, the pond Initial R OI layer are trained, and are based on detection frame regression function Smooth L1 Loss and softmax loss, Initial R PN is trained, first convolutional layer, second convolutional layer, the pond ROI layer after being trained and The RPN.
In the embodiment of the present invention, images to be recognized and benchmark image are obtained;Model is extracted based on fisrt feature, in the base In quasi- image, fisrt feature image is extracted, and model is extracted based on second feature, in the images to be recognized, extracted multiple The location information of the characteristic image of candidate object and each candidate object in the images to be recognized;In the multiple candidate In the characteristic image of object, the determining and highest second feature image of the fisrt feature images match degree;It is special by described second Levy location information of the corresponding candidate object of image in the images to be recognized, be determined as in the images to be recognized with it is described The location information in the region that benchmark image matches, and by the matching of the second feature image and the fisrt feature image Degree is determined as the matching degree of image in the benchmark image and the region.In this way, carrying out image using trained model Match, the calculation amount than calculating characteristic point using Feature Descriptor is small, and the time of consuming is few, therefore, can make the effect of images match Rate improves.
It should be understood that image matching method provided by the above embodiment is in images match, only with above-mentioned each function The division progress of module can according to need and for example, in practical application by above-mentioned function distribution by different function moulds Block is completed, i.e., the internal structure of terminal is divided into different functional modules, to complete all or part of function described above Energy.In addition, image matching apparatus provided by the above embodiment and image matching method embodiment belong to same design, it is specific real Existing process is detailed in embodiment of the method, and which is not described herein again.
Fig. 6 is the structural block diagram of terminal provided in an embodiment of the present invention.The terminal 600 can be portable mobile termianl, Such as: smart phone, tablet computer.Terminal 600 is also possible to referred to as other titles such as user equipment, portable terminal.
In general, terminal 600 includes: processor 601 and memory 602.
Processor 601 may include one or more processing cores, such as 4 core processors, 8 core processors etc..Place Reason device 601 can use DSP (Digital Signal Processing, Digital Signal Processing), FPGA (Field- Programmable Gate Array, field programmable gate array), PLA (Programmable Logic Array, may be programmed Logic array) at least one of example, in hardware realize.Processor 601 also may include primary processor and coprocessor, master Processor is the processor for being handled data in the awake state, also referred to as CPU (Central Processing Unit, central processing unit);Coprocessor is the low power processor for being handled data in the standby state.? In some embodiments, processor 601 can be integrated with GPU (Graphics Processing Unit, image processor), GPU is used to be responsible for the rendering and drafting of content to be shown needed for display screen.In some embodiments, processor 601 can also be wrapped AI (Artificial Intelligence, artificial intelligence) processor is included, the AI processor is for handling related machine learning Calculating operation.
Memory 602 may include one or more computer readable storage mediums, which can To be tangible and non-transient.Memory 602 may also include high-speed random access memory and nonvolatile memory, Such as one or more disk storage equipments, flash memory device.In some embodiments, non-transient in memory 602 Computer readable storage medium for storing at least one instruction, at least one instruction for performed by processor 601 with Realize image matching method provided herein.
In some embodiments, terminal 600 is also optional includes: peripheral device interface 603 and at least one peripheral equipment. Specifically, peripheral equipment includes: radio circuit 604, touch display screen 605, camera 606, voicefrequency circuit 607, positioning component At least one of 608 and power supply 609.
Peripheral device interface 603 can be used for I/O (Input/Output, input/output) is relevant outside at least one Peripheral equipment is connected to processor 601 and memory 602.In some embodiments, processor 601, memory 602 and peripheral equipment Interface 603 is integrated on same chip or circuit board;In some other embodiments, processor 601, memory 602 and outer Any one or two in peripheral equipment interface 603 can realize on individual chip or circuit board, the present embodiment to this not It is limited.
Radio circuit 604 is for receiving and emitting RF (Radio Frequency, radio frequency) signal, also referred to as electromagnetic signal.It penetrates Frequency circuit 604 is communicated by electromagnetic signal with communication network and other communication equipments.Radio circuit 604 turns electric signal It is changed to electromagnetic signal to be sent, alternatively, the electromagnetic signal received is converted to electric signal.Optionally, radio circuit 604 wraps It includes: antenna system, RF transceiver, one or more amplifiers, tuner, oscillator, digital signal processor, codec chip Group, user identity module card etc..Radio circuit 604 can be carried out by least one wireless communication protocol with other terminals Communication.The wireless communication protocol includes but is not limited to: WWW, Metropolitan Area Network (MAN), Intranet, each third generation mobile communication network (2G, 3G, 4G and 5G), WLAN and/or WiFi (Wireless Fidelity, Wireless Fidelity) network.In some embodiments, it penetrates Frequency circuit 604 can also include NFC (Near Field Communication, wireless near field communication) related circuit, this Application is not limited this.
Touch display screen 605 is for showing UI (User Interface, user interface).The UI may include figure, text Sheet, icon, video and its their any combination.Touch display screen 605 also have acquisition touch display screen 605 surface or The ability of the touch signal of surface.The touch signal can be used as control signal and be input to processor 601 and be handled.Touching Display screen 605 is touched for providing virtual push button and/or dummy keyboard, also referred to as soft button and/or soft keyboard.In some embodiments In, touch display screen 605 can be one, and the front panel of terminal 600 is arranged;In further embodiments, touch display screen 605 It can be at least two, be separately positioned on the different surfaces of terminal 600 or in foldover design;In still other embodiments, touch Display screen 605 can be flexible display screen, be arranged on the curved surface of terminal 600 or on fold plane.Even, touch display screen 605 can also be arranged to non-rectangle irregular figure, namely abnormity screen.Touch display screen 605 can use LCD (Liquid Crystal Display, liquid crystal display), OLED (Organic Light-Emitting Diode, Organic Light Emitting Diode) Etc. materials preparation.
CCD camera assembly 606 is for acquiring image or video.Optionally, CCD camera assembly 606 include front camera and Rear camera.In general, front camera is for realizing video calling or self-timer, rear camera is for realizing photo or video Shooting.In some embodiments, rear camera at least two are main camera, depth of field camera, wide-angle imaging respectively Any one in head, to realize that main camera and the fusion of depth of field camera realize background blurring function, main camera and wide-angle Pan-shot and VR (Virtual Reality, virtual reality) shooting function are realized in camera fusion.In some embodiments In, CCD camera assembly 606 can also include flash lamp.Flash lamp can be monochromatic warm flash lamp, be also possible to double-colored temperature flash of light Lamp.Double-colored temperature flash lamp refers to the combination of warm light flash lamp and cold light flash lamp, can be used for the light compensation under different-colour.
Voicefrequency circuit 607 is used to provide the audio interface between user and terminal 600.Voicefrequency circuit 607 may include wheat Gram wind and loudspeaker.Microphone is used to acquire the sound wave of user and environment, and converts sound waves into electric signal and be input to processor 601 are handled, or are input to radio circuit 604 to realize voice communication.For stereo acquisition or the purpose of noise reduction, wheat Gram wind can be it is multiple, be separately positioned on the different parts of terminal 600.Microphone can also be array microphone or omnidirectional's acquisition Type microphone.Loudspeaker is then used to that sound wave will to be converted to from the electric signal of processor 601 or radio circuit 604.Loudspeaker can To be traditional wafer speaker, it is also possible to piezoelectric ceramic loudspeaker.When loudspeaker is piezoelectric ceramic loudspeaker, not only may be used To convert electrical signals to the audible sound wave of the mankind, the sound wave that the mankind do not hear can also be converted electrical signals to survey Away from etc. purposes.In some embodiments, voicefrequency circuit 607 can also include earphone jack.
Positioning component 608 is used for the current geographic position of positioning terminal 600, to realize navigation or LBS (Location Based Service, location based service).Positioning component 608 can be the GPS (Global based on the U.S. Positioning System, global positioning system), China dipper system or Russia Galileo system positioning group Part.
Power supply 609 is used to be powered for the various components in terminal 600.Power supply 609 can be alternating current, direct current, Disposable battery or rechargeable battery.When power supply 609 includes rechargeable battery, which can be wired charging electricity Pond or wireless charging battery.Wired charging battery is the battery to be charged by Wireline, and wireless charging battery is by wireless The battery of coil charges.The rechargeable battery can be also used for supporting fast charge technology.
In some embodiments, terminal 600 further includes having one or more sensors 610.The one or more sensors 610 include but is not limited to: acceleration transducer 611, gyro sensor 612, pressure sensor 613, fingerprint sensor 614, Optical sensor 615 and proximity sensor 616.
The acceleration that acceleration transducer 611 can detecte in three reference axis of the coordinate system established with terminal 600 is big It is small.For example, acceleration transducer 611 can be used for detecting component of the acceleration of gravity in three reference axis.Processor 601 can With the acceleration of gravity signal acquired according to acceleration transducer 611, touch display screen 605 is controlled with transverse views or longitudinal view Figure carries out the display of user interface.Acceleration transducer 611 can be also used for the acquisition of game or the exercise data of user.
Gyro sensor 612 can detecte body direction and the rotational angle of terminal 600, and gyro sensor 612 can To cooperate with acquisition user to act the 3D of terminal 600 with acceleration transducer 611.Processor 601 is according to gyro sensor 612 Following function may be implemented in the data of acquisition: when action induction (for example changing UI according to the tilt operation of user), shooting Image stabilization, game control and inertial navigation.
The lower layer of side frame and/or touch display screen 605 in terminal 600 can be set in pressure sensor 613.Work as pressure When the side frame of terminal 600 is arranged in sensor 613, it can detecte user to the gripping signal of terminal 600, believed according to the gripping Number carry out right-hand man's identification or prompt operation.When the lower layer of touch display screen 605 is arranged in pressure sensor 613, Ke Yigen According to user to the pressure operation of touch display screen 605, realization controls the operability control on the interface UI.Operability Control includes at least one of button control, scroll bar control, icon control, menu control.
Fingerprint sensor 614 is used to acquire the fingerprint of user, according to the identity of collected fingerprint recognition user.Knowing Not Chu the identity of user when being trusted identity, authorize the user to execute relevant sensitive operation, the sensitive operation by processor 601 Including solution lock screen, check encryption information, downloading software, payment and change setting etc..End can be set in fingerprint sensor 614 Front, the back side or the side at end 600.When being provided with physical button or manufacturer Logo in terminal 600, fingerprint sensor 614 can To be integrated with physical button or manufacturer Logo.
Optical sensor 615 is for acquiring ambient light intensity.In one embodiment, processor 601 can be according to optics The ambient light intensity that sensor 615 acquires controls the display brightness of touch display screen 605.Specifically, when ambient light intensity is higher When, the display brightness of touch display screen 605 is turned up;When ambient light intensity is lower, the display for turning down touch display screen 605 is bright Degree.In another embodiment, the ambient light intensity that processor 601 can also be acquired according to optical sensor 615, dynamic adjust The acquisition parameters of CCD camera assembly 606.
Proximity sensor 616, also referred to as range sensor are generally arranged at the front of terminal 600.Proximity sensor 616 is used In the distance between the front of acquisition user and terminal 600.In one embodiment, when proximity sensor 616 detects user When the distance between front of terminal 600 gradually becomes smaller, touch display screen 605 is controlled by processor 601 and is cut from bright screen state It is changed to breath screen state;When proximity sensor 616 detects user and the distance between the front of terminal 600 becomes larger, by Processor 601 controls touch display screen 605 and is switched to bright screen state from breath screen state.
It will be understood by those skilled in the art that the restriction of structure shown in Fig. 6 not structure paired terminal 600, can wrap It includes than illustrating more or fewer components, perhaps combine certain components or is arranged using different components.
In the embodiment of the present invention, images to be recognized and benchmark image are obtained;Model is extracted based on fisrt feature, in the base In quasi- image, fisrt feature image is extracted, and model is extracted based on second feature, in the images to be recognized, extracted multiple The location information of the characteristic image of candidate object and each candidate object in the images to be recognized;In the multiple candidate In the characteristic image of object, the determining and highest second feature image of the fisrt feature images match degree;It is special by described second Levy location information of the corresponding candidate object of image in the images to be recognized, be determined as in the images to be recognized with it is described The location information in the region that benchmark image matches, and by the matching of the second feature image and the fisrt feature image Degree is determined as the matching degree of image in the benchmark image and the region.In this way, carrying out image using trained model Match, the calculation amount than calculating characteristic point using Feature Descriptor is small, and the time of consuming is few, therefore, can make the effect of images match Rate improves.
Fig. 7 is the structural schematic diagram of server provided in an embodiment of the present invention.The server 700 can because configuration or performance not Bigger difference is generated together, may include one or more central processing units (central processing Units, CPU) 722 (for example, one or more processors) and memory 732, one or more storages apply journey The storage medium 730 (such as one or more mass memory units) of sequence 742 or data 744.Wherein, 732 He of memory Storage medium 730 can be of short duration storage or persistent storage.The program for being stored in storage medium 730 may include one or one With upper module (diagram does not mark), each module may include to the series of instructions operation in server.Further, in Central processor 722 can be set to communicate with storage medium 730, execute on server 700 a series of in storage medium 730 Instruction operation.
Server 700 can also include one or more power supplys 726, one or more wired or wireless networks Interface 750, one or more input/output interfaces 758, one or more keyboards 756, and/or, one or one The above operating system 741, such as Windows ServerTM, Mac OS XTM, UnixTM, LinuxTM, FreeBSDTM etc..
Server 700 may include having perhaps one of them or one of more than one program of memory and one Procedure above is stored in memory, and be configured to be executed by one or more than one processor this or one with Upper program executes image matching method described in above-mentioned each embodiment.
In the embodiment of the present invention, images to be recognized and benchmark image are obtained;Model is extracted based on fisrt feature, in the base In quasi- image, fisrt feature image is extracted, and model is extracted based on second feature, in the images to be recognized, extracted multiple The location information of the characteristic image of candidate object and each candidate object in the images to be recognized;In the multiple candidate In the characteristic image of object, the determining and highest second feature image of the fisrt feature images match degree;It is special by described second Levy location information of the corresponding candidate object of image in the images to be recognized, be determined as in the images to be recognized with it is described The location information in the region that benchmark image matches, and by the matching of the second feature image and the fisrt feature image Degree is determined as the matching degree of image in the benchmark image and the region.In this way, carrying out image using trained model Match, the calculation amount than calculating characteristic point using Feature Descriptor is small, and the time of consuming is few, therefore, can make the effect of images match Rate improves.
Those of ordinary skill in the art will appreciate that realizing that all or part of the steps of above-described embodiment can pass through hardware It completes, relevant hardware can also be instructed to complete by program, the program can store in a kind of computer-readable In storage medium, storage medium mentioned above can be read-only memory, disk or CD etc..
The foregoing is merely presently preferred embodiments of the present invention, is not intended to limit the invention, it is all in spirit of the invention and Within principle, any modification, equivalent replacement, improvement and so on be should all be included in the protection scope of the present invention.

Claims (12)

1. a kind of image matching method, which is characterized in that the described method includes:
Obtain images to be recognized and benchmark image;
Model is extracted based on fisrt feature, in the benchmark image, extracts fisrt feature image, and extract based on second feature Model, in the images to be recognized, the characteristic image and each candidate object of the multiple candidate objects of extraction are described wait know Location information in other image;
In the characteristic image of the multiple candidate object, the determining and highest second feature of fisrt feature images match degree Image;
By location information of the corresponding candidate object of the second feature image in the images to be recognized, be determined as it is described to The location information in region to match in identification image with the benchmark image, and by the second feature image and described first The matching degree of characteristic image is determined as the matching degree of image in the benchmark image and the region.
2. the method according to claim 1, wherein it includes the first convolutional layer that the fisrt feature, which extracts model,;
It is described that model is extracted based on fisrt feature, in the benchmark image, extract fisrt feature image, comprising:
The benchmark image is inputted into first convolutional layer, obtains fisrt feature image.
3. method according to claim 1 or 2, which is characterized in that it includes the second convolution that the second feature, which extracts model, Layer, region of interest ROI pond layer and candidate region generation module RPN;
It is described that model is extracted based on second feature, in the images to be recognized, extract the characteristic image of multiple candidate objects with And location information of each candidate object in the images to be recognized, comprising:
The images to be recognized is inputted into second convolutional layer, obtains third feature image;
The third feature image is inputted into the RPN, obtains position letter of multiple candidate objects in the images to be recognized Breath;
Location information of the multiple candidate object in the images to be recognized and the third feature image are inputted into institute The pond ROI layer is stated, the characteristic image of the multiple candidate object is obtained.
4. according to the method described in claim 3, it is characterized in that, described input second convolution for the images to be recognized Layer, before obtaining third feature image, further includes:
Obtain multiple training samples, wherein each training sample includes sample images to be recognized, sample benchmark image, multiple samples This location information and sample matches degree, wherein the multiple sample position information is multiple sample candidate objects in the sample Location information in images to be recognized, the sample matches degree are that the sample benchmark image matches with sample images to be recognized Region in image matching degree;
Inputted the sample images to be recognized and the sample benchmark image as training, the multiple sample position information and The sample matches degree instructs initial first convolutional layer, initial second convolutional layer and Initial R PN as output reference value Practice, first convolutional layer, second convolutional layer and the RPN after being trained.
5. according to the method described in claim 4, it is characterized in that, described by the sample images to be recognized and the sample base Quasi- image is used as output reference value as training input, the multiple sample position information and the sample matches degree, to initial First convolutional layer, initial second convolutional layer and Initial R PN are trained, first convolutional layer after being trained, described Second convolutional layer and the RPN, comprising:
Inputted the sample images to be recognized and the sample benchmark image as training, the multiple sample position information and The sample matches degree is as output reference value, based on comparison loss function contrastive loss, to initial first convolution Layer, initial second convolutional layer are trained, and based on detection frame regression function SmoothL1Loss, are trained to Initial R PN, First convolutional layer, second convolutional layer and the RPN after being trained.
6. a kind of image matching apparatus, which is characterized in that described device includes:
First obtains module, for obtaining images to be recognized and benchmark image;
Extraction module in the benchmark image, extracts fisrt feature image, and base for extracting model based on fisrt feature Model is extracted in second feature, in the images to be recognized, the characteristic image of the multiple candidate objects of extraction and each candidate Location information of the object in the images to be recognized;
First determining module, for determining and the fisrt feature image in the characteristic image of the multiple candidate object With the highest second feature image of degree;
Second determining module, for the position by the corresponding candidate object of the second feature image in the images to be recognized Information, is determined as the location information in the region to match in the images to be recognized with the benchmark image, and by described second The matching degree of characteristic image and the fisrt feature image is determined as the matching of image in the benchmark image and the region Degree.
7. device according to claim 6, which is characterized in that it includes the first convolutional layer that the fisrt feature, which extracts model,;
The extraction module, is used for:
The benchmark image is inputted into first convolutional layer, obtains fisrt feature image.
8. device according to claim 6 or 7, which is characterized in that it includes the second convolution that the second feature, which extracts model, Layer, region of interest ROI pond layer and candidate region generation module RPN;
The extraction module is used for:
The images to be recognized is inputted into second convolutional layer, obtains third feature image;
The third feature image is inputted into the RPN, obtains position letter of multiple candidate objects in the images to be recognized Breath;
Location information of the multiple candidate object in the images to be recognized and the third feature image are inputted into institute The pond ROI layer is stated, the characteristic image of the multiple candidate object is obtained.
9. device according to claim 8, which is characterized in that described device further include:
Second obtains module, for obtaining multiple training samples before obtaining images to be recognized and benchmark image, wherein every A training sample includes sample images to be recognized, sample benchmark image, multiple sample position information and sample matches degree, wherein The multiple sample position information is location information of multiple sample candidate objects in the sample images to be recognized, the sample This matching degree is the matching degree of the image in the region that the sample benchmark image matches with sample images to be recognized;
Training module, it is the multiple for being inputted using the sample images to be recognized and the sample benchmark image as training Sample position information and the sample matches degree as output reference value, to initial first convolutional layer, initial second convolutional layer with And Initial R PN is trained, first convolutional layer, second convolutional layer and the RPN after being trained.
10. device according to claim 9, which is characterized in that the training module is used for:
Inputted the sample images to be recognized and the sample benchmark image as training, the multiple sample position information and The sample matches degree is as output reference value, based on comparison loss function, to initial first convolutional layer, initial second convolutional layer It is trained, based on detection frame regression function Smooth L1Loss and softmax loss, Initial R PN is trained, is obtained First convolutional layer, second convolutional layer and the RPN after to training.
11. a kind of electronic equipment, which is characterized in that the electronic equipment includes processor and memory, is deposited in the memory Contain at least one instruction, at least a Duan Chengxu, code set or instruction set, at least one instruction, an at least Duan Cheng Sequence, the code set or instruction set are loaded by the processor and are executed to realize image as claimed in claim 1 to 5 Matching process.
12. a kind of computer readable storage medium, which is characterized in that be stored at least one instruction, extremely in the storage medium A few Duan Chengxu, code set or instruction set, at least one instruction, an at least Duan Chengxu, the code set or instruction Collection is loaded by the processor and is executed to realize image matching method as claimed in claim 1 to 5.
CN201810144894.XA 2018-02-12 2018-02-12 Image matching method and device Pending CN110147796A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201810144894.XA CN110147796A (en) 2018-02-12 2018-02-12 Image matching method and device

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201810144894.XA CN110147796A (en) 2018-02-12 2018-02-12 Image matching method and device

Publications (1)

Publication Number Publication Date
CN110147796A true CN110147796A (en) 2019-08-20

Family

ID=67589257

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201810144894.XA Pending CN110147796A (en) 2018-02-12 2018-02-12 Image matching method and device

Country Status (1)

Country Link
CN (1) CN110147796A (en)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111461235A (en) * 2020-03-31 2020-07-28 合肥工业大学 Audio and video data processing method and system, electronic equipment and storage medium
CN111950520A (en) * 2020-08-27 2020-11-17 重庆紫光华山智安科技有限公司 Image recognition method and device, electronic equipment and storage medium
CN113435530A (en) * 2021-07-07 2021-09-24 腾讯科技(深圳)有限公司 Image recognition method and device, computer equipment and computer readable storage medium

Citations (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102214291A (en) * 2010-04-12 2011-10-12 云南清眸科技有限公司 Method for quickly and accurately detecting and tracking human face based on video sequence
CN103902960A (en) * 2012-12-28 2014-07-02 北京计算机技术及应用研究所 Real-time face recognition system and method thereof
CN104217225A (en) * 2014-09-02 2014-12-17 中国科学院自动化研究所 A visual target detection and labeling method
CN105631039A (en) * 2016-01-15 2016-06-01 北京邮电大学 Picture browsing method
CN105809089A (en) * 2014-12-29 2016-07-27 中国科学院深圳先进技术研究院 Multi-face detection method and device under complex background
CN105989368A (en) * 2015-02-13 2016-10-05 展讯通信(天津)有限公司 Target detection method and apparatus, and mobile terminal
CN106803057A (en) * 2015-11-25 2017-06-06 腾讯科技(深圳)有限公司 Image information processing method and device
CN106815566A (en) * 2016-12-29 2017-06-09 天津中科智能识别产业技术研究院有限公司 A kind of face retrieval method based on multitask convolutional neural networks
US20170206431A1 (en) * 2016-01-20 2017-07-20 Microsoft Technology Licensing, Llc Object detection and classification in images
CN107145842A (en) * 2017-04-19 2017-09-08 西安电子科技大学 With reference to LBP characteristic patterns and the face identification method of convolutional neural networks

Patent Citations (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102214291A (en) * 2010-04-12 2011-10-12 云南清眸科技有限公司 Method for quickly and accurately detecting and tracking human face based on video sequence
CN103902960A (en) * 2012-12-28 2014-07-02 北京计算机技术及应用研究所 Real-time face recognition system and method thereof
CN104217225A (en) * 2014-09-02 2014-12-17 中国科学院自动化研究所 A visual target detection and labeling method
CN105809089A (en) * 2014-12-29 2016-07-27 中国科学院深圳先进技术研究院 Multi-face detection method and device under complex background
CN105989368A (en) * 2015-02-13 2016-10-05 展讯通信(天津)有限公司 Target detection method and apparatus, and mobile terminal
CN106803057A (en) * 2015-11-25 2017-06-06 腾讯科技(深圳)有限公司 Image information processing method and device
CN105631039A (en) * 2016-01-15 2016-06-01 北京邮电大学 Picture browsing method
US20170206431A1 (en) * 2016-01-20 2017-07-20 Microsoft Technology Licensing, Llc Object detection and classification in images
CN106815566A (en) * 2016-12-29 2017-06-09 天津中科智能识别产业技术研究院有限公司 A kind of face retrieval method based on multitask convolutional neural networks
CN107145842A (en) * 2017-04-19 2017-09-08 西安电子科技大学 With reference to LBP characteristic patterns and the face identification method of convolutional neural networks

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
SHAOQING REN ET AL: ""Faster R-CNN Towards Real-Time Object Detection with Region Proposal Networks"", 《ARXIV》 *

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111461235A (en) * 2020-03-31 2020-07-28 合肥工业大学 Audio and video data processing method and system, electronic equipment and storage medium
US11335096B2 (en) 2020-03-31 2022-05-17 Hefei University Of Technology Method, system and electronic device for processing audio-visual data
CN111950520A (en) * 2020-08-27 2020-11-17 重庆紫光华山智安科技有限公司 Image recognition method and device, electronic equipment and storage medium
CN111950520B (en) * 2020-08-27 2022-12-02 重庆紫光华山智安科技有限公司 Image recognition method and device, electronic equipment and storage medium
CN113435530A (en) * 2021-07-07 2021-09-24 腾讯科技(深圳)有限公司 Image recognition method and device, computer equipment and computer readable storage medium
CN113435530B (en) * 2021-07-07 2023-10-10 腾讯科技(深圳)有限公司 Image recognition method, device, computer equipment and computer readable storage medium

Similar Documents

Publication Publication Date Title
US11710351B2 (en) Action recognition method and apparatus, and human-machine interaction method and apparatus
CN110502954A (en) The method and apparatus of video analysis
CN108594997B (en) Gesture skeleton construction method, device, equipment and storage medium
CN110992493B (en) Image processing method, device, electronic equipment and storage medium
CN110189340A (en) Image partition method, device, electronic equipment and storage medium
CN110083791B (en) Target group detection method and device, computer equipment and storage medium
CN109829456A (en) Image-recognizing method, device and terminal
CN110222551A (en) Method, apparatus, electronic equipment and the storage medium of identification maneuver classification
CN107844781A (en) Face character recognition methods and device, electronic equipment and storage medium
CN108734736A (en) Camera posture method for tracing, device, equipment and storage medium
CN110148178B (en) Camera positioning method, device, terminal and storage medium
CN108305236A (en) Image enhancement processing method and device
CN110222789A (en) Image-recognizing method and storage medium
CN109947886A (en) Image processing method, device, electronic equipment and storage medium
CN109285178A (en) Image partition method, device and storage medium
CN109558837A (en) Face critical point detection method, apparatus and storage medium
CN110059652A (en) Face image processing process, device and storage medium
CN109522863A (en) Ear's critical point detection method, apparatus and storage medium
CN110210573A (en) Fight generation method, device, terminal and the storage medium of image
CN110134744A (en) The methods, devices and systems that Geomagnetism Information is updated
CN110135336A (en) Training method, device and the storage medium of pedestrian's generation model
CN110163160A (en) Face identification method, device, equipment and storage medium
CN108320756A (en) It is a kind of detection audio whether be absolute music audio method and apparatus
CN109360222A (en) Image partition method, device and storage medium
CN109886208A (en) Method, apparatus, computer equipment and the storage medium of object detection

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication
RJ01 Rejection of invention patent application after publication

Application publication date: 20190820