CN111523554A - Image recognition method based on reverse bag-of-words model - Google Patents

Image recognition method based on reverse bag-of-words model Download PDF

Info

Publication number
CN111523554A
CN111523554A CN202010292713.5A CN202010292713A CN111523554A CN 111523554 A CN111523554 A CN 111523554A CN 202010292713 A CN202010292713 A CN 202010292713A CN 111523554 A CN111523554 A CN 111523554A
Authority
CN
China
Prior art keywords
image
words model
descriptors
bag
word vector
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202010292713.5A
Other languages
Chinese (zh)
Inventor
裴云强
吴亚东
王赋攀
侯志伟
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Sichuan University of Science and Engineering
Original Assignee
Sichuan University of Science and Engineering
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Sichuan University of Science and Engineering filed Critical Sichuan University of Science and Engineering
Priority to CN202010292713.5A priority Critical patent/CN111523554A/en
Publication of CN111523554A publication Critical patent/CN111523554A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/46Descriptors for shape, contour or point-related descriptors, e.g. scale invariant feature transform [SIFT] or bags of words [BoW]; Salient regional features
    • G06V10/462Salient features, e.g. scale invariant feature transforms [SIFT]
    • G06V10/464Salient features, e.g. scale invariant feature transforms [SIFT] using a plurality of salient features, e.g. bag-of-words [BoW] representations
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/22Matching criteria, e.g. proximity measures
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/23Clustering techniques
    • G06F18/232Non-hierarchical techniques
    • G06F18/2321Non-hierarchical techniques using statistics or function optimisation, e.g. modelling of probability density functions
    • G06F18/23213Non-hierarchical techniques using statistics or function optimisation, e.g. modelling of probability density functions with fixed number of clusters, e.g. K-means clustering
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/168Feature extraction; Face representation

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • Physics & Mathematics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • General Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • Oral & Maxillofacial Surgery (AREA)
  • Evolutionary Computation (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Artificial Intelligence (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Evolutionary Biology (AREA)
  • Multimedia (AREA)
  • Health & Medical Sciences (AREA)
  • Probability & Statistics with Applications (AREA)
  • General Health & Medical Sciences (AREA)
  • Human Computer Interaction (AREA)
  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)

Abstract

The invention discloses an image identification method based on a reverse bag-of-words model, which comprises the following steps: generating a reverse bag-of-words model: extracting SURF characteristic points of all reference images at a server side, obtaining corresponding 64-dimensional descriptors, and establishing a reverse bag-of-words model by using K-means, wherein leaf nodes correspond to 'visual words'; target image transmission: opening a Web camera to capture an image, directly uploading the image to a server, extracting SURF characteristic points of the image and obtaining a corresponding descriptor; image word vector generation: classifying all SURF descriptors of the reference image and the target image by using the established reverse bag-of-words model, associating the classification result of each descriptor to a visual word, and then calculating a corresponding word vector; target image recognition: and calculating Euclidean distances between the target image word vector and the reference image word vector, and taking the reference image corresponding to the minimum distance as a recognition result. The method can ensure the recognition speed and the recognition matching degree.

Description

Image recognition method based on reverse bag-of-words model
Technical Field
The invention relates to an image recognition method, in particular to an image recognition method based on a reverse bag-of-words model.
Background
Image recognition, which refers to a technique for processing, analyzing and understanding images by a computer to recognize various different patterns of objects and objects, is a practical application of applying a deep learning algorithm. Image recognition technology at present is generally divided into face recognition and commodity recognition, and the face recognition is mainly applied to security inspection, identity verification and mobile payment; the commodity identification is mainly applied to the commodity circulation process, in particular to the field of unmanned retail such as unmanned goods shelves and intelligent retail cabinets.
The basic flow of image recognition is divided into four steps: image acquisition → image preprocessing → feature extraction → image recognition. The traditional image recognition method is mainly based on the traditional technology of vocabulary tree retrieval, takes a word bag model established in the retrieval direction to match images under visual words (leaf nodes) as a main technical route, and is divided into two types: the first type is one-to-one comparison between a target image and a reference image, the reference image with the minimum descriptor distance with all feature points of the target image is obtained through calculation and comparison, the time overhead of a retrieval algorithm is linearly increased, the matching time is increased along with the increase of the feature points, and finally, the matched image with the shortest descriptor distance is found; the second type of image matching method is that a user needs to apply for Token in a server-side image database for each target image, and upload the image to the server-side image database, and when the target image is identified, the corresponding image in the server-side server is found according to the Token applied by the user for image matching, so that redundant time-consuming application steps only reduce the experience of the user in identifying the image through equipment.
In summary, the conventional image recognition method has the following defects: firstly, if the reference image parts in the database are too similar, the result of too many layers and too large data processing delay is caused in the process of building the tree; secondly, if the number of layers is limited to shorten the retrieval time, the result of insufficient accuracy is generated; thirdly, after the tree building operation of each layer, all the feature point descriptors contained in each node of the layer need to be stored, so as to obtain the average node of the next layer, and further, the storage space is far from shortage.
Disclosure of Invention
The present invention is directed to solve the above problems, and an object of the present invention is to provide an image recognition method based on a reverse bag-of-words model, which can ensure recognition speed and also ensure recognition matching degree.
The invention realizes the purpose through the following technical scheme:
an image recognition method based on a reverse bag-of-words model comprises the following steps:
step 1, generating a reverse bag-of-words model: extracting SURF characteristic points of all reference images at a server side, obtaining corresponding 64-dimensional descriptors, and establishing a reverse bag-of-words model by using K-means, wherein leaf nodes correspond to 'visual words';
step 2, target image transmission: opening a Web camera to capture an image, directly uploading the image to a server, extracting SURF characteristic points of the image and obtaining a corresponding descriptor;
step 3, generating image word vectors: classifying all SURF descriptors of the reference image and the target image by using the established reverse bag-of-words model, associating the classification result of each descriptor to a visual word, and then calculating a corresponding word vector;
step 4, target image recognition: and calculating Euclidean distances between the target image word vector and the reference image word vector, and taking the reference image corresponding to the minimum distance as a recognition result.
The SURF is a short hand for Speeded Up Robust Features, is an improved algorithm for SIFT algorithm, is mainly characterized by high speed, and is the prior art. The meaning of K-means is K mean value, and a K mean value clustering algorithm is a clustering analysis algorithm for iterative solution and is the prior art.
Preferably, the method for obtaining the descriptors in the step 1 and the step 2 is as follows: firstly, acquiring a reference or target image, respectively expanding each image to establish an image pyramid based on a box filter, then extracting all SURF feature points by using a SURF algorithm for each scale of the image, and generating a corresponding descriptor for each SURF feature point.
Preferably, in step 1, the method for establishing the database of the inverse bag-of-words model includes the following steps:
step 1.1, defining an N-ary tree structure with the number of layersK, the number of nodes in each layer is NiAnd i represents the ith layer;
step 1.2, determining the number of leaf nodes in the last layer according to the application field of the product and the size of the analyzed database scale;
step 1.3, performing aggregation classification on all descriptors of the database image based on a K-means algorithm to obtain classification results of N sub-nodes, taking cluster centers of all descriptors in each node as the descriptors of the node, and sequencing the nodes according to the position numbers of the nodes;
step 1.4, calculating Euclidean distances between the descriptors of the nodes arranged at the first level and the descriptors of the rest N-1 nodes at the current level, finding out the node with the minimum distance and exchanging the position of the node with the node at the current second position; the matched nodes are not considered, the operation is sequentially carried out on the nodes at the rest odd number positions until the layer is traversed, and then the average descriptor of all the matched nodes of the layer is obtained and used as the descriptor of the father node of the layer;
step 1.5, entering a K-1 layer, repeating the step 1.4 until entering a second layer, and generating an N-ary tree structure model of the image database;
step 1.6, maintaining the tree downwards from the second layer, adjusting the position of each node of each layer to the position where the node is positioned before matching according to Euclidean distance except the second layer until the leaf nodes are correctly returned, wherein each leaf node corresponds to an independent visual word, completing the maintenance of the N-way tree, and generating a reverse bag-of-words model;
and step 1.7, storing the reverse bag-of-words model into a server-side database, and establishing an image database with an N-branch tree structure.
Preferably, in step 3, the generating of the reference image word vector includes the following steps:
step 3.1.1, giving a unique number to the reference image;
step 3.1.2, classifying all SURF descriptors of the reference image by using the established reverse bag-of-words model, calculating a word vector of the reference image corresponding to the descriptor by using TF-IDF according to word frequency after all the descriptors of the reference image are distributed to corresponding leaf nodes, and storing the word vector and the serial number of the corresponding reference image into a database at the server end;
and 3.1.3, performing steps 3.1.1-3.1.2 on each reference image to obtain a word vector of each reference image and storing the word vector into a database at the server side.
Preferably, in step 3, the generating of the target image word vector includes the following steps:
step 3.2.1, opening a Web camera to capture an image by a user, directly uploading the image to a server, extracting SURF characteristic points of a target image and obtaining a corresponding descriptor;
and 3.2.2, classifying all SURF descriptors of the target image by using the established reverse bag-of-words model, associating the classification result of each descriptor to a visual word, and then calculating a word vector of the corresponding target image.
The invention has the beneficial effects that:
the method can ensure the recognition speed and the recognition matching degree; specifically, by establishing a reverse bag-of-words model, compared with establishing a forward bag-of-words model, the generated tree structure is ensured to be a full N-tree, the minimum number of model layers is further ensured, the time overhead of the descriptors distributed to the words is reduced, and after each layer of tree establishment operation, all nodes (all descriptors in the layer) are not required to be stored to obtain the average node of the next layer (the next time of reverse tree establishment is a parent node layer, and the next layer of forward tree establishment is a child node layer), so that the space overhead of the characteristic point of the layer required to be stored when the next layer of nodes is calculated is reduced; after the reverse bag-of-words model is searched, the best matching image is found by utilizing the Euclidean distance, so that the searching precision is greatly improved while the quick searching is ensured. Meanwhile, when the method is adopted and the preprocessed database is matched, the descriptor to be matched and the most similar image of the descriptor can be matched without comparing all reference images one by one, even the time overhead of a fixed constant can be achieved, the time delay of a processing part is greatly reduced, so that the retrieval time is not linearly increased according to the scale of the database, but is fixed to be a constant level and is far smaller than a linear processing task.
Drawings
FIG. 1 is a schematic diagram of the comparison of products using the inverse bag-of-words modeling of the present invention with other products of the same type in terms of time overhead and recognition accuracy.
Detailed Description
The invention will be further illustrated with reference to preferred embodiments and the accompanying drawings in which:
the preferred embodiment:
an image recognition method based on a reverse bag-of-words model comprises the following steps:
step 1, generating a reverse bag-of-words model: extracting SURF characteristic points of all reference images at a server side, obtaining corresponding 64-dimensional descriptors, and establishing a reverse bag-of-words model by using K-means, wherein leaf nodes correspond to 'visual words'; in this step, the specific method for obtaining the descriptor is as follows: firstly, acquiring a reference image, respectively expanding each image to establish an image pyramid based on a box filter, then extracting all SURF feature points by using a SURF algorithm for each scale of the image, and generating a corresponding descriptor for each SURF feature point;
in this step, the method for establishing the database of the reverse bag-of-words model includes the following steps:
step 1.1, defining an N-ary tree structure, wherein the number of layers is K, and the number of nodes in each layer is NiAnd i represents the ith layer;
step 1.2, determining the number of leaf nodes in the last layer according to the application field of the product and the size of the analyzed database scale;
step 1.3, performing aggregation classification on all descriptors of the database image based on a K-means algorithm to obtain classification results of N sub-nodes, taking cluster centers of all descriptors in each node as the descriptors of the node, and sequencing the nodes according to the position numbers of the nodes;
step 1.4, calculating Euclidean distances between the descriptors of the nodes arranged at the first level and the descriptors of the rest N-1 nodes at the current level, finding out the node with the minimum distance and exchanging the position of the node with the node at the current second position; the matched nodes are not considered, the operation is sequentially carried out on the nodes at the rest odd number positions until the layer is traversed, and then the average descriptor of all the matched nodes of the layer is obtained and used as the descriptor of the father node of the layer;
step 1.5, entering a K-1 layer, repeating the step 1.4 until entering a second layer, and generating an N-ary tree structure model of the image database;
step 1.6, maintaining the tree downwards from the second layer, adjusting the position of each node of each layer to the position where the node is positioned before matching according to Euclidean distance except the second layer until the leaf nodes are correctly returned, wherein each leaf node corresponds to an independent visual word, completing the maintenance of the N-way tree, and generating a reverse bag-of-words model;
step 1.7, storing the reverse bag-of-words model into a server-side database, and establishing an image database with an N-branch tree structure;
step 2, target image transmission: opening a Web camera to capture an image, directly uploading the image to a server, extracting SURF characteristic points of the image and obtaining a corresponding descriptor; in this step, the specific method for obtaining the descriptor is as follows: firstly, acquiring a target image, namely an image captured by a Web camera, respectively expanding each image to establish an image pyramid based on a box filter, then extracting all SURF feature points by using a SURF algorithm for each scale of the image, and generating a corresponding descriptor for each SURF feature point;
step 3, generating image word vectors: classifying all SURF descriptors of the reference image and the target image by using the established reverse bag-of-words model, associating the classification result of each descriptor to a visual word, and then calculating a corresponding word vector;
in this step, the generating of the word vector of the reference image includes the following steps:
step 3.1.1, giving a unique number to the reference image;
step 3.1.2, classifying all SURF descriptors of the reference image by using the established reverse bag-of-words model, calculating a word vector of the reference image corresponding to the descriptor by using TF-IDF according to word frequency after all the descriptors of the reference image are distributed to corresponding leaf nodes, and storing the word vector and the serial number of the corresponding reference image into a database at the server end;
step 3.1.3, performing step 3.1.1-3.1.2 on each reference image to obtain a word vector of each reference image and storing the word vector into a server-side database;
the generation of the target image word vector comprises the following steps:
step 3.2.1, opening a Web camera to capture an image by a user, directly uploading the image to a server, extracting SURF characteristic points of a target image and obtaining a corresponding descriptor;
3.2.2, classifying all SURF descriptors of the target image by using the established reverse bag-of-words model, associating the classification result of each descriptor to a visual word, and then calculating a word vector of the corresponding target image;
step 4, target image recognition: and calculating Euclidean distances between the target image word vector and the reference image word vector, and taking the reference image corresponding to the minimum distance as a recognition result.
The advantages of the process of the present invention over other conventional processes are further understood by the following FIG. 1:
FIG. 1 is a schematic diagram of the comparison of products using the inverse bag-of-words modeling of the present invention with other products of the same type in terms of time overhead and recognition accuracy. As can be seen from fig. 1, the method of the present invention, namely the method numbered "invert boww", is significantly reduced in time overhead, significantly improved in accuracy, and has significant advantages compared to other conventional methods. The numbers in parentheses in fig. 1 indicate the total number of reference pictures using different methods.
The above embodiments are only preferred embodiments of the present invention, and are not intended to limit the technical solutions of the present invention, so long as the technical solutions can be realized on the basis of the above embodiments without creative efforts, which should be considered to fall within the protection scope of the patent of the present invention.

Claims (5)

1. An image recognition method based on a reverse bag-of-words model is characterized in that: the method comprises the following steps:
step 1, generating a reverse bag-of-words model: extracting SURF characteristic points of all reference images at a server side, obtaining corresponding 64-dimensional descriptors, and establishing a reverse bag-of-words model by using K-means, wherein leaf nodes correspond to 'visual words';
step 2, target image transmission: opening a Web camera to capture an image, directly uploading the image to a server, extracting SURF characteristic points of the image and obtaining a corresponding descriptor;
step 3, generating image word vectors: classifying all SURF descriptors of the reference image and the target image by using the established reverse bag-of-words model, associating the classification result of each descriptor to a visual word, and then calculating a corresponding word vector;
step 4, target image recognition: and calculating Euclidean distances between the target image word vector and the reference image word vector, and taking the reference image corresponding to the minimum distance as a recognition result.
2. The image recognition method based on the inverse bag-of-words model according to claim 1, characterized in that: the method for obtaining the descriptor in the step 1 and the step 2 comprises the following steps: firstly, acquiring a reference or target image, respectively expanding each image to establish an image pyramid based on a box filter, then extracting all SURF feature points by using a SURF algorithm for each scale of the image, and generating a corresponding descriptor for each SURF feature point.
3. The image recognition method based on the inverse bag-of-words model according to claim 1 or 2, characterized in that: in the step 1, the method for establishing the database of the reverse bag-of-words model comprises the following steps:
step 1.1, defining an N-ary tree structure, wherein the number of layers is K, the number of nodes in each layer is Ni, and i represents the ith layer;
step 1.2, determining the number of leaf nodes in the last layer according to the application field of the product and the size of the analyzed database scale;
step 1.3, performing aggregation classification on all descriptors of the database image based on a K-means algorithm to obtain classification results of N sub-nodes, taking cluster centers of all descriptors in each node as the descriptors of the node, and sequencing the nodes according to the position numbers of the nodes;
step 1.4, calculating Euclidean distances between the descriptors of the nodes arranged at the first level and the descriptors of the rest N-1 nodes at the current level, finding out the node with the minimum distance and exchanging the position of the node with the node at the current second position; the matched nodes are not considered, the operation is sequentially carried out on the nodes at the rest odd number positions until the layer is traversed, and then the average descriptor of all the matched nodes of the layer is obtained and used as the descriptor of the father node of the layer;
step 1.5, entering a K-1 layer, repeating the step 1.4 until entering a second layer, and generating an N-ary tree structure model of the image database;
step 1.6, maintaining the tree downwards from the second layer, adjusting the position of each node of each layer to the position where the node is positioned before matching according to Euclidean distance except the second layer until the leaf nodes are correctly returned, wherein each leaf node corresponds to an independent visual word, completing the maintenance of the N-way tree, and generating a reverse bag-of-words model;
and step 1.7, storing the reverse bag-of-words model into a server-side database, and establishing an image database with an N-branch tree structure.
4. The image recognition method based on the inverse bag-of-words model according to claim 1 or 2, characterized in that: in step 3, the generating of the word vector of the reference image includes the following steps:
step 3.1.1, giving a unique number to the reference image;
step 3.1.2, classifying all SURF descriptors of the reference image by using the established reverse bag-of-words model, calculating a word vector of the reference image corresponding to the descriptor by using TF-IDF according to word frequency after all the descriptors of the reference image are distributed to corresponding leaf nodes, and storing the word vector and the serial number of the corresponding reference image into a database at the server end;
and 3.1.3, performing steps 3.1.1-3.1.2 on each reference image to obtain a word vector of each reference image and storing the word vector into a database at the server side.
5. The image recognition method based on the inverse bag-of-words model according to claim 1 or 2, characterized in that: in step 3, the generating of the target image word vector includes the following steps:
step 3.2.1, opening a Web camera to capture an image by a user, directly uploading the image to a server, extracting SURF characteristic points of a target image and obtaining a corresponding descriptor;
and 3.2.2, classifying all SURF descriptors of the target image by using the established reverse bag-of-words model, associating the classification result of each descriptor to a visual word, and then calculating a word vector of the corresponding target image.
CN202010292713.5A 2020-04-13 2020-04-13 Image recognition method based on reverse bag-of-words model Pending CN111523554A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202010292713.5A CN111523554A (en) 2020-04-13 2020-04-13 Image recognition method based on reverse bag-of-words model

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202010292713.5A CN111523554A (en) 2020-04-13 2020-04-13 Image recognition method based on reverse bag-of-words model

Publications (1)

Publication Number Publication Date
CN111523554A true CN111523554A (en) 2020-08-11

Family

ID=71912035

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202010292713.5A Pending CN111523554A (en) 2020-04-13 2020-04-13 Image recognition method based on reverse bag-of-words model

Country Status (1)

Country Link
CN (1) CN111523554A (en)

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104778284A (en) * 2015-05-11 2015-07-15 苏州大学 Spatial image inquiring method and system
US20170255840A1 (en) * 2014-11-26 2017-09-07 Captricity, Inc. Analyzing content of digital images
CN107193965A (en) * 2017-05-24 2017-09-22 哈尔滨工业大学 A kind of quick indoor orientation method based on BoVW algorithms
CN109753646A (en) * 2017-11-01 2019-05-14 深圳市腾讯计算机系统有限公司 A kind of article attribute recognition approach and electronic equipment

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20170255840A1 (en) * 2014-11-26 2017-09-07 Captricity, Inc. Analyzing content of digital images
CN104778284A (en) * 2015-05-11 2015-07-15 苏州大学 Spatial image inquiring method and system
CN107193965A (en) * 2017-05-24 2017-09-22 哈尔滨工业大学 A kind of quick indoor orientation method based on BoVW algorithms
CN109753646A (en) * 2017-11-01 2019-05-14 深圳市腾讯计算机系统有限公司 A kind of article attribute recognition approach and electronic equipment

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
PEI YUNQIANG: "WA VIS : A Web-based Augmented Reality Text Data Visual Analysis Tool", 《2019 INTERNATIONAL CONFERENCE ON VIRTUAL REALITY AND VISUALIZATION (ICVRV)》, 1 November 2019 (2019-11-01) *
罗升斯: "基于BoW的视觉SLAM闭环检测方法研究", 《中国优秀硕士学位论文全文数据库 信息科技辑》, 15 December 2018 (2018-12-15) *

Similar Documents

Publication Publication Date Title
WO2022068196A1 (en) Cross-modal data processing method and device, storage medium, and electronic device
Melekhov et al. Siamese network features for image matching
CN108694225B (en) Image searching method, feature vector generating method and device and electronic equipment
CN107229757B (en) Video retrieval method based on deep learning and Hash coding
Wang et al. Sketch-based 3d shape retrieval using convolutional neural networks
Gong et al. Web scale photo hash clustering on a single machine
Berretti et al. 3D facial expression recognition using SIFT descriptors of automatically detected keypoints
Leibe et al. Efficient clustering and matching for object class recognition.
Xia et al. Loop closure detection for visual SLAM using PCANet features
CN104794219A (en) Scene retrieval method based on geographical position information
CN105468781A (en) Video query method and device
Julca-Aguilar et al. Symbol detection in online handwritten graphics using faster R-CNN
CN105069457B (en) Image recognition method and device
Melekhov et al. Image patch matching using convolutional descriptors with euclidean distance
CN104281572A (en) Target matching method and system based on mutual information
CN106845375A (en) A kind of action identification method based on hierarchical feature learning
Zhang et al. Loop closure detection via maximization of mutual information
CN109886206B (en) Three-dimensional object identification method and equipment
CN108875828A (en) A kind of fast matching method and system of similar image
Song et al. Deep region hashing for generic instance search from images
CN104778272B (en) A kind of picture position method of estimation excavated based on region with space encoding
CN110083724A (en) A kind of method for retrieving similar images, apparatus and system
CN114519863A (en) Human body weight recognition method, human body weight recognition apparatus, computer device, and medium
JP6793925B2 (en) Verification equipment, methods, and programs
CN116883740A (en) Similar picture identification method, device, electronic equipment and storage medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
WD01 Invention patent application deemed withdrawn after publication
WD01 Invention patent application deemed withdrawn after publication

Application publication date: 20200811