CN110689046A - Image recognition method, image recognition device, computer device, and storage medium - Google Patents

Image recognition method, image recognition device, computer device, and storage medium Download PDF

Info

Publication number
CN110689046A
CN110689046A CN201910792041.1A CN201910792041A CN110689046A CN 110689046 A CN110689046 A CN 110689046A CN 201910792041 A CN201910792041 A CN 201910792041A CN 110689046 A CN110689046 A CN 110689046A
Authority
CN
China
Prior art keywords
query
similarity
image
reference images
reference image
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201910792041.1A
Other languages
Chinese (zh)
Inventor
刘利
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
OneConnect Smart Technology Co Ltd
Original Assignee
OneConnect Smart Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by OneConnect Smart Technology Co Ltd filed Critical OneConnect Smart Technology Co Ltd
Priority to CN201910792041.1A priority Critical patent/CN110689046A/en
Publication of CN110689046A publication Critical patent/CN110689046A/en
Priority to PCT/CN2020/086768 priority patent/WO2021036309A1/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/22Matching criteria, e.g. proximity measures
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/048Activation functions
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Computation (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Artificial Intelligence (AREA)
  • General Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • Computing Systems (AREA)
  • Software Systems (AREA)
  • Molecular Biology (AREA)
  • Computational Linguistics (AREA)
  • Biophysics (AREA)
  • Biomedical Technology (AREA)
  • Mathematical Physics (AREA)
  • General Health & Medical Sciences (AREA)
  • Health & Medical Sciences (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Evolutionary Biology (AREA)
  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
  • Image Analysis (AREA)

Abstract

The invention provides an image recognition method, which comprises the following steps: acquiring a query image and a plurality of reference images; imaging the query image and each reference image, and extracting similarity characteristics of the image pairs; constructing a complete graph by taking the image pairs as nodes; calculating the similarity score of every two reference images, and determining the weight of the corresponding edge of every two reference images in the complete image according to the similarity score of every two reference images; mapping the similarity characteristics of the image pairs into message characteristics through a message network; updating the similarity characteristic of the image pair according to the message characteristic and the weight of the edge; calculating a similarity score of the image pair according to the updated similarity features; and determining a reference image matched with the query image according to the similarity score. The invention also provides an image recognition device, a computer device and a computer readable storage medium. The method and the device utilize the similarity information between the reference images to update the similarity characteristics of the query image and the reference images, and improve the accuracy of image identification.

Description

Image recognition method, image recognition device, computer device, and storage medium
Technical Field
The present invention relates to the field of image recognition technologies, and in particular, to an image recognition method, an image recognition device, a computer device, and a computer-readable storage medium.
Background
At present, when a query image is matched with a reference image, only the similarity between the query image and the reference image is considered, and the similarity between the reference images is ignored. If the similarity between the query image and the reference image is not calculated well, the accuracy of image matching is affected.
Disclosure of Invention
In view of the foregoing, it is desirable to provide an image recognition method, an image recognition apparatus, a computer apparatus, and a computer-readable storage medium, which improve the accuracy of image recognition by performing image recognition using similarity information between reference images.
A first aspect of the present application provides an image recognition method, the method comprising:
acquiring a query image and a plurality of reference images;
forming a query-reference image pair by the query image and each reference image, and extracting similarity characteristics of each query-reference image pair;
constructing a complete graph by taking the query-reference image pairs as nodes, wherein each reference image corresponds to one node;
calculating similarity scores between every two reference images in the multiple reference images, and determining the weight values of edges corresponding to the two reference images in the complete image according to the similarity scores between the two reference images;
mapping the similarity features of each query-reference image pair into message features through a message network;
updating the similarity characteristic of each query-reference image pair according to the message characteristic of each query-reference image pair and the weight of each edge in the complete graph;
calculating a similarity score for each query-reference image pair based on the updated similarity features for each query-reference image pair;
determining images from the plurality of reference images that match the query image based on the similarity score for each query-reference image pair.
In another possible implementation, the extracting the similarity feature of each query-reference image pair includes:
inputting the query-reference image pair into a twin neural network based on a residual error network to obtain a feature map of the query image and a feature map of a reference image in the query-reference image pair;
subtracting the feature map of the query image from the feature map of the reference image to obtain a first feature map;
performing element-by-element squaring operation on the first feature map to obtain a second feature map;
and carrying out batch normalization processing on the second feature graph to obtain the similarity feature of the query-reference image pair.
In another possible implementation manner, the weights of the edges corresponding to the two reference images in the complete graph are:
Figure BDA0002179793950000021
wherein S (g)i,gj) Is the similarity of the reference images i, j.
In another possible implementation manner, the mapping the similarity feature of each query-reference image pair to a message feature through a message network includes:
classifying the similarity features of the query-reference image pair through a full connection layer of the message network to obtain the classified similarity features;
carrying out batch normalization processing on the classified similarity features through a batch normalization layer of the message network to obtain similarity features after normalization processing;
and converting the linear factors in the similarity features after the normalization processing into nonlinear factors through an activation layer of the message network to obtain the message features of the query-reference image pair.
In another possible implementation, the determining, from the plurality of reference images, an image that matches the query image according to the similarity score of each query-reference image pair includes:
determining a reference image with the highest similarity score in the plurality of reference images as an image matching the query image; or
And determining the reference images with similarity scores higher than a preset value in the plurality of reference images as the images matched with the query image.
In another possible implementation manner, the calculating the similarity score between each two reference images of the plurality of reference images includes:
performing region division on the two reference images according to the same division method;
calculating the logarithm relative RGB coordinates of each pixel point of each area of the two reference images;
clustering the pixel points in each region of the two reference images according to the logarithm relative RGB coordinates of each pixel point in each region of the two reference images to obtain a clustering center of each region of the two reference images;
determining a similarity score between the two reference images according to the clustering center of each region of the two reference images.
In another possible implementation manner, the calculating the similarity score between each two reference images of the plurality of reference images includes:
respectively inputting the two reference images into a first depth residual error network and a second depth residual error network, obtaining the overall characteristics of the two reference images from the first depth residual error network, and obtaining the local characteristics of the two reference images from the second depth residual error network;
and calculating a similarity score between the two reference images according to the overall characteristics and the local characteristics of the two reference images.
A second aspect of the present application provides an image recognition apparatus, the apparatus comprising:
an acquisition module for acquiring a query image and a plurality of reference images;
the extraction module is used for forming the query image and each reference image into a query-reference image pair and extracting the similarity characteristic of each query-reference image pair;
a construction module for constructing a complete graph with the query-reference image pairs as nodes, each reference image corresponding to a node;
a first determining module, configured to calculate a similarity score between each two reference images in the multiple reference images, and determine a weight of an edge corresponding to the two reference images in the complete graph according to the similarity score between the two reference images;
a mapping module for mapping the similarity features of each query-reference image pair to message features over a message network;
the updating module is used for updating the similarity characteristic of each query-reference image pair according to the message characteristic of each query-reference image pair and the weight of each edge in the complete graph;
a calculation module for calculating a similarity score for each query-reference image pair based on the updated similarity features for each query-reference image pair;
a second determination module to determine images from the plurality of reference images that match the query image based on the similarity score for each query-reference image pair.
In another possible implementation manner, the extracting, by the extraction module, the similarity feature of each query-reference image pair includes:
inputting the query-reference image pair into a twin neural network based on a residual error network to obtain a feature map of the query image and a feature map of a reference image in the query-reference image pair;
subtracting the feature map of the query image from the feature map of the reference image to obtain a first feature map;
performing element-by-element squaring operation on the first feature map to obtain a second feature map;
and carrying out batch normalization processing on the second feature graph to obtain the similarity feature of the query-reference image pair.
In another possible implementation manner, the weights of the edges corresponding to the two reference images in the complete graph are:
Figure BDA0002179793950000051
wherein S (g)i,gj) Is the similarity of the reference images i, j.
In another possible implementation manner, the mapping module is specifically configured to:
classifying the similarity features of the query-reference image pair through a full connection layer of the message network to obtain the classified similarity features;
carrying out batch normalization processing on the classified similarity features through a batch normalization layer of the message network to obtain similarity features after normalization processing;
and converting the linear factors in the similarity features after the normalization processing into nonlinear factors through an activation layer of the message network to obtain the message features of the query-reference image pair.
In another possible implementation manner, the second determining module is specifically configured to:
determining a reference image with the highest similarity score in the plurality of reference images as an image matching the query image; or
And determining the reference images with similarity scores higher than a preset value in the plurality of reference images as the images matched with the query image.
In another possible implementation manner, the calculation module is specifically configured to:
performing region division on the two reference images according to the same division method;
calculating the logarithm relative RGB coordinates of each pixel point of each area of the two reference images;
clustering the pixel points in each region of the two reference images according to the logarithm relative RGB coordinates of each pixel point in each region of the two reference images to obtain a clustering center of each region of the two reference images;
determining a similarity score between the two reference images according to the clustering center of each region of the two reference images.
In another possible implementation manner, the calculation module is specifically configured to:
respectively inputting the two reference images into a first depth residual error network and a second depth residual error network, obtaining the overall characteristics of the two reference images from the first depth residual error network, and obtaining the local characteristics of the two reference images from the second depth residual error network;
and calculating a similarity score between the two reference images according to the overall characteristics and the local characteristics of the two reference images.
A third aspect of the application provides a computer apparatus comprising a memory for storing at least one instruction and a processor for executing the at least one instruction to implement the image recognition method.
A fourth aspect of the present application provides a computer-readable storage medium having stored thereon at least one instruction which, when executed by a processor, implements the image recognition method.
It is seen from the above technical solutions that the present invention provides an image recognition method, apparatus, computer apparatus and computer readable storage medium, the method obtains a query image and a plurality of reference images; forming a query-reference image pair by the query image and each reference image, and extracting similarity characteristics of each query-reference image pair; constructing a complete graph by taking the query-reference image pairs as nodes, wherein each reference image corresponds to one node; calculating similarity scores between every two reference images in the multiple reference images, and determining the weight values of edges corresponding to the two reference images in the complete image according to the similarity scores between the two reference images; mapping the similarity features of each query-reference image pair into message features through a message network; updating the similarity characteristic of each query-reference image pair according to the message characteristic of each query-reference image pair and the weight of each edge in the complete graph; calculating a similarity score for each query-reference image pair based on the updated similarity features for each query-reference image pair; determining images from the plurality of reference images that match the query image based on the similarity score for each query-reference image pair. The method and the device utilize the similarity information between the reference images to update the similarity characteristics of the query-reference image pair, and improve the accuracy of image identification.
Drawings
In order to more clearly illustrate the embodiments of the present invention or the technical solutions in the prior art, the drawings used in the description of the embodiments or the prior art will be briefly described below, it is obvious that the drawings in the following description are only embodiments of the present invention, and for those skilled in the art, other drawings can be obtained according to the provided drawings without creative efforts.
FIG. 1 is a flow chart of an image recognition method provided by an embodiment of the invention;
FIG. 2 is a functional block diagram of an image recognition apparatus according to an embodiment of the present invention;
FIG. 3 is a schematic structural diagram of a computer device according to a preferred embodiment of the present invention.
Detailed Description
The technical solutions in the embodiments of the present invention will be clearly and completely described below with reference to the drawings in the embodiments of the present invention, and it is obvious that the described embodiments are only a part of the embodiments of the present invention, and not all of the embodiments. All other embodiments, which can be derived by a person skilled in the art from the embodiments given herein without making any creative effort, shall fall within the protection scope of the present invention.
In order to make the aforementioned objects, features and advantages of the present invention comprehensible, embodiments accompanied with figures are described in further detail below.
Fig. 1 is a flowchart of an image recognition method according to an embodiment of the present invention. The image identification method is applied to a computer device and used for matching a query image with a reference image. The order of the steps in the flow chart may be changed and some steps may be omitted according to different needs.
S11, a query image and a plurality of reference images are obtained.
The query image is an image to be identified and the reference image is a known image. The method identifies an image from the plurality of reference images that contains the same content (e.g., an object or person) as the query image. For example, when object identification is required, the query image is an image containing an unknown object, the reference images are images containing known objects, and the method identifies a reference image containing the same object as the query image from the reference images. For another example, when person identification is required, the query image is an image containing an unknown person, the reference images are images containing known persons, and the method identifies an image containing the same person as the query image from the reference images.
The query image may be received from an external device. For example, a monitoring image captured by an external camera is acquired, and the monitoring image captured by the external camera is used as the query image.
Or, the computer device may include a camera, and may control a built-in camera of the computer device to capture an image, where the image captured by the built-in camera is used as the query image.
Alternatively, an image stored in advance by the computer device may be read, and the read image stored in advance may be used as the query image.
Alternatively, an image may be downloaded from a network, and the downloaded image may be used as the query image.
The plurality of reference images may be obtained from a preset image library. For example, when person identification is performed, the plurality of reference images may be acquired from a person image library.
And S12, forming a query-reference image pair by the query image and each reference image, and extracting the similarity characteristic of each query-reference image pair.
In this embodiment, the query image is paired with each reference image to form a plurality of query-reference image pairs.
A twin neural network based on a residual network may be utilized to extract similarity features for each query-reference image pair.
Preferably, said extracting similarity features for each query-reference image pair comprises:
inputting the query-reference image pair into a twin neural network based on a residual error network to obtain a feature map of the query image and a feature map of a reference image in the query-reference image pair;
subtracting the feature map of the query image from the feature map of the reference image to obtain a first feature map;
performing element-by-element squaring operation on the first feature map to obtain a second feature map;
and carrying out batch normalization processing on the second feature graph to obtain the similarity feature of the query-reference image pair.
The twin neural network based on the residual error network is two connected neural networks sharing a weight value, wherein one of the two connected neural networks takes the query image as input, and the other one takes a reference image in the query-reference image pair as input.
In this embodiment, the twin neural network is trained using a query-reference sample image pair in advance. The query-reference sample image pair is an image pair of the query sample image and the reference sample image. Each query-reference sample image pair has a label indicating whether the query sample image and the reference sample image in the query-reference sample image pair contain the same content. If the query sample image and the reference sample image contain the same content (e.g., the same person), the tag may be 1. The label may be 0 if the query sample image and the reference sample image contain different content (e.g., different people).
When the twin neural network is trained, extracting the similarity characteristics of the query-reference sample image pair, inputting the similarity characteristics of the query-reference sample image pair into a linear classifier to obtain the similarity score of the query-reference sample image pair, and obtaining the similarity score according to the similarity scoreCalculating a loss function with the labels of the query-reference sample image pair, and adjusting parameters of the twin neural network to minimize the loss function. Wherein the linear classifier may be a non-linear function, i.e. sigmoid function, of the formula f (x) 1/(1+ e)-x). The loss function may be:
wherein d isiIs a similarity feature of the ith query-reference sample image pair, F () represents a linear classifier, yiA label representing the ith query-reference sample image pair.
And S13, constructing a complete graph by taking the query-reference image pairs as nodes, wherein each reference image corresponds to one node.
The complete graph is a simple graph with an edge connected between each pair of nodes, that is, the nodes in the complete graph are connected two by two.
In this embodiment, each node of the full graph represents a query-reference image pair, and each edge of the full graph corresponds to two reference images and represents a relationship between the two reference images.
S14, calculating similarity scores between every two reference images in the multiple reference images, and determining the weight values of edges corresponding to the two reference images in the complete graph according to the similarity scores between the two reference images.
In this embodiment, the calculating the similarity score between each two reference images of the plurality of reference images includes:
performing region division on the two reference images according to the same division method;
calculating the logarithm relative RGB coordinates of each pixel point of each area of the two reference images;
clustering the pixel points in each region of the two reference images according to the logarithm relative RGB coordinates of each pixel point in each region of the two reference images to obtain a clustering center of each region of the two reference images;
determining a similarity score between the two reference images according to the clustering center of each region of the two reference images.
The two reference images may be each divided into upper and lower two regions or left and right two regions. It is also possible to divide the two reference pictures into more than two regions each, for example into three regions or four regions each.
The red component being RiGreen component is GiAnd the logarithm of the pixel point i with the blue component of Bi is (x) relative to the RGB coordinatei,yi) Which is
Figure BDA0002179793950000101
In (1),
Figure BDA0002179793950000102
can take the logarithm of the base e, i.e.
Figure BDA0002179793950000103
Figure BDA0002179793950000104
Alternatively, the base logarithm may be taken to be other values, such as base 10 logarithm.
The pixels in each region of the two reference images may be clustered using GMM (Gaussian Mixture Model) or K-Means algorithm to obtain a cluster center of each region of the two reference images.
The distance of the cluster centers of each of the two reference images may be calculated, and the similarity between the two reference images may be determined based on the distance of the cluster centers of each of the two reference images. The weighted sum of the distances of the cluster centers of each region of the two reference images may be taken as the similarity between the two reference images. The distance of the cluster centers of each region of the two reference images may be euclidean distance, manhattan distance, mahalanobis distance, or the like.
In another embodiment, the two reference images may be input into a neural network to extract features, and a similarity score between the two reference images may be calculated according to the features of the two reference images. For example, the two reference images are respectively input into a first depth residual error network and a second depth residual error network, the overall characteristics of the two reference images are obtained from the first depth residual error network, the local characteristics of the two reference images are obtained from the second depth residual error network, and the similarity score between the two reference images is calculated according to the overall characteristics and the local characteristics of the two reference images.
In this embodiment, the weight of the edge corresponding to the two reference images in the complete graph may be represented as:
Figure BDA0002179793950000111
wherein, S (g)i,gj) Is the similarity of the reference images i, j.
In another embodiment, each two reference images in the plurality of reference images may be combined into a reference image pair, the similarity feature of each reference image pair is extracted, and the similarity feature of each reference image pair is input into a linear classifier to obtain the similarity score of each reference image pair. Extracting similarity features for each reference image pair may be referred to S12.
And S15, mapping the similarity characteristic of each query-reference image pair into a message characteristic through a message network.
The messaging network is a neural network. In this embodiment, the message network is composed of a fully connected layer, a bulk normalization layer, and an active layer.
The similarity feature of the ith query-reference image pair is denoted as diThe message characteristic of the ith query-reference image pair is denoted as ti,ti=F(di) I is 1,2, … N (indicating that there are N reference pictures).
Preferably, the mapping the similarity feature of each query-reference image pair to a message feature through a message network comprises:
the full connection layer of the message network classifies the similarity characteristics of the query-reference image pair to obtain the classified similarity characteristics;
the batch normalization layer of the message network performs batch normalization processing on the classified similarity features to obtain the similarity features after normalization processing;
and the activation layer of the message network converts the linear factors in the similarity characteristics after the normalization processing into nonlinear factors to obtain the message characteristics of the query-reference image pair.
In this embodiment, the fully-connected layer of the message network classifies the similarity features of the query-reference image pair, so that the similarity features with high similarity enter the next layer of the message network, that is, the batch normalization layer. The batch normalization layer in the message network continuously adjusts the intermediate output of the message network by using the mean value and the standard deviation on the small batch, so that the numerical value of the intermediate output of the whole message network in each layer is more stable. The activation layer in the message network converts the linear factors in the similarity characteristics transferred by the previous layer (namely the batch normalization layer) into the nonlinear factors through the activation function, so that the problem that the linear factors cannot be solved is solved.
In this embodiment, two message networks may be used to map the similarity features of each query-reference image pair to message features. For example, a message network consisting of a fully connected layer, a batch normalization layer, and an active layer is connected after a message network consisting of a fully connected layer, a batch normalization layer, and an active layer. More accurate depth characteristic information can be extracted through the two-layer message network.
And S16, updating the similarity feature of each query-reference image pair according to the message feature of each query-reference image pair and the weight value of each edge in the complete graph.
And updating the similarity feature of each query-reference image pair according to the message feature of each query-reference image pair and the weight value of each edge in the complete graph, namely updating the similarity feature of each query-reference image pair according to the connection relation between each node and other nodes of the complete graph. For each node in the complete graph, using the message characteristics of the query-reference image pair corresponding to other nodes connected with the node as the input characteristics of the node, and updating the similarity characteristics of the query-reference image pair corresponding to the node into weighted fusion of all the input characteristics and the original similarity characteristics, namely
Figure BDA0002179793950000121
Wherein
Figure BDA0002179793950000122
Indicating the updated ith similarity characteristic,
Figure BDA0002179793950000123
indicating the i-th similarity characteristic before updating,
Figure BDA0002179793950000124
representing the message characteristics from node j and alpha representing a weighting parameter that balances the fused characteristics and the original characteristics.
The similarity features of the query-reference image pair may be iteratively updated as follows:
and S17, calculating a similarity score of each query-reference image pair according to the updated similarity characteristics of each query-reference image pair.
In this embodiment, the updated similarity features of each query-reference image pair may be input into a linear classifier to obtain a similarity score of each query-reference image pair.
The linear classifier may be a non-linear function, i.e. sigmoid function, of the formula f (x) 1/(1+ e)-x)。
S18, determining images from the plurality of reference images that match the query image based on the similarity score for each query-reference image pair.
Preferably, the determining, from the plurality of reference images, an image that matches the query image according to the similarity score for each query-reference image pair comprises:
determining a reference image with the highest similarity score in the plurality of reference images as an image matching the query image; or
And determining the reference images with similarity scores higher than a preset value in the plurality of reference images as the images matched with the query image.
For example, there are 20 reference images, and the reference image having the highest similarity score among the 20 reference images is determined as the image matching the query image, or the reference image having a similarity score higher than 0.9 among the 20 reference images is determined as the image matching the query image.
The image identification method of the invention obtains a query image and a plurality of reference images; forming a query-reference image pair by the query image and each reference image, and extracting similarity characteristics of each query-reference image pair; constructing a complete graph by taking the query-reference image pairs as nodes, wherein each reference image corresponds to one node; calculating similarity scores between every two reference images in the multiple reference images, and determining the weight values of edges corresponding to the two reference images in the complete image according to the similarity scores between the two reference images; mapping the similarity features of each query-reference image pair into message features through a message network; updating the similarity characteristic of each query-reference image pair according to the message characteristic of each query-reference image pair and the weight of each edge in the complete graph; calculating a similarity score for each query-reference image pair based on the updated similarity features for each query-reference image pair; determining images from the plurality of reference images that match the query image based on the similarity score for each query-reference image pair. The method utilizes the similarity information between the reference images to update the similarity characteristics of the query-reference image pair, and improves the accuracy of image identification.
As shown in fig. 2, fig. 2 is a functional block diagram of an image recognition apparatus according to an embodiment of the present invention. The image recognition device 20 includes an acquisition module 210, an extraction module 220, a construction module 230, a first determination module 240, a mapping module 250, an update module 260, a calculation module 270, and a second determination module 280. The module referred to in the present invention refers to a series of computer program segments capable of being executed by a processor of a computer device and performing a fixed function, which are stored in a memory of the computer device.
The obtaining module 210 is configured to obtain a query image and a plurality of reference images.
The query image is an image to be identified and the reference image is a known image. The method identifies an image from the plurality of reference images that contains the same content (e.g., an object or person) as the query image. For example, when object identification is required, the query image is an image containing an unknown object, the reference images are images containing known objects, and the method identifies a reference image containing the same object as the query image from the reference images. For another example, when person identification is required, the query image is an image containing an unknown person, the reference images are images containing known persons, and the method identifies an image containing the same person as the query image from the reference images.
The query image may be received from an external device. For example, a monitoring image captured by an external camera is acquired, and the monitoring image captured by the external camera is used as the query image.
Or, the computer device may include a camera, and may control a built-in camera of the computer device to capture an image, where the image captured by the built-in camera is used as the query image.
Alternatively, an image stored in advance by the computer device may be read, and the read image stored in advance may be used as the query image.
Alternatively, an image may be downloaded from a network, and the downloaded image may be used as the query image.
The plurality of reference images may be obtained from a preset image library. For example, when person identification is performed, the plurality of reference images may be acquired from a person image library.
The extracting module 220 is configured to combine the query image and each reference image into a query-reference image pair, and extract a similarity feature of each query-reference image pair.
In this embodiment, the query image is paired with each reference image to form a plurality of query-reference image pairs.
A twin neural network based on a residual network may be utilized to extract similarity features for each query-reference image pair.
Preferably, the extraction module extracts similarity features of each query-reference image pair, in particular for:
inputting the query-reference image pair into a twin neural network based on a residual error network to obtain a feature map of the query image and a feature map of a reference image in the query-reference image pair;
subtracting the feature map of the query image from the feature map of the reference image to obtain a first feature map;
performing element-by-element squaring operation on the first feature map to obtain a second feature map;
and carrying out batch normalization processing on the second feature graph to obtain the similarity feature of the query-reference image pair.
The twin neural network based on the residual error network is two connected neural networks sharing a weight value, wherein one of the two connected neural networks takes the query image as input, and the other one takes a reference image in the query-reference image pair as input.
In this embodiment, the twin neural network is trained using a query-reference sample image pair in advance. The query-reference sample image pair is an image pair of the query sample image and the reference sample image. Each query-reference sample image pair has a label indicating whether the query sample image and the reference sample image in the query-reference sample image pair contain the same content. If the query sample image and the reference sample image contain the same content (e.g., the same person), the tag may be 1. The label may be 0 if the query sample image and the reference sample image contain different content (e.g., different people).
When the twin neural network is trained, extracting the similarity characteristics of the query-reference sample image pair, inputting the similarity characteristics of the query-reference sample image pair into a linear classifier to obtain the similarity score of the query-reference sample image pair, calculating a loss function according to the similarity score and the label of the query-reference sample image pair, and adjusting the parameters of the twin neural network to minimize the loss function. Wherein the linear classifier may be a non-linear function, i.e. sigmoid function, of the formula f (x) 1/(1+ e)-x). The loss function may be:
Figure BDA0002179793950000161
wherein d isiIs a similarity feature of the ith query-reference sample image pair, F () represents a linear classifier, yiA label representing the ith query-reference sample image pair.
The constructing module 230 is configured to construct a complete graph with the query-reference image pairs as nodes, where each reference image corresponds to a node.
The complete graph is a simple graph with an edge connected between each pair of nodes, that is, the nodes in the complete graph are connected two by two.
In this embodiment, each node of the full graph represents a query-reference image pair, and each edge of the full graph corresponds to two reference images and represents a relationship between the two reference images.
The first determining module 240 is configured to calculate a similarity score between each two reference images in the plurality of reference images, and determine a weight of an edge corresponding to the two reference images in the complete graph according to the similarity score between the two reference images.
In this embodiment, the calculating module calculates a similarity score between each two reference images in the plurality of reference images, and is specifically configured to:
performing region division on the two reference images according to the same division method;
calculating the logarithm relative RGB coordinates of each pixel point of each area of the two reference images;
clustering the pixel points in each region of the two reference images according to the logarithm relative RGB coordinates of each pixel point in each region of the two reference images to obtain a clustering center of each region of the two reference images;
determining a similarity score between the two reference images according to the clustering center of each region of the two reference images.
The two reference images may be each divided into upper and lower two regions or left and right two regions. It is also possible to divide the two reference pictures into more than two regions each, for example into three regions or four regions each.
The red component being RiGreen component is GiAnd the logarithm of the pixel point i with the blue component of Bi is (x) relative to the RGB coordinatei,yi) Which is
Figure BDA0002179793950000162
In (1),can take the logarithm of the base e, i.e.
Figure BDA0002179793950000164
Figure BDA0002179793950000171
Alternatively, the base logarithm may be taken to be other values, such as base 10 logarithm.
The pixels in each region of the two reference images may be clustered using GMM (Gaussian Mixture Model) or K-Means algorithm to obtain a cluster center of each region of the two reference images.
The distance of the cluster centers of each of the two reference images may be calculated, and the similarity between the two reference images may be determined based on the distance of the cluster centers of each of the two reference images. The weighted sum of the distances of the cluster centers of each region of the two reference images may be taken as the similarity between the two reference images. The distance of the cluster centers of each region of the two reference images may be euclidean distance, manhattan distance, mahalanobis distance, or the like.
In another embodiment, the two reference images may be input into a neural network to extract features, and a similarity score between the two reference images may be calculated according to the features of the two reference images. For example, the two reference images are respectively input into a first depth residual error network and a second depth residual error network, the overall characteristics of the two reference images are obtained from the first depth residual error network, the local characteristics of the two reference images are obtained from the second depth residual error network, and the similarity score between the two reference images is calculated according to the overall characteristics and the local characteristics of the two reference images.
In this embodiment, the weight of the edge corresponding to the two reference images in the complete graph may be represented as:
Figure BDA0002179793950000172
wherein, S (g)i,gj) Is the similarity of the reference images i, j.
In another embodiment, each two reference images in the plurality of reference images may be combined into a reference image pair, the similarity feature of each reference image pair is extracted, and the similarity feature of each reference image pair is input into a linear classifier to obtain the similarity score of each reference image pair. Extracting the similarity features for each reference image pair may extract the similarity features for each query-reference image pair with reference to the extraction module 220.
The mapping module 250 is configured to map the similarity features of each query-reference image pair into message features through a message network.
The messaging network is a neural network. In this embodiment, the message network is composed of a fully connected layer, a bulk normalization layer, and an active layer.
The similarity feature of the ith query-reference image pair is denoted as diThe message characteristic of the ith query-reference image pair is denoted as ti,ti=F(di) I is 1,2, … N (indicating that there are N reference pictures).
Preferably, the mapping module maps the similarity feature of each query-reference image pair to a message feature through a message network, and is specifically configured to:
the full connection layer of the message network classifies the similarity characteristics of the query-reference image pair to obtain the classified similarity characteristics;
the batch normalization layer of the message network performs batch normalization processing on the classified similarity features to obtain the similarity features after normalization processing;
and the activation layer of the message network converts the linear factors in the similarity characteristics after the normalization processing into nonlinear factors to obtain the message characteristics of the query-reference image pair.
In this embodiment, the fully-connected layer of the message network classifies the similarity features of the query-reference image pair, so that the similarity features with high similarity enter the next layer of the message network, that is, the batch normalization layer. The batch normalization layer in the message network continuously adjusts the intermediate output of the message network by using the mean value and the standard deviation on the small batch, so that the numerical value of the intermediate output of the whole message network in each layer is more stable. The activation layer in the message network converts the linear factors in the similarity characteristics transferred by the previous layer (namely the batch normalization layer) into the nonlinear factors through the activation function, so that the problem that the linear factors cannot be solved is solved.
In this embodiment, two message networks may be used to map the similarity features of each query-reference image pair to message features. For example, a message network consisting of a fully connected layer, a batch normalization layer, and an active layer is connected after a message network consisting of a fully connected layer, a batch normalization layer, and an active layer. More accurate depth characteristic information can be extracted through the two-layer message network.
The updating module 260 is configured to update the similarity feature of each query-reference image pair according to the message feature of each query-reference image pair and the weight of each edge in the complete graph.
And updating the similarity feature of each query-reference image pair according to the message feature of each query-reference image pair and the weight value of each edge in the complete graph, namely updating the similarity feature of each query-reference image pair according to the connection relation between each node and other nodes of the complete graph. For each node in the complete graph, using the message characteristics of the query-reference image pair corresponding to other nodes connected with the node as the input characteristics of the node, and updating the similarity characteristics of the query-reference image pair corresponding to the node into weighted fusion of all the input characteristics and the original similarity characteristics, namely
Figure BDA0002179793950000191
Wherein
Figure BDA0002179793950000192
Indicating the updated ith similarity characteristic,
Figure BDA0002179793950000193
indicating the i-th similarity characteristic before updating,
Figure BDA0002179793950000194
representing the message characteristics from node j and alpha representing a weighting parameter that balances the fused characteristics and the original characteristics.
The similarity features of the query-reference image pair may be iteratively updated as follows:
Figure BDA0002179793950000195
the calculating module 270 is configured to calculate a similarity score for each query-reference image pair according to the updated similarity features of each query-reference image pair.
In this embodiment, the updated similarity features of each query-reference image pair may be input into a linear classifier to obtain a similarity score of each query-reference image pair.
The linear classifier may be a non-linear function, i.e. sigmoid function, of the formula f (x) 1/(1+ e)-x)。
The second determining module 280 is configured to determine images matching the query image from the plurality of reference images according to the similarity score of each query-reference image pair.
Preferably, the second determination module determines, from the plurality of reference images, an image matching the query image according to the similarity score of each query-reference image pair, in particular for:
determining a reference image with the highest similarity score in the plurality of reference images as an image matching the query image; or
And determining the reference images with similarity scores higher than a preset value in the plurality of reference images as the images matched with the query image.
For example, there are 20 reference images, and the reference image having the highest similarity score among the 20 reference images is determined as the image matching the query image, or the reference image having a similarity score higher than 0.9 among the 20 reference images is determined as the image matching the query image.
The image recognition apparatus 20 of the present invention acquires a query image and a plurality of reference images; forming a query-reference image pair by the query image and each reference image, and extracting similarity characteristics of each query-reference image pair; constructing a complete graph by taking the query-reference image pairs as nodes, wherein each reference image corresponds to one node; calculating similarity scores between every two reference images in the multiple reference images, and determining the weight values of edges corresponding to the two reference images in the complete image according to the similarity scores between the two reference images; mapping the similarity features of each query-reference image pair into message features through a message network; updating the similarity characteristic of each query-reference image pair according to the message characteristic of each query-reference image pair and the weight of each edge in the complete graph; calculating a similarity score for each query-reference image pair based on the updated similarity features for each query-reference image pair; determining images from the plurality of reference images that match the query image based on the similarity score for each query-reference image pair. The image recognition device 20 of the present invention updates the similarity characteristics of the query-reference image pair using the similarity information between the reference images, improving the accuracy of image recognition.
The integrated unit implemented in the form of a software functional module may be stored in a computer-readable storage medium. The software functional module is stored in a computer readable storage medium and includes several instructions for causing a computer device (which may be a personal computer, a server, or a network device) or a processor (processor) to execute some steps of the methods according to the embodiments of the present invention.
Fig. 3 is a schematic structural diagram of a computer device 3 according to a preferred embodiment of the present invention for implementing an image recognition display method. In the present embodiment, the computer device 3 comprises at least one transmitting device 31, at least one memory 32, at least one processor 33, at least one receiving device 34 and at least one communication bus. Wherein the communication bus is used for realizing connection communication among the components.
The computer device 3 is a device capable of automatically performing numerical calculation and/or information processing according to a preset or stored instruction, and its hardware includes, but is not limited to, a microprocessor, an Application Specific Integrated Circuit (ASIC), a Programmable Gate Array (FPGA), a Digital Signal Processor (DSP), an embedded device, and the like. The computer arrangement 3 may also comprise network equipment and/or user equipment. Wherein the network device includes, but is not limited to, a single network server, a server group consisting of a plurality of network servers, or a Cloud Computing (Cloud Computing) based Cloud consisting of a large number of hosts or network servers, wherein Cloud Computing is one of distributed Computing, a super virtual computer consisting of a collection of loosely coupled computers.
The computer device 3 may be, but is not limited to, any electronic product that can perform human-computer interaction with a user through a keyboard, a touch pad, or a voice control device, for example, a tablet computer, a smart phone, a monitoring device, and other terminals.
The Network where the computer device 3 is located includes, but is not limited to, the internet, a wide area Network, a metropolitan area Network, a local area Network, a Virtual Private Network (VPN), and the like.
The receiving device 34 and the transmitting device 31 may be wired transmitting ports, or may be wireless devices, for example, including antenna devices, for performing data communication with other devices.
The memory 32 is used to store program code. The memory 32 may be a memory bank, a TF Card (Trans-flash Card), a smart media Card (smart media Card), a secure digital Card (secure digital Card), a flash memory Card (flash Card), or other storage devices.
The processor 33 may comprise one or more microprocessors, digital processors. The processor 33 may call program code stored in the memory 32 to perform the associated functions. For example, the modules described in fig. 2 are program codes stored in the memory 32 and executed by the processor 33 to implement an image recognition display method. The processor 33 is also called a Central Processing Unit (CPU), and is an ultra-large scale integrated circuit, which is an operation Core (Core) and a Control Core (Control Unit).
In the embodiments provided in the present invention, it should be understood that the disclosed system, apparatus and method may be implemented in other ways. For example, the above-described apparatus embodiments are merely illustrative, and for example, the division of the modules is only one logical functional division, and other divisions may be realized in practice.
The modules described as separate parts may or may not be physically separate, and parts displayed as modules may or may not be physical units, may be located in one place, or may be distributed on a plurality of network units. Some or all of the modules may be selected according to actual needs to achieve the purpose of the solution of the present embodiment.
In addition, functional modules in the embodiments of the present invention may be integrated into one processing unit, or each unit may exist alone physically, or two or more units are integrated into one unit. The integrated unit can be realized in a form of hardware, or in a form of hardware plus a software functional module.
It will be evident to those skilled in the art that the invention is not limited to the details of the foregoing illustrative embodiments, and that the present invention may be embodied in other specific forms without departing from the spirit or essential attributes thereof. The present embodiments are therefore to be considered in all respects as illustrative and not restrictive, the scope of the invention being indicated by the appended claims rather than by the foregoing description, and all changes which come within the meaning and range of equivalency of the claims are therefore intended to be embraced therein. Any reference signs in the claims shall not be construed as limiting the claim concerned. Furthermore, it is obvious that the word "comprising" does not exclude other elements or steps, and the singular does not exclude the plural. A plurality of units or means recited in the system claims may also be implemented by one unit or means in software or hardware. The terms second, etc. are used to denote names, but not any particular order.
Finally, it should be noted that the above embodiments are only for illustrating the technical solutions of the present invention and not for limiting, and although the present invention is described in detail with reference to the preferred embodiments, it should be understood by those skilled in the art that modifications or equivalent substitutions may be made on the technical solutions of the present invention without departing from the spirit and scope of the technical solutions of the present invention.

Claims (10)

1. An image recognition method, characterized in that the method comprises:
acquiring a query image and a plurality of reference images;
forming a query-reference image pair by the query image and each reference image, and extracting similarity characteristics of each query-reference image pair;
constructing a complete graph by taking the query-reference image pairs as nodes, wherein each reference image corresponds to one node;
calculating similarity scores between every two reference images in the multiple reference images, and determining the weight values of edges corresponding to the two reference images in the complete image according to the similarity scores between the two reference images;
mapping the similarity features of each query-reference image pair into message features through a message network;
updating the similarity characteristic of each query-reference image pair according to the message characteristic of each query-reference image pair and the weight of each edge in the complete graph;
calculating a similarity score for each query-reference image pair based on the updated similarity features for each query-reference image pair;
determining images from the plurality of reference images that match the query image based on the similarity score for each query-reference image pair.
2. The method of claim 1, wherein said extracting similarity features for each query-reference image pair comprises:
inputting the query-reference image pair into a twin neural network based on a residual error network to obtain a feature map of the query image and a feature map of a reference image in the query-reference image pair;
subtracting the feature map of the query image from the feature map of the reference image to obtain a first feature map;
performing element-by-element squaring operation on the first feature map to obtain a second feature map;
and carrying out batch normalization processing on the second feature graph to obtain the similarity feature of the query-reference image pair.
3. The method of claim 1, wherein the weights of the edges corresponding to the two reference images in the complete graph are:
Figure FDA0002179793940000021
wherein S (g)i,gj) Is the similarity of the reference images i, j.
4. The method of claim 1, wherein mapping the similarity features of each query-reference image pair to message features over a message network comprises:
classifying the similarity features of the query-reference image pair through a full connection layer of the message network to obtain the classified similarity features;
carrying out batch normalization processing on the classified similarity features through a batch normalization layer of the message network to obtain similarity features after normalization processing;
and converting the linear factors in the similarity features after the normalization processing into nonlinear factors through an activation layer of the message network to obtain the message features of the query-reference image pair.
5. The method of claim 1, wherein determining images from the plurality of reference images that match the query image based on the similarity score for each query-reference image pair comprises:
determining a reference image with the highest similarity score in the plurality of reference images as an image matching the query image; or
And determining the reference images with similarity scores higher than a preset value in the plurality of reference images as the images matched with the query image.
6. The method of any one of claims 1-5, wherein the calculating the similarity score between each two of the plurality of reference images comprises:
performing region division on the two reference images according to the same division method;
calculating the logarithm relative RGB coordinates of each pixel point of each area of the two reference images;
clustering the pixel points in each region of the two reference images according to the logarithm relative RGB coordinates of each pixel point in each region of the two reference images to obtain a clustering center of each region of the two reference images;
determining a similarity score between the two reference images according to the clustering center of each region of the two reference images.
7. The method of any one of claims 1-5, wherein the calculating the similarity score between each two of the plurality of reference images comprises:
respectively inputting the two reference images into a first depth residual error network and a second depth residual error network, obtaining the overall characteristics of the two reference images from the first depth residual error network, and obtaining the local characteristics of the two reference images from the second depth residual error network;
and calculating a similarity score between the two reference images according to the overall characteristics and the local characteristics of the two reference images.
8. An image recognition apparatus, characterized in that the apparatus comprises:
an acquisition module for acquiring a query image and a plurality of reference images;
the extraction module is used for forming the query image and each reference image into a query-reference image pair and extracting the similarity characteristic of each query-reference image pair;
a construction module for constructing a complete graph with the query-reference image pairs as nodes, each reference image corresponding to a node;
a first determining module, configured to calculate a similarity score between each two reference images in the multiple reference images, and determine a weight of an edge corresponding to the two reference images in the complete graph according to the similarity score between the two reference images;
a mapping module for mapping the similarity features of each query-reference image pair to message features over a message network;
the updating module is used for updating the similarity characteristic of each query-reference image pair according to the message characteristic of each query-reference image pair and the weight of each edge in the complete graph;
a calculation module for calculating a similarity score for each query-reference image pair based on the updated similarity features for each query-reference image pair;
a second determination module to determine images from the plurality of reference images that match the query image based on the similarity score for each query-reference image pair.
9. A computer device, characterized in that the computer device comprises a memory for storing at least one instruction and a processor for executing the at least one instruction to implement the image recognition method according to any one of claims 1 to 7.
10. A computer-readable storage medium having stored thereon computer instructions, which when executed by a processor, implement the image recognition method of any one of claims 1 to 7.
CN201910792041.1A 2019-08-26 2019-08-26 Image recognition method, image recognition device, computer device, and storage medium Pending CN110689046A (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
CN201910792041.1A CN110689046A (en) 2019-08-26 2019-08-26 Image recognition method, image recognition device, computer device, and storage medium
PCT/CN2020/086768 WO2021036309A1 (en) 2019-08-26 2020-04-24 Image recognition method and apparatus, computer apparatus, and storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910792041.1A CN110689046A (en) 2019-08-26 2019-08-26 Image recognition method, image recognition device, computer device, and storage medium

Publications (1)

Publication Number Publication Date
CN110689046A true CN110689046A (en) 2020-01-14

Family

ID=69108626

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910792041.1A Pending CN110689046A (en) 2019-08-26 2019-08-26 Image recognition method, image recognition device, computer device, and storage medium

Country Status (2)

Country Link
CN (1) CN110689046A (en)
WO (1) WO2021036309A1 (en)

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111950493A (en) * 2020-08-20 2020-11-17 华北电力大学 Image recognition method and device, terminal equipment and readable storage medium
CN112307934A (en) * 2020-10-27 2021-02-02 深圳市商汤科技有限公司 Image detection method, and training method, device, equipment and medium of related model
WO2021036309A1 (en) * 2019-08-26 2021-03-04 深圳壹账通智能科技有限公司 Image recognition method and apparatus, computer apparatus, and storage medium
WO2024061311A1 (en) * 2022-09-21 2024-03-28 北京沃东天骏信息技术有限公司 Model training method and apparatus, and image classification method and apparatus

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN116188805B (en) * 2023-04-26 2023-08-04 青岛尘元科技信息有限公司 Image content analysis method and device for massive images and image information network

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2011021605A1 (en) * 2009-08-18 2011-02-24 公立大学法人大阪府立大学 Object detection method
CN107871143A (en) * 2017-11-15 2018-04-03 深圳云天励飞技术有限公司 Image-recognizing method and device, computer installation and computer-readable recording medium
CN107895021A (en) * 2017-11-15 2018-04-10 深圳云天励飞技术有限公司 Image-recognizing method and device, computer installation and computer-readable recording medium
CN108985190A (en) * 2018-06-28 2018-12-11 北京市商汤科技开发有限公司 Target identification method and device, electronic equipment, storage medium, program product

Family Cites Families (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
IL236598A0 (en) * 2015-01-05 2015-05-31 Superfish Ltd Image similarity as a function of weighted descriptor similarities derived from neural networks
CN110689046A (en) * 2019-08-26 2020-01-14 深圳壹账通智能科技有限公司 Image recognition method, image recognition device, computer device, and storage medium

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2011021605A1 (en) * 2009-08-18 2011-02-24 公立大学法人大阪府立大学 Object detection method
CN107871143A (en) * 2017-11-15 2018-04-03 深圳云天励飞技术有限公司 Image-recognizing method and device, computer installation and computer-readable recording medium
CN107895021A (en) * 2017-11-15 2018-04-10 深圳云天励飞技术有限公司 Image-recognizing method and device, computer installation and computer-readable recording medium
CN108985190A (en) * 2018-06-28 2018-12-11 北京市商汤科技开发有限公司 Target identification method and device, electronic equipment, storage medium, program product

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
庄小芳;滕忠坚;洪亲;: "基于图像混合图结构的区域匹配和图像约化算法", 计算机与现代化, no. 02 *

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2021036309A1 (en) * 2019-08-26 2021-03-04 深圳壹账通智能科技有限公司 Image recognition method and apparatus, computer apparatus, and storage medium
CN111950493A (en) * 2020-08-20 2020-11-17 华北电力大学 Image recognition method and device, terminal equipment and readable storage medium
CN111950493B (en) * 2020-08-20 2024-03-08 华北电力大学 Image recognition method, device, terminal equipment and readable storage medium
CN112307934A (en) * 2020-10-27 2021-02-02 深圳市商汤科技有限公司 Image detection method, and training method, device, equipment and medium of related model
WO2022088411A1 (en) * 2020-10-27 2022-05-05 深圳市商汤科技有限公司 Image detection method and apparatus, related model training method and apparatus, and device, medium and program
WO2024061311A1 (en) * 2022-09-21 2024-03-28 北京沃东天骏信息技术有限公司 Model training method and apparatus, and image classification method and apparatus

Also Published As

Publication number Publication date
WO2021036309A1 (en) 2021-03-04

Similar Documents

Publication Publication Date Title
CN110662484B (en) System and method for whole body measurement extraction
CN110689046A (en) Image recognition method, image recognition device, computer device, and storage medium
CN110363091B (en) Face recognition method, device and equipment under side face condition and storage medium
WO2021082635A1 (en) Region of interest detection method and apparatus, readable storage medium and terminal device
WO2019119505A1 (en) Face recognition method and device, computer device and storage medium
EP4394726A1 (en) Image classification method and apparatus, computer device, and storage medium
CN111814620A (en) Face image quality evaluation model establishing method, optimization method, medium and device
CN113222942A (en) Training method of multi-label classification model and method for predicting labels
CN112598643A (en) Depth counterfeit image detection and model training method, device, equipment and medium
EP4113376A1 (en) Image classification model training method and apparatus, computer device, and storage medium
CN113490947A (en) Detection model training method and device, detection model using method and storage medium
CN116580257A (en) Feature fusion model training and sample retrieval method and device and computer equipment
CN111291887A (en) Neural network training method, image recognition method, device and electronic equipment
CN110765882A (en) Video tag determination method, device, server and storage medium
CN112036284B (en) Image processing method, device, equipment and storage medium
CN110489659A (en) Data matching method and device
CN111694954A (en) Image classification method and device and electronic equipment
CN113128526B (en) Image recognition method and device, electronic equipment and computer-readable storage medium
CN111340213A (en) Neural network training method, electronic device, and storage medium
CN114332993A (en) Face recognition method and device, electronic equipment and computer readable storage medium
CN112329666A (en) Face recognition method and device, electronic equipment and storage medium
CN111709473A (en) Object feature clustering method and device
CN112232890B (en) Data processing method, device, equipment and storage medium
CN113780148A (en) Traffic sign image recognition model training method and traffic sign image recognition method
CN111275183A (en) Visual task processing method and device and electronic system

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination