WO2020006961A1 - Procédé et dispositif d'extraction d'image - Google Patents

Procédé et dispositif d'extraction d'image Download PDF

Info

Publication number
WO2020006961A1
WO2020006961A1 PCT/CN2018/116334 CN2018116334W WO2020006961A1 WO 2020006961 A1 WO2020006961 A1 WO 2020006961A1 CN 2018116334 W CN2018116334 W CN 2018116334W WO 2020006961 A1 WO2020006961 A1 WO 2020006961A1
Authority
WO
WIPO (PCT)
Prior art keywords
image
matched
feature vector
position information
training
Prior art date
Application number
PCT/CN2018/116334
Other languages
English (en)
Chinese (zh)
Inventor
周恺卉
王长虎
Original Assignee
北京字节跳动网络技术有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 北京字节跳动网络技术有限公司 filed Critical 北京字节跳动网络技术有限公司
Publication of WO2020006961A1 publication Critical patent/WO2020006961A1/fr

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/22Matching criteria, e.g. proximity measures

Definitions

  • Embodiments of the present application relate to the field of computer technology, and in particular, to a method and an apparatus for extracting an image.
  • image recognition models to identify images is a common method of image recognition technology.
  • An image recognition model is usually a model that is trained using a large number of training samples.
  • a target image such as a watermark image, a person image, an object image, etc.
  • the sample images are trained to obtain an image recognition model.
  • the embodiments of the present application provide a method and a device for extracting an image.
  • an embodiment of the present application provides a method for extracting an image.
  • the method includes: obtaining a reference object image and a set of images to be matched; and inputting the reference object image into a first subnetwork included in a pre-trained image recognition model.
  • the to-be-matched images in the to-be-matched image set perform the following extraction steps: input the to-be-matched image into a second sub-network included in the image recognition model, and obtain at least one position information The feature vector to be matched corresponding to the position information, wherein the feature vector to be matched is the feature vector of the region image included in the image to be matched, and the position information is used to characterize the position of the region image in the to-be-matched image; The distance between the matching feature vector and the reference feature vector; in response to determining that there is a distance less than or equal to a preset distance threshold in the determined distance, the image to be matched is extracted as an image matching the reference object image.
  • the extracting step further includes: determining position information of an area image corresponding to a distance less than or equal to a distance threshold, and outputting the determined position information.
  • the extracting step further includes: generating a matched image including a position marker based on the output position information and the image to be matched, where the position marker is used to mark the image of the region to be matched corresponding to the output position information during matching. Position in the back image.
  • the second sub-network includes a dimension transformation layer for transforming the feature vector to a target dimension; and inputting the image to be matched into the second sub-network included in the image recognition model to obtain at least one feature vector to be matched,
  • the method includes: inputting the to-be-matched image into a second sub-network included in the image recognition model to obtain at least one to-be-matched feature vector having the same dimension as the reference feature vector.
  • the image recognition model is obtained by training in the following steps: obtaining a training sample set, where the training samples include sample object images, sample matching images, and labeled position information of the sample matching images.
  • the labeled position information indicates that the sample matching image includes Position of the regional image; select training samples from the training sample set, and perform the following training steps: input the sample object images included in the selected training samples into the first subnetwork included in the initial model, obtain the first feature vector, and input the sample matching images
  • the second sub-network included in the initial model obtains at least one position information and a second feature vector corresponding to the position information; and from the obtained at least one position information, the position information representing the target region image in the sample matching image is determined as a target Position information, and determine the second feature vector corresponding to the target position information as the target second feature vector; based on the first loss value representing the error of the target position information and the second feature vector representing the distance between the target second feature vector and the first feature vector Two loss values, Whether training is complete initial model; in response to determining
  • determining whether the initial model training is completed based on the first loss value representing the error of the target position information and the second loss value representing the difference between the distance between the target second feature vector and the first feature vector includes: The preset weight value uses the weighted summation result of the first loss value and the second loss value as the total loss value, and compares the total loss value with the target value, and determines whether the initial model is completed according to the comparison result.
  • the step of training to obtain an image recognition model further includes: in response to determining that the initial model is not trained, adjusting parameters of the initial model, and selecting training samples from unselected training samples in the training sample set, Use the adjusted initial model as the initial model and continue with the training steps.
  • an embodiment of the present application provides an apparatus for extracting an image.
  • the apparatus includes: an acquiring unit configured to acquire a reference object image and a set of images to be matched; and a generating unit configured to input the reference object image.
  • the first sub-network included in the pre-trained image recognition model obtains the feature vector of the reference object image as the reference feature vector;
  • the extraction unit is configured to perform the following extraction steps on the to-be-matched images in the to-be-matched image set:
  • the matching image is input to a second sub-network included in the image recognition model to obtain at least one position information and a feature vector corresponding to the position information, where the feature vector to be matched is a feature vector of an area image included in the image to be matched.
  • the matched image is an image that matches the reference target image.
  • the extraction unit includes: an output module configured to determine position information of a region image corresponding to a distance less than or equal to a distance threshold, and output the determined position information.
  • the extraction unit further includes a generating module configured to generate a matched image including a position marker based on the output position information and the image to be matched, where the position marker is used to mark the output position information corresponding to The position of the image to be matched in the matched image.
  • the second sub-network includes a dimensional transformation layer for transforming the feature vector to a target dimension; and the extraction unit is further configured to: input the image to be matched into the second sub-network included in the image recognition model to obtain At least one feature vector to be matched having the same dimension as the reference feature vector.
  • the image recognition model is obtained by training in the following steps: obtaining a training sample set, where the training samples include sample object images, sample matching images, and labeled position information of the sample matching images.
  • the labeled position information indicates that the sample matching image includes Position of the regional image; select training samples from the training sample set, and perform the following training steps: input the sample object images included in the selected training samples into the first subnetwork included in the initial model, obtain the first feature vector, and input the sample matching images
  • the second sub-network included in the initial model obtains at least one position information and a second feature vector corresponding to the position information; and from the obtained at least one position information, the position information representing the target region image in the sample matching image is determined as a target Position information, and determine the second feature vector corresponding to the target position information as the target second feature vector; based on the first loss value representing the error of the target position information and the second feature vector representing the distance between the target second feature vector and the first feature vector Two loss values, It is determined whether the initial model is trained; in response to
  • determining whether the initial model training is completed based on the first loss value representing the error of the target position information and the second loss value representing the difference between the distance between the target second feature vector and the first feature vector includes: The preset weight value uses the weighted summation result of the first loss value and the second loss value as the total loss value, and compares the total loss value with the target value, and determines whether the initial model is completed according to the comparison result.
  • the step of training to obtain an image recognition model further includes: in response to determining that the initial model is not trained, adjusting parameters of the initial model, and selecting training samples from unselected training samples in the training sample set, Use the adjusted initial model as the initial model and continue with the training steps.
  • an embodiment of the present application provides an electronic device.
  • the electronic device includes: one or more processors; a storage device on which one or more programs are stored; when one or more programs are read by one or more Each processor executes such that one or more processors implement the method as described in any implementation of the first aspect.
  • an embodiment of the present application provides a computer-readable medium having stored thereon a computer program that, when executed by a processor, implements the method as described in any implementation manner of the first aspect.
  • the method and device for extracting images obtained by the embodiments of the present application obtain a reference feature vector of a reference image and at least one feature vector to be matched of an image to be matched by using a pre-trained image recognition model, and then compare the reference feature vector and The distance of the feature vector to be matched is used to obtain an image matching the reference image, so that when the training sample required for training the image recognition model does not include the reference image, the image recognition model is used to extract the image that matches the reference image, improving It increases the flexibility of image recognition and enriches the means of image recognition.
  • FIG. 1 is an exemplary system architecture diagram to which an embodiment of the present application can be applied;
  • FIG. 1 is an exemplary system architecture diagram to which an embodiment of the present application can be applied;
  • FIG. 2 is a flowchart of an embodiment of a method for extracting an image according to the present application
  • FIG. 3 is a flowchart of an image recognition model obtained by training of a method for extracting an image according to the present application
  • FIG. 4 is a schematic diagram of an application scenario of a method for extracting an image according to the present application
  • FIG. 5 is a flowchart of still another embodiment of a method for extracting an image according to the present application.
  • FIG. 6 is a schematic structural diagram of an embodiment of an apparatus for extracting an image according to the present application.
  • FIG. 7 is a schematic structural diagram of a computer system suitable for implementing an electronic device according to an embodiment of the present application.
  • FIG. 1 illustrates an exemplary system architecture 100 to which a method for extracting an image or an apparatus for extracting an image of an embodiment of the present application can be applied.
  • the system architecture 100 may include terminal devices 101, 102, and 103, a network 104, and a server 105.
  • the network 104 is a medium for providing a communication link between the terminal devices 101, 102, 103 and the server 105.
  • the network 104 may include various connection types, such as wired, wireless communication links, or fiber optic cables, among others.
  • the user can use the terminal devices 101, 102, 103 to interact with the server 105 through the network 104 to receive or send messages and the like.
  • Various communication client applications can be installed on the terminal devices 101, 102, and 103, such as image processing applications, shooting applications, social platform software, and the like.
  • the terminal devices 101, 102, and 103 may be hardware or software.
  • the terminal devices 101, 102, and 103 can be various electronic devices with a display screen, including but not limited to smart phones, tablet computers, laptop computers, and desktop computers.
  • the terminal devices 101, 102, and 103 are software, they can be installed in the electronic devices listed above. It can be implemented as multiple software or software modules (such as software or software modules used to provide distributed services), or it can be implemented as a single software or software module. It is not specifically limited here.
  • the server 105 may be a server that provides various services, such as a background server that provides support for various applications on the terminal devices 101, 102, and 103.
  • the background server may perform processing such as analysis on the acquired image, and output the processing result (for example, the extracted image matching the reference image).
  • the method for extracting an image provided in the embodiment of the present application may be executed by the server 105, or may be executed by the terminal devices 101, 102, and 103.
  • the device for extracting an image may be provided in the server 105 or in the terminal devices 101, 102, 103.
  • the server may be hardware or software.
  • the server can be implemented as a distributed server cluster consisting of multiple servers or as a single server.
  • the server can be implemented as multiple software or software modules (such as software or software modules used to provide distributed services), or it can be implemented as a single software or software module. It is not specifically limited here.
  • terminal devices, networks, and servers in FIG. 1 are merely exemplary. Depending on the implementation needs, there can be any number of terminal devices, networks, and servers.
  • a flowchart 200 of one embodiment of a method for extracting an image according to the present application is shown.
  • the method for extracting an image includes the following steps:
  • Step 201 Obtain a reference object image and a set of images to be matched.
  • an execution subject of the method for extracting an image may obtain a reference object image and a set of images to be matched remotely or locally through a wired connection or a wireless connection.
  • the reference object image may be an image to be compared with other images, and the reference object image may be an image characterizing an object.
  • Objects can be various things, such as watermarks, signs, faces, objects, and so on.
  • the set of images to be matched may be a set of certain types of images (for example, images containing a trademark) stored in advance.
  • Step 202 Input a reference object image into a first sub-network included in a pre-trained image recognition model, and obtain a feature vector of the reference object image as a reference feature vector.
  • the execution subject may input the reference object image into a first sub-network included in a pre-trained image recognition model, and obtain a feature vector of the reference object image as the reference feature vector.
  • the first sub-network is used to characterize the correspondence between the image and the feature vector of the image.
  • the image recognition model may be various neural network models created based on machine learning technology.
  • the neural network model may have a structure of various neural networks (for example, DenseBox, VGGNet, ResNet, SegNet, etc.).
  • the above reference feature vector may be a feature (e.g., shape, color, texture) extracted from a first sub-network included in a neural network model (e.g., a network composed of one or some convolutional layers included in the neural network model), which characterizes an image And other characteristics).
  • a feature e.g., shape, color, texture
  • a neural network model e.g., a network composed of one or some convolutional layers included in the neural network model
  • Step 203 For the to-be-matched images in the to-be-matched image set, perform the following extraction steps: input the to-be-matched images into a second sub-network included in the image recognition model to obtain at least one position information and a feature vector to be matched corresponding to the location information; Determine the distance between the obtained feature vector to be matched and the reference feature vector; and in response to determining that there is a distance less than or equal to a preset distance threshold in the determined distance, extract the to-be-matched image as an image matching the reference object image.
  • the execution subject may perform the following extraction step on the to-be-matched image:
  • Step 2031 Input the image to be matched into the second sub-network included in the image recognition model, and obtain at least one location information and a feature vector to be matched corresponding to the location information.
  • the second sub-network is used to characterize the correspondence between the image and the position information of the image and the feature vector to be matched of the image.
  • the position information is used to characterize the position of the area image corresponding to the feature vector to be matched in the to-be-matched image.
  • the feature vector to be matched is a feature vector of an area image included in the image to be matched.
  • the second sub-network may determine each position information from the images to be matched according to the determined at least one position information. Characterize the region images, and determine the feature vector of each region image.
  • the area image may be an image characterizing an object (eg, a watermark, a logo, etc.).
  • the location information may include coordinate information and identification information. Among them, the coordinate information (such as the coordinates of the corner points of the regional image, the size of the regional image, etc.) is used to indicate the position of the regional image in the image to be matched, and the identification information (such as the serial number of the regional image and the category of the regional image) is used to identify the area image.
  • an image to be matched includes two watermarked images
  • the position information determined by the second subnet is "(1, x1, y1, w1, h1)" and (2, x2, y2, w2, h2)
  • 1, 2 are the serial numbers of the two watermarked images
  • (x1, y1), (x2, y2) are the coordinates of the upper left corner of the two watermarked images
  • w1 and w2 are the widths of the two watermarked images.
  • H1, h2 are the heights of the two watermark images, respectively.
  • the execution subject can extract feature vectors of the to-be-matched images, and extract feature vectors corresponding to the two position information as feature-to-be-matched feature vectors from the feature vectors of the to-be-matched images, respectively.
  • the second sub-network may be a neural network based on an existing target detection network (for example, SSD (Single Shot MultiBox Detector), R-CNN (Region-based Convolutional Neural Networks, Faster R-CNN, etc.).
  • SSD Single Shot MultiBox Detector
  • R-CNN Registered-based Convolutional Neural Networks
  • Faster R-CNN etc.
  • the foregoing second sub-network includes a dimension transformation layer for transforming a feature vector to a target dimension.
  • the dimension transformation layer may process the feature vector (for example, the values of some dimensions included in the feature vector are combined by taking an average value); it may also be a pooling layer included in the second sub-network.
  • the pooling layer can be used to down-sample or up-sample the input data to compress or increase the amount of data.
  • the above target dimensions may be various dimensions set by a technician, for example, the same dimensions as those of the reference feature vector.
  • the execution subject may input the image to be matched into a second sub-network included in the image recognition model, and the second sub-network extracts at least one feature vector of the image to be matched, and then extracts the dimensional transformation layer pair included in the second sub-network.
  • Each feature vector of the dimensional transform is performed to obtain at least one feature vector to be matched that has the same dimension as the reference feature vector.
  • a ROI Pooling (Region Of Interest, Pooling) layer can be used, so that each feature vector to be matched has the same dimension as the reference feature vector.
  • the ROI Pooling layer is a well-known technology that is widely studied and applied at present, and will not be repeated here.
  • Step 2032 Determine the distance between the obtained feature vector to be matched and the reference feature vector.
  • the execution body may determine a distance between each of the obtained at least one to-be-matched feature vector and a reference feature vector.
  • the distance may be any of the following: Euclidean distance, Mahalanobis distance, and the like.
  • the preset distance may be any value greater than or equal to 0.
  • the distance can represent the degree of similarity between two feature vectors, that is, the degree of similarity between two images. As an example, if the distance between the two feature vectors is larger, the images that are corresponding to the two feature vectors are less similar.
  • Step 2033 In response to determining that a distance smaller than or equal to a preset distance threshold exists in the determined distance, extract the image to be matched as an image matching the reference object image.
  • the distance threshold may be a value set by a technician according to experience, or may be a value calculated (for example, by calculating an average value) calculated by the execution subject based on historical data (for example, a recorded historical distance threshold). Specifically, if there is a distance smaller than the distance threshold in each of the determined distances, it indicates that an area image similar to the reference object image exists in the image to be matched, that is, the image to be matched matches the reference object image.
  • the image recognition model can be obtained by training in advance through the following steps:
  • Step 301 Obtain a training sample set.
  • the training samples include sample object images, sample matching images, and sample matching image labeling position information.
  • the labeling position information represents the position of the region image included in the sample matching image.
  • the sample object image may be an image characterizing an object (for example, a watermark, a logo, a human face, an object, etc.).
  • the number of labeled position information may be at least one, and each labeled information may correspond to a region image, and each region image includes a region image in which the characterized object and the sample object image are the same.
  • Step 302 Select training samples from the training sample set.
  • the selection method and the number of training samples are not limited in this application.
  • the training samples may be selected from the training sample set in a random selection or in the order of the number of training samples.
  • Step 303 input the sample object image included in the selected training sample into the first sub-network included in the initial model to obtain a first feature vector, and input the sample matching image into the second sub-network included in the initial model to obtain at least one position information and the The second feature vector corresponding to the position information.
  • the initial model may be various existing neural network models created based on machine learning technology.
  • the neural network model may have various existing neural network structures (for example, DenseBox, VGGNet, ResNet, SegNet, etc.).
  • Each of the above feature vectors may be a vector composed of data extracted from certain layers (such as a convolution layer) included in the neural network model.
  • the above-mentioned first sub-network and second sub-network are the same as the first sub-network and the second sub-network described in step 202 and step 203, respectively, and details are not described herein again.
  • Step 304 From the obtained at least one position information, determine position information representing the target region image in the sample matching image as the target position information, and determine a second feature vector corresponding to the target position information as the target second feature vector.
  • the above-mentioned target area image may be an area image in which the characterized object is the same as the object characterized by the sample object image.
  • the execution subject performing this step may use the position information specified by the technician as the target position information, the area image characterized by the target position information as the target area image, and the second feature vector of the target area image as the target second feature vector; or The execution subject executing this step may determine the similarity between the area image corresponding to each position information and the sample object image and determine the area image with the greatest similarity to the sample object image as the target area image according to the obtained position information. , Determining the position information of the target area image as the target position information, and determining the second feature vector of the target area image as the target second feature vector.
  • Step 305 Determine whether the initial model training is completed based on the first loss value representing the error of the target position information and the second loss value representing the difference between the distance between the target second feature vector and the first feature vector.
  • the first loss value may represent a gap between the target position information and the labeled position information corresponding to the target area image.
  • the smaller the first loss value the smaller the difference between the target position information and the labeled position information corresponding to the target area image, that is, the closer the target position information is to the labeled position information.
  • the first loss value can be obtained according to any of the following loss functions: Softmax loss function, Smooth L1 (smooth L1 norm) loss function, and the like.
  • the second loss value may represent a distance between the target second feature vector and the first feature vector.
  • the larger the second loss value the greater the distance between the target second feature vector and the first feature vector, that is, the less similar the target area image and the sample object image.
  • the second loss value may be a distance between the target second feature vector and the first feature vector (eg, Euclidean distance, Mahalanobis distance, etc.).
  • the second loss value can be obtained by the Triplet loss function, where the Triplet error function is as follows:
  • L is the second loss value
  • is the sum sign
  • i is the serial number of each training sample selected this time
  • a represents the sample object image
  • p represents the positive sample image (that is, the target area image).
  • n represents a negative sample image (that is, an area image other than the target area image in the sample matching image; or a preset, characterized image where the object is different from the object characterized by the sample object image).
  • a feature vector representing a sample object image included in the training sample with the sequence number i A feature vector representing a positive sample image (such as an image of a target region) corresponding to the training sample with the number i, Characterize the feature vector of a negative sample image (for example, a region image other than the target region image in the sample matching image) corresponding to the training sample with the number i.
  • threshold represents a preset distance
  • Characterizing the first distance that is, the distance between the first feature vector and the feature vector of the positive sample image
  • Represent the first distance that is, the distance between the first feature vector and the feature vector of the negative sample image).
  • the "+" on the lower right side of the square brackets in the above formula means taking a positive value, that is, when the expression of the expression in the square brackets is positive, the positive value is taken, and when it is negative, 0 is taken.
  • the parameters of the initial model can be adjusted according to the back-propagation algorithm so that the L value is minimum or the L value converges, indicating that the training is complete.
  • the execution subject performing this step may obtain the total loss value based on the first loss value and the second loss value, compare the total loss value with the target value, and determine whether the initial model is completed according to the comparison result.
  • the target value may be a preset loss value threshold. When the difference between the total loss value and the target value is less than or equal to the loss value threshold, it is determined that the training is completed.
  • an executing subject executing this step may use a weighted summation result of the first loss value and the second loss value as a total loss value according to a preset weight value, and use the total loss value.
  • the loss value is compared with the target value, and whether the initial model is trained is determined based on the comparison result.
  • the above weight value can adjust the ratio of the first loss value and the second loss value to the total loss value, so that the image recognition model can achieve different functions in different application scenarios (for example, some scenes focus on extracting position information, and some scenes focus on Compare distances of feature vectors).
  • Step 306 In response to determining that the training is completed, determine the initial model as an image recognition model.
  • the execution subject of the image recognition model that is trained may respond to determining that the initial model is not trained, adjusting the parameters of the initial model, and unselected training samples from the training sample set.
  • a training sample is selected, and an initial model adjusted by parameters is used as an initial model, and the training step is continued.
  • the initial model is a convolutional neural network
  • the backpropagation algorithm can be used to adjust the weights in each convolutional layer in the initial model.
  • a training sample may be selected from the unselected training samples in the training sample set, and the initial model adjusted by parameters is used as the initial model, and steps 303 to 306 are continuously performed.
  • the execution subject of the image recognition model obtained through the training may be the same as or different from the execution subject of the method for extracting an image. If they are the same, the training subject who has obtained the image recognition model may store the structure information and parameter values of the parameters of the trained image recognition model locally after obtaining the image recognition model. If they are different, the execution subject trained in the image recognition model may send the structure information and parameter values of the trained image recognition model to the execution subject of the method for extracting the image after the image recognition model is trained.
  • FIG. 4 is a schematic diagram of an application scenario of a method for extracting an image according to this embodiment.
  • the server 401 first obtains a watermark image 402 (ie, a reference object image) uploaded by the terminal device 408, and obtains a set of images 403 to be matched locally.
  • the server 401 inputs the watermark image 402 into the first sub-network 4041 included in the pre-trained image recognition model 404, and obtains a feature vector of the watermark image 402 as a reference feature vector 405.
  • the server 401 selects one to-be-matched image 4031 from the to-be-matched image set 403, inputs the to-be-matched image 4031 into the second sub-network 4041 included in the image recognition model 404, and obtains the position information 4061, 4062, and 4063, and the corresponding position information.
  • the feature vectors to be matched 4071, 4072, and 4073 are the feature vectors of the watermark images 40311, 40312, and 40313 included in the image 4031 to be matched, respectively.
  • the server 401 determines that the distance between the feature vector to be matched 4071 and the reference feature vector 405 is less than or equal to a preset distance threshold, extracts the to-be-matched image 4031 as an image matching the reference object image, and sends the matched image to the terminal device 408.
  • the server 401 repeatedly selects the image to be matched from the set of images to be matched 403 and the watermarked image 402 to match, thereby extracting multiple images from the set of images to be matched 403 that match the watermarked image 402.
  • a pre-trained image recognition model is used to obtain a reference feature vector of a reference image and at least one feature vector to be matched of the image to be matched, and then compare the reference feature vector and the feature vector to be matched.
  • Distance to obtain an image matching the reference image thereby improving the pertinence of matching with the reference image, and realizing that when the training samples required for training the image recognition model do not include the reference image, the image recognition model is used to extract the The reference image matches the image, which improves the flexibility of image recognition and enriches the means of image recognition.
  • a flowchart 500 of still another embodiment of a method for extracting an image is shown.
  • the process 500 of the method for extracting an image includes the following steps:
  • Step 501 Obtain a reference object image and a set of images to be matched.
  • step 501 is substantially the same as step 501 in the embodiment corresponding to FIG. 2, and details are not described herein again.
  • Step 502 Input a reference object image into a first sub-network included in a pre-trained image recognition model, and obtain a feature vector of the reference object image as a reference feature vector.
  • step 502 is substantially the same as step 502 in the embodiment corresponding to FIG. 2, and details are not described herein again.
  • Step 503 for the to-be-matched images in the to-be-matched image set, perform the following extraction step: input the to-be-matched images into a second sub-network included in the image recognition model to obtain at least one position information and a feature vector to be matched corresponding to the position information; Determining the distance between the obtained feature vector to be matched and the reference feature vector; in response to determining that there is a distance less than or equal to a preset distance threshold in the determined distance, extracting the image to be matched as an image matching the reference object image; Position information of a region image corresponding to a distance equal to a distance threshold, and output the determined position information.
  • the execution subject may perform the following extraction step on the to-be-matched image:
  • Step 5031 Input the image to be matched into the second sub-network included in the image recognition model to obtain at least one position information and a feature vector to be matched corresponding to the position information.
  • Step 5031 is basically the same as step 2031 in the embodiment corresponding to FIG. 2, and details are not described herein again.
  • Step 5032 Determine the distance between the obtained feature vector to be matched and the reference feature vector. Step 5032 is basically the same as step 2032 in the embodiment corresponding to FIG. 2, and details are not described herein again.
  • Step 5033 In response to determining that a distance smaller than or equal to a preset distance threshold exists in the determined distance, extract the image to be matched as an image matching the reference object image. Step 5033 is basically the same as step 2033 in the embodiment corresponding to FIG. 2, and details are not described herein again.
  • Step 5034 Determine the position information of the area image corresponding to the distance less than or equal to the distance threshold, and output the determined position information.
  • an execution subject of the method for extracting an image may be obtained from step 5031 based on the distance determined in step 5032 and equal to or less than a preset distance threshold.
  • position information corresponding to a distance less than or equal to the distance threshold is determined, and position information corresponding to a distance less than or equal to the distance threshold is output.
  • the execution subject may output position information in various ways.
  • the display body connected to the execution subject may display information such as identification information, coordinate information, and the like of a region image included in the position information.
  • the above-mentioned execution subject may generate a matched image including a position mark based on the output location information and the image to be matched after outputting the location information.
  • the position marker is used to mark the position of the region to be matched image corresponding to the output position information in the matched image.
  • the execution subject may draw a frame of a preset shape in the image to be matched according to the output position information, use the drawn frame as a position mark, and use the to-be-matched image including the position mark as a matched image.
  • the process 500 of the method for extracting an image in this embodiment highlights the steps of outputting position information. Therefore, the solution described in this embodiment can further determine the position of the target region image included in the image to be matched, and improve the specificity of image recognition.
  • this application provides an embodiment of an apparatus for extracting an image.
  • the apparatus embodiment corresponds to the method embodiment shown in FIG. 2.
  • the device can be specifically applied to various electronic devices.
  • the apparatus 600 for extracting an image in this embodiment includes: an acquiring unit 601 configured to acquire a reference object image and a set of images to be matched; and a generating unit 602 configured to input a reference object image into a pre-training A first sub-network included in the image recognition model of FIG.
  • an extraction unit 603 is configured to perform the following extraction step on the to-be-matched images in the to-be-matched image set:
  • the image is input to a second sub-network included in the image recognition model to obtain at least one position information and a feature vector to be matched corresponding to the position information, where the feature vector to be matched is a feature vector of an area image included in the image to be matched, and the position information is used for Characterizing the position of the area image in the image to be matched; determining the distance between the obtained feature vector to be matched and the reference feature vector; and in response to determining that a distance less than or equal to a preset distance threshold exists in the determined distance, extracting the to be matched
  • the image is an image matching the reference target image.
  • the obtaining unit 601 may obtain the reference object image and the set of images to be matched from a remote or local source through a wired connection method or a wireless connection method.
  • the reference object image may be an image to be compared with other images, and the reference object image is an image representing an object.
  • Objects can be various things, such as watermarks, signs, faces, objects, and so on.
  • the set of images to be matched may be a set of certain types of images (for example, images containing a trademark) stored in advance.
  • the generating unit 602 may input the reference object image into a first sub-network included in a pre-trained image recognition model, and obtain a feature vector of the reference object image as a reference feature vector.
  • the first sub-network is used to characterize the correspondence between the image and the feature vector of the image.
  • the image recognition model may be various neural network models created based on machine learning technology.
  • the neural network model may have a structure of various neural networks (for example, DenseBox, VGGNet, ResNet, SegNet, etc.).
  • the above reference feature vector may be a feature (e.g., shape, color, texture) extracted from a first sub-network included in a neural network model (e.g., a network composed of one or some convolutional layers included in the neural network model), which characterizes an image And other characteristics).
  • a feature e.g., shape, color, texture
  • a neural network model e.g., a network composed of one or some convolutional layers included in the neural network model
  • the extraction unit 603 may perform the following steps on the image to be matched:
  • the image to be matched is input to a second sub-network included in the image recognition model, and at least one position information and a feature vector corresponding to the position information are obtained.
  • the second sub-network is used to characterize the correspondence between the image and the position information of the image and the feature vector to be matched of the image.
  • the position information is used to characterize the position of the area image corresponding to the feature vector to be matched in the to-be-matched image.
  • the feature vector to be matched is a feature vector of an area image included in the image to be matched.
  • a distance between the obtained feature vector to be matched and the reference feature vector is determined.
  • the above-mentioned extraction unit 603 may determine a distance between each of the obtained at least one feature vector to be matched and the reference feature vector.
  • the distance may be any of the following: Euclidean distance, Mahalanobis distance, and the like.
  • the image to be matched is extracted as an image matching the reference object image.
  • the distance threshold may be a value set by a technician based on experience, or may be a value calculated (for example, by calculating an average value) calculated by the extraction unit 603 according to historical data (for example, a recorded historical distance threshold).
  • the extraction unit 603 may include an output module configured to determine position information of a region image corresponding to a distance less than or equal to a distance threshold, and output the determined position information.
  • the extraction unit 603 may further include: a generating module configured to generate a matched image including a position marker based on the output position information and the image to be matched, where the position marker It is used to mark the position of the region to be matched image corresponding to the output position information in the matched image.
  • a generating module configured to generate a matched image including a position marker based on the output position information and the image to be matched, where the position marker It is used to mark the position of the region to be matched image corresponding to the output position information in the matched image.
  • the second sub-network may include a dimension transformation layer for transforming a feature vector to a target dimension; and the extraction unit 603 may be further configured to: input the image to be matched into an image Identify the second sub-network included in the model, and obtain at least one feature vector to be matched that has the same dimension as the reference feature vector.
  • the image recognition model is obtained by training in the following steps: obtaining a training sample set, where the training sample includes labeled object information of the sample object image, the sample matched image, and the sample matched image, and the labeled position
  • the information characterizes the position of the region image included in the sample matching image; selects training samples from the training sample set, and executes the following training steps: inputting the sample object image included in the selected training sample into the first sub-network included in the initial model to obtain the first feature Vector, and input the sample matching image into the second sub-network included in the initial model to obtain at least one location information and a second feature vector corresponding to the location information; and determine the target characterizing the sample matching image from the obtained at least one location information
  • the position information of the area image is used as the target position information
  • the second feature vector corresponding to the target position information is determined as the target second feature vector; based on the first loss value representing the error of the target position information and the target feature second feature vector and the first feature Distance of vector Loss
  • the execution subject of the image recognition model trained to obtain the weighted summation result of the first loss value and the second loss value as the total loss value according to a preset weight value and The total loss value is compared with the target value, and whether the initial model is trained is determined based on the comparison result.
  • the step of training to obtain an image recognition model may further include: in response to determining that the initial model is not trained, adjusting parameters of the initial model, and selecting unselected data from the training sample set. Among the training samples, a training sample is selected, and the initial model adjusted by parameters is used as the initial model, and the training step is continued.
  • the apparatus obtains a reference feature vector of a reference image and at least one feature vector to be matched of an image to be matched by using a pre-trained image recognition model, and then compares the reference feature vector and the feature vector to be matched by Distance to obtain an image matching the reference image, thereby improving the pertinence of matching with the reference image, and realizing that when the training samples required for training the image recognition model do not include the reference image, the image recognition model is used to extract the
  • the reference image matches the image, which improves the flexibility of image recognition and enriches the means of image recognition.
  • FIG. 7 illustrates a schematic structural diagram of a computer system 700 suitable for implementing an electronic device (such as a server or a terminal device shown in FIG. 1) in the embodiment of the present application.
  • an electronic device such as a server or a terminal device shown in FIG. 1
  • the electronic device shown in FIG. 7 is only an example, and should not impose any limitation on the functions and scope of use of the embodiments of the present application.
  • the computer system 700 includes a central processing unit (CPU) 701, which can be based on a program stored in a read-only memory (ROM) 702 or a program loaded from a storage section 708 into a random access memory (RAM) 703. Instead, perform various appropriate actions and processes.
  • ROM read-only memory
  • RAM random access memory
  • various programs and data required for the operation of the system 700 are also stored.
  • the CPU 701, ROM 702, and RAM 703 are connected to each other through a bus 704.
  • An input / output (I / O) interface 705 is also connected to the bus 704.
  • the following components are connected to the I / O interface 705: an input section 706 including a keyboard, a mouse, etc .; an output section 707 including a cathode ray tube (CRT), a liquid crystal display (LCD), etc .; and a speaker; And a communication section 709 including a network interface card such as a LAN card, a modem, and the like.
  • the communication section 709 performs communication processing via a network such as the Internet.
  • the driver 710 is also connected to the I / O interface 705 as needed.
  • a removable medium 711 such as a magnetic disk, an optical disk, a magneto-optical disk, a semiconductor memory, etc., is installed on the drive 710 as needed, so that a computer program read out therefrom is installed into the storage section 708 as needed.
  • the process described above with reference to the flowchart may be implemented as a computer software program.
  • embodiments of the present disclosure include a computer program product including a computer program carried on a computer-readable medium, the computer program containing program code for performing a method shown in a flowchart.
  • the computer program may be downloaded and installed from a network through the communication section 709, and / or installed from a removable medium 711.
  • CPU central processing unit
  • the computer-readable medium described in this application may be a computer-readable signal medium or a computer-readable medium or any combination of the foregoing.
  • the computer-readable medium may be, for example, but not limited to, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, or device, or any combination thereof. More specific examples of computer-readable media may include, but are not limited to: electrical connections with one or more wires, portable computer disks, hard disks, random access memory (RAM), read-only memory (ROM), erasable programmable Read-only memory (EPROM or flash memory), optical fiber, portable compact disk read-only memory (CD-ROM), optical storage device, magnetic storage device, or any suitable combination of the foregoing.
  • a computer-readable medium may be any tangible medium that contains or stores a program that can be used by or in combination with an instruction execution system, apparatus, or device.
  • a computer-readable signal medium may include a data signal that is included in baseband or propagated as part of a carrier wave, and which carries computer-readable program code. Such a propagated data signal may take many forms, including but not limited to electromagnetic signals, optical signals, or any suitable combination of the foregoing.
  • the computer-readable signal medium can also be any computer-readable medium other than a computer-readable medium, which can send, propagate, or transmit a program for use by or in connection with an instruction execution system, apparatus, or device.
  • Program code embodied on a computer-readable medium may be transmitted using any appropriate medium, including but not limited to: wireless, wire, optical fiber cable, RF, etc., or any suitable combination of the foregoing.
  • Computer program code for performing the operations of this application may be written in one or more programming languages, or a combination thereof, including programming languages such as Java, Smalltalk, C ++, and also conventional Procedural programming language—such as "C" or a similar programming language.
  • the program code can be executed entirely on the user's computer, partly on the user's computer, as an independent software package, partly on the user's computer, partly on a remote computer, or entirely on a remote computer or server.
  • the remote computer can be connected to the user's computer through any kind of network, including a local area network (LAN) or a wide area network (WAN), or it can be connected to an external computer (such as through an Internet service provider) Internet connection).
  • LAN local area network
  • WAN wide area network
  • Internet service provider Internet service provider
  • each block in the flowchart or block diagram may represent a module, a program segment, or a part of code, which contains one or more functions to implement a specified logical function Executable instructions.
  • the functions labeled in the blocks may also occur in a different order than those labeled in the drawings. For example, two blocks represented one after the other may actually be executed substantially in parallel, and they may sometimes be executed in the reverse order, depending on the functions involved.
  • each block in the block diagrams and / or flowcharts, and combinations of blocks in the block diagrams and / or flowcharts can be implemented by a dedicated hardware-based system that performs the specified function or operation , Or it can be implemented with a combination of dedicated hardware and computer instructions.
  • the units described in the embodiments of the present application may be implemented by software or hardware.
  • the described unit may also be provided in a processor, for example, it may be described as: a processor includes an acquisition unit, a generation unit, and an extraction unit. Wherein, the names of these units do not constitute a limitation on the unit itself in some cases, for example, the obtaining unit may also be described as a “unit for obtaining a reference target image and an image set to be matched”.
  • the present application also provides a computer-readable medium, which may be included in the electronic device described in the above embodiments; or may exist separately without being assembled into the electronic device in.
  • the computer-readable medium carries one or more programs, and when the one or more programs are executed by the electronic device, the electronic device: obtains a reference object image and a set of images to be matched; and inputs the reference object image into a pre-trained
  • the first sub-network included in the image recognition model obtains the feature vector of the reference object image as the reference feature vector; for the to-be-matched images in the to-be-matched image set, the following extraction step is performed: input the to-be-matched image into the first Two sub-networks to obtain at least one location information and a feature vector corresponding to the location information, wherein the feature vector to be matched is a feature vector of an area image included in the image to be matched, and the location information is used to characterize the area image in the to-be-matched image.

Landscapes

  • Engineering & Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • Theoretical Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Artificial Intelligence (AREA)
  • Evolutionary Biology (AREA)
  • Evolutionary Computation (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Image Analysis (AREA)

Abstract

La présente invention concerne, selon des modes de réalisation, un procédé et un dispositif d'extraction d'image. Un mode de réalisation selon le procédé consiste : à acquérir une image d'un objet de référence et un ensemble d'images à mettre en correspondance ; et à entrer l'image d'objet de référence dans un premier sous-réseau compris dans un modèle de reconnaissance d'image pré-appris, et à acquérir un vecteur de caractéristique de l'image d'objet de référence en tant que vecteur de caractéristique de référence, les étapes d'extraction suivantes étant exécutées sur des images à mettre en correspondance dans l'ensemble d'images à mettre en correspondance : l'entrée de l'image à mettre en correspondance dans un deuxième sous-réseau compris dans le modèle de reconnaissance d'image de manière acquérir au moins des informations de position et un vecteur de caractéristique à mettre en correspondance correspondant aux informations de position ; la détermination de la distance entre le vecteur de caractéristique acquis à mettre en correspondance et le vecteur de caractéristique de référence ; et l'extraction, en fonction de la détermination que les distances déterminées comprennent une distance inférieure ou égale à un seuil de distance prédéfini, de l'image à mettre en correspondance en tant que l'image correspondant à l'image d'objet de référence. Les modes de réalisation de l'invention permettent d'améliorer la flexibilité de la reconnaissance d'image et d'enrichir les moyens de la reconnaissance d'image.
PCT/CN2018/116334 2018-07-03 2018-11-20 Procédé et dispositif d'extraction d'image WO2020006961A1 (fr)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN201810715195.6A CN108898186B (zh) 2018-07-03 2018-07-03 用于提取图像的方法和装置
CN201810715195.6 2018-07-03

Publications (1)

Publication Number Publication Date
WO2020006961A1 true WO2020006961A1 (fr) 2020-01-09

Family

ID=64347534

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2018/116334 WO2020006961A1 (fr) 2018-07-03 2018-11-20 Procédé et dispositif d'extraction d'image

Country Status (2)

Country Link
CN (1) CN108898186B (fr)
WO (1) WO2020006961A1 (fr)

Cited By (14)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111507289A (zh) * 2020-04-22 2020-08-07 上海眼控科技股份有限公司 视频匹配方法、计算机设备和存储介质
CN111783872A (zh) * 2020-06-30 2020-10-16 北京百度网讯科技有限公司 训练模型的方法、装置、电子设备及计算机可读存储介质
CN111783869A (zh) * 2020-06-29 2020-10-16 杭州海康威视数字技术股份有限公司 训练数据筛选方法、装置、电子设备及存储介质
CN111984814A (zh) * 2020-08-10 2020-11-24 广联达科技股份有限公司 一种建筑图纸中的箍筋匹配方法和装置
CN112183627A (zh) * 2020-09-28 2021-01-05 中星技术股份有限公司 生成预测密度图网络的方法和车辆年检标数量检测方法
CN112488943A (zh) * 2020-12-02 2021-03-12 北京字跳网络技术有限公司 模型训练和图像去雾方法、装置、设备
CN112560958A (zh) * 2020-12-17 2021-03-26 北京赢识科技有限公司 基于人像识别的人员接待方法、装置及电子设备
CN112613386A (zh) * 2020-12-18 2021-04-06 宁波大学科学技术学院 一种基于脑电波的监控方法及装置
CN112950563A (zh) * 2021-02-22 2021-06-11 深圳中科飞测科技股份有限公司 检测方法及装置、检测设备和存储介质
CN113033557A (zh) * 2021-04-16 2021-06-25 北京百度网讯科技有限公司 用于训练图像处理模型和检测图像的方法、装置
CN113095129A (zh) * 2021-03-01 2021-07-09 北京迈格威科技有限公司 姿态估计模型训练方法、姿态估计方法、装置和电子设备
CN113537309A (zh) * 2021-06-30 2021-10-22 北京百度网讯科技有限公司 一种对象识别方法、装置及电子设备
CN113657406A (zh) * 2021-07-13 2021-11-16 北京旷视科技有限公司 模型训练和特征提取方法、装置、电子设备及存储介质
WO2023213233A1 (fr) * 2022-05-06 2023-11-09 墨奇科技(北京)有限公司 Procédé de traitement de tâche, procédé d'entraînement de réseau neuronal, appareil, dispositif et support

Families Citing this family (14)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109939439B (zh) * 2019-03-01 2022-04-05 腾讯科技(深圳)有限公司 虚拟角色的卡住检测方法、模型训练方法、装置及设备
CN111723926B (zh) * 2019-03-22 2023-09-12 北京地平线机器人技术研发有限公司 用于确定图像视差的神经网络模型的训练方法和训练装置
CN110021052B (zh) * 2019-04-11 2023-05-30 北京百度网讯科技有限公司 用于生成眼底图像生成模型的方法和装置
CN112036421A (zh) * 2019-05-16 2020-12-04 搜狗(杭州)智能科技有限公司 一种图像处理方法、装置和电子设备
CN110660103B (zh) * 2019-09-17 2020-12-25 北京三快在线科技有限公司 一种无人车定位方法及装置
CN110969183B (zh) * 2019-09-20 2023-11-21 北京方位捷讯科技有限公司 一种根据图像数据确定目标对象受损程度的方法及系统
CN110825904B (zh) * 2019-10-24 2022-05-06 腾讯科技(深圳)有限公司 一种图像匹配方法、装置、电子设备和存储介质
CN110766081B (zh) * 2019-10-24 2022-09-13 腾讯科技(深圳)有限公司 一种界面图像检测的方法、模型训练的方法以及相关装置
CN111353526A (zh) * 2020-02-19 2020-06-30 上海小萌科技有限公司 一种图像匹配方法、装置以及相关设备
CN111767940A (zh) * 2020-05-14 2020-10-13 北京迈格威科技有限公司 目标物体识别方法、装置、设备和存储介质
CN111597993B (zh) * 2020-05-15 2023-09-05 北京百度网讯科技有限公司 数据处理的方法及装置
CN111797790B (zh) * 2020-07-10 2021-11-05 北京字节跳动网络技术有限公司 图像处理方法和装置、存储介质和电子设备
CN113590857A (zh) * 2021-08-10 2021-11-02 北京有竹居网络技术有限公司 键值匹配方法、装置、可读介质及电子设备
CN114972909A (zh) * 2022-05-16 2022-08-30 北京三快在线科技有限公司 一种模型训练的方法、构建地图的方法及装置

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20140105509A1 (en) * 2012-10-15 2014-04-17 Canon Kabushiki Kaisha Systems and methods for comparing images
CN104376326A (zh) * 2014-11-02 2015-02-25 吉林大学 一种用于图像场景识别的特征提取方法
CN107239535A (zh) * 2017-05-31 2017-10-10 北京小米移动软件有限公司 相似图片检索方法及装置
CN107679466A (zh) * 2017-09-21 2018-02-09 百度在线网络技术(北京)有限公司 信息输出方法和装置
CN108038880A (zh) * 2017-12-20 2018-05-15 百度在线网络技术(北京)有限公司 用于处理图像的方法和装置
CN108154196A (zh) * 2018-01-19 2018-06-12 百度在线网络技术(北京)有限公司 用于输出图像的方法和装置

Family Cites Families (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106780612B (zh) * 2016-12-29 2019-09-17 浙江大华技术股份有限公司 一种图像中的物体检测方法及装置
CN106951484B (zh) * 2017-03-10 2020-10-30 百度在线网络技术(北京)有限公司 图片检索方法及装置、计算机设备及计算机可读介质
CN107944395B (zh) * 2017-11-27 2020-08-18 浙江大学 一种基于神经网络验证人证合一的方法及系统

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20140105509A1 (en) * 2012-10-15 2014-04-17 Canon Kabushiki Kaisha Systems and methods for comparing images
CN104376326A (zh) * 2014-11-02 2015-02-25 吉林大学 一种用于图像场景识别的特征提取方法
CN107239535A (zh) * 2017-05-31 2017-10-10 北京小米移动软件有限公司 相似图片检索方法及装置
CN107679466A (zh) * 2017-09-21 2018-02-09 百度在线网络技术(北京)有限公司 信息输出方法和装置
CN108038880A (zh) * 2017-12-20 2018-05-15 百度在线网络技术(北京)有限公司 用于处理图像的方法和装置
CN108154196A (zh) * 2018-01-19 2018-06-12 百度在线网络技术(北京)有限公司 用于输出图像的方法和装置

Cited By (21)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111507289A (zh) * 2020-04-22 2020-08-07 上海眼控科技股份有限公司 视频匹配方法、计算机设备和存储介质
CN111783869A (zh) * 2020-06-29 2020-10-16 杭州海康威视数字技术股份有限公司 训练数据筛选方法、装置、电子设备及存储介质
CN111783872A (zh) * 2020-06-30 2020-10-16 北京百度网讯科技有限公司 训练模型的方法、装置、电子设备及计算机可读存储介质
CN111783872B (zh) * 2020-06-30 2024-02-02 北京百度网讯科技有限公司 训练模型的方法、装置、电子设备及计算机可读存储介质
CN111984814A (zh) * 2020-08-10 2020-11-24 广联达科技股份有限公司 一种建筑图纸中的箍筋匹配方法和装置
CN111984814B (zh) * 2020-08-10 2024-04-12 广联达科技股份有限公司 一种建筑图纸中的箍筋匹配方法和装置
CN112183627A (zh) * 2020-09-28 2021-01-05 中星技术股份有限公司 生成预测密度图网络的方法和车辆年检标数量检测方法
CN112488943A (zh) * 2020-12-02 2021-03-12 北京字跳网络技术有限公司 模型训练和图像去雾方法、装置、设备
CN112488943B (zh) * 2020-12-02 2024-02-02 北京字跳网络技术有限公司 模型训练和图像去雾方法、装置、设备
CN112560958A (zh) * 2020-12-17 2021-03-26 北京赢识科技有限公司 基于人像识别的人员接待方法、装置及电子设备
CN112613386B (zh) * 2020-12-18 2023-12-19 宁波大学科学技术学院 一种基于脑电波的监控方法及装置
CN112613386A (zh) * 2020-12-18 2021-04-06 宁波大学科学技术学院 一种基于脑电波的监控方法及装置
CN112950563A (zh) * 2021-02-22 2021-06-11 深圳中科飞测科技股份有限公司 检测方法及装置、检测设备和存储介质
CN113095129A (zh) * 2021-03-01 2021-07-09 北京迈格威科技有限公司 姿态估计模型训练方法、姿态估计方法、装置和电子设备
CN113095129B (zh) * 2021-03-01 2024-04-26 北京迈格威科技有限公司 姿态估计模型训练方法、姿态估计方法、装置和电子设备
CN113033557A (zh) * 2021-04-16 2021-06-25 北京百度网讯科技有限公司 用于训练图像处理模型和检测图像的方法、装置
CN113537309A (zh) * 2021-06-30 2021-10-22 北京百度网讯科技有限公司 一种对象识别方法、装置及电子设备
CN113537309B (zh) * 2021-06-30 2023-07-28 北京百度网讯科技有限公司 一种对象识别方法、装置及电子设备
CN113657406A (zh) * 2021-07-13 2021-11-16 北京旷视科技有限公司 模型训练和特征提取方法、装置、电子设备及存储介质
CN113657406B (zh) * 2021-07-13 2024-04-23 北京旷视科技有限公司 模型训练和特征提取方法、装置、电子设备及存储介质
WO2023213233A1 (fr) * 2022-05-06 2023-11-09 墨奇科技(北京)有限公司 Procédé de traitement de tâche, procédé d'entraînement de réseau neuronal, appareil, dispositif et support

Also Published As

Publication number Publication date
CN108898186B (zh) 2020-03-06
CN108898186A (zh) 2018-11-27

Similar Documents

Publication Publication Date Title
WO2020006961A1 (fr) Procédé et dispositif d'extraction d'image
CN108509915B (zh) 人脸识别模型的生成方法和装置
US10902245B2 (en) Method and apparatus for facial recognition
CN108898086B (zh) 视频图像处理方法及装置、计算机可读介质和电子设备
CN108229296B (zh) 人脸皮肤属性识别方法和装置、电子设备、存储介质
CN108197618B (zh) 用于生成人脸检测模型的方法和装置
US20190080148A1 (en) Method and apparatus for generating image
WO2020000879A1 (fr) Procédé et appareil de reconnaissance d'image
WO2021190115A1 (fr) Procédé et appareil de recherche de cible
US20190102605A1 (en) Method and apparatus for generating information
CN107679466B (zh) 信息输出方法和装置
CN109101919B (zh) 用于生成信息的方法和装置
WO2019242222A1 (fr) Procédé et dispositif à utiliser lors de la génération d'informations
WO2020024484A1 (fr) Procédé et dispositif de production de données
CN109993150B (zh) 用于识别年龄的方法和装置
WO2020019591A1 (fr) Procédé et dispositif utilisés pour la génération d'informations
CN107507153B (zh) 图像去噪方法和装置
WO2020062493A1 (fr) Procédé et appareil de traitement d'image
CN111275784B (zh) 生成图像的方法和装置
CN108388889B (zh) 用于分析人脸图像的方法和装置
WO2021083069A1 (fr) Procédé et dispositif de formation de modèle d'échange de visages
CN108229375B (zh) 用于检测人脸图像的方法和装置
CN108509921B (zh) 用于生成信息的方法和装置
WO2022105118A1 (fr) Procédé et appareil d'identification d'état de santé basés sur une image, dispositif et support de stockage
CN110490959B (zh) 三维图像处理方法及装置、虚拟形象生成方法以及电子设备

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 18925299

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

32PN Ep: public notification in the ep bulletin as address of the adressee cannot be established

Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205A DATED 19/04/2021)

122 Ep: pct application non-entry in european phase

Ref document number: 18925299

Country of ref document: EP

Kind code of ref document: A1