CN114424253A - Model training method and device, storage medium and electronic equipment - Google Patents

Model training method and device, storage medium and electronic equipment Download PDF

Info

Publication number
CN114424253A
CN114424253A CN201980100619.0A CN201980100619A CN114424253A CN 114424253 A CN114424253 A CN 114424253A CN 201980100619 A CN201980100619 A CN 201980100619A CN 114424253 A CN114424253 A CN 114424253A
Authority
CN
China
Prior art keywords
image
loss function
classified
class
neural network
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201980100619.0A
Other languages
Chinese (zh)
Inventor
高洪涛
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Guangdong Oppo Mobile Telecommunications Corp Ltd
Shenzhen Huantai Technology Co Ltd
Original Assignee
Guangdong Oppo Mobile Telecommunications Corp Ltd
Shenzhen Huantai Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Guangdong Oppo Mobile Telecommunications Corp Ltd, Shenzhen Huantai Technology Co Ltd filed Critical Guangdong Oppo Mobile Telecommunications Corp Ltd
Publication of CN114424253A publication Critical patent/CN114424253A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Evolutionary Biology (AREA)
  • Evolutionary Computation (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Artificial Intelligence (AREA)
  • Image Analysis (AREA)

Abstract

A model training method, a device, a storage medium and an electronic device are provided, wherein the method comprises the following steps: acquiring a sample image set; inputting the sample image set into a deep neural network for training; if the input image is a classified image, calculating a loss value based on a first loss function; if the input image is the target detection image, calculating a loss value based on the first loss function and the second loss function; and performing back propagation based on the loss value to update the network parameters until convergence to obtain the image recognition model. The method can improve the accuracy of the deep neural network in image category prediction and target detection.

Description

Model training method and device, storage medium and electronic equipment Technical Field
The application relates to the technical field of image processing, in particular to a model training method, a model training device, a storage medium and electronic equipment.
Background
Image processing is a technique that uses a computer to analyze an image to achieve a desired result. In the field of image processing technology, image category prediction has become an important research subject. As research on neural network models advances, methods for obtaining a prediction type of an image by performing type prediction on the image by using the models have been widely accepted. Therefore, it is particularly important how to train through the model to improve the accuracy of the prediction of the subsequent image classes.
Disclosure of Invention
The embodiment of the application provides a model training method and device, a storage medium and electronic equipment, which can improve the accuracy of a deep neural network on image category prediction.
In a first aspect, an embodiment of the present application provides a model training method, including:
obtaining a sample image set, wherein the sample image set comprises a target detection image and a classification image, and the target detection image carries position information and a first class label;
inputting the sample images in the sample image set into a preset deep neural network for training;
if the sample image input into the deep neural network is the classified image, calculating a loss value based on a first loss function;
if the sample image input into the deep neural network is the target detection image, calculating a loss value based on the first loss function and the second loss function;
performing back propagation based on the calculated loss value to update the network parameters until convergence to obtain an image recognition model, wherein the image recognition model is used for recognizing the category of the input image and the position of the category object
In a second aspect, an embodiment of the present application provides a model training apparatus, including:
the image acquisition module is used for acquiring a sample image set, wherein the sample image set comprises a target detection image and a classification image, and the target detection image carries position information and a first class label;
the image input module is used for inputting the sample images in the sample image set into a preset deep neural network for training;
the first calculation module is used for calculating a loss value based on a first loss function if the sample image input into the deep neural network is the classified image;
a second calculation module, configured to calculate a loss value based on the first loss function and a second loss function if the sample image input to the deep neural network is the target detection image;
an iterative training module for performing back propagation to update network parameters until convergence based on the calculated loss value to obtain an image recognition model, wherein the image recognition model is used for recognizing the category of the input image and the position of the category object
In a third aspect, an embodiment of the present application provides a storage medium having a computer program stored thereon, which, when run on a computer, causes the computer to perform:
obtaining a sample image set, wherein the sample image set comprises a target detection image and a classification image, and the target detection image carries position information and a first class label;
inputting the sample images in the sample image set into a preset deep neural network for training;
if the sample image input into the deep neural network is the classified image, calculating a loss value based on a first loss function;
if the sample image input into the deep neural network is the target detection image, calculating a loss value based on the first loss function and the second loss function;
and performing back propagation on the basis of the calculated loss value to update the network parameters until convergence to obtain an image recognition model, wherein the image recognition model is used for recognizing the category of the input image and the position of the category object.
In a fourth aspect, an embodiment of the present application provides an electronic device, including a processor and a memory, where the memory has a computer program, and the processor, by calling the computer program, is configured to perform:
obtaining a sample image set, wherein the sample image set comprises a target detection image and a classification image, and the target detection image carries position information and a first class label;
inputting the sample images in the sample image set into a preset deep neural network for training;
if the sample image input into the deep neural network is the classified image, calculating a loss value based on a first loss function;
if the sample image input into the deep neural network is the target detection image, calculating a loss value based on the first loss function and the second loss function;
and performing back propagation on the basis of the calculated loss value to update the network parameters until convergence to obtain an image recognition model, wherein the image recognition model is used for recognizing the category of the input image and the position of the category object.
In the scheme provided by the embodiment of the application, when a deep neural network is trained, a sample image set comprising a target detection image and a classification image is obtained, the sample image in the sample image set is used for training a preset deep neural network for training, in the training process, when the sample image input into the deep neural network is the classification image, a loss value is calculated based on a first loss function, when the sample image input into the deep neural network is the target detection image, the loss value is calculated based on the first loss function and a second loss function, and back propagation is performed based on the loss value so as to update network parameters until convergence, in the training scheme, the preset deep neural network is trained by combining the target detection image and the classification image, and as the target detection image carries position information and a first class label, the position information indicates the specific position of a class object in the image, in the network training process, the network can more accurately extract the characteristics of the class objects, and the accuracy of the image recognition model obtained by image recognition model training on image class prediction is improved.
Drawings
In order to more clearly illustrate the technical solutions in the embodiments of the present application, the drawings needed to be used in the description of the embodiments are briefly introduced below, and it is obvious that the drawings in the following description are only some embodiments of the present application, and it is obvious for those skilled in the art to obtain other drawings based on these drawings without creative efforts.
Fig. 1 is a schematic flowchart of a first method for training a model according to an embodiment of the present disclosure.
Fig. 2 is a schematic flowchart of a second method for training a model according to an embodiment of the present disclosure.
Fig. 3 is a schematic structural diagram of a model training device according to an embodiment of the present application.
Fig. 4 is a schematic structural diagram of an electronic device according to an embodiment of the present application.
Fig. 5 is a schematic structural diagram of a model training circuit of an electronic device according to an embodiment of the present disclosure.
Detailed Description
The technical solutions in the embodiments of the present application will be clearly and completely described below with reference to the drawings in the embodiments of the present application. It is to be understood that the embodiments described are only a few embodiments of the present application and not all embodiments. All other embodiments, which can be derived by a person skilled in the art from the embodiments given herein without inventive step, are within the scope of the present application.
Reference herein to "an embodiment" means that a particular feature, structure, or characteristic described in connection with the embodiment can be included in at least one embodiment of the application. The appearances of the phrase in various places in the specification are not necessarily all referring to the same embodiment, nor are separate or alternative embodiments mutually exclusive of other embodiments. It is explicitly and implicitly understood by one skilled in the art that the embodiments described herein can be combined with other embodiments.
The embodiment of the application provides a model training method, which comprises the following steps:
obtaining a sample image set, wherein the sample image set comprises a target detection image and a classification image, and the target detection image carries position information and a first class label;
inputting the sample images in the sample image set into a preset deep neural network for training;
if the sample image input into the deep neural network is the classified image, calculating a loss value based on a first loss function;
if the sample image input into the deep neural network is the target detection image, calculating a loss value based on the first loss function and the second loss function;
and performing back propagation on the basis of the calculated loss value to update the network parameters until convergence to obtain an image recognition model, wherein the image recognition model is used for recognizing the category of the input image and the position of the category object.
In some embodiments, the classification image carries a second class label, the target detection image carries position information and a first class label, and the first class labels carried by all the target detection images form a first class label set;
before the calculating the loss value based on the first loss function, the method further includes:
if the sample image input into the deep neural network is the classified image, judging whether a second class label corresponding to the input classified image is contained in the first class label set;
if yes, calculating a loss value based on the first loss function;
and if not, calculating a loss value based on a third loss function, wherein when the input sample images are the same, the loss value calculated by the first loss function is smaller than the loss value calculated by the third loss function.
In some embodiments, the third loss function is k x the first loss function, where k > 1.
In some embodiments, the first loss function is m f and the third loss function is n f, where f is a base loss function, 0 < m < 1, and n > 1.
In some embodiments, the deep neural network is a convolutional neural network; and forming a second class label set by second class labels carried by all classified images, wherein the label types in the first class label set are less than the label types in the second class label set.
In some embodiments, the deep neural network is a convolutional neural network; and forming a second class label set by second class labels carried by all classified images, wherein the label types in the first class label set are less than the label types in the second class label set.
In some embodiments, the back-propagating based on the calculated loss value to update the network parameter until after convergence further comprises:
acquiring an image to be classified;
and carrying out image recognition on the image to be classified according to the image recognition model so as to determine a target class corresponding to the image to be classified and the position of an object belonging to the target class in the image to be classified.
An execution subject of the model training method may be the model training apparatus provided in the embodiment of the present application, or an electronic device integrated with the model training apparatus, where the model training apparatus may be implemented in a hardware or software manner. The electronic device may be a smart phone, a tablet computer, a palm computer, a notebook computer, or a desktop computer.
Referring to fig. 1, fig. 1 is a first flowchart of a model training method according to an embodiment of the present disclosure. The specific process of the model training method provided by the embodiment of the application can be as follows:
in 101, a sample image set is obtained, where the sample image set includes a target detection image and a classification image, where the target detection image carries position information and a first class label.
The image multi-classification based on target detection belongs to strong supervision, and position information of each class object in the image needs to be provided, but when the training sample size of the classification model is large, the labeling of the position information is high in labor cost. The general image multi-classification belongs to a weakly supervised image classification method, and the classification method only needs to label the class name of an image, but cannot identify the position of a class object in the image.
The model training scheme of the embodiment of the application can be applied to an image classification and positioning model, and the model can identify the category of the image and the position of a category object in the image. For example, the position of the category object can be marked by the target box. The model may be constructed based on a deep neural network, such as a Back Propagation (BP) neural network, a convolutional neural network, and the like.
The method and the device adopt two training samples to form a sample image set in a mixed mode, wherein the two sample images comprise a target detection image and a classification image, the target detection image carries a class label and has position information, and the position information indicates the position of a class object in the image. The classified images carry class labels. For convenience of the following description, the class label carried by the target detection image is denoted as a first class label, and the class label carried by the classification image is denoted as a second class label. The first class labels carried by all the target detection images form a first class label set; the second class labels carried by all classified images constitute a second class label set. In some embodiments, the label categories in the second set of category labels may partially coincide with the category labels in the first set of category labels.
In 102, the sample images in the sample image set are input into a preset deep neural network for training.
The two training samples are mixed to form a sample image training model, and the model is essentially strong supervision algorithm and direct classification combined training. In the training process, sample pictures in a sample image set mixed with a target detection image and a classification image are randomly input into a preset neural network for calculation. And calculating loss values by adopting different loss functions according to different types of input sample images.
In 103, if the sample image input to the deep neural network is a classified image, a loss value is calculated based on the first loss function.
At 104, if the sample image input to the deep neural network is the target detection image, a loss value is calculated based on the first loss function and the second loss function.
When training the network using the classified images, the loss function in the network is constituted by a first loss function for calculating a loss value generated when the images are classified. At this time, because the training data has no target frame, when the error information is reversely propagated, only the network parameters related to the classification training part are updated, and the network parameters related to the target detection part are not updated. At this time, since the training data carries the target frame, when the error information is reversely propagated, the network parameters related to the classification training part and the network parameters related to the target detection part are updated, that is, all the network parameters are updated.
When the network is trained by using the target detection image, the loss function in the network is composed of a first loss function and a second loss function, the second loss function is used for calculating the loss value generated when the target detection is carried out on the image, and the first loss function is used for calculating the loss value generated when the image is classified.
Therefore, two loss functions are involved in the training process of the deep neural network. The total loss function can be expressed as L ═ Lp+L clsWherein L isclsIs a first loss function, LpIs a second loss function. If the sample image input into the deep neural network is a classified image, Lp=0。
In this embodiment, the loss function may be selected according to the deep neural network used. For example, a mean square error function, a cross entropy function, or the like may be employed as the loss function.
At 105, the network parameters are updated based on the calculated loss values to converge, resulting in an image recognition model that identifies the class of the input image and the location of the class object.
And in the training process of the network, calculating a loss value based on the loss function and the calculation mode, and performing back propagation based on the calculated loss value to update the network parameters until the network converges. For example, until the number of iterative training reaches a preset value, or until the loss value reaches a minimum, or until the loss value is less than a preset value. And after the training is carried out to be convergent, determining network parameters, and taking the deep neural network after the network parameters are determined as an image number recognition model.
In the network training process, the target detection image carries the position information, and the position information indicates the specific position of the class object in the image, so that the network can more accurately extract the features of the class object in the network training process. By the method, when the sample image input to the network is the classified image, even if the classified image does not carry the position information, the capability of the network for identifying the features of the class objects is enhanced due to the training of the target detection image, the features of the class objects in the classified image can be identified more accurately, and the positions of the class objects can be determined with higher accuracy. It is understood that the category object in the present application refers to an object corresponding to the category label corresponding to the sample image.
For example, taking a preset deep neural network as a convolutional neural network as an example, a cross entropy function is used as a loss function, training data is input, a loss value is calculated according to the loss function, and back propagation is performed based on the loss value, so as to optimize the weight in each convolutional kernel in each convolutional layer of the network.
In particular implementation, the present application is not limited by the execution sequence of the described steps, and some steps may be performed in other sequences or simultaneously without conflict.
From the above, in the model training method provided in the embodiment of the present application, when training the deep neural network, a sample image set including a target detection image and a classification image is obtained, a preset deep neural network is trained by using the sample images in the sample image set, in the training process, when the sample image input to the deep neural network is the classification image, a loss value is calculated based on a first loss function, when the sample image input to the deep neural network is the target detection image, the loss value is calculated based on the first loss function and a second loss function, and back propagation is performed based on the loss value to update a network parameter until convergence, in the training scheme, the preset deep neural network is trained by combining the target detection image and the classification image, and since the target detection image carries position information and a first class label, the position information indicates a specific position of a class object in the image, in the process of training the network, the network can more accurately extract the characteristics of the class objects, and the accuracy of the image recognition model obtained by training the image recognition model on the image class prediction is improved.
The model training method of the present application will be described in further detail below on the basis of the methods described in the above embodiments. Referring to fig. 2, fig. 2 is a second flow chart of the model training method according to the embodiment of the invention. The method comprises the following steps:
in 201, a sample image set is obtained, where the sample image set includes a target detection image and a classification image, where the target detection image carries position information and a first class label, and the first class labels carried by all the target detection images form a first class label set.
In the embodiment, two training samples are mixed to form a sample image set, wherein the two sample images include a target detection image and a classification image, the target detection image carries a class label and has position information, and the position information indicates the position of a class object in the image. The classified images carry class labels. For convenience of the following description, the class label carried by the target detection image is denoted as a first class label, and the class label carried by the classification image is denoted as a second class label. The first class labels carried by all the target detection images form a first class label set; the second class labels carried by all classified images constitute a second class label set. In some embodiments, the label categories in the second set of category labels may partially coincide with the category labels in the first set of category labels.
For example, the deep neural network is used for classification of animals. The sample images are animal images, wherein the target detection images not only carry class labels of the animals, but also identify the positions of the class animals corresponding to the images in the form of target frames in each image. However, the animal categories in the target detection image are only of a large animal class, e.g., dog, cat, deer, etc., but there is no finer category classification, e.g., dogs are not classified as golden retrievers, huskies, shepherds, etc. Meanwhile, the classified images only carry the class labels of the animals, and the specific positions of the animals in the images are not identified, but the classified images have wider and deeper class labels. For example, the classification image includes a large class that is not included in the target detection image, for example, the target detection image does not include the large class, but the classification image includes the large class. The classification image may also include small classes that are not included in the target detection image, for example, small classes such as golden retrievers, huskies, shepherds, etc., are not included in the target detection image, but these classes are included in the classification image. That is, the number of categories of category labels in the second set of category labels may be greater than the number of categories of category labels in the first set of category labels.
Based on the scheme of the embodiment, the two sample images are mixed together to serve as a training sample, the deep neural network is trained in a joint training mode, and the trained network can output high-accuracy position information for small-category dogs which do not appear in the target detection image.
In 202, the sample images in the sample image set are input into a preset deep neural network for training.
The two training samples are mixed to form a sample image training model, and the model is essentially strong supervision algorithm and direct classification combined training. In the training process, sample pictures in a sample image set mixed with a target detection image and a classification image are randomly input into a preset neural network for calculation. And calculating loss values by adopting different loss functions according to different types of input sample images.
At 203, if the sample image input into the deep neural network is a classified image, it is determined whether a second class label corresponding to the input classified image is included in the first class label set.
At 204, if so, a loss value is calculated based on the first loss function.
If not, a loss value is calculated based on a third loss function, wherein the loss value calculated by the first loss function is smaller than the loss value calculated by the third loss function when the input sample images are the same, in 205.
Based on the above example, although the deep neural network is trained by means of joint training, the trained network can output position information with higher accuracy for small types of dogs that do not appear in the target detection image. However, in the training process, when the classification image includes a large class that is not included in the target detection image, for example, the classification image includes a large class that is not included in the target detection image, the accuracy of the position detection is poor. In the present embodiment, this problem is solved in a new loss value calculation manner.
When the sample image input to the deep neural network is a classified image, whether a second class label corresponding to the input classified image is included in the first class label set is judged. If so, a loss value is calculated based on the first loss function. And if not, calculating a loss value based on a third loss function, wherein when the input sample images are the same, the loss value calculated by the first loss function is smaller than the loss value calculated by the third loss function. That is, when the category of the classification image includes a category that is not included in the target detection image, in order to improve the accuracy of network target detection, the loss value is calculated by using a third loss function that is different from another case (a case where the category of the classification image is included in the category of the target detection image), so that the calculated loss value is larger, the network is more sensitive to the category, the features of the category image can be learned more accurately, so as to optimize the model parameters, and further improve the accuracy of detection of the category and the target.
For example, in some implementations, the third loss function is k x the first loss function, where k > 1. In this embodiment, a third loss function is obtained by multiplying a weight coefficient, which is a constant greater than 1, by a formula of the first loss function, for example, k is 1 to 3 in some embodiments; for another example, in some embodiments, k is 1-1.5; for another example, in some embodiments, k is 1.5 ~ 2.
For another example, in some embodiments, the first loss function is m f and the third loss function is n f, where f is the base loss function, 0 < m < 1, and n > 1. For example, f is a cross-entropy loss function, the calculation formula of the first loss function is obtained by multiplying the calculation formula of the cross-entropy loss function by a positive number smaller than 1, and the calculation formula of the second loss function is obtained by multiplying the calculation formula of the cross-entropy loss function by a constant larger than 1.
At 206, if the sample image input to the deep neural network is the target detection image, a loss value is calculated based on the first loss function and the second loss function.
When the network is trained by using the target detection image, the loss function in the network is composed of a first loss function and a second loss function, the second loss function is used for calculating the loss value generated when the target detection is carried out on the image, and the first loss function is used for calculating the loss value generated when the image is classified.
In 207, the network parameters are updated based on the calculated loss values back-propagated until convergence, resulting in an image recognition model that identifies the class of the input image and the location of the class object.
In some embodiments, the back-propagating based on the calculated loss value to update the network parameter until after convergence further comprises: acquiring an image to be classified; and carrying out image recognition on the image to be classified according to the image recognition model so as to determine a target class corresponding to the image to be classified and the position of an object belonging to the target class in the image to be classified.
In this embodiment, the image recognition model obtained by training is used to recognize the image category, and the image to be classified is input into the image recognition model to be calculated, so as to obtain the category label corresponding to the image to be classified and the position of the corresponding category object in the image.
As can be seen from the above, in the model training method provided in the embodiment of the present invention, on the basis of the joint training of the classification data and the target detection data, when the sample image input to the deep neural network is a classification image, if the class label corresponding to the classification image is not included in the class label of the target detection image, the back propagation is performed with a large loss value, the recognition capability of the model for multiple classes is extended, and the accuracy of multiple classes is improved.
The embodiment of the present application further provides a model training device, including:
the image acquisition module is used for acquiring a sample image set, wherein the sample image set comprises a target detection image and a classification image, and the target detection image carries position information and a first class label;
the image input module is used for inputting the sample images in the sample image set into a preset deep neural network for training;
the first calculation module is used for calculating a loss value based on a first loss function if the sample image input into the deep neural network is the classified image;
a second calculation module, configured to calculate a loss value based on the first loss function and a second loss function if the sample image input to the deep neural network is the target detection image;
and the iterative training module is used for carrying out back propagation on the basis of the calculated loss value so as to update the network parameters until convergence, so as to obtain an image recognition model, and the image recognition model is used for recognizing the category of the input image and the position of the category object.
In some embodiments, the classified images carry second class labels, and the first class labels carried by all the target detection images form a first class label set; the device further comprises:
the label detection module is used for judging whether a second class label corresponding to the input classified image is contained in the first class label set or not if the sample image input into the deep neural network is the classified image;
the first computing module is further to:
if the second class label corresponding to the input classified image is contained in the first class label set, calculating a loss value based on the first loss function;
and if the second class label corresponding to the input classified image is not included in the first class label set, calculating a loss value based on a third loss function, wherein when the input sample images are the same, the loss value calculated by the first loss function is smaller than the loss value calculated by the third loss function.
In some embodiments, the third loss function is k x the first loss function, where k > 1.
In some embodiments, the first loss function is m f and the third loss function is n f, where f is a base loss function, 0 < m < 1, and n > 1.
In some embodiments, the deep neural network is a convolutional neural network; and forming a second class label set by second class labels carried by all classified images, wherein the label types in the first class label set are less than the label types in the second class label set.
In some embodiments, the apparatus further comprises an image classification module to:
acquiring an image to be classified;
and carrying out image recognition on the image to be classified according to the image recognition model so as to determine a target class corresponding to the image to be classified and the position of an object belonging to the target class in the image to be classified.
In one embodiment, a model training apparatus is also provided. Referring to fig. 3, fig. 3 is a schematic structural diagram of a model training apparatus 300 according to an embodiment of the present disclosure. The model training apparatus 300 is applied to an electronic device, and the model training apparatus 300 includes an image obtaining module 301, an image input module 302, a first calculating module 303, a second calculating module 304, and an iterative training module 305, as follows:
the image obtaining module 301 is configured to obtain a sample image set, where the sample image set includes a target detection image and a classification image, and the target detection image carries position information and a first class label;
an image input module 302, configured to input a sample image in the sample image set into a preset deep neural network for training;
a first calculating module 303, configured to calculate a loss value based on a first loss function if the sample image input to the deep neural network is the classified image;
a second calculating module 304, configured to calculate a loss value based on the first loss function and the second loss function if the sample image input to the deep neural network is the target detection image;
and the iterative training module 305 is configured to perform back propagation to update the network parameters until convergence based on the calculated loss value, so as to obtain an image recognition model, where the image recognition model is used to recognize the category of the input image and the position of the category object.
In some embodiments, the classification image carries a second class label, the target detection image carries position information and a first class label, and the first class labels carried by all the target detection images form a first class label set;
the model training apparatus 300 further includes a label detection module, configured to determine whether a second class label corresponding to the input classified image is included in the first class label set if the sample image input to the deep neural network is the classified image;
the first calculation module 303 is further configured to: if the second class label corresponding to the input classified image is contained in the first class label set, calculating a loss value based on the first loss function;
and calculating a loss value based on a third loss function if a second class label corresponding to the input classified image is not included in the first class label set, wherein the loss value calculated by the first loss function is smaller than the loss value calculated by the third loss function when the input sample images are the same.
In some embodiments, the third loss function is k x the first loss function, where k > 1.
In some embodiments, the first loss function is m f and the third loss function is n f, where f is a base loss function, 0 < m < 1, and n > 1.
In some embodiments, the deep neural network is a convolutional neural network; and forming a second class label set by second class labels carried by all classified images, wherein the label types in the first class label set are less than the label types in the second class label set.
In some embodiments, the model training apparatus 300 further comprises an image classification module for: acquiring an image to be classified; and carrying out image recognition on the image to be classified according to the image recognition model so as to determine a target class corresponding to the image to be classified and the position of an object belonging to the target class in the image to be classified.
In specific implementation, the above modules may be implemented as independent entities, or may be combined arbitrarily to be implemented as the same or several entities, and specific implementation of the above modules may refer to the foregoing method embodiments, which are not described herein again.
It should be noted that the model training apparatus provided in the embodiment of the present application and the model training method in the above embodiment belong to the same concept, and any method provided in the embodiment of the model training method may be run on the model training apparatus, and the specific implementation process thereof is described in detail in the embodiment of the model training method, and is not described herein again.
As can be seen from the above, in the model training apparatus provided in this embodiment of the application, when training a deep neural network, a sample image set including a target detection image and a classification image is obtained, a preset deep neural network is trained by using sample images in the sample image set, in the training process, when a sample image input to the deep neural network is a classification image, a loss value is calculated based on a first loss function, when a sample image input to the deep neural network is a target detection image, a loss value is calculated based on the first loss function and a second loss function, and back propagation is performed based on the loss value to update a network parameter until convergence, in the training scheme, the preset deep neural network is trained by combining the target detection image and the classification image, and since the target detection image carries position information and a first class label, the position information indicates a specific position of a class object in the image, in the process of training the network, the network can more accurately extract the characteristics of the class objects, and the accuracy of the image recognition model obtained by training the image recognition model on the image class prediction is improved.
The embodiment of the application further provides an electronic device, and the electronic device can be a mobile terminal such as a tablet computer or a smart phone. Referring to fig. 4, fig. 4 is a schematic structural diagram of an electronic device according to an embodiment of the present disclosure. The electronic device 800 may include a camera module 801, a memory 802, a processor 803, a touch display 804, a speaker 805, a microphone 806, and the like.
The camera module 801 may include model training circuitry, which may be implemented using hardware and/or software components, and may include various Processing units that define an Image Signal Processing (Image Signal Processing) pipeline. The model training circuit may include at least: a camera, an Image Signal Processor (ISP Processor), control logic, an Image memory, and a display. Wherein the camera may comprise at least one or more lenses and an image sensor. The image sensor may include an array of color filters (e.g., Bayer filters). The image sensor may acquire light intensity and wavelength information captured with each imaging pixel of the image sensor and provide a set of raw image data that may be processed by an image signal processor.
The image signal processor may process the raw image data pixel by pixel in a variety of formats. For example, each image pixel may have a bit depth of 8, 10, 12, or 14 bits, and the image signal processor may perform one or more model training operations on the raw image data, gathering statistical information about the image data. Wherein the model training operations may be performed with the same or different bit depth precision. The raw image data can be stored in an image memory after being processed by an image signal processor. The image signal processor may also receive image data from an image memory.
The image Memory may be part of a Memory device, a storage device, or a separate dedicated Memory within the electronic device, and may include a DMA (Direct Memory Access) feature.
When image data is received from the image memory, the image signal processor may perform one or more model training operations, such as temporal filtering. The processed image data may be sent to an image memory for additional processing before being displayed. The image signal processor may also receive processed data from the image memory and perform image data processing on the processed data in the raw domain and in the RGB and YCbCr color spaces. The processed image data may be output to a display for viewing by a user and/or further processed by a Graphics Processing Unit (GPU). Further, the output of the image signal processor may also be sent to an image memory, and the display may read image data from the image memory. In one embodiment, the image memory may be configured to implement one or more frame buffers.
The statistical data determined by the image signal processor may be sent to the control logic. For example, the statistical data may include statistical information of the image sensor such as auto exposure, auto white balance, auto focus, flicker detection, black level compensation, lens shading correction, and the like.
The control logic may include a processor and/or microcontroller that executes one or more routines (e.g., firmware). One or more routines may determine camera control parameters and ISP control parameters based on the received statistics. For example, the control parameters of the camera may include camera flash control parameters, control parameters of the lens (e.g., focal length for focusing or zooming), or a combination of these parameters. The ISP control parameters may include gain levels and color correction matrices for automatic white balance and color adjustment (e.g., during RGB processing), etc.
Referring to fig. 5, fig. 5 is a schematic structural diagram of a model training circuit in the present embodiment. For ease of illustration, only aspects of model training techniques related to embodiments of the present invention are shown.
For example, the model training circuit may include: camera, image signal processor, control logic ware, image memory, display. The camera may include one or more lenses and an image sensor, among others. In some embodiments, the camera may be either a tele camera or a wide camera.
And the image collected by the camera is transmitted to an image signal processor for processing. After the image signal processor processes the image, statistical data of the image (such as brightness of the image, contrast value of the image, color of the image, etc.) may be sent to the control logic. The control logic device can determine the control parameters of the camera according to the statistical data, so that the camera can carry out operations such as automatic focusing and automatic exposure according to the control parameters. The image can be stored in the image memory after being processed by the image signal processor. The image signal processor may also read the image stored in the image memory for processing. In addition, the image can be directly sent to a display for displaying after being processed by the image signal processor. The display may also read the image in the image memory for display.
In addition, not shown in the figure, the electronic device may further include a CPU and a power supply module. The CPU is connected with the logic controller, the image signal processor, the image memory and the display, and is used for realizing global control. The power supply module is used for supplying power to each module.
The memory 802 stores applications containing executable code. The application programs may constitute various functional modules. The processor 803 executes various functional applications and data processing by running the application programs stored in the memory 802.
The processor 803 is a control center of the electronic device, connects various parts of the entire electronic device using various interfaces and lines, and performs various functions of the electronic device and processes data by running or executing an application program stored in the memory 802 and calling data stored in the memory 802, thereby integrally monitoring the electronic device.
The touch display screen 804 may be used to receive user touch control operations for the electronic device. Speaker 805 may play sound signals. The microphone 806 may be used to pick up sound signals.
In this embodiment, the processor 803 in the electronic device loads the executable code corresponding to the processes of one or more application programs into the memory 802 according to the following instructions, and the processor 803 runs the application programs stored in the memory 802, so as to execute:
obtaining a sample image set, wherein the sample image set comprises a target detection image and a classification image, and the target detection image carries position information and a first class label;
inputting the sample images in the sample image set into a preset deep neural network for training;
if the sample image input into the deep neural network is the classified image, calculating a loss value based on a first loss function;
if the sample image input into the deep neural network is the target detection image, calculating a loss value based on the first loss function and the second loss function;
and performing back propagation on the basis of the calculated loss value to update the network parameters until convergence to obtain an image recognition model, wherein the image recognition model is used for recognizing the category of the input image and the position of the category object.
In some embodiments, the classification image carries a second class label, the target detection image carries position information and a first class label, and the first class labels carried by all the target detection images form a first class label set; the processor 803 also performs:
if the sample image input into the deep neural network is the classified image, judging whether a second class label corresponding to the input classified image is contained in the first class label set;
if yes, calculating a loss value based on the first loss function;
and if not, calculating a loss value based on a third loss function, wherein when the input sample images are the same, the loss value calculated by the first loss function is smaller than the loss value calculated by the third loss function.
In some embodiments, the processor 803 also performs:
acquiring an image to be classified; and carrying out image recognition on the image to be classified according to the image recognition model so as to determine a target class corresponding to the image to be classified and the position of an object belonging to the target class in the image to be classified.
From the above, an embodiment of the present application provides an electronic device, where the electronic device obtains a sample image set including a target detection image and a classification image when training a deep neural network, trains a preset deep neural network by using the sample images in the sample image set to perform training, calculates a loss value based on a first loss function when a sample image input to the deep neural network is a classification image, calculates a loss value based on the first loss function and a second loss function when the sample image input to the deep neural network is a target detection image, and performs back propagation based on the loss value to update network parameters until convergence, in the above training scheme, the preset deep neural network is trained by combining the target detection image and the classification image, and since the target detection image carries position information and a first class label, the position information indicates the specific position of the class object in the image, so that in the process of training the network, the network can more accurately extract the characteristics of the class object, and the accuracy of the image recognition model obtained by training the image recognition model for image class prediction is improved.
An embodiment of the present application further provides a storage medium, where a computer program is stored in the storage medium, and when the computer program runs on a computer, the computer executes the model training method according to any of the above embodiments.
It should be noted that, all or part of the steps in the methods of the above embodiments may be implemented by hardware related to instructions of a computer program, which may be stored in a computer-readable storage medium, which may include, but is not limited to: read Only Memory (ROM), Random Access Memory (RAM), magnetic or optical disks, and the like.
Furthermore, the terms "first", "second", and "third", etc. in this application are used to distinguish different objects, and are not used to describe a particular order. Furthermore, the terms "include" and "have," as well as any variations thereof, are intended to cover non-exclusive inclusions. For example, a process, method, system, article, or apparatus that comprises a list of steps or modules is not limited to only those steps or modules listed, but rather, some embodiments may include other steps or modules not listed or inherent to such process, method, article, or apparatus.
The model training method, the model training device, the storage medium, and the electronic device provided in the embodiments of the present application are described in detail above. The principle and the implementation of the present application are explained herein by applying specific examples, and the above description of the embodiments is only used to help understand the method and the core idea of the present application; meanwhile, for those skilled in the art, according to the idea of the present application, there may be variations in the specific embodiments and the application scope, and in summary, the content of the present specification should not be construed as a limitation to the present application.

Claims (20)

  1. A method of model training, comprising:
    obtaining a sample image set, wherein the sample image set comprises a target detection image and a classification image, and the target detection image carries position information and a first class label;
    inputting the sample images in the sample image set into a preset deep neural network for training;
    if the sample image input into the deep neural network is the classified image, calculating a loss value based on a first loss function;
    if the sample image input into the deep neural network is the target detection image, calculating a loss value based on the first loss function and the second loss function;
    and performing back propagation on the basis of the calculated loss value to update the network parameters until convergence to obtain an image recognition model, wherein the image recognition model is used for recognizing the category of the input image and the position of the category object.
  2. The model training method of claim 1, wherein the classification images carry second class labels, the object detection images carry position information and first class labels, and the first class labels carried by all the object detection images form a first class label set;
    before the calculating the loss value based on the first loss function, the method further includes:
    if the sample image input into the deep neural network is the classified image, judging whether a second class label corresponding to the input classified image is contained in the first class label set;
    if yes, calculating a loss value based on the first loss function;
    and if not, calculating a loss value based on a third loss function, wherein when the input sample images are the same, the loss value calculated by the first loss function is smaller than the loss value calculated by the third loss function.
  3. The model training method of claim 2, wherein the third loss function is k first loss function, where k > 1.
  4. The model training method of claim 2, wherein the first loss function is m f and the third loss function is n f, wherein f is a base loss function, 0 < m < 1, and n > 1.
  5. The model training method of claim 2, wherein the deep neural network is a convolutional neural network; and forming a second class label set by second class labels carried by all classified images, wherein the label types in the first class label set are less than the label types in the second class label set.
  6. The model training method of claim 1, wherein said back-propagating based on the calculated loss value to update the network parameters until after convergence further comprises:
    acquiring an image to be classified;
    and carrying out image recognition on the image to be classified according to the image recognition model so as to determine a target class corresponding to the image to be classified and the position of an object belonging to the target class in the image to be classified.
  7. A model training apparatus, comprising:
    the image acquisition module is used for acquiring a sample image set, wherein the sample image set comprises a target detection image and a classification image, and the target detection image carries position information and a first class label;
    the image input module is used for inputting the sample images in the sample image set into a preset deep neural network for training;
    the first calculation module is used for calculating a loss value based on a first loss function if the sample image input into the deep neural network is the classified image;
    a second calculation module, configured to calculate a loss value based on the first loss function and a second loss function if the sample image input to the deep neural network is the target detection image;
    and the iterative training module is used for carrying out back propagation on the basis of the calculated loss value so as to update the network parameters until convergence, so as to obtain an image recognition model, and the image recognition model is used for recognizing the category of the input image and the position of the category object.
  8. The model training apparatus of claim 7, wherein the classification images carry second class labels, and the first class labels carried by all the target detection images form a first class label set; the device further comprises:
    the label detection module is used for judging whether a second class label corresponding to the input classified image is contained in the first class label set or not if the sample image input into the deep neural network is the classified image;
    the first computing module is further to:
    if the second class label corresponding to the input classified image is contained in the first class label set, calculating a loss value based on the first loss function;
    and if the second class label corresponding to the input classified image is not included in the first class label set, calculating a loss value based on a third loss function, wherein when the input sample images are the same, the loss value calculated by the first loss function is smaller than the loss value calculated by the third loss function.
  9. The model training apparatus of claim 8, wherein the third loss function is k x the first loss function, where k > 1.
  10. The model training apparatus of claim 8 wherein the first loss function is m f and the third loss function is n f, wherein f is a base loss function, 0 < m < 1, and n > 1.
  11. The model training apparatus of claim 8, wherein the deep neural network is a convolutional neural network; and forming a second class label set by second class labels carried by all classified images, wherein the label types in the first class label set are less than the label types in the second class label set.
  12. The model training apparatus of claim 7, wherein the apparatus further comprises an image classification module to:
    acquiring an image to be classified;
    and carrying out image recognition on the image to be classified according to the image recognition model so as to determine a target class corresponding to the image to be classified and the position of an object belonging to the target class in the image to be classified.
  13. A storage medium having a computer program stored thereon, which, when run on a computer, causes the computer to perform:
    obtaining a sample image set, wherein the sample image set comprises a target detection image and a classification image, and the target detection image carries position information and a first class label;
    inputting the sample images in the sample image set into a preset deep neural network for training;
    if the sample image input into the deep neural network is the classified image, calculating a loss value based on a first loss function;
    if the sample image input into the deep neural network is the target detection image, calculating a loss value based on the first loss function and the second loss function;
    and performing back propagation on the basis of the calculated loss value to update the network parameters until convergence to obtain an image recognition model, wherein the image recognition model is used for recognizing the category of the input image and the position of the category object.
  14. The storage medium of claim 13, wherein the classified images carry second category labels, the object detection images carry location information and first category labels, and the first category labels carried by all the object detection images form a first category label set;
    the computer program, when executed on a computer, may further cause the computer to perform:
    if the sample image input into the deep neural network is the classified image, judging whether a second class label corresponding to the input classified image is contained in the first class label set;
    if yes, calculating a loss value based on the first loss function;
    and if not, calculating a loss value based on a third loss function, wherein when the input sample images are the same, the loss value calculated by the first loss function is smaller than the loss value calculated by the third loss function.
  15. An electronic device comprising a processor and a memory, the memory storing a computer program, wherein the processor, by invoking the computer program, is configured to perform:
    obtaining a sample image set, wherein the sample image set comprises a target detection image and a classification image, and the target detection image carries position information and a first class label;
    inputting the sample images in the sample image set into a preset deep neural network for training;
    if the sample image input into the deep neural network is the classified image, calculating a loss value based on a first loss function;
    if the sample image input into the deep neural network is the target detection image, calculating a loss value based on the first loss function and the second loss function;
    and performing back propagation on the basis of the calculated loss value to update the network parameters until convergence to obtain an image recognition model, wherein the image recognition model is used for recognizing the category of the input image and the position of the category object.
  16. The electronic device of claim 15, wherein the classified images carry second category labels, the object detection images carry location information and first category labels, and the first category labels carried by all the object detection images form a first category label set; the processor may further be configured to, by invoking the computer program, perform:
    if the sample image input into the deep neural network is the classified image, judging whether a second class label corresponding to the input classified image is contained in the first class label set;
    if yes, calculating a loss value based on the first loss function;
    and if not, calculating a loss value based on a third loss function, wherein when the input sample images are the same, the loss value calculated by the first loss function is smaller than the loss value calculated by the third loss function.
  17. The electronic device of claim 16, wherein the third loss function is k x the first loss function, where k > 1.
  18. The electronic device of claim 16, wherein the first loss function is m f and the third loss function is n f, wherein f is a base loss function, 0 < m < 1, and n > 1.
  19. The electronic device of claim 16, wherein the deep neural network is a convolutional neural network; and forming a second class label set by second class labels carried by all classified images, wherein the label types in the first class label set are less than the label types in the second class label set.
  20. The electronic device of claim 15, wherein the processor is further operable by invoking the computer program to perform:
    acquiring an image to be classified;
    and carrying out image recognition on the image to be classified according to the image recognition model so as to determine a target class corresponding to the image to be classified and the position of an object belonging to the target class in the image to be classified.
CN201980100619.0A 2019-11-08 2019-11-08 Model training method and device, storage medium and electronic equipment Pending CN114424253A (en)

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
PCT/CN2019/116710 WO2021087985A1 (en) 2019-11-08 2019-11-08 Model training method and apparatus, storage medium, and electronic device

Publications (1)

Publication Number Publication Date
CN114424253A true CN114424253A (en) 2022-04-29

Family

ID=75849227

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201980100619.0A Pending CN114424253A (en) 2019-11-08 2019-11-08 Model training method and device, storage medium and electronic equipment

Country Status (2)

Country Link
CN (1) CN114424253A (en)
WO (1) WO2021087985A1 (en)

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114821207A (en) * 2022-06-30 2022-07-29 浙江凤凰云睿科技有限公司 Image classification method and device, storage medium and terminal
CN115439699A (en) * 2022-10-25 2022-12-06 北京鹰瞳科技发展股份有限公司 Training method of target detection model, target detection method and related product
CN116486134A (en) * 2023-03-02 2023-07-25 哈尔滨市科佳通用机电股份有限公司 Train brake hose hook falling-out fault detection method based on deep neural network
CN116935102A (en) * 2023-06-30 2023-10-24 上海蜜度信息技术有限公司 Lightweight model training method, device, equipment and medium

Families Citing this family (35)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113298156A (en) * 2021-05-28 2021-08-24 有米科技股份有限公司 Neural network training method and device for image gender classification
CN113516053B (en) * 2021-05-28 2024-05-14 西安空间无线电技术研究所 Ship target refinement detection method with rotation invariance
CN113282927B (en) * 2021-05-31 2024-02-02 平安国际智慧城市科技股份有限公司 Malicious code detection method, device, equipment and computer readable storage medium
CN113221837B (en) * 2021-06-01 2024-06-07 北京金山云网络技术有限公司 Object segmentation method, training method and device of object segmentation model
CN113837216B (en) * 2021-06-01 2024-05-10 腾讯科技(深圳)有限公司 Data classification method, training device, medium and electronic equipment
CN113364792B (en) * 2021-06-11 2022-07-12 奇安信科技集团股份有限公司 Training method of flow detection model, flow detection method, device and equipment
CN113537286A (en) * 2021-06-11 2021-10-22 浙江智慧视频安防创新中心有限公司 Image classification method, device, equipment and medium
CN113505820B (en) * 2021-06-23 2024-02-06 北京阅视智能技术有限责任公司 Image recognition model training method, device, equipment and medium
CN113496256B (en) * 2021-06-24 2024-04-09 中汽创智科技有限公司 Image annotation model training method, annotation method, device, equipment and medium
CN113378833B (en) * 2021-06-25 2023-09-01 北京百度网讯科技有限公司 Image recognition model training method, image recognition device and electronic equipment
CN113591918B (en) * 2021-06-29 2024-02-06 北京百度网讯科技有限公司 Training method of image processing model, image processing method, device and equipment
CN113505800A (en) * 2021-06-30 2021-10-15 深圳市慧鲤科技有限公司 Image processing method and training method, device, equipment and medium of model thereof
CN113408662A (en) * 2021-07-19 2021-09-17 北京百度网讯科技有限公司 Image recognition method and device, and training method and device of image recognition model
CN113836338B (en) * 2021-07-21 2024-05-24 北京邮电大学 Fine granularity image classification method, device, storage medium and terminal
CN113780101A (en) * 2021-08-20 2021-12-10 京东鲲鹏(江苏)科技有限公司 Obstacle avoidance model training method and device, electronic equipment and storage medium
CN113657523A (en) * 2021-08-23 2021-11-16 科大讯飞股份有限公司 Image target classification method, device, equipment and storage medium
CN113449704B (en) * 2021-08-31 2022-03-25 北京的卢深视科技有限公司 Face recognition model training method and device, electronic equipment and storage medium
CN113962383A (en) * 2021-10-15 2022-01-21 北京百度网讯科技有限公司 Model training method, target tracking method, device, equipment and storage medium
CN113947701B (en) * 2021-10-18 2024-02-23 北京百度网讯科技有限公司 Training method, object recognition method, device, electronic equipment and storage medium
CN113962965B (en) * 2021-10-26 2023-06-09 腾讯科技(深圳)有限公司 Image quality evaluation method, device, equipment and storage medium
CN113963148B (en) * 2021-10-29 2023-08-08 北京百度网讯科技有限公司 Object detection method, object detection model training method and device
CN113780480B (en) * 2021-11-11 2022-02-22 深圳佑驾创新科技有限公司 Method for constructing multi-target detection and category identification model based on YOLOv5
CN114972725B (en) * 2021-12-30 2023-05-23 华为技术有限公司 Model training method, readable medium and electronic device
CN114332547B (en) * 2022-03-17 2022-07-08 浙江太美医疗科技股份有限公司 Medical object classification method and apparatus, electronic device, and storage medium
CN114549938B (en) * 2022-04-25 2022-09-09 广州市玄武无线科技股份有限公司 Model training method, image information management method, image recognition method and device
WO2023216251A1 (en) * 2022-05-13 2023-11-16 华为技术有限公司 Map generation method, model training method, readable medium, and electronic device
CN115270848B (en) * 2022-06-17 2023-09-29 合肥心之声健康科技有限公司 PPG and ECG automatic conversion intelligent algorithm, storage medium and computer system
CN115294396B (en) * 2022-08-12 2024-04-23 北京百度网讯科技有限公司 Backbone network training method and image classification method
CN115529159B (en) * 2022-08-16 2024-03-08 中国电信股份有限公司 Training method, device, equipment and storage medium of encryption traffic detection model
CN115331062B (en) * 2022-08-29 2023-08-08 北京达佳互联信息技术有限公司 Image recognition method, image recognition device, electronic device and computer-readable storage medium
CN115601618B (en) * 2022-11-29 2023-03-10 浙江华是科技股份有限公司 Magnetic core defect detection method and system and computer storage medium
CN115793490B (en) * 2023-02-06 2023-04-11 南通弈匠智能科技有限公司 Intelligent household energy-saving control method based on big data
CN116663650B (en) * 2023-06-06 2023-12-19 北京百度网讯科技有限公司 Training method of deep learning model, target object detection method and device
CN116468973B (en) * 2023-06-09 2023-10-10 深圳比特微电子科技有限公司 Training method and device for target detection model of low-illumination image
CN117282687B (en) * 2023-10-18 2024-05-28 广州市普理司科技有限公司 Automatic mark picking and supplementing control system for visual inspection of printed matter

Family Cites Families (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107134144B (en) * 2017-04-27 2019-07-12 武汉理工大学 A kind of vehicle checking method for traffic monitoring
CN110490177A (en) * 2017-06-02 2019-11-22 腾讯科技(深圳)有限公司 A kind of human-face detector training method and device
CN109522967A (en) * 2018-11-28 2019-03-26 广州逗号智能零售有限公司 A kind of commodity attribute recognition methods, device, equipment and storage medium
CN110189317A (en) * 2019-05-30 2019-08-30 上海卡罗网络科技有限公司 A kind of road image intelligent acquisition and recognition methods based on deep learning
CN110298266B (en) * 2019-06-10 2023-06-06 天津大学 Deep neural network target detection method based on multiscale receptive field feature fusion
CN110349147B (en) * 2019-07-11 2024-02-02 腾讯医疗健康(深圳)有限公司 Model training method, fundus macular region lesion recognition method, device and equipment

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114821207A (en) * 2022-06-30 2022-07-29 浙江凤凰云睿科技有限公司 Image classification method and device, storage medium and terminal
CN114821207B (en) * 2022-06-30 2022-11-04 浙江凤凰云睿科技有限公司 Image classification method and device, storage medium and terminal
CN115439699A (en) * 2022-10-25 2022-12-06 北京鹰瞳科技发展股份有限公司 Training method of target detection model, target detection method and related product
CN116486134A (en) * 2023-03-02 2023-07-25 哈尔滨市科佳通用机电股份有限公司 Train brake hose hook falling-out fault detection method based on deep neural network
CN116935102A (en) * 2023-06-30 2023-10-24 上海蜜度信息技术有限公司 Lightweight model training method, device, equipment and medium
CN116935102B (en) * 2023-06-30 2024-02-20 上海蜜度科技股份有限公司 Lightweight model training method, device, equipment and medium

Also Published As

Publication number Publication date
WO2021087985A1 (en) 2021-05-14

Similar Documents

Publication Publication Date Title
CN114424253A (en) Model training method and device, storage medium and electronic equipment
US11457138B2 (en) Method and device for image processing, method for training object detection model
WO2020192483A1 (en) Image display method and device
WO2019134504A1 (en) Method and device for blurring image background, storage medium, and electronic apparatus
CN111178183B (en) Face detection method and related device
CN107316035A (en) Object identifying method and device based on deep learning neutral net
CN111771226A (en) Electronic device, image processing method thereof, and computer-readable recording medium
CN108229673B (en) Convolutional neural network processing method and device and electronic equipment
CN110519485B (en) Image processing method, image processing device, storage medium and electronic equipment
WO2023160312A1 (en) Person re-identification method and apparatus based on self-supervised learning, and device and storage medium
CN110572636B (en) Camera contamination detection method and device, storage medium and electronic equipment
CN110728705B (en) Image processing method, image processing device, storage medium and electronic equipment
CN110866872B (en) Pavement crack image preprocessing intelligent selection method and device and electronic equipment
CN112602319B (en) Focusing device, method and related equipment
CN112418327A (en) Training method and device of image classification model, electronic equipment and storage medium
CN115100732A (en) Fishing detection method and device, computer equipment and storage medium
WO2021134485A1 (en) Method and device for scoring video, storage medium and electronic device
CN112906810B (en) Target detection method, electronic device, and storage medium
CN111597937B (en) Fish gesture recognition method, device, equipment and storage medium
CN113837257A (en) Target detection method and device
CN111611835A (en) Ship detection method and device
CN116977256A (en) Training method, device, equipment and storage medium for defect detection model
CN113223083B (en) Position determining method and device, electronic equipment and storage medium
WO2021179819A1 (en) Photo processing method and apparatus, and storage medium and electronic device
CN113298122A (en) Target detection method and device and electronic equipment

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination