WO2021102655A1 - 网络模型训练方法、图像属性识别方法、装置及电子设备 - Google Patents
网络模型训练方法、图像属性识别方法、装置及电子设备 Download PDFInfo
- Publication number
- WO2021102655A1 WO2021102655A1 PCT/CN2019/120749 CN2019120749W WO2021102655A1 WO 2021102655 A1 WO2021102655 A1 WO 2021102655A1 CN 2019120749 W CN2019120749 W CN 2019120749W WO 2021102655 A1 WO2021102655 A1 WO 2021102655A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- image
- model
- neural network
- loss function
- network model
- Prior art date
Links
- 238000012549 training Methods 0.000 title claims abstract description 250
- 238000000034 method Methods 0.000 title claims abstract description 99
- 230000006870 function Effects 0.000 claims description 140
- 230000000306 recurrent effect Effects 0.000 claims description 88
- 238000013527 convolutional neural network Methods 0.000 claims description 62
- 238000003062 neural network model Methods 0.000 claims description 50
- 238000013528 artificial neural network Methods 0.000 claims description 38
- 238000004590 computer program Methods 0.000 claims description 19
- 238000011176 pooling Methods 0.000 claims description 19
- 238000004364 calculation method Methods 0.000 claims description 2
- 238000003672 processing method Methods 0.000 description 12
- 238000010586 diagram Methods 0.000 description 11
- 238000012545 processing Methods 0.000 description 8
- 229920000742 Cotton Polymers 0.000 description 6
- 238000010276 construction Methods 0.000 description 6
- 241001465754 Metazoa Species 0.000 description 4
- 239000011521 glass Substances 0.000 description 4
- 238000004422 calculation algorithm Methods 0.000 description 3
- 125000004122 cyclic group Chemical group 0.000 description 3
- 230000001815 facial effect Effects 0.000 description 3
- 230000003287 optical effect Effects 0.000 description 3
- 238000004891 communication Methods 0.000 description 2
- 238000002474 experimental method Methods 0.000 description 2
- 238000000605 extraction Methods 0.000 description 2
- 210000000697 sensory organ Anatomy 0.000 description 2
- 230000009286 beneficial effect Effects 0.000 description 1
- 238000013135 deep learning Methods 0.000 description 1
- 230000008034 disappearance Effects 0.000 description 1
- 238000007599 discharging Methods 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 238000005516 engineering process Methods 0.000 description 1
- 230000003631 expected effect Effects 0.000 description 1
- 238000003384 imaging method Methods 0.000 description 1
- 230000002452 interceptive effect Effects 0.000 description 1
- 239000004973 liquid crystal related substance Substances 0.000 description 1
- 230000001960 triggered effect Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
Definitions
- the embodiments of the application relate to computer technology, and in particular to a network model training method, image attribute recognition method, device, and electronic equipment.
- Image recognition refers to the use of computers to process, analyze, and understand images to identify targets and objects in various patterns. It is a practical application of deep learning algorithms.
- the current image recognition method is to directly feed the image into the convolutional neural network for feature extraction, and process the extracted features in the fully connected layer of the convolutional neural network to obtain the final prediction result of the image.
- the image recognition results obtained through such image recognition will ignore many image attributes and the correlation and sequence between image attributes. For example, when recognizing a human body image, if there are hats, glasses, tops, For bags, bottoms, shoes, etc., there will be relevance and order between tops and bottoms. If they are directly identified through the above-mentioned prior art, the relevance and order between tops and bottoms will be ignored, leading to inconsistencies. Accurate recognition effect.
- This application provides a network model training method, image attribute recognition method, device, and electronic equipment, which can accurately recognize image attributes and the correlation between each attribute.
- an embodiment of the present application provides a network model training method, and the method includes:
- the image sample set including a plurality of initial values of image attributes
- the basic model including a convolutional neural network model and a recurrent neural network model;
- the convergent basic model is used as a recognition model for recognizing image attributes.
- an embodiment of the present application also provides an image attribute recognition method, which includes:
- the image attribute recognition model adopts the image attribute recognition model obtained by training the network model training method provided in the embodiment of the present application.
- an embodiment of the present application provides a network model training device, including:
- the first obtaining module is configured to obtain an image sample set, the image sample set including a plurality of initial values of image attributes
- the first recognition module is configured to input the image sample set into the basic model for image attribute recognition, so as to obtain the first training result obtained according to the recurrent neural network model and the first training result according to the convolutional neural network model The second training result obtained;
- a training module configured to perform joint training on the convolutional neural network model and the recurrent neural network model according to the first training result, the second training result, the initial value of the image attribute, and the target loss function , Until the basic model converges;
- the determining module is used to use the converged basic model as a recognition model for recognizing image attributes.
- an image attribute recognition device including:
- the receiving module is used to receive the image attribute recognition request
- the second obtaining module is configured to obtain the image to be recognized according to the image attribute recognition request
- the calling module is used to call the pre-trained image attribute recognition model
- the second recognition module is used to input the image to be recognized into a pre-trained image attribute recognition model, and to recognize the image attributes of the image to be recognized to obtain an image attribute recognition result;
- the image attribute recognition model adopts the image attribute recognition model obtained by the network model training method provided in the embodiment of the present application.
- an embodiment of the present application provides a storage medium on which a computer program is stored, wherein when the computer program is executed on a computer, the computer is caused to execute the network model training method provided in this embodiment or Image attribute recognition method.
- an embodiment of the present application provides an electronic device including a memory and a processor, the memory stores a computer program, and the processor invokes the computer program stored in the memory to execute:
- the image sample set including a plurality of initial values of image attributes
- the basic model including a convolutional neural network model and a recurrent neural network model;
- the convergent basic model is used as a recognition model for recognizing image attributes.
- an embodiment of the present application provides an electronic device including a memory and a processor, the memory stores a computer program, and the processor invokes the computer program stored in the memory to execute:
- the image attribute recognition model adopts the image attribute recognition model obtained by training the network model training method provided in the embodiment of the present application.
- FIG. 1 is a schematic diagram of the first process of a network model training method provided by an embodiment of the present application.
- FIG. 2 is a schematic diagram of the second process of the network model training method provided by an embodiment of the present application.
- FIG. 3 is a schematic flowchart of an image attribute recognition method provided by an embodiment of the present application.
- Fig. 4 is a schematic structural diagram of a network model training device provided by an embodiment of the present application.
- Fig. 5 is a schematic structural diagram of an image attribute recognition device provided by an embodiment of the present application.
- Fig. 6 is a first schematic structural diagram of an electronic device provided by an embodiment of the present application.
- FIG. 7 is a schematic diagram of a second structure of an electronic device provided by an embodiment of the present application.
- FIG. 1 is a schematic flowchart of the first network model training method provided by an embodiment of the present application.
- the process of the network model training method may include:
- the image sample set includes a variety of images, such as human body images, animal images, plant images, etc.
- images such as human body images, animal images, plant images, etc.
- multiple human images in the image sample set may be used as the training of the network model image.
- the images in the image sample set can choose to include a variety of different clothing combinations Human body image.
- the correlation between facial features or the correlation between limbs can also be selected as the direction of model training.
- the images containing facial features in multiple human images can be cropped into different images of the same size.
- the preset number of feature points of the face in the human image can be obtained.
- the image containing facial features in the human body image is disassembled and constructed.
- the human body image is considered not to be an image sample set.
- each training image in the image sample set has its own corresponding initial value of image attributes.
- the human body image contains attributes such as tops, bottoms, hats, shoes, etc.
- the correlation between the top and bottom is 10%, and the 10% correlation between the top and bottom can be used as the initial value of the image attribute.
- the initial value of the image attribute can be multiple or one, depending on specific needs. It is determined according to the number of attributes in the training image and the degree of association between the attributes.
- the basic model can be jointly created by using different types of network models.
- a convolutional neural network model Convolutional Neural Networks, CNN
- a recurrent neural network Recurrent Neural Network, RNN
- CNN convolutional Neural Networks
- RNN recurrent Neural Network
- an input layer can be set, the input layer is used to input the training images in the image sample set to the basic model, and then the input layer is connected with the convolutional layer, and the convolutional layer is respectively connected with the pooling layer and the recurrent neural network Connection, the cyclic neural network is connected to the first fully connected layer, the pooling layer is connected to the second fully connected layer, and the first fully connected layer and the second fully connected layer are used as the output layer of the basic model.
- the input layer, the convolutional layer, the pooling layer, and the second fully connected layer are connected in sequence to form a convolutional neural network, and the recurrent neural network is arranged between the convolutional layer and the first fully connected layer.
- the training sample input is a continuous sequence
- the length of the sequence is different, such as a time-based sequence: a continuous speech and a continuous handwritten text.
- the recurrent neural network can handle the problem of uncertain input training values.
- the recurrent neural network also has the problem of gradient disappearance, it is difficult to process long sequences of data. Therefore, the recurrent neural network can adopt gated recurrent unit networks (Gated Recurrent Unit networks, GRU), and the recurrent unit of the gated recurrent unit network is only Contains two gates: update gate and reset gate. The two gates of update gate and reset gate do not form a self-loop, but directly recurse between system states.
- the gated recurrent unit network corresponds to the first loss function
- the convolutional neural network corresponds to the second loss function
- the first loss function and the second loss function It can be a loss function of different types, or it can be a loss function of the same type, and the target loss function corresponding to the basic model is obtained from the first loss function and the second loss function.
- the first loss function can be multiplied by the loss coefficient, and then the second loss function can be added to obtain the objective function corresponding to the basic model, where the loss coefficient can be a parameter obtained through experiments and can be set to 0.8 to 1. between.
- the training image in the image sample set is input into the input layer, where the training image can be a human body image, and the attributes of the image are identified through the constructed basic model and the objective function corresponding to the basic model.
- the training image is first calculated by the convolutional layer to obtain the first feature value, and the first feature value output by the last layer of the convolutional layer is input to the pooling layer and the recurrent neural network, respectively, where the second feature output by the recurrent neural network The value is input to the first fully connected layer to obtain the first training result, and the third feature value input from the pooling layer is input to the second fully connected layer to obtain the second training result.
- first training result and the second training result are not completely the same, and the final training result obtained by the basic model is obtained based on the first training result and the second training result.
- the intersection of the first training result and the second training result can be taken, and the target training result in the intersection is the final training result obtained by the basic model.
- the final training result you can also choose to add the first training result and the second training result to get the final training result, and you can also select part of the training result from the first training result and the second training result according to a preset rule. The final training result.
- multiple first training results and second training results are obtained according to the basic model, and then the final training result is obtained according to the first training result and the second training result, and the final training result is obtained by And the initial value of the image attribute is input into the target loss function to get the target loss value.
- the input training image is a human body image, where the person in the human body image is wearing a coat, a bottom coat, shoes, a hat, and other wearing objects.
- Each wearing object can be regarded as one of the attributes of the human body image.
- the two are input into the target loss function to obtain the corresponding target loss value. It can be judged whether the target loss value is close to the preset loss value. If the target loss value is If the distance preset loss value is within the preset range, it is considered that the basic model has been trained and is in a state of convergence.
- the electronic device inputs the human body image into the basic model, outputs the first training result in the first fully connected layer, and inputs the second training result in the second fully connected layer.
- the output result of the first fully connected layer contains the correlation between the attributes.
- the first training result has three attributes: cotton clothes, gloves, and shorts. Among them, there is a correlation between cotton clothes and gloves, and cotton clothes and shorts are related. There is no correlation between them.
- the second training result has three attributes: cotton clothes, gloves and short skirts. The intersection of the first training result and the second training result can be taken as the final training result.
- the final training result contains the two attributes of cotton clothes and gloves and the correlation between cotton clothes and gloves.
- the preset range of the distance from the preset loss value is 1-10
- the first training result and the initial value of the image attribute can be directly input to the first loss function
- the second training result and the initial value of the image attribute can be input to the second loss function. Since the target loss function is based on the first loss function, The loss function and the second loss function are obtained, and the target loss value can be directly calculated through the loss function when the first training result and the second training result are obtained. Compare the target loss value with the preset loss value to judge whether the basic model converges. For example, when the target loss value is less than or equal to the preset loss value, it is considered that the target loss value reaches the preset condition, and the basic model is considered to have converged at this time.
- the number of trainings can be multiple times.
- the basic model is finally converged and the expected effect of the basic model training is achieved.
- the convergent basic model is used as the image attribute recognition model for image attribute recognition.
- the image attribute recognition model can be applied to electronic devices to identify the image attributes stored in the electronic device by the user according to the image attribute recognition model.
- the attribute recognition results obtain the correlation between the image attributes, which can also improve the accuracy of image attribute recognition.
- the network model training method obtains an image sample set, which includes multiple initial values of image attributes; constructs a basic model and a target loss function corresponding to the basic model, and the basic model includes convolutional neural Network model and recurrent neural network model; input the image sample set into the basic model for image attribute recognition to obtain the first training result obtained according to the recurrent neural network model and the second training result obtained according to the convolutional neural network model; According to the first training result, the second training result, the initial value of the image attribute and the target loss function, the convolutional neural network model and the recurrent neural network model are jointly trained until the basic model converges; the converged basic model is used as the recognition of the recognition image attribute model.
- the image attribute recognition model obtained in this way can improve the accuracy of image attribute recognition and can also identify the correlation between image attributes.
- FIG. 2 is a schematic diagram of the second process of the network model training method provided by an embodiment of the present application.
- the network model training method may include:
- the image sample set includes a variety of images, such as human body images, animal images, plant images, etc.
- images such as human body images, animal images, plant images, etc.
- multiple human images in the image sample set may be used as training images for the network model.
- multiple human body images are used as training images, and the attributes of each human body image and the degree of association between the attributes can be extracted.
- a human body image there are image attributes such as hats, glasses, tops, bottoms, shoes, etc.
- the correlations between different attributes are different. For example, there is no close correlation between wearing glasses and wearing tops. There is no close correlation between wearing shoes and wearing a hat.
- the corresponding relevance between the image attribute and the image attribute can be used as an initial value of the image attribute.
- the input layer is used to input the training image, and then a convolution layer is set, and the convolution layer is used to perform preliminary image feature extraction on the input training image to obtain the first feature value, and then set the first feature value.
- An eigenvalue is input into the next layer of the basic model structure.
- the recurrent neural network can process the first eigenvalue output by the convolutional layer and input the second eigenvalue.
- the other side of the recurrent neural network is connected to the first fully connected layer.
- the fully connected layer can be used as an output layer of the basic model to process the second feature value and output the first training result.
- the input layer, convolutional layer, pooling layer, and second fully connected layer are sequentially connected to form a convolutional neural network.
- the first eigenvalue output by the convolutional layer is processed by the pooling layer to obtain the third eigenvalue, and finally the second
- the fully connected layer processes the third feature value and outputs the second training result.
- the entire basic model can be seen as a combination of convolutional neural network model and recurrent neural network.
- the convolutional layer can be set to multiple layers.
- the convolutional layer includes conv3, conv6, conv9, etc.
- the input human body image is processed by the multi-layer convolutional layer to obtain a feature map in a certain dimension (Feature Map) , Use the feature map as the first feature value.
- the recurrent neural network can be a gated recurrent unit network, that is, a GRU neural network.
- first loss function and the second loss function may be the same type of loss function, or may be different types of loss functions, for example, both the first loss function and the second loss function may be cross-entropy loss functions.
- the output of the last layer of the convolutional neural network can be processed by the softmax algorithm. This step is usually to obtain the probability that the output belongs to a certain class. For a single sample, the output is a vector.
- the formula of softmax is:
- y 'i represents the actual value of the label of the i-th
- y i is the output vector softmax [Y1, Y2, Y3 ...] in the i-th element.
- the first loss function can be multiplied by the loss coefficient, and then the second loss function can be added to obtain the objective function corresponding to the basic model, where the loss coefficient can be a parameter obtained through experiments and can be set to 0.8 to 1. between.
- the setting of the target loss function also needs to be adjusted according to the training direction 2.
- Setting the loss coefficient before the first loss function to multiply the first loss function is to adjust the target function a method.
- the training image in the image sample set is input into the input layer, where the training image can be a human body image, and the attributes of the image are identified through the constructed basic model and the objective function corresponding to the basic model.
- the training image is first calculated by the convolutional layer to obtain the first feature value, and the first feature value output by the last layer of the convolutional layer is input to the pooling layer and the recurrent neural network, respectively, where the second feature output by the recurrent neural network The value is input to the first fully connected layer to obtain the first training result, and the third feature value input from the pooling layer is input to the second fully connected layer to obtain the second training result.
- first training result and the second training result are not completely the same, and the final training result obtained by the basic model is obtained based on the first training result and the second training result.
- the intersection of the first training result and the second training result can be taken, and the target training result in the intersection is the final training result obtained by the basic model.
- the final training result you can also choose to add the first training result and the second training result to get the final training result, and you can also select part of the training result from the first training result and the second training result according to a preset rule. The final training result.
- multiple first training results and second training results are obtained according to the basic model, and then the final training result is obtained according to the first training result and the second training result, and the final training result is obtained by And the initial value of the image attribute is input into the target loss function to get the target loss value.
- the first training result and the initial value of the image attribute can be directly input to the first loss function
- the second training result and the initial value of the image attribute can be input to the second loss function. Since the target loss function is based on the first loss function, The loss function and the second loss function are derived, and the target loss value can be directly calculated through the loss function when the first training result and the second training result are obtained. Compare the target loss value with the preset loss value to judge whether the basic model converges. For example, when the target loss value is less than or equal to the preset loss value, you are deemed to have reached the preset condition for the target loss value, and the basic model is considered to converge.
- the target loss value when the target loss value does not meet the preset condition, for example, the target loss value is not within the preset range, or the target loss value does not reach the preset loss value, it can be regarded as basic model training and If it is not completed, the training results output by the basic model cannot reach the expected results, so the model parameters of the basic model need to be adjusted.
- the parameters of the convolutional neural network model and the cyclic neural network model can be adjusted. Some model parameters are adjusted. Among them, the parameters of the model can be adjusted through the back-propagation algorithm.
- the convergent basic model is used as the image attribute recognition model for image attribute recognition.
- the image attribute recognition model can be applied to electronic devices to identify the image attributes stored in the electronic device by the user according to the image attribute recognition model.
- the attribute recognition results obtain the correlation between the image attributes, which can also improve the accuracy of image attribute recognition.
- inputting a random human body image into the convergent basic model can accurately identify the type of clothing worn in the human body image and the association between the clothes; or can accurately identify the features of the human body's five sense organs and the relationship between the five sense organs. It shows that the convergent basic model has been able to accurately identify the attributes of the input image and the correlation between the attributes, and the convergent basic model can be used as an image attribute recognition model.
- the network model training method obtains the image sample set and the initial value of the image attributes included in the image sample set, and then constructs a basic model based on the convolutional neural network and the recurrent neural network.
- Set the first loss function set the second loss function for the convolutional neural network, and obtain the target loss function according to the first loss function and the second loss function; input the image sample set into the basic model to train the basic model to obtain the first The training result and the second training result.
- the parameters of the basic model are adjusted according to the initial value of the image attribute, the target loss function, the first training result and the second training result, until the basic model converges, and the converged model is recognized as the image attribute
- the model is used to accurately identify the image attributes and the degree of association between each attribute.
- FIG. 3 is a schematic flowchart of an image attribute recognition method provided by an embodiment of the present application.
- the image attribute recognition method may include the following processes:
- the image attribute recognition request may be triggered by the electronic device receiving a touch operation, a voice operation, and receiving a start instruction of a corresponding target application.
- a variety of clothing or accessories can be input into the virtual human body image, where tops, bottoms, shoes, hats, earrings, necklaces, etc. can all be attributes of the human body image Information, the user can input these items into the virtual human body image and wear them in the corresponding position to obtain the new virtual human body image.
- the user can choose to recognize, and the electronic device receives the image attribute recognition request for the new virtual human body image. Human body image for recognition.
- the user when the user browsing interface contains multiple images, the user can click on a specific location on the electronic device or use a finger to divide the area to select the image to be recognized. At this time, the electronic device can select the location according to the user's selection. Or the region acquires the image to be recognized.
- the electronic device can actively obtain the image that the user needs to recognize according to the image recognition request. For example, when the user browses a picture, the electronic device can actively search for the image to be recognized according to the image recognition request. image.
- the image attribute recognition request contains the specific type of the target subject.
- the electronic device receives the image attribute recognition request, it can obtain the target subject in the image to be recognized. For example, how much is a group photo of the user? For personal objects, you can extract the subject of the person that needs to be identified; a landscape photo contains a variety of plants or animals, and you can extract the subject of the animal image that needs to be identified. In the process of identifying image attributes, you need to exclude non-recognition Object, keep the target subject.
- the image where the target subject is located can be cropped to obtain the target image.
- image attribute recognition it is possible to prevent subjects that do not need to be recognized from interfering with the recognition of the target subject. In the process, the recognition speed is faster and the recognition result is more accurate.
- the image attribute recognition model adopts the image attribute recognition model trained by the network model training method provided in this embodiment.
- the network model training method provided in this embodiment.
- the target image is input into the image attribute recognition model and then the image attribute recognition is performed to obtain the recognition results of multiple attributes in the image and the correlation between the attributes.
- the image attribute recognition model can recognize the correlation between human clothing, for example, the correlation between shorts and short sleeves is 100%, and the correlation between jeans and sports shoes is 80% , The correlation between hats and glasses is 50%, etc., so as to get the correlation between various wearables and clothes, and users can better refer to how to match clothes.
- the image attribute recognition method provided by the embodiment of the application receives the image attribute recognition request, obtains the image to be recognized according to the image attribute recognition request, and then calls the pre-trained image attribute recognition model to input the image to be recognized into the pre-training
- the image attribute recognition model is used to identify the image attributes of the image to be recognized to obtain the result of image attribute recognition, so as to obtain the correlation between the attributes of the image.
- the network model training device 400 may include: a first acquisition module 410, a construction module 420, a first recognition module 430, a training module 440, and a determination module 450.
- the first obtaining module 410 is configured to obtain an image sample set, the image sample set including a plurality of initial values of image attributes;
- the construction module 420 is configured to construct a basic model and a target loss function corresponding to the basic model, where the basic model includes a convolutional neural network model and a recurrent neural network model;
- the first recognition module 430 is configured to input the image sample set into the basic model for image attribute recognition, so as to obtain the first training result obtained according to the recurrent neural network model and according to the convolutional neural network The second training result obtained by the model;
- the training module 440 is configured to combine the convolutional neural network model and the recurrent neural network model according to the first training result, the second training result, the initial value of the image attribute, and the target loss function Training until the basic model converges;
- the determining module 450 is configured to use the converged basic model as a recognition model for recognizing image attributes.
- the construction module 420 includes a setting sub-module 421, a first connection sub-module 422, and a second connection sub-module 423.
- the setting sub-module 421 is used to set a convolutional layer, a pooling layer, and a first global A connection layer and a second fully connected layer; a first connection sub-module 422 for connecting the convolutional layer, the pooling layer, and the second fully connected layer to obtain the convolutional neural network model;
- the second connection sub-module 423 is configured to connect the recurrent neural network to the convolutional layer and to connect the recurrent neural network to the first fully connected layer to obtain the recurrent neural network model.
- the construction module 420 is specifically configured to construct a first loss function corresponding to the convolutional neural network model, and construct a second loss function corresponding to the recurrent neural network model, according to the first loss function and the second loss function Obtain the target loss function corresponding to the basic model.
- the second loss function may be multiplied by a loss coefficient to obtain a target second loss function, and the target second loss function and the first loss function may be added to obtain the target loss function.
- the training module 440 is specifically configured to input the first training result, the second training result, and the initial value of the image attribute into the target loss function to obtain a target loss value;
- the target loss value adjusts the parameters of the basic model.
- the determining module 450 is specifically configured to input the image sample set into the convolutional layer to obtain the first feature value; input the first feature value to the recurrent neural network to Obtain the second eigenvalue; input the second eigenvalue to the first fully connected layer to obtain the first training result.
- the first acquiring module 410 acquires an image sample set, the image sample set includes a plurality of image attribute initial values; the building module 420 builds a basic model and the target loss function corresponding to the basic model, the basic model includes a convolutional neural network model And the recurrent neural network model; the first recognition module 430 inputs the image sample set into the basic model for image attribute recognition, so as to obtain the first training result obtained according to the recurrent neural network model and the first training result obtained according to the convolutional neural network model The second training result; the training module 440 performs joint training on the convolutional neural network model and the recurrent neural network model according to the first training result, the second training result, the initial value of the image attribute and the target loss function, until the basic model converges; the determination module 450
- the convergent basic model is used as the recognition model for recognizing image attributes.
- the basic network model after training can improve the accuracy of image attribute recognition and identify the correlation between various image attributes.
- the network model training device provided in the embodiments of this application and the network model training method in the above embodiments belong to the same concept, and the network model training method can be run on the network model training device provided in the embodiment of the training method
- the network model training method can be run on the network model training device provided in the embodiment of the training method
- any method of refer to the embodiment of the training method of the network model for the specific implementation process, which will not be repeated here.
- the image attribute recognition device 500 may include: a receiving module 510, a second acquiring module 520, a calling module 530, and a second recognition module 540.
- the receiving module 510 is configured to receive an image attribute recognition request
- the second obtaining module 520 is configured to obtain the image to be recognized according to the image attribute recognition request
- the calling module 530 is used to call a pre-trained image attribute recognition model
- the second recognition module 540 is configured to input the image to be recognized into a pre-trained image attribute recognition model, and to recognize the image attributes of the image to be recognized to obtain an image attribute recognition result.
- the second obtaining module 520 is specifically configured to recognize the target subject in the image to be recognized according to the image attribute recognition request, and obtain the target image in the image to be recognized according to the target subject.
- the image attribute recognition device 500 receives the image attribute recognition request through the receiving module 510; the second obtaining module 520 obtains the image to be recognized according to the image attribute recognition request; the calling module 530 calls the pre-trained image attribute recognition Model; the second recognition module 540 inputs the image to be recognized into the pre-trained image attribute recognition model, and recognizes the image attributes of the image to be recognized to obtain the image attribute recognition result.
- the image attribute recognition device trained by the above network model training method can accurately recognize each attribute in the image and the correlation between each attribute, and improve the accuracy of image attribute recognition.
- the image attribute recognition device provided in this embodiment of the application belongs to the same concept as the image attribute recognition method in the above embodiment.
- the image attribute that can be run on the image attribute recognition device is any of the methods provided in the method embodiments.
- For the specific implementation process of the method please refer to the image processing method embodiment, which will not be repeated here.
- the embodiment of the present application provides a computer-readable storage medium on which a computer program is stored.
- the computer executes the network model training method or image provided in the embodiment of the present application. ⁇ Treatment methods.
- the storage medium may be a magnetic disk, an optical disc, a read only memory (Read Only Memory, ROM,), or a random access device (Random Access Memory, RAM), etc.
- An embodiment of the present application also provides an electronic device, including a memory, a processor, and a computer program stored in the memory.
- the processor is configured to execute the computer program stored in the memory by calling the computer program stored in the memory.
- Example provides the training method of the network model or the image attribute recognition method.
- the above-mentioned electronic device may be a mobile terminal such as a tablet computer or a smart phone.
- FIG. 6 is a schematic diagram of the first structure of an electronic device provided by an embodiment of this application.
- the electronic device 500 may include components such as a memory 601 and a processor 602. Those skilled in the art can understand that the structure of the electronic device shown in FIG. 7 does not constitute a limitation on the electronic device, and may include more or fewer components than those shown in the figure, or a combination of certain components, or different component arrangements.
- the memory 601 may be used to store software programs and modules.
- the processor 602 executes various functional applications and data processing by running the computer programs and modules stored in the memory 601.
- the memory 601 may mainly include a storage program area and a storage data area.
- the storage program area may store an operating system, a computer program required by at least one function (such as a sound playback function, an image playback function, etc.), etc.; Data created by the use of electronic equipment, etc.
- the processor 602 is the control center of the electronic device. It uses various interfaces and lines to connect the various parts of the entire electronic device, and executes the electronic device by running or executing the application program stored in the memory 601 and calling the data stored in the memory 601
- the various functions and processing data of the electronic equipment can be used to monitor the electronic equipment as a whole.
- the memory 601 may include a high-speed random access memory, and may also include a non-volatile memory, such as at least one magnetic disk storage device, a flash memory device, or other volatile solid-state storage devices.
- the memory 601 may further include a memory controller to provide the processor 602 with access to the memory 601.
- the processor 602 in the electronic device will load the executable code corresponding to the process of one or more application programs into the memory 601 according to the following instructions, and the processor 602 will run and store the executable code in the memory 601.
- the image sample set including a plurality of initial values of image attributes
- the basic model including a convolutional neural network model and a recurrent neural network model;
- the convergent basic model is used as a recognition model for recognizing image attributes.
- the processor 602 when the processor 602 executes the construction of the target loss function corresponding to the basic model, it may execute:
- the processor 602 when the processor 602 executes to obtain the target loss function corresponding to the basic model according to the first loss function and the second loss function, it may execute:
- the target loss function is obtained by adding the target second loss function and the first loss function.
- the processor 602 executes the calculation of the convolutional neural network model and the recurrent neural network according to the first training result, the second training result, the initial value of the image attribute, and the target loss function.
- the network model is jointly trained, until the basic model converges, you can execute:
- the parameters of the basic model are adjusted according to the target loss value.
- processor 602 when the processor 602 executes the construction of the basic model, it may execute:
- the recurrent neural network is connected to the convolutional layer, and the recurrent neural network is connected to the first fully connected layer to obtain the recurrent neural network model.
- the processor 602 when the processor 602 executes inputting the image sample set into the basic model for image attribute recognition to obtain the first training result obtained according to the recurrent neural network model, it may execute:
- the second feature value is input to the first fully connected layer to obtain the first training result.
- the processor 602 in the electronic device will load the executable code corresponding to the process of one or more application programs into the memory 601 according to the following instructions, and the processor 501 will run and store the executable code in the memory.
- the image to be recognized is input to a pre-trained image attribute recognition model, and the image attributes of the image to be recognized are recognized to obtain an image attribute recognition result.
- processor 602 when the processor 602 executes to obtain the image to be identified according to the image attribute identification request, it may execute:
- FIG. 7 is a schematic diagram of a second structure of an electronic device provided by an embodiment of the application.
- the electronic device further includes: a camera component 603, a radio frequency circuit 604, an audio circuit 605, and Power supply 606.
- the display 603, the radio frequency circuit 604, the audio circuit 605, and the power supply 606 are electrically connected to the processor 602, respectively.
- the display 603 can be used to display information input by the user or information provided to the user, and various graphical user interfaces. These graphical user interfaces can be composed of graphics, text, icons, videos, and any combination thereof.
- the display 603 may include a display panel.
- the display panel may be configured in the form of a liquid crystal display (LCD) or an organic light-emitting diode (OLED).
- LCD liquid crystal display
- OLED organic light-emitting diode
- the radio frequency circuit 604 may be used to transmit and receive radio frequency signals to establish wireless communication with network equipment or other electronic equipment through wireless communication, and to transmit and receive signals with the network equipment or other electronic equipment.
- the audio circuit 605 can be used to provide an audio interface between the user and the electronic device through a speaker or a microphone.
- the power supply 606 can be used to power various components of the electronic device 600.
- the power supply 606 may be logically connected to the processor 602 through a power management system, so that functions such as management of charging, discharging, and power consumption management can be realized through the power management system.
- the electronic device 600 may also include a camera component, a Bluetooth module, etc.
- the camera component may include an image processing circuit, which may be implemented by hardware and/or software components, and may include defining image signal processing (Image Signal Processing) various processing units of the pipeline.
- the image processing circuit may at least include: multiple cameras, an image signal processor (Image Signal Processor, ISP processor), a control logic, an image memory, a display, and the like.
- Each camera may include at least one or more lenses and image sensors.
- the image sensor may include a color filter array (such as a Bayer filter). The image sensor can obtain the light intensity and wavelength information captured by each imaging pixel of the image sensor, and provide a set of raw image data that can be processed by the image signal processor.
- the network model training method/image processing method and device provided in the embodiments of the present application belong to the same concept as the network model training method/image processing method in the above embodiments.
- the device can run any of the methods provided in the network model training method/image processing method embodiment.
- For the specific implementation process please refer to the network model training method/image processing method embodiment. I won't repeat it here.
- the computer program can be stored in a computer readable storage medium, such as in a memory, and executed by at least one processor. May include the flow of the embodiment of the training method of the network model/the image processing method.
- the storage medium may be a magnetic disk, an optical disc, a read only memory (ROM, Read Only Memory), a random access memory (RAM, Random Access Memory), etc.
- each functional module can be integrated in a processing chip, or each module can exist alone physically, or two or more The modules are integrated in one module.
- the above-mentioned integrated modules can be implemented in the form of hardware or software functional modules. If the integrated module is implemented in the form of a software function module and sold or used as an independent product, it can also be stored in a computer readable storage medium, such as a read-only memory, a magnetic disk or an optical disk, etc. .
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Data Mining & Analysis (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Bioinformatics & Computational Biology (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Life Sciences & Earth Sciences (AREA)
- Evolutionary Biology (AREA)
- Evolutionary Computation (AREA)
- Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Artificial Intelligence (AREA)
- Image Analysis (AREA)
Abstract
Description
Claims (20)
- 一种网络模型训练方法,其中,所述方法包括:获取图像样本集,所述图像样本集包括多个图像属性初始值;构建基础模型及所述基础模型对应的目标损失函数,所述基础模型包括卷积神经网络模型和循环神经网络模型;将所述图像样本集输入至所述基础模型之中进行图像属性识别,以获取根据所述循环神经网络模型得到的第一训练结果和根据所述卷积神经网络模型得到的第二训练结果;根据所述第一训练结果、所述第二训练结果、所述图像属性初始值和所述目标损失函数对所述卷积神经网络模型和所述循环神经网络模型进行联合训练,直至所述基础模型收敛;将收敛的所述基础模型作为识别图像属性的识别模型。
- 根据权利要求1所述的网络模型训练方法,其中,所述构建所述基础模型对应的目标损失函数,包括:构建所述卷积神经网络模型对应的第一损失函数;构建所述循环神经网络模型对应的第二损失函数;根据所述第一损失函数和所述第二损失函数得到所述基础模型对应的目标损失函数。
- 根据权利要求2所述的网络模型训练方法,其中,所述根据所述第一损失函数和所述第二损失函数得到所述基础模型对应的目标损失函数,包括:将所述第二损失函数乘以损失系数以得到目标第二损失函数;将所述目标第二损失函数与所述第一损失函数相加得到所述目标损失函数。
- 根据权利要求3所述的网络模型训练方法,其中,所述根据所述第一训练结果、所述第二训练结果、所述图像属性初始值和所述目标损失函数对所述卷积神经网络模型和所述循环神经网络模型进行联合训练,直至所述基础模型收敛,包括:将所述第一训练结果、所述第二训练结果和所述图像属性初始值输入至所述目标损失函数,以得到目标损失值;根据所述目标损失值对所述基础模型的参数进行调整。
- 根据权利要求1所述的图像属性识别方法,其中,所述构建基础模型,包括:设置卷积层、池化层、第一全连接层及第二全连接层;将所述卷积层、所述池化层及所述第二全连接层连接,以得到所述卷积神经网络模型;将所述循环神经网络与所述卷积层连接,以及将所述循环神经网络与所述第一全连接层连接,以得到所述循环神经网络模型。
- 根据权利要求5所述的网络模型,其中,所述将所述图像样本集输入至所述基础模型之中进行图像属性识别,以获取根据所述循环神经网络模型得到的第一训练结果,包括:将所述图像样本集输入至所述卷积层以得到所述第一特征值;将所述第一特征值输入至所述循环神经网络以得到所述第二特征值;将所述第二特征值输入至所述第一全连接层以得到所述第一训练结果。
- 一种图像属性识别方法,其中,所述方法包括:接收图像属性识别请求;根据所述图像属性识别请求获取待识别图像;调用预先训练的图像属性识别模型;将所述待识别图像输入至所述预先训练的图像属性识别模型,对所述待识别图像的图像属性进行识别,以得到图像属性识别结果;其中,所述图像属性识别模型为采用权利要求1至6任一项所述的所述网络模型的训练方法训练得到的图像属性识别模型。
- 根据权利要求7所述的图像属性识别方法,其中,所述根据所述图像属性识别请求获取待识别图像,包括:根据图像属性识别请求识别所述待识别图像中的目标主体;根据所述目标主体获取所述待识别图像中的目标图像。
- 一种网络模型的训练装置,其中,包括:第一获取模块,用于获取图像样本集,所述图像样本集包括多个图像属性初始值;构建模块,用于构建基础模型及所述基础模型对应的目标损失函数,所述基础模型包括卷积神经网络模型和循环神经网络模型;第一识别模块,用于将所述图像样本集输入至所述基础模型之中进行图像属性识别,以获取根据所述循环神经网络模型得到的第一训练结果和根据所述卷积神经网络模型得到的第二训练结果;训练模块,用于根据所述第一训练结果、所述第二训练结果、所述图像属性初始值和所述目标损失函数对所述卷积神经网络模型和所述循环神经网络模型进行联合训练,直至所述基础模型收敛;确定模块,用于将收敛的所述基础模型作为识别图像属性的识别模型。
- 根据权利要求9所述的训练装置,其中,所述构建模块包括:设置子模块,用于设置卷积层、池化层、第一全连接层及第二全连接层;第一连接子模块,用于将所述卷积层、所述池化层及所述第二全连接层连接,以得到所述卷积神经网络模型;第二连接子模块,用于将所述循环神经网络与所述卷积层连接,以及将所述循环神经网络与所述第一全连接层连接,以得到所述循环神经网络模型。
- 一种图像属性的识别装置,其中,包括:接收模块,用于接收图像属性识别请求;第二获取模块,用于根据所述图像属性识别请求获取待识别图像;调用模块,用于调用预先训练的图像属性识别模型;第二识别模块,用于将所述待识别图像输入至所述预先训练的图像属性识别模型,对所述待识别图像的图像属性进行识别,以得到图像属性识别结果;其中,所述图像属性识别模型为采用权利要求1至6任一项所述的网络模型训练方法得到的图像属性识别模型。
- 一种存储介质,其中,所述存储介质中存储有计算机程序,当所述计算机程序在计算机上运行时,使得所述计算机执行权利要求1至6任一项所述的网络模型训练方法或权利要求7、8所述的图像属性识别方法。
- 一种电子设备,其中,所述电子设备包括处理器和存储器,所述存储器中存储有计算机程序,所述处理器通过调用所述存储器中存储的所述计算机程序,用于执行:获取图像样本集,所述图像样本集包括多个图像属性初始值;构建基础模型及所述基础模型对应的目标损失函数,所述基础模型包括卷积神经网络模型和循环神经网络模型;将所述图像样本集输入至所述基础模型之中进行图像属性识别,以获取根据所述循环神经网络模型得到的第一训练结果和根据所述卷积神经网络模型得到的第二训练结果;根据所述第一训练结果、所述第二训练结果、所述图像属性初始值和所述目标损失函数对所述卷积神经网络模型和所述循环神经网络模型进行联合训练,直至所述基础模型收敛;将收敛的所述基础模型作为识别图像属性的识别模型。
- 根据权利要求13所述的电子设备,其中,所述处理器用于执行:构建所述卷积神经网络模型对应的第一损失函数;构建所述循环神经网络模型对应的第二损失函数;根据所述第一损失函数和所述第二损失函数得到所述基础模型对应的目标损失函数。
- 根据权利要求14所述的电子设备,其中,所述处理器用于执行:将所述第二损失函数乘以损失系数以得到目标第二损失函数;将所述目标第二损失函数与所述第一损失函数相加得到所述目标损失函数。
- 根据权利要求15所述的电子设备,其中,所述处理器用于执行:将所述第一训练结果、所述第二训练结果和所述图像属性初始值输入至所述目标损失函数,以得到目标损失值;根据所述目标损失值对所述基础模型的参数进行调整。
- 根据权利要求13所述的电子设备,其中,所述处理器用于执行:设置卷积层、池化层、第一全连接层及第二全连接层;将所述卷积层、所述池化层及所述第二全连接层连接,以得到所述卷积神经网络模型;将所述循环神经网络与所述卷积层连接,以及将所述循环神经网络与所述第一全连接 层连接,以得到所述循环神经网络模型,所述循环神经网络模型和所述卷积神经网络模型组合形成所述基础模型。
- 根据权利要求17所述的电子设备,其中,所述处理器用于执行:将所述图像样本集输入至所述卷积层以得到所述第一特征值;将所述第一特征值输入至所述循环神经网络以得到所述第二特征值;将所述第二特征值输入至所述第一全连接层以得到所述第一训练结果。
- 一种电子设备,其中,所述电子设备包括处理器和存储器,所述存储器中存储有计算机程序,所述处理器通过调用所述存储器中存储的所述计算机程序,用于执行:接收图像属性识别请求;根据所述图像属性识别请求获取待识别图像;调用预先训练的图像属性识别模型;将所述待识别图像输入至所述预先训练的图像属性识别模型,对所述待识别图像的图像属性进行识别,以得到图像属性识别结果;其中,所述图像属性识别模型为采用权利要求1至6任一项所述的所述网络模型的训练方法训练得到的图像属性识别模型。
- 根据权利要求19所述的电子设备,其中,所述处理器用于执行:根据图像属性识别请求识别所述待识别图像中的目标主体;根据所述目标主体获取所述待识别图像中的目标图像。
Priority Applications (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201980100863.7A CN114450690A (zh) | 2019-11-25 | 2019-11-25 | 网络模型训练方法、图像属性识别方法、装置及电子设备 |
PCT/CN2019/120749 WO2021102655A1 (zh) | 2019-11-25 | 2019-11-25 | 网络模型训练方法、图像属性识别方法、装置及电子设备 |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
PCT/CN2019/120749 WO2021102655A1 (zh) | 2019-11-25 | 2019-11-25 | 网络模型训练方法、图像属性识别方法、装置及电子设备 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2021102655A1 true WO2021102655A1 (zh) | 2021-06-03 |
Family
ID=76129757
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/CN2019/120749 WO2021102655A1 (zh) | 2019-11-25 | 2019-11-25 | 网络模型训练方法、图像属性识别方法、装置及电子设备 |
Country Status (2)
Country | Link |
---|---|
CN (1) | CN114450690A (zh) |
WO (1) | WO2021102655A1 (zh) |
Cited By (11)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN113505848A (zh) * | 2021-07-27 | 2021-10-15 | 京东科技控股股份有限公司 | 模型训练方法和装置 |
CN113570707A (zh) * | 2021-07-30 | 2021-10-29 | 集美大学 | 三维人体重建方法、装置、计算机设备及存储介质 |
CN113673498A (zh) * | 2021-07-28 | 2021-11-19 | 浙江大华技术股份有限公司 | 目标检测方法、装置、设备和计算机可读存储介质 |
CN113822199A (zh) * | 2021-09-23 | 2021-12-21 | 浙江大华技术股份有限公司 | 对象属性识别方法、装置、存储介质和电子装置 |
CN114092746A (zh) * | 2021-11-29 | 2022-02-25 | 北京易华录信息技术股份有限公司 | 一种多属性识别方法、装置、存储介质及电子设备 |
CN114358692A (zh) * | 2022-01-07 | 2022-04-15 | 拉扎斯网络科技(上海)有限公司 | 配送时长调整方法、装置及电子设备 |
CN114463559A (zh) * | 2022-01-29 | 2022-05-10 | 新疆爱华盈通信息技术有限公司 | 图像识别模型的训练方法、装置、网络和图像识别方法 |
CN114494055A (zh) * | 2022-01-19 | 2022-05-13 | 西安交通大学 | 基于循环神经网络的鬼成像方法、系统、设备及存储介质 |
CN114528977A (zh) * | 2022-01-24 | 2022-05-24 | 北京智源人工智能研究院 | 一种等变网络训练方法、装置、电子设备及存储介质 |
US20220172462A1 (en) * | 2020-02-13 | 2022-06-02 | Tencent Technology (Shenzhen) Company Limited | Image processing method, apparatus, and device, and storage medium |
CN116698408A (zh) * | 2023-06-09 | 2023-09-05 | 西安理工大学 | 轴承润滑状态检测方法、装置、电子设备及存储介质 |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN108596090A (zh) * | 2018-04-24 | 2018-09-28 | 北京达佳互联信息技术有限公司 | 人脸图像关键点检测方法、装置、计算机设备及存储介质 |
CN109522942A (zh) * | 2018-10-29 | 2019-03-26 | 中国科学院深圳先进技术研究院 | 一种图像分类方法、装置、终端设备和存储介质 |
CN109871444A (zh) * | 2019-01-16 | 2019-06-11 | 北京邮电大学 | 一种文本分类方法及系统 |
CN110298266A (zh) * | 2019-06-10 | 2019-10-01 | 天津大学 | 基于多尺度感受野特征融合的深度神经网络目标检测方法 |
-
2019
- 2019-11-25 CN CN201980100863.7A patent/CN114450690A/zh active Pending
- 2019-11-25 WO PCT/CN2019/120749 patent/WO2021102655A1/zh active Application Filing
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN108596090A (zh) * | 2018-04-24 | 2018-09-28 | 北京达佳互联信息技术有限公司 | 人脸图像关键点检测方法、装置、计算机设备及存储介质 |
CN109522942A (zh) * | 2018-10-29 | 2019-03-26 | 中国科学院深圳先进技术研究院 | 一种图像分类方法、装置、终端设备和存储介质 |
CN109871444A (zh) * | 2019-01-16 | 2019-06-11 | 北京邮电大学 | 一种文本分类方法及系统 |
CN110298266A (zh) * | 2019-06-10 | 2019-10-01 | 天津大学 | 基于多尺度感受野特征融合的深度神经网络目标检测方法 |
Cited By (15)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20220172462A1 (en) * | 2020-02-13 | 2022-06-02 | Tencent Technology (Shenzhen) Company Limited | Image processing method, apparatus, and device, and storage medium |
US12033374B2 (en) * | 2020-02-13 | 2024-07-09 | Tencent Technology (Shenzhen) Company Limited | Image processing method, apparatus, and device, and storage medium |
CN113505848A (zh) * | 2021-07-27 | 2021-10-15 | 京东科技控股股份有限公司 | 模型训练方法和装置 |
CN113505848B (zh) * | 2021-07-27 | 2023-09-26 | 京东科技控股股份有限公司 | 模型训练方法和装置 |
CN113673498A (zh) * | 2021-07-28 | 2021-11-19 | 浙江大华技术股份有限公司 | 目标检测方法、装置、设备和计算机可读存储介质 |
CN113570707A (zh) * | 2021-07-30 | 2021-10-29 | 集美大学 | 三维人体重建方法、装置、计算机设备及存储介质 |
CN113822199A (zh) * | 2021-09-23 | 2021-12-21 | 浙江大华技术股份有限公司 | 对象属性识别方法、装置、存储介质和电子装置 |
CN114092746A (zh) * | 2021-11-29 | 2022-02-25 | 北京易华录信息技术股份有限公司 | 一种多属性识别方法、装置、存储介质及电子设备 |
CN114358692A (zh) * | 2022-01-07 | 2022-04-15 | 拉扎斯网络科技(上海)有限公司 | 配送时长调整方法、装置及电子设备 |
CN114494055A (zh) * | 2022-01-19 | 2022-05-13 | 西安交通大学 | 基于循环神经网络的鬼成像方法、系统、设备及存储介质 |
CN114494055B (zh) * | 2022-01-19 | 2024-02-06 | 西安交通大学 | 基于循环神经网络的鬼成像方法、系统、设备及存储介质 |
CN114528977A (zh) * | 2022-01-24 | 2022-05-24 | 北京智源人工智能研究院 | 一种等变网络训练方法、装置、电子设备及存储介质 |
CN114463559A (zh) * | 2022-01-29 | 2022-05-10 | 新疆爱华盈通信息技术有限公司 | 图像识别模型的训练方法、装置、网络和图像识别方法 |
CN114463559B (zh) * | 2022-01-29 | 2024-05-10 | 芯算一体(深圳)科技有限公司 | 图像识别模型的训练方法、装置、网络和图像识别方法 |
CN116698408A (zh) * | 2023-06-09 | 2023-09-05 | 西安理工大学 | 轴承润滑状态检测方法、装置、电子设备及存储介质 |
Also Published As
Publication number | Publication date |
---|---|
CN114450690A (zh) | 2022-05-06 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2021102655A1 (zh) | 网络模型训练方法、图像属性识别方法、装置及电子设备 | |
CN109902548B (zh) | 一种对象属性识别方法、装置、计算设备及系统 | |
WO2019174439A1 (zh) | 图像识别方法、装置、终端和存储介质 | |
KR102425578B1 (ko) | 객체를 인식하는 방법 및 장치 | |
US9875258B1 (en) | Generating search strings and refinements from an image | |
WO2021227726A1 (zh) | 面部检测、图像检测神经网络训练方法、装置和设备 | |
WO2022016556A1 (zh) | 一种神经网络蒸馏方法以及装置 | |
US20180088677A1 (en) | Performing operations based on gestures | |
CN110009052A (zh) | 一种图像识别的方法、图像识别模型训练的方法及装置 | |
KR20230141688A (ko) | 디바이스가 이미지를 보정하는 방법 및 그 디바이스 | |
US10346893B1 (en) | Virtual dressing room | |
WO2020078119A1 (zh) | 模拟用户穿戴服装饰品的方法、装置和系统 | |
CN107679447A (zh) | 面部特征点检测方法、装置及存储介质 | |
KR102056806B1 (ko) | 영상 통화 서비스를 제공하는 단말과 서버 | |
Zhang et al. | Facial smile detection based on deep learning features | |
WO2021092808A1 (zh) | 网络模型的训练方法、图像的处理方法、装置及电子设备 | |
CN113449573A (zh) | 一种动态手势识别方法及设备 | |
WO2019231130A1 (ko) | 전자 장치 및 그의 제어방법 | |
WO2021047587A1 (zh) | 手势识别方法、电子设备、计算机可读存储介质和芯片 | |
WO2019120031A1 (zh) | 服饰搭配推荐方法、装置、存储介质及移动终端 | |
CN112257645B (zh) | 人脸的关键点定位方法和装置、存储介质及电子装置 | |
KR102637342B1 (ko) | 대상 객체를 추적하는 방법과 장치 및 전자 장치 | |
CN111104911A (zh) | 一种基于大数据训练的行人重识别方法及装置 | |
CN115223239A (zh) | 一种手势识别方法、系统、计算机设备以及可读存储介质 | |
KR102476619B1 (ko) | 전자 장치 및 이의 제어 방법 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 19953924 Country of ref document: EP Kind code of ref document: A1 |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 19953924 Country of ref document: EP Kind code of ref document: A1 |
|
32PN | Ep: public notification in the ep bulletin as address of the adressee cannot be established |
Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205A DATED 19.07.2022) |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 19953924 Country of ref document: EP Kind code of ref document: A1 |