WO2020155518A1 - Object detection method and device, computer device and storage medium - Google Patents

Object detection method and device, computer device and storage medium Download PDF

Info

Publication number
WO2020155518A1
WO2020155518A1 PCT/CN2019/091100 CN2019091100W WO2020155518A1 WO 2020155518 A1 WO2020155518 A1 WO 2020155518A1 CN 2019091100 W CN2019091100 W CN 2019091100W WO 2020155518 A1 WO2020155518 A1 WO 2020155518A1
Authority
WO
WIPO (PCT)
Prior art keywords
object detection
loss
module
model
training
Prior art date
Application number
PCT/CN2019/091100
Other languages
French (fr)
Chinese (zh)
Inventor
巢中迪
庄伯金
王少军
Original Assignee
平安科技(深圳)有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 平安科技(深圳)有限公司 filed Critical 平安科技(深圳)有限公司
Publication of WO2020155518A1 publication Critical patent/WO2020155518A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • G06N3/084Backpropagation, e.g. using gradient descent
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/10Terrestrial scenes
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V2201/00Indexing scheme relating to image or video recognition or understanding
    • G06V2201/07Target detection

Definitions

  • This application relates to the field of artificial intelligence, and in particular to an object detection method, device, computer equipment and storage medium.
  • Object detection is one of the classic problems in computer vision. Its task is to use a box to mark the position of the object in the image and give the object category. From the traditional artificially designed feature plus shallow classifier framework to the end-to-end detection framework based on deep learning, object detection is improving step by step. However, currently commonly used object detection methods such as YOLO (You Only Look Once) detection Methods, SSD (Single Shot Multi-Box Detection) and other detection methods still generally have the problem of low object detection accuracy.
  • YOLO You Only Look Once
  • SSD Single Shot Multi-Box Detection
  • the embodiments of the present application provide an object detection method, device, computer equipment, and storage medium to solve the problem that the object detection accuracy rate is still low.
  • an object detection method including:
  • the object detection model includes a detection module, a classification module, and a discrimination module
  • the object detection model is updated according to the detection loss, the classification loss, and the discrimination loss to obtain a target object detection model.
  • an object detection model training device including:
  • the image acquisition module to be detected is used to acquire the image to be detected
  • the object detection result acquisition module is used to input the to-be-detected image into a target object detection model for object detection to obtain the object detection result of the to-be-detected image, wherein the target object detection model adopts a training sample acquisition module,
  • the model training module, the loss acquisition module and the target object detection model acquisition module obtain:
  • the training sample acquisition module is used to acquire training samples
  • a model training module is used to input the training samples into an object detection model for model training, where the object detection model includes a detection module, a classification module, and a discrimination module;
  • a loss acquisition module configured to acquire the detection loss generated by the detection module, the classification loss generated by the classification module, and the discrimination loss generated by the discrimination module during the model training process;
  • the target object detection model acquisition module is used to update the object detection model according to the detection loss, the classification loss and the discrimination loss to obtain a target object detection model.
  • a computer device in a third aspect, includes a memory, a processor, and computer-readable instructions that are stored in the memory and can run on the processor.
  • the processor executes the computer-readable instructions, the foregoing The steps of the object detection method.
  • an embodiment of the present application provides a computer non-volatile readable storage medium, including: computer readable instructions, which implement the steps of the above object detection method when the computer readable instructions are executed by a processor.
  • the image to be detected is first obtained; then the image to be detected is input into the target object detection model for object detection, and the object detection result of the image to be detected is obtained ,
  • the target object detection model combines the detection loss, classification loss and discrimination loss to update the object detection model, which has better detection and classification effects, and can obtain detection results with higher accuracy.
  • FIG. 1 is a flowchart of an object-based detection method in an embodiment of the present application
  • Figure 2 is a schematic diagram of an object-based detection device in an embodiment of the present application.
  • Fig. 3 is a schematic diagram of a computer device in an embodiment of the present application.
  • first, second, third, etc. may be used in the embodiments of the present application to describe the preset range, etc., these preset ranges should not be limited to these terms. These terms are only used to distinguish the preset ranges from each other.
  • the first preset range may also be referred to as the second preset range, and similarly, the second preset range may also be referred to as the first preset range.
  • the word “if” as used herein can be interpreted as “when” or “when” or “in response to determination” or “in response to detection”.
  • the phrase “if determined” or “if detected (statement or event)” can be interpreted as “when determined” or “in response to determination” or “when detected (statement or event) )” or “in response to detection (statement or event)”.
  • Fig. 1 shows a flow chart of the object detection method in this embodiment.
  • the object detection method can be applied to an object detection system, and the object detection system can be used to realize the detection and classification of objects, and the object detection system can be specifically applied to computer equipment.
  • the computer device is a device that can perform human-computer interaction with the user, including but not limited to devices such as computers, smart phones, and tablets.
  • the object detection method includes the following steps:
  • step S2 Input the image to be detected into the target object detection model for object detection, and obtain the object detection result of the image to be detected.
  • the model training steps adopted by the target object detection model specifically include:
  • training samples required for model training are obtained.
  • images related to a certain type of scene can be selected as training samples according to the needs of object detection.
  • the images saved in the driving recorder can be used as training samples.
  • the pictures saved in the driving recorder can reflect the road conditions ahead during the driving of the vehicle.
  • the image can be used as a training sample to train the target object detection model, so that the trained target object detection model can be The object is detected, so that the vehicle makes a preset response according to the received detection result. Understandably, it is necessary to pre-label the objects appearing in the image saved in the driving recorder before the model training (label the objects that need to be detected, and the objects that are not required for detection may not be labeled).
  • deep neural A network such as a convolutional neural network extracts deep features of images belonging to the same category as the object to be annotated to identify the category of the object when the object detection model (including the corresponding deep neural network for extracting image features) detects.
  • S20 Input training samples into the object detection model for model training, where the object detection model includes a detection module, a classification module, and a discrimination module.
  • model training refers to the training of the target object detection model.
  • the detection module is used to detect objects in the image, and the classification module is used to identify and classify the detected objects.
  • the judgment module includes a first judgment module and/or a second judgment module.
  • the first judgment module is used to judge the output of the detection module. Whether the result is correct or not, the second judgment module is used to judge whether the output result of the classification module is correct.
  • the first judgment module and the second judgment module can exist at the same time, or only the second judgment module exists, and the second judgment module is used as the judgment Module.
  • the training samples are input into the object detection model for model training, where the object detection model includes not only a detection model and a classification model, but also a discriminant model. Understandably, model training with training samples is the process of inputting training samples into the object detection model for detection.
  • step S20 it further includes:
  • S211 Obtain a detection model of the object to be processed, which includes a detection module and a classification module.
  • the detection model of the object to be processed is obtained. It is understandable that the detection model of the object to be processed may specifically be a detection model such as YOLO (You Only Look Once) detection model and SSD (Single Shot Multi-Box Detection). These models include detection modules and classification modules. This embodiment is an improvement based on these object detection models to be processed.
  • S212 Add a discrimination module to the detection model of the object to be processed, where the discrimination module is used to discriminate the results output by the detection module and/or the classification module.
  • a discrimination module is added on the original basis of the object detection model to be processed, so as to determine the output result of the object detection model to be processed. Adding a discrimination module can help to know the accuracy of the detection model of the object to be processed, so as to update the model according to the detection error of the object detection model to improve the accuracy of detection.
  • S213 Perform model initialization operation on the object detection model to be processed after adding the discrimination module to obtain the object detection model.
  • the initialization operation of the model refers to the initialization of the network parameters in the model, and the initial values of the network parameters may be preset based on experience.
  • the network parameters in the detection module and the classification module in the object detection model to be processed have actually been updated through multiple trainings.
  • the discrimination module is then based on the detection module and/or classification
  • the result of the module output will be discriminated and updated, because the detection module and the classification module have been learning for a long time at the beginning of the training. It is more thorough if the discrimination module is used for updating in a short time; on the contrary, the initialization After the operation, the discrimination module will make a judgment every time the detection module and/or classification module outputs a result during the training phase, and can update the network parameters according to the output result in time with the training process, so as to achieve better detection accuracy.
  • steps S211-S213 an implementation method for obtaining an object detection model is provided. Specifically, a discrimination module is added to the object detection model to be processed, and the initialization operation of the model is performed, which is beneficial to improve the subsequent training and update of the object detection model. The detection accuracy of the target object detection model.
  • step S20 inputting the training samples into the object detection model for model training includes:
  • S221 Input the training sample, and extract the feature vector of the training sample through the object detection model.
  • the object detection model includes a deep neural network for extracting feature vectors of training samples, which may specifically be a convolutional neural network.
  • the object detection model will use the deep neural network to extract the feature vectors of the training samples to provide a technical basis for model training.
  • the feature value in the feature vector specifically refers to the pixel value.
  • the feature vector is normalized, that is, the feature value in the feature vector is normalized to the interval of [0,1].
  • the pixel values of the image are 28, 212 and 216 the pixel value level, etc., a large number of different images may be contained in the pixel value, which makes the computational efficiency is low, so the use of the normalized manner
  • the eigenvalues in the eigenvectors are compressed in the same range, so that the calculation efficiency is improved and the model training time is shorter.
  • S223 Perform model training on the object detection model according to the normalized feature vector.
  • steps S221-S223 an implementation method for inputting training samples into the object detection model for model training is provided, and the extracted training sample features are normalized, and the feature values in the feature vector are compressed in the same range Within the interval, training time can be significantly shortened and training efficiency improved.
  • the detection module, the classification module, and the discrimination module respectively represent a function.
  • the detection loss generated by the detection module the classification
  • the classification loss generated by the module and the discrimination loss generated by the discrimination module can be used for reference to help adjust the object detection model, so that the target object detection model can be used as much as possible when the detection module, classification module and discrimination module are used again to achieve functions. Errors to improve the accuracy of the target object detection model detection.
  • step S30 the detection loss generated by the detection module, the classification loss generated by the classification module, and the discrimination loss generated by the discrimination module are obtained during the model training process, which specifically include:
  • the first training feature vector is the result output by the detection module
  • the first label vector is a feature vector used to verify whether the first training feature vector is correct, and represents the real result.
  • a preset detection loss function is used to calculate the loss between the first training feature vector and the pre-stored first label vector to obtain the detection loss, so as to update the network parameters of the model according to the detection loss.
  • the detection loss function may include a loss function for the predicted center coordinates, which is expressed as: Among them, ⁇ represents the adjustment factor, which is a preset parameter value, i represents the grid unit divided during detection, I represents the total number of grid units, j represents the predicted value of the bounding box, and J represents the total number of predicted values of the bounding box.
  • object detection models such as yolo need to perform image segmentation on the input training samples to obtain I grid units.
  • J prediction bounding boxes are obtained.
  • the obj stands for object, which means to detect the object.
  • the detection loss function may also include a loss function about the width and height of the predicted bounding box, expressed as: among them, Represents the square root of the predicted width and the square root of the predicted height, Represents the true value of the square root of the width and the square root of the height output by the training sample (other repeated parameters will not be explained, so as not to repeat them).
  • the above provides two aspects of the center coordinates predicted by the model and the width and height of the predicted bounding box to measure the loss during detection, where the first training feature vector output by the detection module specifically includes (x i , y i ) and The first label vector specifically includes with Through the detection loss function, the network parameters of the object detection model can be updated more accurately.
  • the second training feature vector is the result output by the classification module
  • the second label vector is a feature vector used to verify whether the second training feature vector is correct, and represents the real result.
  • a preset classification loss function is used to calculate the loss between the second training feature vector and the pre-stored second label vector to obtain the classification loss, so as to update the network parameters of the model according to the classification loss.
  • the classification loss function can be expressed as: Among them, i represents the grid unit divided during detection, I represents the total number of grid units, Indicates that when there is a target in the i-th grid cell, Take 1, otherwise Take 0, p i represents the predicted classification, Represents the true situation of the classification output by the training sample.
  • the second training feature vector output by the classification module specifically includes p i
  • the second label vector specifically includes Through the classification loss function, the network parameters of the object classification model can be updated more accurately.
  • S33 In the model training process, obtain the third training feature vector output by the discrimination module, and calculate the discrimination loss by using the preset discriminant loss function according to the third training feature vector.
  • the third training feature vector is the result output by the discrimination module.
  • a preset discriminant loss function is used to calculate the discriminant loss, so as to update the network parameters of the model according to the discriminant loss.
  • the discrimination loss function can be specifically expressed as: Wherein, I represents the total number of grid cells, i denotes grid cells obtained divided detection, classification D (p i) denotes the prediction result of the discrimination output of the module.
  • the discriminant loss function can reflect the loss generated by the discriminant module during training, so as to more accurately update the network parameters of the object discriminant model.
  • steps S31-S33 take the loss of one training sample as an example when calculating the detection loss, classification loss, and discrimination loss.
  • the values of each training sample will be The detection loss, classification loss, and discrimination loss arithmetic are added to obtain the total detection loss, classification loss, and discrimination loss, and the model is updated according to the total detection loss, classification loss and discrimination loss.
  • Steps S31-S33 provide specific implementations for obtaining detection loss, classification loss, and discrimination loss.
  • the obtained detection loss, classification loss, and discrimination loss can accurately describe the loss generated during the training process, so that the model can be updated more accurately. accurate.
  • S40 Update the object detection model according to the detection loss, classification loss and discrimination loss to obtain the target object detection model.
  • step S40 specifically includes:
  • the back-propagation algorithm is a learning algorithm suitable for multi-layer neural networks under the guidance of a tutor. It is based on the gradient descent method.
  • updating the object detection model using a back propagation algorithm can speed up the update and improve the training efficiency of model training.
  • the total loss of detection loss, classification loss and discrimination loss is large, the use of backpropagation algorithm has better results.
  • the update process can be stopped, the training ends, and the detection accuracy rate is higher. High target object detection model.
  • Steps S41-S42 provide an implementation manner for updating the object detection model, which can quickly complete the update process and obtain a target object detection model with higher detection accuracy.
  • the image to be detected is first acquired, and then the image to be detected is input into the target object detection model for object detection, and the object detection result of the image to be detected is obtained, wherein the target object detection model combines The detection loss, classification loss and discrimination loss jointly update the object detection model, so that the target object detection model obtained by training has better detection and classification effects.
  • the embodiments of the present application further provide device embodiments that implement the steps and methods in the foregoing method embodiments.
  • Fig. 2 shows a principle block diagram of an object detection device corresponding to the object detection method in the embodiment one to one.
  • the object detection device includes a to-be-detected image acquisition module 10, an object detection result acquisition module 20, and also includes a training sample acquisition module 30, a model training module 40, a loss acquisition module 50, and a target object detection model acquisition module 60 .
  • the realization functions of the to-be-detected image acquisition module 10, the object detection result acquisition module 20, the training sample acquisition module 30, the model training module 40, the loss acquisition module 50, and the target object detection model acquisition module 60 correspond to the object detection method in the embodiment
  • the steps of are one-to-one correspondence, in order to avoid redundant description, this embodiment will not describe them one by one.
  • the to-be-detected image acquisition module 10 is used to acquire the to-be-detected image.
  • the object detection result acquisition module 20 is used to input the image to be detected into the target object detection model for object detection, and obtain the object detection result of the image to be detected.
  • the training sample acquisition module 30 is used to acquire training samples.
  • the model training module 40 is used to input training samples into the object detection model for model training, where the object detection model includes a detection module, a classification module and a discrimination module.
  • the loss acquisition module 50 is used to obtain the detection loss generated by the detection module, the classification loss generated by the classification module, and the discrimination loss generated by the discrimination module during the model training process.
  • the target object detection model acquisition module 60 is used to update the object detection model according to the detection loss, classification loss and discrimination loss to obtain the target object detection model.
  • the object detection device further includes a detection model acquisition unit of the object to be processed, a discrimination module adding unit and an initialization unit.
  • the to-be-processed object detection model acquisition unit is used to acquire the to-be-processed object detection model.
  • the to-be-processed object detection model includes a detection module and a classification module.
  • the discrimination module adding unit is used to add a discrimination module to the object detection model to be processed, wherein the discrimination module is used to discriminate the results output by the detection module and/or the classification module.
  • the initialization unit is used to initialize the object detection model to be processed after adding the discrimination module to obtain the object detection model.
  • the model training module 40 includes a feature vector extraction unit, a normalized feature vector acquisition unit, and a model training unit.
  • the feature vector extraction unit is used to input training samples, and extract feature vectors of the training samples through the object detection model.
  • the model training unit is used to perform model training on the object detection model according to the normalized feature vector.
  • the loss acquisition module 50 includes a detection loss acquisition unit, a classification loss acquisition unit, and a discrimination loss acquisition unit.
  • the detection loss acquisition unit is used to obtain the first training feature vector output by the detection module during the model training process, and calculate the loss between the first training feature vector and the pre-stored first label vector using a preset detection loss function, Get detection loss.
  • the classification loss acquisition unit is used to obtain the second training feature vector output by the classification module during the model training process, and calculate the loss between the second training feature vector and the pre-stored second label vector by using a preset classification loss function, Get classification loss.
  • the discrimination loss acquisition unit is used to obtain the third training feature vector output by the discrimination module during the model training process, and calculate the discrimination loss by using the preset discriminant loss function according to the third training feature vector.
  • the target object detection model acquisition module 60 includes a network parameter update unit and a target object detection model acquisition unit.
  • the network parameter update unit is used to update the network parameters in the object detection model by using the back propagation algorithm according to the detection loss, classification loss and discrimination loss.
  • the target object detection model acquisition unit is used to stop updating the network parameters when the change values of the network parameters are less than the iterative stop threshold to obtain the target object detection model.
  • the image to be detected is first acquired, and then the image to be detected is input into the target object detection model for object detection, and the object detection result of the image to be detected is obtained, wherein the target object detection model combines The detection loss, classification loss and discrimination loss jointly update the object detection model, so that the target object detection model obtained by training has better detection and classification effects.
  • This embodiment provides a computer non-volatile readable storage medium.
  • the computer non-volatile readable storage medium stores computer readable instructions.
  • the object detection method in the embodiment is implemented. To avoid repetition, I won’t repeat them here.
  • the computer-readable instructions realize the functions of the modules/units in the object detection device in the embodiment when they are executed by the processor. To avoid repetition, details are not repeated here.
  • Fig. 3 is a schematic diagram of a computer device provided by an embodiment of the present application.
  • the computer device 70 of this embodiment includes: a processor 71, a memory 72, and computer-readable instructions 73 stored in the memory 72 and running on the processor 71, and the computer-readable instructions 73 are processed
  • the object detection method in the embodiment is implemented when the device 71 is executed. To avoid repetition, it will not be repeated here.
  • the computer-readable instruction 73 is executed by the processor 71, the function of each model/unit in the object detection device in the embodiment is realized. In order to avoid repetition, it will not be repeated here.
  • the computer device 70 may be a computing device such as a desktop computer, a notebook, a palmtop computer, and a cloud server.
  • the computer device 70 may include, but is not limited to, a processor 71 and a memory 72.
  • FIG. 3 is only an example of the computer device 70, and does not constitute a limitation on the computer device 70. It may include more or less components than those shown in the figure, or a combination of certain components, or different components.
  • computer equipment may also include input and output devices, network access devices, buses, and so on.
  • the so-called processor 71 may be a central processing unit (Central Processing Unit, CPU), other general-purpose processors, digital signal processors (Digital Signal Processor, DSP), application specific integrated circuits (Application Specific Integrated Circuit, ASIC), Field-Programmable Gate Array (FPGA) or other programmable logic devices, discrete gates or transistor logic devices, discrete hardware components, etc.
  • the general-purpose processor may be a microprocessor or the processor may also be any conventional processor or the like.
  • the memory 72 may be an internal storage unit of the computer device 70, such as a hard disk or memory of the computer device 70.
  • the memory 72 may also be an external storage device of the computer device 70, such as a plug-in hard disk equipped on the computer device 70, a smart memory card (Smart Media Card, SMC), a Secure Digital (SD) card, and a flash memory card (Flash). Card) and so on.
  • the memory 72 may also include both an internal storage unit of the computer device 70 and an external storage device.
  • the memory 72 is used to store computer readable instructions and other programs and data required by the computer equipment.
  • the memory 72 can also be used to temporarily store data that has been output or will be output.

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • General Health & Medical Sciences (AREA)
  • Molecular Biology (AREA)
  • Biophysics (AREA)
  • Computational Linguistics (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Computation (AREA)
  • Artificial Intelligence (AREA)
  • Biomedical Technology (AREA)
  • Computing Systems (AREA)
  • General Engineering & Computer Science (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Health & Medical Sciences (AREA)
  • Multimedia (AREA)
  • Image Analysis (AREA)

Abstract

An object detection method and device, a computer device and a storage medium, which relate to the field of artificial intelligence. The object detection method comprises: acquiring an image to be detected (S1); and inputting the image to be detected into a target object detection model and carrying out object detection, so as to obtain an object detection result of the image to be detected, wherein model training steps used by the target object detection model comprise: acquiring a training sample; inputting the training sample into the object detection model and carrying out model training, the object detection model comprising a detection module, a classification module and a discrimination module; obtaining a detection loss generated by the detection module, a classification loss generated by the classification module and a discrimination loss generated by the discrimination module in the model training process; and updating the object detection model according to the detection loss, the classification loss and the discrimination loss, so as to obtain the target object detection model (S2). The accuracy of object detection may be effectively improved by using the described object detection method.

Description

物体检测方法、装置、计算机设备及存储介质Object detection method, device, computer equipment and storage medium
本申请以2019年2月3日提交的申请号为201910108522.6,名称为“物体检测模型训练方法、装置、计算机设备及存储介质”的中国发明专利申请为基础,并要求其优先权。This application is based on the Chinese invention patent application filed on February 3, 2019 with the application number 201910108522.6, titled "Object Detection Model Training Method, Device, Computer Equipment and Storage Medium", and claims its priority.
【技术领域】【Technical Field】
本申请涉及人工智能领域,尤其涉及一种物体检测方法、装置、计算机设备及存储介质。This application relates to the field of artificial intelligence, and in particular to an object detection method, device, computer equipment and storage medium.
【背景技术】【Background technique】
物体检测是计算机视觉中的经典问题之一,其任务是用框去标出图像中物体的位置,并给出物体的类别。从传统的人工设计特征加浅层分类器的框架,到基于深度学习的端到端的检测框架,物体检测在一步步地改进,但是,目前常用的物体检测方法如YOLO(You Only Look Once)检测方法、SSD(Single Shot Multi-Box Detection)等检测方法仍普遍存在物体检测准确率较低的问题。Object detection is one of the classic problems in computer vision. Its task is to use a box to mark the position of the object in the image and give the object category. From the traditional artificially designed feature plus shallow classifier framework to the end-to-end detection framework based on deep learning, object detection is improving step by step. However, currently commonly used object detection methods such as YOLO (You Only Look Once) detection Methods, SSD (Single Shot Multi-Box Detection) and other detection methods still generally have the problem of low object detection accuracy.
【发明内容】[Content of the invention]
有鉴于此,本申请实施例提供了一种物体检测方法、装置、计算机设备及存储介质,用以解决目前仍普遍存在物体检测准确率较低的问题。In view of this, the embodiments of the present application provide an object detection method, device, computer equipment, and storage medium to solve the problem that the object detection accuracy rate is still low.
第一方面,本申请实施例提供了一种物体检测方法,包括:In the first aspect, an embodiment of the present application provides an object detection method, including:
获取待检测图像;Obtain the image to be detected;
将所述待检测图像输入到目标物体检测模型中进行物体检测,得到所述待检测图像的物体检测结果,其中,所述目标物体检测模型采用的模型训练步骤包括:Input the to-be-detected image into a target object detection model for object detection to obtain an object detection result of the to-be-detected image, wherein the model training step adopted by the target object detection model includes:
获取训练样本;Obtain training samples;
将所述训练样本输入到物体检测模型中进行模型训练,其中,所述物体检测模型包括检测模块、分类模块和判别模块;Inputting the training samples into an object detection model for model training, where the object detection model includes a detection module, a classification module, and a discrimination module;
在模型训练过程中得到由所述检测模块产生的检测损失、由所述分类模块产生的分类损失 和由所述判别模块产生的判别损失;Obtaining the detection loss generated by the detection module, the classification loss generated by the classification module, and the discrimination loss generated by the discrimination module in the model training process;
根据所述检测损失、所述分类损失和所述判别损失更新所述物体检测模型,得到目标物体检测模型。The object detection model is updated according to the detection loss, the classification loss, and the discrimination loss to obtain a target object detection model.
第二方面,本申请实施例提供了一种物体检测模型训练装置,包括:In the second aspect, an embodiment of the present application provides an object detection model training device, including:
待检测图像获取模块,用于获取待检测图像;The image acquisition module to be detected is used to acquire the image to be detected;
物体检测结果获取模块,用于将所述待检测图像输入到目标物体检测模型中进行物体检测,得到所述待检测图像的物体检测结果,其中,所述目标物体检测模型采用训练样本获取模块、模型训练模块、损失获取模块和目标物体检测模型获取模块得到:The object detection result acquisition module is used to input the to-be-detected image into a target object detection model for object detection to obtain the object detection result of the to-be-detected image, wherein the target object detection model adopts a training sample acquisition module, The model training module, the loss acquisition module and the target object detection model acquisition module obtain:
训练样本获取模块,用于获取训练样本;The training sample acquisition module is used to acquire training samples;
模型训练模块,用于将所述训练样本输入到物体检测模型中进行模型训练,其中,所述物体检测模型包括检测模块、分类模块和判别模块;A model training module is used to input the training samples into an object detection model for model training, where the object detection model includes a detection module, a classification module, and a discrimination module;
损失获取模块,用于在模型训练过程中得到由所述检测模块产生的检测损失、由所述分类模块产生的分类损失和由所述判别模块产生的判别损失;A loss acquisition module, configured to acquire the detection loss generated by the detection module, the classification loss generated by the classification module, and the discrimination loss generated by the discrimination module during the model training process;
目标物体检测模型获取模块,用于根据所述检测损失、所述分类损失和所述判别损失更新所述物体检测模型,得到目标物体检测模型。The target object detection model acquisition module is used to update the object detection model according to the detection loss, the classification loss and the discrimination loss to obtain a target object detection model.
第三方面,一种计算机设备,包括存储器、处理器以及存储在所述存储器中并可在所述处理器上运行的计算机可读指令,所述处理器执行所述计算机可读指令时实现上述物体检测方法的步骤。In a third aspect, a computer device includes a memory, a processor, and computer-readable instructions that are stored in the memory and can run on the processor. When the processor executes the computer-readable instructions, the foregoing The steps of the object detection method.
第四方面,本申请实施例提供了一种计算机非易失性可读存储介质,包括:计算机可读指令,所述计算机可读指令被处理器执行时实现上述物体检测方法的步骤。In a fourth aspect, an embodiment of the present application provides a computer non-volatile readable storage medium, including: computer readable instructions, which implement the steps of the above object detection method when the computer readable instructions are executed by a processor.
在本申请提供的物体检测方法、装置、计算机设备及存储介质中,首先获取待检测图像;然后将所述待检测图像输入到目标物体检测模型中进行物体检测,得到待检测图像的物体检测结果,其中,所述目标物体检测模型结合了检测损失、分类损失和判别损失共同更新物体检测模型,具备更优的检测和分类效果,能够得到准确率较高的检测结果。In the object detection method, device, computer equipment and storage medium provided in this application, the image to be detected is first obtained; then the image to be detected is input into the target object detection model for object detection, and the object detection result of the image to be detected is obtained , Wherein the target object detection model combines the detection loss, classification loss and discrimination loss to update the object detection model, which has better detection and classification effects, and can obtain detection results with higher accuracy.
【附图说明】【Explanation of drawings】
为了更清楚地说明本申请实施例的技术方案,下面将对实施例中所需要使用的附图作简单地介绍,显而易见地,下面描述中的附图仅仅是本申请的一些实施例,对于本领域普通技术人员来讲,在不付出创造性劳动性的前提下,还可以根据这些附图获得其它的附图。In order to explain the technical solutions of the embodiments of the present application more clearly, the following will briefly introduce the drawings needed in the embodiments. Obviously, the drawings in the following description are only some embodiments of the present application. For those of ordinary skill in the art, without creative labor, other drawings can be obtained from these drawings.
图1是本申请一实施例中基于物体检测方法的一流程图;FIG. 1 is a flowchart of an object-based detection method in an embodiment of the present application;
图2是本申请一实施例中基于物体检测装置的一示意图;Figure 2 is a schematic diagram of an object-based detection device in an embodiment of the present application;
图3是本申请一实施例中计算机设备的一示意图。Fig. 3 is a schematic diagram of a computer device in an embodiment of the present application.
【具体实施方式】【detailed description】
为了更好的理解本申请的技术方案,下面结合附图对本申请实施例进行详细描述。In order to better understand the technical solutions of the present application, the embodiments of the present application will be described in detail below with reference to the accompanying drawings.
应当明确,所描述的实施例仅仅是本申请一部分实施例,而不是全部的实施例。基于本申请中的实施例,本领域普通技术人员在没有做出创造性劳动前提下所获得的所有其它实施例,都属于本申请保护的范围。It should be clear that the described embodiments are only a part of the embodiments of the present application, rather than all the embodiments. Based on the embodiments in this application, all other embodiments obtained by those of ordinary skill in the art without creative work shall fall within the protection scope of this application.
在本申请实施例中使用的术语是仅仅出于描述特定实施例的目的,而非旨在限制本申请。在本申请实施例和所附权利要求书中所使用的单数形式的“一种”、“所述”和“该”也旨在包括多数形式,除非上下文清楚地表示其他含义。The terms used in the embodiments of the present application are only for the purpose of describing specific embodiments, and are not intended to limit the present application. The singular forms of "a", "said" and "the" used in the embodiments of the present application and the appended claims are also intended to include plural forms, unless the context clearly indicates other meanings.
应当理解,本文中使用的术语“和/或”仅仅是一种描述关联对象的相同的字段,表示可以存在三种关系,例如,A和/或B,可以表示:单独存在A,同时存在A和B,单独存在B这三种情况。另外,本文中字符“/”,一般表示前后关联对象是一种“或”的关系。It should be understood that the term "and/or" used in this article is only a description of the same field of the associated object, indicating that there can be three relationships, for example, A and/or B can mean that A exists alone and A exists at the same time. And B, there are three cases of B alone. In addition, the character "/" in this text generally indicates that the associated objects before and after are in an "or" relationship.
应当理解,尽管在本申请实施例中可能采用术语第一、第二、第三等来描述预设范围等,但这些预设范围不应限于这些术语。这些术语仅用来将预设范围彼此区分开。例如,在不脱离本申请实施例范围的情况下,第一预设范围也可以被称为第二预设范围,类似地,第二预设范围也可以被称为第一预设范围。It should be understood that although the terms first, second, third, etc. may be used in the embodiments of the present application to describe the preset range, etc., these preset ranges should not be limited to these terms. These terms are only used to distinguish the preset ranges from each other. For example, without departing from the scope of the embodiments of the present application, the first preset range may also be referred to as the second preset range, and similarly, the second preset range may also be referred to as the first preset range.
取决于语境,如在此所使用的词语“如果”可以被解释成为“在……时”或“当……时”或“响应于确定”或“响应于检测”。类似地,取决于语境,短语“如果确定”或“如果检测(陈述的条件或事件)”可以被解释成为“当确定时”或“响应于确定”或“当检测(陈述的条件或事件)时”或“响应于检测(陈述的条件或事件)”。Depending on the context, the word "if" as used herein can be interpreted as "when" or "when" or "in response to determination" or "in response to detection". Similarly, depending on the context, the phrase "if determined" or "if detected (statement or event)" can be interpreted as "when determined" or "in response to determination" or "when detected (statement or event) )" or "in response to detection (statement or event)".
图1示出本实施例中物体检测方法的一流程图。该物体检测方法可应用在物体检测系统中,在对物体进行检测分类时可采用该物体检测系统实现,该物体检测系统具体可应用在计算机设备上。其中,该计算机设备是可与用户进行人机交互的设备,包括但不限于电脑、智能手机和平板等设备。如图1所示,该物体检测方法包括如下步骤:Fig. 1 shows a flow chart of the object detection method in this embodiment. The object detection method can be applied to an object detection system, and the object detection system can be used to realize the detection and classification of objects, and the object detection system can be specifically applied to computer equipment. Among them, the computer device is a device that can perform human-computer interaction with the user, including but not limited to devices such as computers, smart phones, and tablets. As shown in Figure 1, the object detection method includes the following steps:
S1:获取待检测图像。S1: Obtain the image to be detected.
S2:将待检测图像输入到目标物体检测模型中进行物体检测,得到待检测图像的物体检测结果,其中,步骤S2中,目标物体检测模型采用的模型训练步骤具体包括:S2: Input the image to be detected into the target object detection model for object detection, and obtain the object detection result of the image to be detected. In step S2, the model training steps adopted by the target object detection model specifically include:
S10:获取训练样本。S10: Obtain training samples.
在一实施例中,获取模型训练所需的训练样本。具体地,根据物体检测的需要可以选择与某一类场景相关的图像作为训练样本。例如,可以将行车记录仪中保存的图像作为训练样本。行车记录仪中保存的图片能够反映车辆行驶过程中前方的道路状况,可以将该图像作为训练样本进行目标物体检测模型的训练,使得训练得到的目标物体检测模型可以对车辆行驶过程中前方的出现的物体进行检测,从而车辆根据接收到的检测结果做出预设的反应。可以理解地,在进行模型训练前需要预先对行车记录仪中保存的图像中出现的物体进行标注(标注需要检测的物体,没有检测需要的物体可以不标注),此外,还需要预先采用深度神经网络(如卷积神经网络)提取与进行标注的物体属于同一类别图像的深层特征,以在物体检测模型(包括相对应的深度神经网络,用以提取图像特征)检测时识别物体的类别。In one embodiment, training samples required for model training are obtained. Specifically, images related to a certain type of scene can be selected as training samples according to the needs of object detection. For example, the images saved in the driving recorder can be used as training samples. The pictures saved in the driving recorder can reflect the road conditions ahead during the driving of the vehicle. The image can be used as a training sample to train the target object detection model, so that the trained target object detection model can be The object is detected, so that the vehicle makes a preset response according to the received detection result. Understandably, it is necessary to pre-label the objects appearing in the image saved in the driving recorder before the model training (label the objects that need to be detected, and the objects that are not required for detection may not be labeled). In addition, it is also necessary to use deep neural A network (such as a convolutional neural network) extracts deep features of images belonging to the same category as the object to be annotated to identify the category of the object when the object detection model (including the corresponding deep neural network for extracting image features) detects.
S20:将训练样本输入到物体检测模型中进行模型训练,其中,物体检测模型包括检测模块、分类模块和判别模块。S20: Input training samples into the object detection model for model training, where the object detection model includes a detection module, a classification module, and a discrimination module.
其中,模型训练是指目标物体检测模型的训练。检测模块用以检测图像中的物体,分类模块用以对检测到的物体进行识别及分类,判别模块包括第一判别模块和/或第二判别模块,第一判别模块用以判别检测模块输出的结果是否正确,第二判别模块用以判别分类模块输出的结果是否正确,其中,第一判别模块和第二判别模块可以同时存在,也可以只存在第二判别模块,将第二判别模块作为判别模块。Among them, model training refers to the training of the target object detection model. The detection module is used to detect objects in the image, and the classification module is used to identify and classify the detected objects. The judgment module includes a first judgment module and/or a second judgment module. The first judgment module is used to judge the output of the detection module. Whether the result is correct or not, the second judgment module is used to judge whether the output result of the classification module is correct. Among them, the first judgment module and the second judgment module can exist at the same time, or only the second judgment module exists, and the second judgment module is used as the judgment Module.
在一实施例中,将训练样本输入到物体检测模型中进行模型训练,其中,该物体检测模型除了包括检测模型和分类模型,还包括判别模型。可以理解地,训练样本进行模型训练即把训练样本输入到物体检测模型进行检测的过程。In an embodiment, the training samples are input into the object detection model for model training, where the object detection model includes not only a detection model and a classification model, but also a discriminant model. Understandably, model training with training samples is the process of inputting training samples into the object detection model for detection.
进一步地,在步骤S20之前,还包括:Further, before step S20, it further includes:
S211:获取待处理物体检测模型,待处理物体检测模型包括检测模块和分类模块。S211: Obtain a detection model of the object to be processed, which includes a detection module and a classification module.
在一实施例中,获取待处理物体检测模型,可以理解地,该待处理物体检测模型具体可以是YOLO(You Only Look Once)检测模型和SSD(Single Shot Multi-Box Detection)等检测模型。这些模型包括检测模块和分类模块。本实施例是基于这些待处理物体检测模型进行的改进。In one embodiment, the detection model of the object to be processed is obtained. It is understandable that the detection model of the object to be processed may specifically be a detection model such as YOLO (You Only Look Once) detection model and SSD (Single Shot Multi-Box Detection). These models include detection modules and classification modules. This embodiment is an improvement based on these object detection models to be processed.
S212:在待处理物体检测模型中加入判别模块,其中,判别模块用于对检测模块和/或分类模块输出的结果进行判别。S212: Add a discrimination module to the detection model of the object to be processed, where the discrimination module is used to discriminate the results output by the detection module and/or the classification module.
在一实施例中,在待处理物体检测模型原有的基础上添加了判别模块,以对待处理物体检测模型输出的结果进行判别。添加判别模块能够帮助获知待处理物体检测模型检测 的准确率情况,以根据待处理物体检测模型检测出错的情况更新模型,提高检测的准确率。In one embodiment, a discrimination module is added on the original basis of the object detection model to be processed, so as to determine the output result of the object detection model to be processed. Adding a discrimination module can help to know the accuracy of the detection model of the object to be processed, so as to update the model according to the detection error of the object detection model to improve the accuracy of detection.
S213:对加入判别模块后的待处理物体检测模型进行模型的初始化操作,得到物体检测模型。S213: Perform model initialization operation on the object detection model to be processed after adding the discrimination module to obtain the object detection model.
其中,模型的初始化操作是指初始化模型中的网络参数,网络参数的初始值可以是根据经验预先设定好的。Among them, the initialization operation of the model refers to the initialization of the network parameters in the model, and the initial values of the network parameters may be preset based on experience.
可以理解地,如果不对模型进行初始化操作,那么待处理物体检测模型中检测模块和分类模块中的网络参数实际上已经是经过多次训练更新的,此时判别模块再根据检测模块和/或分类模块输出的结果进行判别并更新的效果会比较差,因为检测模块和分类模块在一开始训练时已经了长时间的学习,短时间内采用判别模块进行更新不能更新地比较彻底;相反,进行初始化操作后,在训练阶段每检测模块和/或分类模块输出一个结果时判别模块就会进行判别,能够随着训练过程及时根据输出的结果更新网络参数,从而达到更好的检测准确率。Understandably, if the model is not initialized, the network parameters in the detection module and the classification module in the object detection model to be processed have actually been updated through multiple trainings. At this time, the discrimination module is then based on the detection module and/or classification The result of the module output will be discriminated and updated, because the detection module and the classification module have been learning for a long time at the beginning of the training. It is more thorough if the discrimination module is used for updating in a short time; on the contrary, the initialization After the operation, the discrimination module will make a judgment every time the detection module and/or classification module outputs a result during the training phase, and can update the network parameters according to the output result in time with the training process, so as to achieve better detection accuracy.
步骤S211-S213中,提供了一种得到物体检测模型的实施方式,具体地,在待处理物体检测模型中加入判别模块,并进行模型的初始化操作,有利于提高后续根据物体检测模型训练更新得到的目标物体检测模型检测的准确率。In steps S211-S213, an implementation method for obtaining an object detection model is provided. Specifically, a discrimination module is added to the object detection model to be processed, and the initialization operation of the model is performed, which is beneficial to improve the subsequent training and update of the object detection model. The detection accuracy of the target object detection model.
进一步地,在步骤S20中,将训练样本输入到物体检测模型中进行模型训练,具体包括:Further, in step S20, inputting the training samples into the object detection model for model training includes:
S221:输入训练样本,通过物体检测模型提取训练样本的特征向量。S221: Input the training sample, and extract the feature vector of the training sample through the object detection model.
在一实施例中,物体检测模型中包括用于提取训练样本的特征向量的深度神经网络,具体可以是卷积神经网络。当输入训练样本到物体检测模型时,物体检测模型将采用深度神经网络提取训练样本的特征向量,为模型训练提供技术基础。In an embodiment, the object detection model includes a deep neural network for extracting feature vectors of training samples, which may specifically be a convolutional neural network. When the training samples are input to the object detection model, the object detection model will use the deep neural network to extract the feature vectors of the training samples to provide a technical basis for model training.
S222:将特征向量进行归一化处理,得到归一化特征向量,其中,归一化处理的表达式为:y=(x-MinValue)/(MaxValue-MinValue),y为归一化特征向量,x为特征向量,MaxValue为特征向量中特征值的最大值,MinValue为特征向量中特征值的最小值。S222: Perform normalization processing on the feature vector to obtain a normalized feature vector, where the expression of the normalization process is: y=(x-MinValue)/(MaxValue-MinValue), y is the normalized feature vector , X is the feature vector, MaxValue is the maximum value of the feature value in the feature vector, and MinValue is the minimum value of the feature value in the feature vector.
其中,特征向量中特征值具体是指像素值。Among them, the feature value in the feature vector specifically refers to the pixel value.
在一实施例中,对特征向量进行归一化处理,也即将特征向量中特征值归一化到[0,1]的区间中。可以理解地,图像的像素值有2 8、2 12和2 16等像素值级别,一副图像中可以包含大量不同的像素值,这使得计算效率较低,因此采用将归一化的方式把特征向量中特征值都压缩在同一个范围区间内,使得计算效率提高,模型训练的时间也会更短。 In one embodiment, the feature vector is normalized, that is, the feature value in the feature vector is normalized to the interval of [0,1]. Be appreciated that the pixel values of the image are 28, 212 and 216 the pixel value level, etc., a large number of different images may be contained in the pixel value, which makes the computational efficiency is low, so the use of the normalized manner The eigenvalues in the eigenvectors are compressed in the same range, so that the calculation efficiency is improved and the model training time is shorter.
S223:根据归一化特征向量对物体检测模型进行模型训练。S223: Perform model training on the object detection model according to the normalized feature vector.
在步骤S221-S223中,提供了一种训练样本输入到物体检测模型中进行模型训练的实施方式,将提取的训练样本特征进行归一化处理,把特征向量中特征值都压缩在同一个范围区间内,能够明显缩短训练时长,提高训练效率。In steps S221-S223, an implementation method for inputting training samples into the object detection model for model training is provided, and the extracted training sample features are normalized, and the feature values in the feature vector are compressed in the same range Within the interval, training time can be significantly shortened and training efficiency improved.
S30:在模型训练过程中得到由检测模块产生的检测损失、由分类模块产生的分类损失和由判别模块产生的判别损失。S30: Obtain the detection loss generated by the detection module, the classification loss generated by the classification module, and the discrimination loss generated by the discrimination module in the model training process.
可以理解地,检测模块、分类模块和判别模块分别代表一种功能,每一种功能在实现的时候是有可能出错的,也即生成的损失,根据这些由检测模块产生的检测损失、由分类模块产生的分类损失和由判别模块产生的判别损失,可以借鉴帮助对物体检测模型进行调整,以使得到的目标物体检测模型能够在再次采用检测模块、分类模块和判别模块实现功能时尽可能不出错,以提高目标物体检测模型检测的准确率。Understandably, the detection module, the classification module, and the discrimination module respectively represent a function. When each function is implemented, it is possible to make a mistake, that is, the loss generated. According to the detection loss generated by the detection module, the classification The classification loss generated by the module and the discrimination loss generated by the discrimination module can be used for reference to help adjust the object detection model, so that the target object detection model can be used as much as possible when the detection module, classification module and discrimination module are used again to achieve functions. Errors to improve the accuracy of the target object detection model detection.
进一步地,在步骤S30中,在模型训练过程中得到由检测模块产生的检测损失、由分类模块产生的分类损失和由判别模块产生的判别损失,具体包括:Further, in step S30, the detection loss generated by the detection module, the classification loss generated by the classification module, and the discrimination loss generated by the discrimination module are obtained during the model training process, which specifically include:
S31:在模型训练过程中,得到检测模块输出的第一训练特征向量,采用预设的检测损失函数计算第一训练特征向量与预先存储的第一标签向量之间的损失,得到检测损失。S31: During the model training process, the first training feature vector output by the detection module is obtained, and the preset detection loss function is used to calculate the loss between the first training feature vector and the pre-stored first label vector to obtain the detection loss.
其中,第一训练特征向量即检测模块输出的结果,第一标签向量是用于校验第一训练特征向量是否正确的特征向量,代表的是真实结果。Among them, the first training feature vector is the result output by the detection module, and the first label vector is a feature vector used to verify whether the first training feature vector is correct, and represents the real result.
在一实施例中,采用预设的检测损失函数计算第一训练特征向量与预先存储的第一标签向量之间的损失,得到检测损失,以根据该检测损失更新模型的网络参数。具体地,检测损失函数可以包括对预测的中心坐标的损失函数,表示为:
Figure PCTCN2019091100-appb-000001
Figure PCTCN2019091100-appb-000002
其中,λ表示调节因子,是预设的参数值,i表示检测时分割得到的网格单元,I表示网格单元总数,j表示边界框预测值,J表示边界框预测值的总数量,
Figure PCTCN2019091100-appb-000003
表示在第i个网格单元中存在目标时,第j个边界框预测值对该预测有效,
Figure PCTCN2019091100-appb-000004
取1,;若第i个网格单元中不存在目标,
Figure PCTCN2019091100-appb-000005
取0,(x i,y i)表示预测边界框的位置,
Figure PCTCN2019091100-appb-000006
表示由训练样本输出的边界框的真实位置。
In an embodiment, a preset detection loss function is used to calculate the loss between the first training feature vector and the pre-stored first label vector to obtain the detection loss, so as to update the network parameters of the model according to the detection loss. Specifically, the detection loss function may include a loss function for the predicted center coordinates, which is expressed as:
Figure PCTCN2019091100-appb-000001
Figure PCTCN2019091100-appb-000002
Among them, λ represents the adjustment factor, which is a preset parameter value, i represents the grid unit divided during detection, I represents the total number of grid units, j represents the predicted value of the bounding box, and J represents the total number of predicted values of the bounding box.
Figure PCTCN2019091100-appb-000003
Indicates that when there is a target in the i-th grid unit, the j-th bounding box predicted value is valid for the prediction,
Figure PCTCN2019091100-appb-000004
Take 1,; if there is no target in the i-th grid unit,
Figure PCTCN2019091100-appb-000005
Take 0, (x i , y i ) represents the position of the predicted bounding box,
Figure PCTCN2019091100-appb-000006
Represents the true position of the bounding box output by the training sample.
可以理解地,如yolo等物体检测模型,需要对输入的训练样本进行图像分割,得到I个网格单元,在预测训练样本的物体位置时,得到J个预测边界框,上述
Figure PCTCN2019091100-appb-000007
的obj表示object,表示对物体进行检测。
Understandably, object detection models such as yolo need to perform image segmentation on the input training samples to obtain I grid units. When predicting the object positions of the training samples, J prediction bounding boxes are obtained.
Figure PCTCN2019091100-appb-000007
The obj stands for object, which means to detect the object.
进一步地,检测损失函数还可以包括关于预测边界框的宽高的损失函数,表示为:
Figure PCTCN2019091100-appb-000008
其中,
Figure PCTCN2019091100-appb-000009
表示预测宽度的平方根和预测高度的平方根,
Figure PCTCN2019091100-appb-000010
表示由训练样本输出的宽度的平方根和高度的平方根的真实值(其他重复出现的参数将不再作解释,以免赘述)。
Further, the detection loss function may also include a loss function about the width and height of the predicted bounding box, expressed as:
Figure PCTCN2019091100-appb-000008
among them,
Figure PCTCN2019091100-appb-000009
Represents the square root of the predicted width and the square root of the predicted height,
Figure PCTCN2019091100-appb-000010
Represents the true value of the square root of the width and the square root of the height output by the training sample (other repeated parameters will not be explained, so as not to repeat them).
可以理解地,以上提供了从模型预测的中心坐标和预测边界框的宽高两方面来衡量检测时的损失,其中,检测模块输出的第一训练特征向量具体包括(x i,y i)和
Figure PCTCN2019091100-appb-000011
第一标签向量具体包括
Figure PCTCN2019091100-appb-000012
Figure PCTCN2019091100-appb-000013
通过该检测损失函数能够更准确地更新物体检测模型的网络参数。
Understandably, the above provides two aspects of the center coordinates predicted by the model and the width and height of the predicted bounding box to measure the loss during detection, where the first training feature vector output by the detection module specifically includes (x i , y i ) and
Figure PCTCN2019091100-appb-000011
The first label vector specifically includes
Figure PCTCN2019091100-appb-000012
with
Figure PCTCN2019091100-appb-000013
Through the detection loss function, the network parameters of the object detection model can be updated more accurately.
S32:在模型训练过程中,得到分类模块输出的第二训练特征向量,采用预设的分类损失函数计算第二训练特征向量与预先存储的第二标签向量之间的损失,得到分类损失。S32: In the model training process, the second training feature vector output by the classification module is obtained, and the preset classification loss function is used to calculate the loss between the second training feature vector and the pre-stored second label vector to obtain the classification loss.
其中,第二训练特征向量即分类模块输出的结果,第二标签向量是用于校验第二训练特征向量是否正确的特征向量,代表的是真实结果。Among them, the second training feature vector is the result output by the classification module, and the second label vector is a feature vector used to verify whether the second training feature vector is correct, and represents the real result.
在一实施例中,采用预设的分类损失函数计算第二训练特征向量与预先存储的第二标签向量之间的损失,得到分类损失,以根据该分类损失更新模型的网络参数。具体地,分类损失函数具体可以表示为:
Figure PCTCN2019091100-appb-000014
其中,i表示检测时分割得到的网格单元,I表示网格单元总数,
Figure PCTCN2019091100-appb-000015
表示在第i个网格单元中存在目标时,
Figure PCTCN2019091100-appb-000016
取1,否则
Figure PCTCN2019091100-appb-000017
取0,p i表示预测的分类,
Figure PCTCN2019091100-appb-000018
表示由训练样本输出的分类的真实情况。其中,分类模块输出的第二训练特征向量具体包括p i,第二标签向量具体包括
Figure PCTCN2019091100-appb-000019
通过该分类损失函数能够更准确地更新物体分类模型的网络参数。
In an embodiment, a preset classification loss function is used to calculate the loss between the second training feature vector and the pre-stored second label vector to obtain the classification loss, so as to update the network parameters of the model according to the classification loss. Specifically, the classification loss function can be expressed as:
Figure PCTCN2019091100-appb-000014
Among them, i represents the grid unit divided during detection, I represents the total number of grid units,
Figure PCTCN2019091100-appb-000015
Indicates that when there is a target in the i-th grid cell,
Figure PCTCN2019091100-appb-000016
Take 1, otherwise
Figure PCTCN2019091100-appb-000017
Take 0, p i represents the predicted classification,
Figure PCTCN2019091100-appb-000018
Represents the true situation of the classification output by the training sample. Among them, the second training feature vector output by the classification module specifically includes p i , and the second label vector specifically includes
Figure PCTCN2019091100-appb-000019
Through the classification loss function, the network parameters of the object classification model can be updated more accurately.
S33:在模型训练过程中,得到判别模块输出的第三训练特征向量,根据第三训练特征向量,采用预设的判别损失函数计算得到判别损失。S33: In the model training process, obtain the third training feature vector output by the discrimination module, and calculate the discrimination loss by using the preset discriminant loss function according to the third training feature vector.
其中,第三训练特征向量即判别模块输出的结果。Among them, the third training feature vector is the result output by the discrimination module.
在一实施例中,根据第三训练特征向量,采用预设的判别损失函数计算得到判别损失,以根据该判别损失更新模型的网络参数。具体地,以只包括第二判别模块的判别模块为例:判别损失函数具体可以表示为:
Figure PCTCN2019091100-appb-000020
其中,I表示网格单元总数,i表示检测时分割得到的网格单元,D(p i)表示预测的分类在判别模块输出的结果。通过该判别损失函数能够反映判别模块在训练时产生的损失,从而更准确地更新物体判别模型的网络参数。
In an embodiment, according to the third training feature vector, a preset discriminant loss function is used to calculate the discriminant loss, so as to update the network parameters of the model according to the discriminant loss. Specifically, taking the discrimination module that only includes the second discrimination module as an example: the discrimination loss function can be specifically expressed as:
Figure PCTCN2019091100-appb-000020
Wherein, I represents the total number of grid cells, i denotes grid cells obtained divided detection, classification D (p i) denotes the prediction result of the discrimination output of the module. The discriminant loss function can reflect the loss generated by the discriminant module during training, so as to more accurately update the network parameters of the object discriminant model.
可以理解地,步骤S31-S33中的公式在计算检测损失、分类损失和判别损失时,是以一个训练样本的损失为例,在实际有大量样本进行模型训练时,将分别把各个训练样本的检测损失、分类损失和判别损失算数相加得到总的检测损失、分类损失和判别损失,并根据该总的检测损失、分类损失和判别损失进行模型的更新。Understandably, the formulas in steps S31-S33 take the loss of one training sample as an example when calculating the detection loss, classification loss, and discrimination loss. When there are actually a large number of samples for model training, the values of each training sample will be The detection loss, classification loss, and discrimination loss arithmetic are added to obtain the total detection loss, classification loss, and discrimination loss, and the model is updated according to the total detection loss, classification loss and discrimination loss.
步骤S31-S33提供了得到检测损失、分类损失和判别损失的具体实施方式,通过该得到的检测损失、分类损失和判别损失,能够准确地描述训练过程中产生的损失,以使模型更新得更准确。Steps S31-S33 provide specific implementations for obtaining detection loss, classification loss, and discrimination loss. The obtained detection loss, classification loss, and discrimination loss can accurately describe the loss generated during the training process, so that the model can be updated more accurately. accurate.
S40:根据检测损失、分类损失和判别损失更新物体检测模型,得到目标物体检测模型。S40: Update the object detection model according to the detection loss, classification loss and discrimination loss to obtain the target object detection model.
进一步地,步骤S40,具体包括:Further, step S40 specifically includes:
S41:根据检测损失、分类损失和判别损失,采用反向传播算法对物体检测模型中的网络参数进行更新。S41: According to the detection loss, classification loss and discrimination loss, the back propagation algorithm is used to update the network parameters in the object detection model.
其中,反向传播算法是在有导师指导下,适合于多层神经元网络的一种学习算法,它建立在梯度下降法的基础上。Among them, the back-propagation algorithm is a learning algorithm suitable for multi-layer neural networks under the guidance of a tutor. It is based on the gradient descent method.
在一实施例中,采用反向传播算法更新物体检测模型可以加快更新的速度,提高模型训练的训练效率。在检测损失、分类损失和判别损失总损失较多的情况下,采用反向传播算法有较好的效果。In one embodiment, updating the object detection model using a back propagation algorithm can speed up the update and improve the training efficiency of model training. When the total loss of detection loss, classification loss and discrimination loss is large, the use of backpropagation algorithm has better results.
S42:当网络参数的变化值均小于停止迭代阈值时,停止更新网络参数,得到目标物体检测模型。S42: When the change values of the network parameters are all less than the iterative stop threshold, stop updating the network parameters to obtain the target object detection model.
在一实施例中,当网络参数的变化值均小于停止迭代阈值时,也即网络参数的变化值都在可接受的误差范围内时,可以停止更新的过程,训练结束,得到检测准确率较高的目标物体检测模型。In one embodiment, when the change value of the network parameter is less than the stop iteration threshold, that is, when the change value of the network parameter is within the acceptable error range, the update process can be stopped, the training ends, and the detection accuracy rate is higher. High target object detection model.
步骤S41-S42中提供了一种更新物体检测模型的实施方式,能够快速完成更新过程,得到检测准确率较高的目标物体检测模型。Steps S41-S42 provide an implementation manner for updating the object detection model, which can quickly complete the update process and obtain a target object detection model with higher detection accuracy.
在本申请实施例中,首先获取待检测图像,然后将所述待检测图像输入到目标物体检测模型中进行物体检测,得到所述待检测图像的物体检测结果,其中,目标物体检测模型结合了检测损失、分类损失和判别损失共同更新物体检测模型,使训练得到的目标物体检测模型具备更优的检测和分类效果。In the embodiment of the present application, the image to be detected is first acquired, and then the image to be detected is input into the target object detection model for object detection, and the object detection result of the image to be detected is obtained, wherein the target object detection model combines The detection loss, classification loss and discrimination loss jointly update the object detection model, so that the target object detection model obtained by training has better detection and classification effects.
应理解,上述实施例中各步骤的序号的大小并不意味着执行顺序的先后,各过程的执行顺序应以其功能和内在逻辑确定,而不应对本申请实施例的实施过程构成任何限定。It should be understood that the size of the sequence number of each step in the foregoing embodiment does not mean the order of execution, and the execution sequence of each process should be determined by its function and internal logic, and should not constitute any limitation on the implementation process of the embodiment of the present application.
基于实施例中所提供的物体检测方法,本申请实施例进一步给出实现上述方法实施例中各步骤及方法的装置实施例。Based on the object detection methods provided in the embodiments, the embodiments of the present application further provide device embodiments that implement the steps and methods in the foregoing method embodiments.
图2示出与实施例中物体检测方法一一对应的物体检测装置的原理框图。如图2所示,该物体检测装置包括待检测图像获取模块10、物体检测结果获取模块20,还包括训练样本获取模块30、模型训练模块40、损失获取模块50和目标物体检测模型获取模块60。其中,待检测图像获取模块10、物体检测结果获取模块20、训练样本获取模块30、模型训练模块40、损失获取模块50和目标物体检测模型获取模块60的实现功能与实施例中物体检测方法对应的步骤一一对应,为避免赘述,本实施例不一一详述。Fig. 2 shows a principle block diagram of an object detection device corresponding to the object detection method in the embodiment one to one. As shown in FIG. 2, the object detection device includes a to-be-detected image acquisition module 10, an object detection result acquisition module 20, and also includes a training sample acquisition module 30, a model training module 40, a loss acquisition module 50, and a target object detection model acquisition module 60 . Among them, the realization functions of the to-be-detected image acquisition module 10, the object detection result acquisition module 20, the training sample acquisition module 30, the model training module 40, the loss acquisition module 50, and the target object detection model acquisition module 60 correspond to the object detection method in the embodiment The steps of are one-to-one correspondence, in order to avoid redundant description, this embodiment will not describe them one by one.
待检测图像获取模块10,用于获取待检测图像。The to-be-detected image acquisition module 10 is used to acquire the to-be-detected image.
物体检测结果获取模块20,用于将待检测图像输入到目标物体检测模型中进行物体检测,得到待检测图像的物体检测结果。The object detection result acquisition module 20 is used to input the image to be detected into the target object detection model for object detection, and obtain the object detection result of the image to be detected.
训练样本获取模块30,用于获取训练样本。The training sample acquisition module 30 is used to acquire training samples.
模型训练模块40,用于将训练样本输入到物体检测模型中进行模型训练,其中,物体检测模型包括检测模块、分类模块和判别模块。The model training module 40 is used to input training samples into the object detection model for model training, where the object detection model includes a detection module, a classification module and a discrimination module.
损失获取模块50,用于在模型训练过程中得到由检测模块产生的检测损失、由分类模块产生的分类损失和由判别模块产生的判别损失。The loss acquisition module 50 is used to obtain the detection loss generated by the detection module, the classification loss generated by the classification module, and the discrimination loss generated by the discrimination module during the model training process.
目标物体检测模型获取模块60,用于根据检测损失、分类损失和判别损失更新物体检测模型,得到目标物体检测模型。The target object detection model acquisition module 60 is used to update the object detection model according to the detection loss, classification loss and discrimination loss to obtain the target object detection model.
可选地,物体检测装置还包括待处理物体检测模型获取单元、判别模块加入单元和初始化单元。Optionally, the object detection device further includes a detection model acquisition unit of the object to be processed, a discrimination module adding unit and an initialization unit.
待处理物体检测模型获取单元,用于获取待处理物体检测模型,待处理物体检测模型包括检测模块和分类模块。The to-be-processed object detection model acquisition unit is used to acquire the to-be-processed object detection model. The to-be-processed object detection model includes a detection module and a classification module.
判别模块加入单元,用于在待处理物体检测模型中加入判别模块,其中,判别模块用于对检测模块和/或分类模块输出的结果进行判别。The discrimination module adding unit is used to add a discrimination module to the object detection model to be processed, wherein the discrimination module is used to discriminate the results output by the detection module and/or the classification module.
初始化单元,用于对加入判别模块后的待处理物体检测模型进行模型的初始化操作,得到物体检测模型。The initialization unit is used to initialize the object detection model to be processed after adding the discrimination module to obtain the object detection model.
可选地,模型训练模块40包括特征向量提取单元、归一化特征向量获取单元和模型训练单元。Optionally, the model training module 40 includes a feature vector extraction unit, a normalized feature vector acquisition unit, and a model training unit.
特征向量提取单元,用于输入训练样本,通过物体检测模型提取训练样本的特征向量。The feature vector extraction unit is used to input training samples, and extract feature vectors of the training samples through the object detection model.
归一化特征向量获取单元,用于将特征向量进行归一化处理,得到归一化特征向量,其中, 归一化处理的表达式为:y=(x-MinValue)/(MaxValue-MinValue),y为归一化特征向量,x为特征向量,MaxValue为特征向量中特征值的最大值,MinValue为特征向量中特征值的最小值。The normalized feature vector acquiring unit is used to normalize the feature vector to obtain the normalized feature vector, where the expression of the normalized processing is: y=(x-MinValue)/(MaxValue-MinValue) , Y is the normalized feature vector, x is the feature vector, MaxValue is the maximum value of the feature value in the feature vector, and MinValue is the minimum value of the feature value in the feature vector.
模型训练单元,用于根据归一化特征向量对物体检测模型进行模型训练。The model training unit is used to perform model training on the object detection model according to the normalized feature vector.
可选地,损失获取模块50包括检测损失获取单元、分类损失获取单元和判别损失获取单元。Optionally, the loss acquisition module 50 includes a detection loss acquisition unit, a classification loss acquisition unit, and a discrimination loss acquisition unit.
检测损失获取单元,用于在模型训练过程中,得到检测模块输出的第一训练特征向量,采用预设的检测损失函数计算第一训练特征向量与预先存储的第一标签向量之间的损失,得到检测损失。The detection loss acquisition unit is used to obtain the first training feature vector output by the detection module during the model training process, and calculate the loss between the first training feature vector and the pre-stored first label vector using a preset detection loss function, Get detection loss.
分类损失获取单元,用于在模型训练过程中,得到分类模块输出的第二训练特征向量,采用预设的分类损失函数计算第二训练特征向量与预先存储的第二标签向量之间的损失,得到分类损失。The classification loss acquisition unit is used to obtain the second training feature vector output by the classification module during the model training process, and calculate the loss between the second training feature vector and the pre-stored second label vector by using a preset classification loss function, Get classification loss.
判别损失获取单元,用于在模型训练过程中,得到判别模块输出的第三训练特征向量,根据第三训练特征向量,采用预设的判别损失函数计算得到判别损失。The discrimination loss acquisition unit is used to obtain the third training feature vector output by the discrimination module during the model training process, and calculate the discrimination loss by using the preset discriminant loss function according to the third training feature vector.
可选地,目标物体检测模型获取模块60包括网络参数更新单元和目标物体检测模型获取单元。Optionally, the target object detection model acquisition module 60 includes a network parameter update unit and a target object detection model acquisition unit.
网络参数更新单元,用于根据检测损失、分类损失和判别损失,采用反向传播算法对物体检测模型中的网络参数进行更新。The network parameter update unit is used to update the network parameters in the object detection model by using the back propagation algorithm according to the detection loss, classification loss and discrimination loss.
目标物体检测模型获取单元,用于当网络参数的变化值均小于停止迭代阈值时,停止更新网络参数,得到目标物体检测模型。The target object detection model acquisition unit is used to stop updating the network parameters when the change values of the network parameters are less than the iterative stop threshold to obtain the target object detection model.
在本申请实施例中,首先获取待检测图像,然后将所述待检测图像输入到目标物体检测模型中进行物体检测,得到所述待检测图像的物体检测结果,其中,目标物体检测模型结合了检测损失、分类损失和判别损失共同更新物体检测模型,使训练得到的目标物体检测模型具备更优的检测和分类效果。In the embodiment of the present application, the image to be detected is first acquired, and then the image to be detected is input into the target object detection model for object detection, and the object detection result of the image to be detected is obtained, wherein the target object detection model combines The detection loss, classification loss and discrimination loss jointly update the object detection model, so that the target object detection model obtained by training has better detection and classification effects.
本实施例提供一计算机非易失性可读存储介质,该计算机非易失性可读存储介质上存储有计算机可读指令,该计算机可读指令被处理器执行时实现实施例中物体检测方法,为避免重复,此处不一一赘述。或者,该计算机可读指令被处理器执行时实现实施例中物体检测装置中各模块/单元的功能,为避免重复,此处不一一赘述。This embodiment provides a computer non-volatile readable storage medium. The computer non-volatile readable storage medium stores computer readable instructions. When the computer readable instructions are executed by a processor, the object detection method in the embodiment is implemented. To avoid repetition, I won’t repeat them here. Alternatively, the computer-readable instructions realize the functions of the modules/units in the object detection device in the embodiment when they are executed by the processor. To avoid repetition, details are not repeated here.
图3是本申请一实施例提供的计算机设备的示意图。如图3所示,该实施例的计算机设备70包括:处理器71、存储器72以及存储在存储器72中并可在处理器71上运行的计算机可读指令73,该计算机可读指令73被处理器71执行时实现实施例中的物体检测 方法,为避免重复,此处不一一赘述。或者,该计算机可读指令73被处理器71执行时实现实施例中物体检测装置中各模型/单元的功能,为避免重复,此处不一一赘述。Fig. 3 is a schematic diagram of a computer device provided by an embodiment of the present application. As shown in FIG. 3, the computer device 70 of this embodiment includes: a processor 71, a memory 72, and computer-readable instructions 73 stored in the memory 72 and running on the processor 71, and the computer-readable instructions 73 are processed The object detection method in the embodiment is implemented when the device 71 is executed. To avoid repetition, it will not be repeated here. Alternatively, when the computer-readable instruction 73 is executed by the processor 71, the function of each model/unit in the object detection device in the embodiment is realized. In order to avoid repetition, it will not be repeated here.
计算机设备70可以是桌上型计算机、笔记本、掌上电脑及云端服务器等计算设备。计算机设备70可包括,但不仅限于,处理器71、存储器72。本领域技术人员可以理解,图3仅仅是计算机设备70的示例,并不构成对计算机设备70的限定,可以包括比图示更多或更少的部件,或者组合某些部件,或者不同的部件,例如计算机设备还可以包括输入输出设备、网络接入设备、总线等。The computer device 70 may be a computing device such as a desktop computer, a notebook, a palmtop computer, and a cloud server. The computer device 70 may include, but is not limited to, a processor 71 and a memory 72. Those skilled in the art can understand that FIG. 3 is only an example of the computer device 70, and does not constitute a limitation on the computer device 70. It may include more or less components than those shown in the figure, or a combination of certain components, or different components. For example, computer equipment may also include input and output devices, network access devices, buses, and so on.
所称处理器71可以是中央处理单元(Central Processing Unit,CPU),还可以是其他通用处理器、数字信号处理器(Digital Signal Processor,DSP)、专用集成电路(Application Specific Integrated Circuit,ASIC)、现场可编程门阵列(Field-Programmable Gate Array,FPGA)或者其他可编程逻辑器件、分立门或者晶体管逻辑器件、分立硬件组件等。通用处理器可以是微处理器或者该处理器也可以是任何常规的处理器等。The so-called processor 71 may be a central processing unit (Central Processing Unit, CPU), other general-purpose processors, digital signal processors (Digital Signal Processor, DSP), application specific integrated circuits (Application Specific Integrated Circuit, ASIC), Field-Programmable Gate Array (FPGA) or other programmable logic devices, discrete gates or transistor logic devices, discrete hardware components, etc. The general-purpose processor may be a microprocessor or the processor may also be any conventional processor or the like.
存储器72可以是计算机设备70的内部存储单元,例如计算机设备70的硬盘或内存。存储器72也可以是计算机设备70的外部存储设备,例如计算机设备70上配备的插接式硬盘,智能存储卡(Smart Media Card,SMC),安全数字(Secure Digital,SD)卡,闪存卡(Flash Card)等。进一步地,存储器72还可以既包括计算机设备70的内部存储单元也包括外部存储设备。存储器72用于存储计算机可读指令以及计算机设备所需的其他程序和数据。存储器72还可以用于暂时地存储已经输出或者将要输出的数据。The memory 72 may be an internal storage unit of the computer device 70, such as a hard disk or memory of the computer device 70. The memory 72 may also be an external storage device of the computer device 70, such as a plug-in hard disk equipped on the computer device 70, a smart memory card (Smart Media Card, SMC), a Secure Digital (SD) card, and a flash memory card (Flash). Card) and so on. Further, the memory 72 may also include both an internal storage unit of the computer device 70 and an external storage device. The memory 72 is used to store computer readable instructions and other programs and data required by the computer equipment. The memory 72 can also be used to temporarily store data that has been output or will be output.
所属领域的技术人员可以清楚地了解到,为了描述的方便和简洁,仅以上述各功能单元、模块的划分进行举例说明,实际应用中,可以根据需要而将上述功能分配由不同的功能单元、模块完成,即将装置的内部结构划分成不同的功能单元或模块,以完成以上描述的全部或者部分功能。Those skilled in the art can clearly understand that, for the convenience and conciseness of description, only the division of the above functional units and modules is used as an example. In actual applications, the above functions can be allocated to different functional units, Module completion, that is, the internal structure of the device is divided into different functional units or modules to complete all or part of the functions described above.
以上实施例仅用以说明本申请的技术方案,而非对其限制;尽管参照前述实施例对本申请进行了详细的说明,本领域的普通技术人员应当理解:其依然可以对前述各实施例所记载的技术方案进行修改,或者对其中部分技术特征进行等同替换;而这些修改或者替换,并不使相应技术方案的本质脱离本申请各实施例技术方案的精神和范围,均应包含在本申请的保护范围之内。The above embodiments are only used to illustrate the technical solutions of the present application, not to limit them; although the present application has been described in detail with reference to the foregoing embodiments, those of ordinary skill in the art should understand that they can still compare the previous embodiments. The recorded technical solutions are modified, or some of the technical features are equivalently replaced; and these modifications or replacements do not cause the essence of the corresponding technical solutions to deviate from the spirit and scope of the technical solutions of the embodiments of the application, and shall be included in the application Within the scope of protection.

Claims (20)

  1. 一种物体检测方法,其特征在于,所述方法包括:An object detection method, characterized in that the method includes:
    获取待检测图像;Obtain the image to be detected;
    将所述待检测图像输入到目标物体检测模型中进行物体检测,得到所述待检测图像的物体检测结果,其中,所述目标物体检测模型采用的模型训练步骤包括:Input the to-be-detected image into a target object detection model for object detection to obtain an object detection result of the to-be-detected image, wherein the model training step adopted by the target object detection model includes:
    获取训练样本;Obtain training samples;
    将所述训练样本输入到物体检测模型中进行模型训练,其中,所述物体检测模型包括检测模块、分类模块和判别模块;Inputting the training samples into an object detection model for model training, where the object detection model includes a detection module, a classification module, and a discrimination module;
    在模型训练过程中得到由所述检测模块产生的检测损失、由所述分类模块产生的分类损失和由所述判别模块产生的判别损失;Obtaining the detection loss generated by the detection module, the classification loss generated by the classification module, and the discrimination loss generated by the discrimination module during the model training process;
    根据所述检测损失、所述分类损失和所述判别损失更新所述物体检测模型,得到目标物体检测模型。The object detection model is updated according to the detection loss, the classification loss, and the discrimination loss to obtain a target object detection model.
  2. 根据权利要求1所述的方法,其特征在于,在所述将所述训练样本输入到物体检测模型中进行模型训练之前,所述方法还包括:The method according to claim 1, wherein before said inputting said training samples into an object detection model for model training, said method further comprises:
    获取待处理物体检测模型,所述待处理物体检测模型包括所述检测模块和所述分类模块;Acquiring a detection model of the object to be processed, where the detection model of the object to be processed includes the detection module and the classification module;
    在所述待处理物体检测模型中加入所述判别模块,其中,所述判别模块用于对所述检测模块和/或分类模块输出的结果进行判别;Adding the discrimination module to the object detection model to be processed, wherein the discrimination module is used to discriminate the results output by the detection module and/or the classification module;
    对加入所述判别模块后的所述待处理物体检测模型进行模型的初始化操作,得到所述物体检测模型。Perform a model initialization operation on the object detection model to be processed after adding the discrimination module to obtain the object detection model.
  3. 根据权利要求1所述的方法,其特征在于,所述将所述训练样本输入到物体检测模型中进行模型训练,包括:The method according to claim 1, wherein said inputting said training samples into an object detection model for model training comprises:
    输入所述训练样本,通过所述物体检测模型提取所述训练样本的特征向量;Input the training sample, and extract the feature vector of the training sample through the object detection model;
    将所述特征向量进行归一化处理,得到归一化特征向量,其中,归一化处理的表达式为:y=(x-MinValue)/(MaxValue-MinValue),y为所述归一化特征向量,x为所述特征向量,MaxValue为所述特征向量中特征值的最大值,MinValue为所述特征向量中特征值的最小值;The eigenvector is normalized to obtain a normalized eigenvector, where the expression of the normalization is: y=(x-MinValue)/(MaxValue-MinValue), y is the normalization Feature vector, x is the feature vector, MaxValue is the maximum value of the feature value in the feature vector, and MinValue is the minimum value of the feature value in the feature vector;
    根据所述归一化特征向量对所述物体检测模型进行模型训练。Model training is performed on the object detection model according to the normalized feature vector.
  4. 根据权利要求1所述的方法,其特征在于,所述在模型训练过程中得到由所述检测模块产生的检测损失、由所述分类模块产生的分类损失和由所述判别模块产生的判别损失,包括:The method of claim 1, wherein the detection loss generated by the detection module, the classification loss generated by the classification module, and the discrimination loss generated by the discrimination module are obtained in the model training process ,include:
    在模型训练过程中,得到所述检测模块输出的第一训练特征向量,采用预设的检测损失函 数计算所述第一训练特征向量与预先存储的第一标签向量之间的损失,得到所述检测损失;In the process of model training, the first training feature vector output by the detection module is obtained, and the preset detection loss function is used to calculate the loss between the first training feature vector and the pre-stored first label vector to obtain the Detection loss;
    在模型训练过程中,得到所述分类模块输出的第二训练特征向量,采用预设的分类损失函数计算所述第二训练特征向量与预先存储的第二标签向量之间的损失,得到所述分类损失;In the model training process, the second training feature vector output by the classification module is obtained, and the preset classification loss function is used to calculate the loss between the second training feature vector and the pre-stored second label vector to obtain the Classification loss
    在模型训练过程中,得到所述判别模块输出的第三训练特征向量,根据所述第三训练特征向量,采用预设的判别损失函数计算得到所述判别损失。In the model training process, the third training feature vector output by the discrimination module is obtained, and the discrimination loss is calculated by using a preset discriminant loss function according to the third training feature vector.
  5. 根据权利要求1至4任意一项所述的方法,其特征在于,所述根据所述检测损失、所述分类损失和所述判别损失更新所述物体检测模型,得到目标物体检测模型,包括:The method according to any one of claims 1 to 4, wherein said updating said object detection model according to said detection loss, said classification loss and said discrimination loss to obtain a target object detection model, comprising:
    根据所述检测损失、所述分类损失和所述判别损失,采用反向传播算法对所述物体检测模型中的网络参数进行更新;According to the detection loss, the classification loss and the discrimination loss, use a back propagation algorithm to update the network parameters in the object detection model;
    当所述网络参数的变化值均小于停止迭代阈值时,停止更新所述网络参数,得到所述目标物体检测模型。When the change values of the network parameters are all less than the iterative stop threshold, stop updating the network parameters to obtain the target object detection model.
  6. 一种物体检测装置,其特征在于,所述装置包括:An object detection device, characterized in that the device includes:
    待检测图像获取模块,用于获取待检测图像;The image acquisition module to be detected is used to acquire the image to be detected;
    物体检测结果获取模块,用于将所述待检测图像输入到目标物体检测模型中进行物体检测,得到所述待检测图像的物体检测结果,其中,所述目标物体检测模型采用训练样本获取模块、模型训练模块、损失获取模块和目标物体检测模型获取模块得到:The object detection result acquisition module is used to input the to-be-detected image into a target object detection model for object detection to obtain the object detection result of the to-be-detected image, wherein the target object detection model adopts a training sample acquisition module, The model training module, the loss acquisition module and the target object detection model acquisition module obtain:
    训练样本获取模块,用于获取训练样本;The training sample acquisition module is used to acquire training samples;
    模型训练模块,用于将所述训练样本输入到物体检测模型中进行模型训练,其中,所述物体检测模型包括检测模块、分类模块和判别模块;A model training module is used to input the training samples into an object detection model for model training, where the object detection model includes a detection module, a classification module, and a discrimination module;
    损失获取模块,用于在模型训练过程中得到由所述检测模块产生的检测损失、由所述分类模块产生的分类损失和由所述判别模块产生的判别损失;A loss acquisition module, configured to acquire the detection loss generated by the detection module, the classification loss generated by the classification module, and the discrimination loss generated by the discrimination module during the model training process;
    目标物体检测模型获取模块,用于根据所述检测损失、所述分类损失和所述判别损失更新所述物体检测模型,得到目标物体检测模型。The target object detection model acquisition module is used to update the object detection model according to the detection loss, the classification loss and the discrimination loss to obtain a target object detection model.
  7. 根据权利要求6所述的装置,其特征在于,所述物体检测模型训练装置还包括待处理物体检测模型获取单元、判别模块加入单元和初始化单元:The device according to claim 6, characterized in that the object detection model training device further comprises a to-be-processed object detection model acquisition unit, a discrimination module adding unit and an initialization unit:
    待处理物体检测模型获取单元,用于获取待处理物体检测模型,待处理物体检测模型包括检测模块和分类模块;The object detection model acquisition unit to be processed is used to acquire the object detection model to be processed, and the object detection model to be processed includes a detection module and a classification module;
    判别模块加入单元,用于在待处理物体检测模型中加入判别模块,其中,判别模块用于对检测模块和/或分类模块输出的结果进行判别;The discrimination module adding unit is used to add a discrimination module to the object detection model to be processed, wherein the discrimination module is used to discriminate the results output by the detection module and/or the classification module;
    初始化单元,用于对加入判别模块后的待处理物体检测模型进行模型的初始化操作,得到 物体检测模型。The initialization unit is used to initialize the object detection model to be processed after adding the discrimination module to obtain the object detection model.
  8. 根据权利要求6所述的装置,其特征在于,所述模型训练模块包括特征向量提取单元、归一化特征向量获取单元和模型训练单元:The device according to claim 6, wherein the model training module comprises a feature vector extraction unit, a normalized feature vector acquisition unit, and a model training unit:
    特征向量提取单元,用于输入所述训练样本,通过所述物体检测模型提取所述训练样本的特征向量;The feature vector extraction unit is configured to input the training sample, and extract the feature vector of the training sample through the object detection model;
    归一化特征向量获取单元,用于将所述特征向量进行归一化处理,得到归一化特征向量,其中,归一化处理的表达式为:y=(x-MinValue)/(MaxValue-MinValue),y为所述归一化特征向量,x为所述特征向量,MaxValue为所述特征向量中特征值的最大值,MinValue为所述特征向量中特征值的最小值;The normalized feature vector obtaining unit is used to perform normalization processing on the feature vector to obtain a normalized feature vector, where the expression of the normalization processing is: y=(x-MinValue)/(MaxValue- MinValue), y is the normalized feature vector, x is the feature vector, MaxValue is the maximum value of the feature value in the feature vector, and MinValue is the minimum value of the feature value in the feature vector;
    模型训练单元,用于根据所述归一化特征向量对所述物体检测模型进行模型训练。The model training unit is configured to perform model training on the object detection model according to the normalized feature vector.
  9. 根据权利要求6所述的装置,其特征在于,所述损失获取模块包括检测损失获取单元、分类损失获取单元和判别损失获取单元:The device according to claim 6, wherein the loss acquisition module comprises a detection loss acquisition unit, a classification loss acquisition unit, and a discrimination loss acquisition unit:
    检测损失获取单元,用于在模型训练过程中,得到检测模块输出的第一训练特征向量,采用预设的检测损失函数计算第一训练特征向量与预先存储的第一标签向量之间的损失,得到检测损失;The detection loss acquisition unit is used to obtain the first training feature vector output by the detection module during the model training process, and calculate the loss between the first training feature vector and the pre-stored first label vector using a preset detection loss function, Get detection loss;
    分类损失获取单元,用于在模型训练过程中,得到分类模块输出的第二训练特征向量,采用预设的分类损失函数计算第二训练特征向量与预先存储的第二标签向量之间的损失,得到分类损失;The classification loss acquisition unit is used to obtain the second training feature vector output by the classification module during the model training process, and calculate the loss between the second training feature vector and the pre-stored second label vector by using a preset classification loss function, Get classification loss;
    判别损失获取单元,用于在模型训练过程中,得到判别模块输出的第三训练特征向量,根据第三训练特征向量,采用预设的判别损失函数计算得到判别损失。The discrimination loss acquisition unit is used to obtain the third training feature vector output by the discrimination module during the model training process, and calculate the discrimination loss by using the preset discriminant loss function according to the third training feature vector.
  10. 根据权利要求6-9所述的装置,其特征在于,所述目标物体检测模型获取模块包括网络参数更新单元和目标物体检测模型获取单元:The device according to claims 6-9, wherein the target object detection model acquisition module comprises a network parameter update unit and a target object detection model acquisition unit:
    网络参数更新单元,用于根据所述检测损失、所述分类损失和所述判别损失,采用反向传播算法对所述物体检测模型中的网络参数进行更新;A network parameter update unit, configured to use a back propagation algorithm to update the network parameters in the object detection model according to the detection loss, the classification loss, and the discrimination loss;
    目标物体检测模型获取单元,用于当所述网络参数的变化值均小于停止迭代阈值时,停止更新所述网络参数,得到所述目标物体检测模型。The target object detection model acquisition unit is configured to stop updating the network parameters when the change values of the network parameters are less than the iterative stop threshold to obtain the target object detection model.
  11. 一种计算机设备,包括存储器、处理器以及存储在所述存储器中并可在所述处理器上运行的计算机可读指令,其特征在于,所述处理器执行所述计算机可读指令时实现如下步骤:A computer device includes a memory, a processor, and computer-readable instructions stored in the memory and capable of running on the processor, wherein the processor executes the computer-readable instructions as follows step:
    获取待检测图像;Obtain the image to be detected;
    将所述待检测图像输入到目标物体检测模型中进行物体检测,得到所述待检测图像的物体 检测结果,其中,所述目标物体检测模型采用的模型训练步骤包括:Input the to-be-detected image into a target object detection model for object detection, and obtain an object detection result of the to-be-detected image, wherein the model training step adopted by the target object detection model includes:
    获取训练样本;Obtain training samples;
    将所述训练样本输入到物体检测模型中进行模型训练,其中,所述物体检测模型包括检测模块、分类模块和判别模块;Inputting the training samples into an object detection model for model training, where the object detection model includes a detection module, a classification module, and a discrimination module;
    在模型训练过程中得到由所述检测模块产生的检测损失、由所述分类模块产生的分类损失和由所述判别模块产生的判别损失;Obtaining the detection loss generated by the detection module, the classification loss generated by the classification module, and the discrimination loss generated by the discrimination module during the model training process;
    根据所述检测损失、所述分类损失和所述判别损失更新所述物体检测模型,得到目标物体检测模型。The object detection model is updated according to the detection loss, the classification loss, and the discrimination loss to obtain a target object detection model.
  12. 根据权利要求11所述的计算机设备,其特征在于,所述处理器执行所述计算机可读指令时还实现如下步骤:The computer device according to claim 11, wherein the processor further implements the following steps when executing the computer-readable instruction:
    获取待处理物体检测模型,所述待处理物体检测模型包括所述检测模块和所述分类模块;Acquiring a detection model of the object to be processed, where the detection model of the object to be processed includes the detection module and the classification module;
    在所述待处理物体检测模型中加入所述判别模块,其中,所述判别模块用于对所述检测模块和/或分类模块输出的结果进行判别;Adding the discrimination module to the object detection model to be processed, wherein the discrimination module is used to discriminate the results output by the detection module and/or the classification module;
    对加入所述判别模块后的所述待处理物体检测模型进行模型的初始化操作,得到所述物体检测模型。Perform a model initialization operation on the object detection model to be processed after adding the discrimination module to obtain the object detection model.
  13. 根据权利要求11所述的计算机设备,其特征在于,所述处理器执行所述计算机可读指令时还实现如下步骤:The computer device according to claim 11, wherein the processor further implements the following steps when executing the computer-readable instruction:
    输入所述训练样本,通过所述物体检测模型提取所述训练样本的特征向量;Input the training sample, and extract the feature vector of the training sample through the object detection model;
    将所述特征向量进行归一化处理,得到归一化特征向量,其中,归一化处理的表达式为:y=(x-MinValue)/(MaxValue-MinValue),y为所述归一化特征向量,x为所述特征向量,MaxValue为所述特征向量中特征值的最大值,MinValue为所述特征向量中特征值的最小值;The eigenvector is normalized to obtain a normalized eigenvector, where the expression of the normalization is: y=(x-MinValue)/(MaxValue-MinValue), y is the normalization Feature vector, x is the feature vector, MaxValue is the maximum value of the feature value in the feature vector, and MinValue is the minimum value of the feature value in the feature vector;
    根据所述归一化特征向量对所述物体检测模型进行模型训练。Model training is performed on the object detection model according to the normalized feature vector.
  14. 根据权利要求11所述的计算机设备,其特征在于,所述处理器执行所述计算机可读指令时还实现如下步骤:The computer device according to claim 11, wherein the processor further implements the following steps when executing the computer-readable instruction:
    在模型训练过程中,得到所述检测模块输出的第一训练特征向量,采用预设的检测损失函数计算所述第一训练特征向量与预先存储的第一标签向量之间的损失,得到所述检测损失;In the process of model training, the first training feature vector output by the detection module is obtained, and the preset detection loss function is used to calculate the loss between the first training feature vector and the pre-stored first label vector to obtain the Detection loss;
    在模型训练过程中,得到所述分类模块输出的第二训练特征向量,采用预设的分类损失函数计算所述第二训练特征向量与预先存储的第二标签向量之间的损失,得到所述分类损失;In the model training process, the second training feature vector output by the classification module is obtained, and the preset classification loss function is used to calculate the loss between the second training feature vector and the pre-stored second label vector to obtain the Classification loss
    在模型训练过程中,得到所述判别模块输出的第三训练特征向量,根据所述第三训练特征向量,采用预设的判别损失函数计算得到所述判别损失。In the model training process, the third training feature vector output by the discrimination module is obtained, and the discrimination loss is calculated by using a preset discriminant loss function according to the third training feature vector.
  15. 根据权利要求11-14所述的计算机设备,其特征在于,所述处理器执行所述计算机可读指令时还实现如下步骤:14. The computer device according to claims 11-14, wherein the processor further implements the following steps when executing the computer-readable instructions:
    根据所述检测损失、所述分类损失和所述判别损失,采用反向传播算法对所述物体检测模型中的网络参数进行更新;According to the detection loss, the classification loss and the discrimination loss, use a back propagation algorithm to update the network parameters in the object detection model;
    当所述网络参数的变化值均小于停止迭代阈值时,停止更新所述网络参数,得到所述目标物体检测模型。When the change values of the network parameters are all less than the iterative stop threshold, stop updating the network parameters to obtain the target object detection model.
  16. 一种计算机非易失性可读存储介质,所述计算机非易失性可读存储介质存储有计算机可读指令,其特征在于,所述计算机可读指令被处理器执行时实现如下步骤:A computer non-volatile readable storage medium, the computer non-volatile readable storage medium storing computer readable instructions, characterized in that, when the computer readable instructions are executed by a processor, the following steps are implemented:
    获取待检测图像;Obtain the image to be detected;
    将所述待检测图像输入到目标物体检测模型中进行物体检测,得到所述待检测图像的物体检测结果,其中,所述目标物体检测模型采用的模型训练步骤包括:Input the to-be-detected image into a target object detection model for object detection to obtain an object detection result of the to-be-detected image, wherein the model training step adopted by the target object detection model includes:
    获取训练样本;Obtain training samples;
    将所述训练样本输入到物体检测模型中进行模型训练,其中,所述物体检测模型包括检测模块、分类模块和判别模块;Inputting the training samples into an object detection model for model training, where the object detection model includes a detection module, a classification module, and a discrimination module;
    在模型训练过程中得到由所述检测模块产生的检测损失、由所述分类模块产生的分类损失和由所述判别模块产生的判别损失;Obtaining the detection loss generated by the detection module, the classification loss generated by the classification module, and the discrimination loss generated by the discrimination module during the model training process;
    根据所述检测损失、所述分类损失和所述判别损失更新所述物体检测模型,得到目标物体检测模型。The object detection model is updated according to the detection loss, the classification loss, and the discrimination loss to obtain a target object detection model.
  17. 根据权利要求16所述的计算机非易失性可读存储介质,其特征在于,所述计算机可读指令被一个或多个处理器执行时,使得所述一个或多个处理器还实现如下步骤:The computer non-volatile readable storage medium according to claim 16, wherein when the computer readable instructions are executed by one or more processors, the one or more processors further implement the following steps :
    获取待处理物体检测模型,所述待处理物体检测模型包括所述检测模块和所述分类模块;Acquiring a detection model of the object to be processed, where the detection model of the object to be processed includes the detection module and the classification module;
    在所述待处理物体检测模型中加入所述判别模块,其中,所述判别模块用于对所述检测模块和/或分类模块输出的结果进行判别;Adding the discrimination module to the object detection model to be processed, wherein the discrimination module is used to discriminate the results output by the detection module and/or the classification module;
    对加入所述判别模块后的所述待处理物体检测模型进行模型的初始化操作,得到所述物体检测模型。Perform a model initialization operation on the object detection model to be processed after adding the discrimination module to obtain the object detection model.
  18. 根据权利要求16所述的计算机非易失性可读存储介质,其特征在于,所述计算机可读指令被一个或多个处理器执行时,使得所述一个或多个处理器还实现如下步骤:The computer non-volatile readable storage medium according to claim 16, wherein when the computer readable instructions are executed by one or more processors, the one or more processors further implement the following steps :
    输入所述训练样本,通过所述物体检测模型提取所述训练样本的特征向量;Input the training sample, and extract the feature vector of the training sample through the object detection model;
    将所述特征向量进行归一化处理,得到归一化特征向量,其中,归一化处理的表达式为:y=(x-MinValue)/(MaxValue-MinValue),y为所述归一化特征向量,x为所述特征向量,MaxValue 为所述特征向量中特征值的最大值,MinValue为所述特征向量中特征值的最小值;The eigenvector is normalized to obtain a normalized eigenvector, where the expression of the normalization is: y=(x-MinValue)/(MaxValue-MinValue), y is the normalization Feature vector, x is the feature vector, MaxValue is the maximum value of the feature value in the feature vector, and MinValue is the minimum value of the feature value in the feature vector;
    根据所述归一化特征向量对所述物体检测模型进行模型训练。Model training is performed on the object detection model according to the normalized feature vector.
  19. 根据权利要求16所述的计算机非易失性可读存储介质,其特征在于,所述计算机可读指令被一个或多个处理器执行时,使得所述一个或多个处理器还实现如下步骤:The computer non-volatile readable storage medium according to claim 16, wherein when the computer readable instructions are executed by one or more processors, the one or more processors further implement the following steps :
    在模型训练过程中,得到所述检测模块输出的第一训练特征向量,采用预设的检测损失函数计算所述第一训练特征向量与预先存储的第一标签向量之间的损失,得到所述检测损失;In the process of model training, the first training feature vector output by the detection module is obtained, and the preset detection loss function is used to calculate the loss between the first training feature vector and the pre-stored first label vector to obtain the Detection loss;
    在模型训练过程中,得到所述分类模块输出的第二训练特征向量,采用预设的分类损失函数计算所述第二训练特征向量与预先存储的第二标签向量之间的损失,得到所述分类损失;In the model training process, the second training feature vector output by the classification module is obtained, and the preset classification loss function is used to calculate the loss between the second training feature vector and the pre-stored second label vector to obtain the Classification loss
    在模型训练过程中,得到所述判别模块输出的第三训练特征向量,根据所述第三训练特征向量,采用预设的判别损失函数计算得到所述判别损失。In the model training process, the third training feature vector output by the discrimination module is obtained, and the discrimination loss is calculated by using a preset discriminant loss function according to the third training feature vector.
  20. 根据权利要求16-19所述的计算机非易失性可读存储介质,其特征在于,所述计算机可读指令被一个或多个处理器执行时,使得所述一个或多个处理器还实现如下步骤:The computer non-volatile readable storage medium according to claims 16-19, wherein when the computer readable instructions are executed by one or more processors, the one or more processors also implement The following steps:
    根据所述检测损失、所述分类损失和所述判别损失,采用反向传播算法对所述物体检测模型中的网络参数进行更新;According to the detection loss, the classification loss and the discrimination loss, use a back propagation algorithm to update the network parameters in the object detection model;
    当所述网络参数的变化值均小于停止迭代阈值时,停止更新所述网络参数,得到所述目标物体检测模型。When the change values of the network parameters are all less than the iterative stop threshold, stop updating the network parameters to obtain the target object detection model.
PCT/CN2019/091100 2019-02-03 2019-06-13 Object detection method and device, computer device and storage medium WO2020155518A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN201910108522.6A CN110020592B (en) 2019-02-03 2019-02-03 Object detection model training method, device, computer equipment and storage medium
CN201910108522.6 2019-02-03

Publications (1)

Publication Number Publication Date
WO2020155518A1 true WO2020155518A1 (en) 2020-08-06

Family

ID=67188871

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2019/091100 WO2020155518A1 (en) 2019-02-03 2019-06-13 Object detection method and device, computer device and storage medium

Country Status (2)

Country Link
CN (1) CN110020592B (en)
WO (1) WO2020155518A1 (en)

Cited By (15)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112102194A (en) * 2020-09-15 2020-12-18 北京金山云网络技术有限公司 Face restoration model training method and device
CN112183358A (en) * 2020-09-29 2021-01-05 新石器慧拓(北京)科技有限公司 Training method and device for target detection model
CN112418480A (en) * 2020-10-14 2021-02-26 上海眼控科技股份有限公司 Meteorological image prediction method, meteorological image prediction device, computer equipment and storage medium
CN112508097A (en) * 2020-12-08 2021-03-16 深圳市优必选科技股份有限公司 Image conversion model training method and device, terminal equipment and storage medium
CN112561885A (en) * 2020-12-17 2021-03-26 中国矿业大学 YOLOv 4-tiny-based gate valve opening detection method
CN112634245A (en) * 2020-12-28 2021-04-09 广州绿怡信息科技有限公司 Loss detection model training method, loss detection method and device
CN112633355A (en) * 2020-12-18 2021-04-09 北京迈格威科技有限公司 Image data processing method and device and target detection model training method and device
CN112633351A (en) * 2020-12-17 2021-04-09 博彦多彩数据科技有限公司 Detection method, detection device, storage medium and processor
CN112966565A (en) * 2021-02-05 2021-06-15 深圳市优必选科技股份有限公司 Object detection method and device, terminal equipment and storage medium
CN113033579A (en) * 2021-03-31 2021-06-25 北京有竹居网络技术有限公司 Image processing method, image processing device, storage medium and electronic equipment
CN113298122A (en) * 2021-04-30 2021-08-24 北京迈格威科技有限公司 Target detection method and device and electronic equipment
CN113591839A (en) * 2021-06-28 2021-11-02 北京有竹居网络技术有限公司 Feature extraction model construction method, target detection method and device
CN113627298A (en) * 2021-07-30 2021-11-09 北京百度网讯科技有限公司 Training method of target detection model and method and device for detecting target object
CN116935102A (en) * 2023-06-30 2023-10-24 上海蜜度信息技术有限公司 Lightweight model training method, device, equipment and medium
CN116958607A (en) * 2023-09-20 2023-10-27 中国人民解放军火箭军工程大学 Data processing method and device for target damage prediction

Families Citing this family (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110442804A (en) * 2019-08-13 2019-11-12 北京市商汤科技开发有限公司 A kind of training method, device, equipment and the storage medium of object recommendation network
CN112417955B (en) * 2020-10-14 2024-03-05 国能大渡河沙坪发电有限公司 Method and device for processing tour inspection video stream
CN112580731B (en) * 2020-12-24 2022-06-24 深圳市对庄科技有限公司 Jadeite product identification method, system, terminal, computer equipment and storage medium

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20100202693A1 (en) * 2009-02-09 2010-08-12 Samsung Electronics Co., Ltd. Apparatus and method for recognizing hand shape in portable terminal
US20130034263A1 (en) * 2011-08-04 2013-02-07 Yuanyuan Ding Adaptive Threshold for Object Detection
CN107038448A (en) * 2017-03-01 2017-08-11 中国科学院自动化研究所 Target detection model building method
CN107944443A (en) * 2017-11-16 2018-04-20 深圳市唯特视科技有限公司 One kind carries out object consistency detection method based on end-to-end deep learning

Family Cites Families (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP5394959B2 (en) * 2010-03-23 2014-01-22 富士フイルム株式会社 Discriminator generating apparatus and method, and program
CN106845522B (en) * 2016-12-26 2020-01-31 华北理工大学 classification discrimination system in metallurgical balling process
GB2564668B (en) * 2017-07-18 2022-04-13 Vision Semantics Ltd Target re-identification
CN108009524B (en) * 2017-12-25 2021-07-09 西北工业大学 Lane line detection method based on full convolution network
CN108376235A (en) * 2018-01-15 2018-08-07 深圳市易成自动驾驶技术有限公司 Image detecting method, device and computer readable storage medium

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20100202693A1 (en) * 2009-02-09 2010-08-12 Samsung Electronics Co., Ltd. Apparatus and method for recognizing hand shape in portable terminal
US20130034263A1 (en) * 2011-08-04 2013-02-07 Yuanyuan Ding Adaptive Threshold for Object Detection
CN107038448A (en) * 2017-03-01 2017-08-11 中国科学院自动化研究所 Target detection model building method
CN107944443A (en) * 2017-11-16 2018-04-20 深圳市唯特视科技有限公司 One kind carries out object consistency detection method based on end-to-end deep learning

Cited By (22)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112102194A (en) * 2020-09-15 2020-12-18 北京金山云网络技术有限公司 Face restoration model training method and device
CN112183358A (en) * 2020-09-29 2021-01-05 新石器慧拓(北京)科技有限公司 Training method and device for target detection model
CN112183358B (en) * 2020-09-29 2024-04-23 新石器慧通(北京)科技有限公司 Training method and device for target detection model
CN112418480A (en) * 2020-10-14 2021-02-26 上海眼控科技股份有限公司 Meteorological image prediction method, meteorological image prediction device, computer equipment and storage medium
CN112508097A (en) * 2020-12-08 2021-03-16 深圳市优必选科技股份有限公司 Image conversion model training method and device, terminal equipment and storage medium
CN112508097B (en) * 2020-12-08 2024-01-19 深圳市优必选科技股份有限公司 Image conversion model training method and device, terminal equipment and storage medium
CN112561885A (en) * 2020-12-17 2021-03-26 中国矿业大学 YOLOv 4-tiny-based gate valve opening detection method
CN112561885B (en) * 2020-12-17 2023-04-18 中国矿业大学 YOLOv 4-tiny-based gate valve opening detection method
CN112633351A (en) * 2020-12-17 2021-04-09 博彦多彩数据科技有限公司 Detection method, detection device, storage medium and processor
CN112633355A (en) * 2020-12-18 2021-04-09 北京迈格威科技有限公司 Image data processing method and device and target detection model training method and device
CN112634245A (en) * 2020-12-28 2021-04-09 广州绿怡信息科技有限公司 Loss detection model training method, loss detection method and device
CN112966565A (en) * 2021-02-05 2021-06-15 深圳市优必选科技股份有限公司 Object detection method and device, terminal equipment and storage medium
CN113033579A (en) * 2021-03-31 2021-06-25 北京有竹居网络技术有限公司 Image processing method, image processing device, storage medium and electronic equipment
CN113033579B (en) * 2021-03-31 2023-03-21 北京有竹居网络技术有限公司 Image processing method, image processing device, storage medium and electronic equipment
CN113298122A (en) * 2021-04-30 2021-08-24 北京迈格威科技有限公司 Target detection method and device and electronic equipment
CN113591839B (en) * 2021-06-28 2023-05-09 北京有竹居网络技术有限公司 Feature extraction model construction method, target detection method and device
CN113591839A (en) * 2021-06-28 2021-11-02 北京有竹居网络技术有限公司 Feature extraction model construction method, target detection method and device
CN113627298A (en) * 2021-07-30 2021-11-09 北京百度网讯科技有限公司 Training method of target detection model and method and device for detecting target object
CN116935102A (en) * 2023-06-30 2023-10-24 上海蜜度信息技术有限公司 Lightweight model training method, device, equipment and medium
CN116935102B (en) * 2023-06-30 2024-02-20 上海蜜度科技股份有限公司 Lightweight model training method, device, equipment and medium
CN116958607A (en) * 2023-09-20 2023-10-27 中国人民解放军火箭军工程大学 Data processing method and device for target damage prediction
CN116958607B (en) * 2023-09-20 2023-12-22 中国人民解放军火箭军工程大学 Data processing method and device for target damage prediction

Also Published As

Publication number Publication date
CN110020592A (en) 2019-07-16
CN110020592B (en) 2024-04-09

Similar Documents

Publication Publication Date Title
WO2020155518A1 (en) Object detection method and device, computer device and storage medium
EP4148622A1 (en) Neural network training method, image classification system, and related device
CN108038474B (en) Face detection method, convolutional neural network parameter training method, device and medium
WO2018108129A1 (en) Method and apparatus for use in identifying object type, and electronic device
US20210056293A1 (en) Face detection method
WO2017096753A1 (en) Facial key point tracking method, terminal, and nonvolatile computer readable storage medium
CN111476284A (en) Image recognition model training method, image recognition model training device, image recognition method, image recognition device and electronic equipment
WO2021136027A1 (en) Similar image detection method and apparatus, device and storage medium
CN109919002B (en) Yellow stop line identification method and device, computer equipment and storage medium
WO2019200702A1 (en) Descreening system training method and apparatus, descreening method and apparatus, device, and medium
JP2022141931A (en) Method and device for training living body detection model, method and apparatus for living body detection, electronic apparatus, storage medium, and computer program
US9129152B2 (en) Exemplar-based feature weighting
WO2021051497A1 (en) Pulmonary tuberculosis determination method and apparatus, computer device, and storage medium
WO2022218396A1 (en) Image processing method and apparatus, and computer readable storage medium
US11893773B2 (en) Finger vein comparison method, computer equipment, and storage medium
CN111914908B (en) Image recognition model training method, image recognition method and related equipment
CN115797736B (en) Training method, device, equipment and medium for target detection model and target detection method, device, equipment and medium
WO2019232861A1 (en) Handwriting model training method and apparatus, text recognition method and apparatus, and device and medium
WO2023221608A1 (en) Mask recognition model training method and apparatus, device, and storage medium
CN107862680A (en) A kind of target following optimization method based on correlation filter
WO2023160666A1 (en) Target detection method and apparatus, and target detection model training method and apparatus
WO2023109361A1 (en) Video processing method and system, device, medium and product
CN108399430A (en) A kind of SAR image Ship Target Detection method based on super-pixel and random forest
WO2023088174A1 (en) Target detection method and apparatus
Meus et al. Embedded vision system for pedestrian detection based on HOG+ SVM and use of motion information implemented in Zynq heterogeneous device

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 19914141

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 19914141

Country of ref document: EP

Kind code of ref document: A1