WO2020062433A1 - 一种神经网络模型训练及通用接地线的检测方法 - Google Patents
一种神经网络模型训练及通用接地线的检测方法 Download PDFInfo
- Publication number
- WO2020062433A1 WO2020062433A1 PCT/CN2018/113661 CN2018113661W WO2020062433A1 WO 2020062433 A1 WO2020062433 A1 WO 2020062433A1 CN 2018113661 W CN2018113661 W CN 2018113661W WO 2020062433 A1 WO2020062433 A1 WO 2020062433A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- neural network
- ground
- image
- network model
- road
- Prior art date
Links
- 238000000034 method Methods 0.000 title claims abstract description 37
- 238000012549 training Methods 0.000 title claims abstract description 34
- 238000003062 neural network model Methods 0.000 title claims description 25
- 238000013528 artificial neural network Methods 0.000 claims abstract description 25
- 230000011218 segmentation Effects 0.000 claims abstract description 25
- 238000001514 detection method Methods 0.000 claims abstract description 23
- 238000004364 calculation method Methods 0.000 claims description 8
- 238000004422 calculation algorithm Methods 0.000 abstract description 5
- 238000013527 convolutional neural network Methods 0.000 description 24
- 230000006870 function Effects 0.000 description 23
- 230000004913 activation Effects 0.000 description 5
- 238000010586 diagram Methods 0.000 description 5
- 238000012545 processing Methods 0.000 description 5
- 238000005457 optimization Methods 0.000 description 4
- 230000008569 process Effects 0.000 description 4
- 230000003068 static effect Effects 0.000 description 4
- 230000004927 fusion Effects 0.000 description 3
- 238000010606 normalization Methods 0.000 description 3
- 238000000605 extraction Methods 0.000 description 2
- 230000006872 improvement Effects 0.000 description 2
- 238000005259 measurement Methods 0.000 description 2
- 230000004048 modification Effects 0.000 description 2
- 238000012986 modification Methods 0.000 description 2
- 238000003909 pattern recognition Methods 0.000 description 2
- 238000011176 pooling Methods 0.000 description 2
- 230000001629 suppression Effects 0.000 description 2
- ORILYTVJVMAKLC-UHFFFAOYSA-N Adamantane Natural products C1C(C2)CC3CC1CC2C3 ORILYTVJVMAKLC-UHFFFAOYSA-N 0.000 description 1
- 230000007547 defect Effects 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 230000018109 developmental process Effects 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 238000002372 labelling Methods 0.000 description 1
- 238000013178 mathematical model Methods 0.000 description 1
- 210000005036 nerve Anatomy 0.000 description 1
- 230000009466 transformation Effects 0.000 description 1
- 238000013519 translation Methods 0.000 description 1
- 230000000007 visual effect Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/20—Image preprocessing
- G06V10/26—Segmentation of patterns in the image field; Cutting or merging of image elements to establish the pattern region, e.g. clustering-based techniques; Detection of occlusion
Definitions
- the invention belongs to the field of intelligent driving, and more particularly, relates to a universal ground wire detection method.
- one aspect of the present invention is to provide a method for training a neural network model, which is characterized by:
- the training method includes the following training steps: Step 11: Obtain a sample of a road, a pair of ground points labeled with a drivable area and a dynamic object at a boundary in a road sample image;
- Step 12 input the road sample image into an initialized neural network model
- Step 13 Use the labeled road sample image training to initialize the neural network model; the loss function of the neural network is:
- L represents the loss function
- p i , t i , s i respectively represent the predicted values of the pixels at the same location in the ground point classification map, ground point distance map, and driveable area segmentation map.
- the corresponding label values are respectively;
- L cls is the loss function of the classification, preferably the cross entropy loss is used; Is to normalize all pixels involved in the calculation;
- L reg is the loss function of the regression, preferably using the mean square error,
- L seg is the cross-entropy loss function, Normalize the pixels involved in the calculation of regression and segmentation, respectively.
- ⁇ and ⁇ represent different coefficients.
- the road sample image and the labeled drivable area map are scaled to a preset size.
- the objects in the image that have overlapping parts with the road pavement, the boundary lines between the connected parts of these objects and the real road pavement are ground lines, and the two ends of the ground line are ground points.
- step 12 includes steps 121 and 122:
- Step 121 the road sample image is input into an encoder portion of the initialized neural network model
- Step 122 The image features obtained by the encoder are input to a decoder of the initialized neural network model, and a driving area segmentation result and a ground line detection result are obtained.
- the decoder includes a drivable area segmentation branch and a ground line detection branch.
- the present invention provides a neural network model obtained by using the training method described in any one of the above.
- a method for detecting a universal ground wire by using a training method of a neural network model according to any one of claims 1-5 is provided, wherein the detection method includes the following steps:
- Step 1 Single-target the image acquired by the camera device, and record and store the internal parameters and distortion parameters of the camera device;
- Step 2 The image obtained in step 1 is input into the trained neural network to obtain a driving area segmentation map, a ground point, and a ground line.
- the step 1 further includes scaling the image to a preset size using bilinear interpolation.
- the invention of the present invention lies in the following aspects, but is not limited to the following aspects:
- This method uses a pre-trained driveable area detection model to identify the current road image captured by the vehicle.
- the model can extract the current road image features and learn. While dividing the current road image into a drivable area and an obstacle area, it also detects the ground line and its corresponding object category. Subsequent cases of dynamic objects can be distinguished to generate different types of ground lines, and then the physical boundaries of the object, such as the speed of the object, can be estimated from the changes in the continuous multiple frames of the object.
- FIG. 1 is an example diagram of a structured boundary of a drivable area provided by an embodiment of the present invention
- FIG. 2 is a schematic structural diagram of a driveable area segmentation network in an encoder structure and a decoder structure in a network according to an embodiment of the present invention
- FIG. 3 is a schematic diagram of a structure for feature fusion in an encoder in a network according to an embodiment of the present invention.
- the present invention obtains a structured representation of the boundary of the drivable area in the current road image by processing the current road image.
- the vehicle can plan a driving strategy based on the structured representation.
- FIG. 1 shows an example diagram of a structured boundary of a drivable area provided by an embodiment of the present invention, and a general ground line detection result is given in the example diagram.
- the shaded area is the identified driveable area, and the other areas are non-driveable areas.
- the embodiment of the present invention also gives a structured representation.
- the boundary objects of the drivable area can be divided into static objects and dynamic objects.
- Static objects include static obstacles such as road shoulders, fences, and triangular cones.
- Dynamic objects include freely movable objects such as cars, electric vehicles, bicycles, and pedestrians.
- the embodiment of the present invention is represented by a ground line.
- the ground wire is divided into different categories according to the category of the object or different directions of the same object.
- the line segments on the ground of the electric vehicle and the wheel parts of the vehicle respectively represent the grounding wires of the electric vehicle, the grounding wires of the front and rear sides of the motor vehicle, and the grounding wires of the left and right sides of the motor vehicle.
- the two points of the ground wire are ground points. There are two types of ground points: visible ground point and speculative ground point.
- Step 1 Obtain a road sample image.
- the road sample area should be labeled with a pair of ground points of the dynamic objects at the boundary.
- the road sample image can be regarded as a sample image for training the detection model.
- the training model uses a supervised training method, so the sample images used need to have corresponding labels.
- the drivable area needs to be labeled for each pixel in the image.
- the ground points of dynamic objects at the borders of the drivable area in the sample image also need to be labeled.
- the two ground points at both ends of the ground line need to be labeled as a pair, and the type of ground point also needs to be labeled.
- the sample image is derived from a video stream collected by a camera.
- the camera needs to perform single-targeting and record the internal parameters and distortion parameters of the camera.
- the single-frame image is corrected according to the camera parameters, so that the picture is transformed into an undistorted or near-undistorted state.
- a sample library can be constructed for model training.
- Step 2 Enter the road sample pictures into the initialized neural network model.
- road sample images need to be input into the neural network.
- the road sample image needs to be scaled to a preset size.
- Step 3 Train the initialized neural network with labeled road sample images.
- a neural network is a network system formed by a large number of simple processing units that are widely connected to each other. It has a strong learning ability because it has a large number of adjustable parameters.
- the neural network model is a mathematical model based on the neural network. Based on the powerful learning ability of the neural network model, the neural network model has been widely used in many fields.
- convolutional neural network models are often used for pattern recognition. Due to the characteristics of local connection and weight sharing of convolutional layers in the convolutional neural network model, the parameters that need to be trained are greatly reduced, the network model is simplified, and the training efficiency is improved.
- a convolutional neural network can be used as the initialization neural network model.
- Features of road sample images were extracted using part of the convolutional neural network layer.
- the subsequent convolutional neural network layer maps the relevant features to obtain the recognition result of the drivable area.
- the neural network can also use these characteristics to get the information of the ground point pair.
- the results of the recognition of the drivable area output from the neural network and the detection results of the ground object point pairs of the dynamic objects at the borders of the drivable area are compared with the pre-labeled drivable areas and ground point pairs of the road sample image, so that the parameters of the initial neural network model Optimization is performed.
- a trained driving area and ground point detection model can be obtained.
- the present application provides a training method for a dynamic object ground point detection model of a drivable region and a border of the drivable region.
- the road sample images are labeled with the ground points and categories of dynamic objects in the drivable area and the boundary of the area.
- the road sample images are input into the pre-established initial neural network model.
- the road sample images are used to train the initial nerve in a supervised learning manner Network model.
- Step 1 The road sample image is input to the convolutional neural network encoder.
- the input road sample image is an RGB image.
- the convolutional neural network encoder is composed of a convolutional layer, batch normalization, a ReLU activation function, and a pooling layer.
- the convolution layer uses the same convolution kernel for different areas of an image to extract a feature of the image, such as the edge along a certain direction, and the weights are shared between different areas, which can greatly reduce the training parameters. . Further, by using a variety of convolution kernels to perform feature extraction on different regions of the image, various features of the image can be obtained.
- RelU activation function is a commonly used activation function in convolutional neural networks, which provides non-linear modeling capabilities for the entire neural network.
- the pooling layer reduces the size of the feature and reduces the amount of calculation.
- Existing convolutional neural network models include VGG Net (Visual Geometry Group), AlexNet, Network Network, ResNet deep residual network model, and so on. These networks differ in terms of network depth, calculation volume, and accuracy of feature extraction.
- the model selection of the convolutional neural network may be selected according to the computing power of the equipped equipment and the accuracy of the required drivable area and ground line detection.
- the convolutional neural network encoder in the embodiment of the present application further includes a structure for feature fusion. Unlike the classification of convolutional neural network to extract abstract features, because subsequent segmentation of driving areas and detection of ground points require more accurate positioning, the features extracted by the encoder must not only include abstract semantic features, but also some Specific details. Therefore, the convolutional neural network encoder uses a structure for fusing features of each layer. As shown in Figure 3, this structure achieves the fusion of features at different depths by adding the features of convolutional neural networks at different depths.
- Step 2 The image features obtained by the convolutional neural network encoder are input into the driving area segmentation branch of the convolutional neural network decoder to obtain the driving area segmentation result.
- the convolutional neural network decoder includes a convolutional layer, batch normalization, ReLU activation function, and an upsampling section, where the upsampling section is to expand the size of the feature map and make the final output driveable.
- the region segmentation result is the same size as the original road sample image.
- the upsampling uses the deconvolution method. Deconvolution is the process of convolution backpropagation. It can control the size of the output feature map by controlling the step size of the deconvolution. Therefore, the deconvolution operation can be used to implement upsampling.
- each pixel has a dimension of 2.
- each pixel is classified into a driveable area and a non-drivable area. Probability, the category with the larger probability is the classification of the pixel, so that the road image can be divided into driveable and non-drivable areas.
- Step 3 The image features obtained by the convolutional neural network encoder are input to the ground wire detection branch of the convolutional neural network decoder.
- the input of this branch is the same as that of the segmentation branch of the drivable area, and also has a similar structure, except that the number of output channels is different.
- Image features are convolved in this branch, normalized by batch processing, ReLU activation function and upsampling, and finally a feature map of the same size as the original picture is obtained.
- the feature map will have two convolutional layer branches.
- One of the branches obtains a score map of the same size as the original picture, and the number of channels of the score map is C + 1.
- C represents the type of ground point, including the ground points visible on the left and right sides of the vehicle, the ground points not visible on the left and right sides of the vehicle, the ground points visible from the front and rear measurements of the vehicle, the ground points not visible from the front and rear measurements of the vehicle, the ground points visible from the electric vehicle, and electric The car cannot see the ground point, etc .; 1 means the background is not the ground point.
- the scores of KxK pixels around each pixel are added to vote, and then the softmax function is input to obtain the score of the point classification.
- the other branch obtains a distance map of the same size as the original image.
- the number of channels in the distance map is 4, which represents the abscissa distance ⁇ x1 and ordinate distance ⁇ y1 of the point from the center of the ground point and the center of another ground point of the ground line.
- the center of the ground point is obtained by voting according to the distance map of pixels in the surrounding KxK range. In this way, a candidate ground point can be obtained, and then a non-maximum suppression algorithm is used to obtain a final ground point.
- the specific method is to select the grounding point with the highest score for a certain type of grounding point, and remove it when the other grounding points that are classified into this category are less than d from the grounding point with the highest score, so that the grounding point is treated. Do this for other unprocessed ground points. After all types of ground points have undergone this operation, the non-maximum suppression process is completed. This determines all ground points.
- For the detected ground point you also need to connect to determine the ground line. For a certain ground point, find another pixel according to the distance obtained by voting around KxK pixels around it. When the pixel is less than c from another ground point and the When the ground point types meet the corresponding relationship, connect the two ground points. After traversing all ground points, if there are still unconnected ground points, these points are discarded.
- the road sample image is outputted by the convolutional neural network to drive the segmentation map, ground point classification map and ground point distance map.
- the driveable area and the boundary of the driveable road sample image can be obtained. Obstacle ground wire.
- the driving area segmentation map, grounding point classification map, grounding point distance map and road sample image itself need to be compared to train the neural network. The comparison and training methods are explained in detail next.
- Step 1 Scale the road sample image and the labeled map of the drivable area to a preset size.
- the preset size is 448x448.
- the ground point segmentation map and ground point distance map are calculated according to the ground point coordinates and types in the annotation.
- Step 2 The road sample image is input to the above convolutional neural network to obtain a driving area segmentation map, a ground point classification map, and a ground point distance map. For each pixel in the drivable region segmentation map, calculate the cross-entropy loss between the labeled and drivable region segmentation map. Calculate the cross-entropy loss between the ground point classification map output from the neural network and the ground point classification map obtained from the labeled information. For a pixel marked as a ground point, calculate the mean square error loss between the distance value of the pixel in the ground point distance map and the distance value obtained according to the labeled information. These three losses are then added as a loss function for the training of the neural network.
- L represents the loss function
- p i , t i , s i represent the predicted values of the pixels at the same location in the ground point classification map, ground point distance map, and driveable area segmentation map.
- the corresponding label values are respectively;
- L cls is the loss function of the classification, here cross-entropy loss is used, Is to normalize all pixels involved in the calculation; similarly, L reg is the loss function of regression, you can use the mean square error, etc.
- L seg is also a cross entropy loss function, Normalize the pixels involved in computing regression and segmentation, respectively.
- the three loss functions are combined with different coefficients ⁇ and ⁇ .
- Step 3 In an implementation of the embodiment of the present application, a convolutional neural network initialized with labeled road sample images needs to train 16 epochs according to the above-mentioned loss function, in which the learning rate is set to 0.00001, the optimization used The algorithm is Adam's algorithm. In other methods in the embodiments of the present application, the number of trainings and the learning rate may be adjusted according to the amount of data, and the optimization method may also use other optimization methods based on gradient descent.
- Step 1 Perform a single target on the camera, record and store the internal parameters and distortion parameters of the camera. After obtaining the video stream from the camera, a single video frame is taken to correct the distortion of the video frame according to the above-mentioned calibration parameters, and then the picture is scaled to a preset size using bilinear interpolation.
- Step 2 Input the picture obtained in Step 1 into the convolutional neural network. After the convolutional neural network and the above-mentioned subsequent processing, a segmentable map of the drivable area and a grounding point of obstacles at the border of the drivable area can be obtained, as shown in FIG. 1.
- Step 3 Based on the ground lines of the dynamic objects in different video frames, the speed of the objects can be easily estimated.
- the ground line is also easily projected into 3D space to estimate the pose and distance of the object.
- the intelligent system can plan the driving route more accurately based on the structured information provided by the ground line and the information of the driveable area.
- Each module or each step of the embodiments of the present invention may be implemented by a general-purpose computing device. They may be concentrated on a single computing device or distributed on a network composed of multiple computing devices. Alternatively, they may be calculated using computing.
- the device executable program code is implemented so that they can be stored in a storage device and executed by a computing device, and in some cases, the steps shown or described can be performed in a different order than here, or They are respectively made into individual integrated circuit modules, or multiple modules or steps in them are made into a single integrated circuit module for implementation. In this way, the embodiments of the present invention are not limited to any specific combination of hardware and software.
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- General Physics & Mathematics (AREA)
- General Health & Medical Sciences (AREA)
- Molecular Biology (AREA)
- Biophysics (AREA)
- Computational Linguistics (AREA)
- Data Mining & Analysis (AREA)
- Evolutionary Computation (AREA)
- Artificial Intelligence (AREA)
- Biomedical Technology (AREA)
- Computing Systems (AREA)
- General Engineering & Computer Science (AREA)
- Life Sciences & Earth Sciences (AREA)
- Mathematical Physics (AREA)
- Software Systems (AREA)
- Health & Medical Sciences (AREA)
- Multimedia (AREA)
- Image Analysis (AREA)
Abstract
Description
Claims (7)
- 一种神经网络模型的训练方法,其特征在于:所述训练方法包括如下训练步骤:步骤11:获取道路的样本,道路样本图像中标注可行驶区域以及边界处动态物体的一对接地点;步骤12:将所述道路样本图像输入初始化的神经网络模型;步骤13:利用经过标注的道路样本图像训练初始化的神经网络模型;其中所述神经网络的损失函数为:其中,L表示损失函数,p i,t i,s i分别表示同一位置的像素分别在接地点分类图、接地点距离图和可行驶区域分割图中的预测值, 分别为相应的label值;L cls是分类的损失函数,优选使用交叉熵损失; 是对所有参与计算的像素点做归一化;L reg是回归的损失函数,优选使用均方误差,L seg是交叉熵损失函数, 分别是对参与计算回归和分割的像素点做归一化,λ,γ表示不同的系数;所述图像中与道路路面具有重叠部分的物体,这些物体与真实道路路面连接部分的边界线为接地线,所述接地线的两端点为接地点。
- 如权利要求1所述的训练方法,其特征在于,所述步骤12包括步骤121和步骤122:步骤121:将所述道路样本图像输入所述初始化的神经网络模型的编码器部分;步骤122:将所述编码器获得的图像特征输入到所述初始化的神经网络模型的解码器,获得可行驶区域分割结果和接地线检测结果。
- 如权利要求2所述的训练方法,其特征在于,所述解码器包括可行驶区域分割分支和接地线检测分支。
- 如权利要求1所述的训练方法,其特征在于,在步骤11中,将所述道路样本图像和标注的可行驶区域图缩放至预设尺寸。
- 一种神经网络模型,其采用权利要求1-4中任一项所述的训练方法得到。
- 利用权利要求1-4中任一项的神经网络模型的训练方法检测通用接地线的方法,其特征在于,检测方法包括以下步骤:步骤1:对摄像装置获取的图像进行单目标定,记录并存储所述摄像装置的内部参数和畸变参数;步骤2:将步骤1中获得的所述图像输入到所述训练好的所述神经网络中,得到可行驶区域分割图以及接地点、接地线。
- 如权利要求6所述的检测通用接地线的方法,其特征在于,在所述步骤1中还包括使用双线性插值将所述图像缩放为预设尺寸。
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201811143635.1A CN109726627B (zh) | 2018-09-29 | 2018-09-29 | 一种神经网络模型训练及通用接地线的检测方法 |
CN201811143635.1 | 2018-09-29 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2020062433A1 true WO2020062433A1 (zh) | 2020-04-02 |
Family
ID=66295410
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/CN2018/113661 WO2020062433A1 (zh) | 2018-09-29 | 2018-11-02 | 一种神经网络模型训练及通用接地线的检测方法 |
Country Status (2)
Country | Link |
---|---|
CN (1) | CN109726627B (zh) |
WO (1) | WO2020062433A1 (zh) |
Cited By (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111597913A (zh) * | 2020-04-23 | 2020-08-28 | 浙江大学 | 一种基于语义分割模型的车道线图片检测分割方法 |
CN111968088A (zh) * | 2020-08-14 | 2020-11-20 | 西安电子科技大学 | 一种基于像素和区域分割决策融合的建筑物检测方法 |
CN112036231A (zh) * | 2020-07-10 | 2020-12-04 | 武汉大学 | 一种基于车载视频的车道线和路面指示标志检测与识别方法 |
CN112949097A (zh) * | 2021-04-19 | 2021-06-11 | 合肥工业大学 | 一种基于深度迁移学习的轴承剩余寿命预测模型和方法 |
CN113781492A (zh) * | 2020-06-10 | 2021-12-10 | 阿里巴巴集团控股有限公司 | 目标元素含量测量方法、训练方法、相关装置及存储介质 |
CN114648534A (zh) * | 2022-05-24 | 2022-06-21 | 成都理工大学 | 基于视频帧聚类的管网缺陷智能识别方法及装置、介质 |
CN116400605A (zh) * | 2023-06-08 | 2023-07-07 | 成都航空职业技术学院 | 一种机器人自动控制方法及系统 |
CN116863429A (zh) * | 2023-07-26 | 2023-10-10 | 小米汽车科技有限公司 | 检测模型的训练方法、可行使区域的确定方法和装置 |
CN117523318A (zh) * | 2023-12-26 | 2024-02-06 | 宁波微科光电股份有限公司 | 一种抗光干扰的地铁屏蔽门异物检测方法、装置及介质 |
Families Citing this family (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110956069B (zh) * | 2019-05-30 | 2022-06-21 | 魔门塔(苏州)科技有限公司 | 一种行人3d位置的检测方法及装置、车载终端 |
CN110705470A (zh) * | 2019-09-30 | 2020-01-17 | 的卢技术有限公司 | 一种实时检测可行使路面的方法及系统 |
CN110956214B (zh) * | 2019-12-03 | 2023-10-13 | 北京车和家信息技术有限公司 | 一种自动驾驶视觉定位模型的训练方法及装置 |
CN111368794B (zh) * | 2020-03-19 | 2023-09-19 | 北京百度网讯科技有限公司 | 障碍物检测方法、装置、设备和介质 |
CN111950478B (zh) * | 2020-08-17 | 2021-07-23 | 浙江东鼎电子股份有限公司 | 一种动态平板秤称重区域汽车s型行驶行为检测方法 |
CN113239960B (zh) * | 2021-04-09 | 2024-05-28 | 中用科技有限公司 | 融合ai视觉算法的道路防护智能预警方法和系统 |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN106815566A (zh) * | 2016-12-29 | 2017-06-09 | 天津中科智能识别产业技术研究院有限公司 | 一种基于多任务卷积神经网络的人脸检索方法 |
JP2018060296A (ja) * | 2016-10-03 | 2018-04-12 | グローリー株式会社 | 画像処理装置、画像処理システム及び画像処理方法 |
CN108171136A (zh) * | 2017-12-21 | 2018-06-15 | 浙江银江研究院有限公司 | 一种多任务卡口车辆以图搜图的系统及方法 |
CN108230329A (zh) * | 2017-12-18 | 2018-06-29 | 孙颖 | 基于多尺度卷积神经网络的语义分割方法 |
CN108550259A (zh) * | 2018-04-19 | 2018-09-18 | 何澜 | 道路拥堵判断方法、终端设备及计算机可读存储介质 |
Family Cites Families (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN108256413B (zh) * | 2017-11-27 | 2022-02-25 | 科大讯飞股份有限公司 | 可通行区域检测方法及装置、存储介质、电子设备 |
CN108345875B (zh) * | 2018-04-08 | 2020-08-18 | 北京初速度科技有限公司 | 可行驶区域检测模型训练方法、检测方法和装置 |
-
2018
- 2018-09-29 CN CN201811143635.1A patent/CN109726627B/zh active Active
- 2018-11-02 WO PCT/CN2018/113661 patent/WO2020062433A1/zh active Application Filing
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2018060296A (ja) * | 2016-10-03 | 2018-04-12 | グローリー株式会社 | 画像処理装置、画像処理システム及び画像処理方法 |
CN106815566A (zh) * | 2016-12-29 | 2017-06-09 | 天津中科智能识别产业技术研究院有限公司 | 一种基于多任务卷积神经网络的人脸检索方法 |
CN108230329A (zh) * | 2017-12-18 | 2018-06-29 | 孙颖 | 基于多尺度卷积神经网络的语义分割方法 |
CN108171136A (zh) * | 2017-12-21 | 2018-06-15 | 浙江银江研究院有限公司 | 一种多任务卡口车辆以图搜图的系统及方法 |
CN108550259A (zh) * | 2018-04-19 | 2018-09-18 | 何澜 | 道路拥堵判断方法、终端设备及计算机可读存储介质 |
Cited By (15)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111597913A (zh) * | 2020-04-23 | 2020-08-28 | 浙江大学 | 一种基于语义分割模型的车道线图片检测分割方法 |
CN113781492A (zh) * | 2020-06-10 | 2021-12-10 | 阿里巴巴集团控股有限公司 | 目标元素含量测量方法、训练方法、相关装置及存储介质 |
CN112036231A (zh) * | 2020-07-10 | 2020-12-04 | 武汉大学 | 一种基于车载视频的车道线和路面指示标志检测与识别方法 |
CN112036231B (zh) * | 2020-07-10 | 2022-10-21 | 武汉大学 | 一种基于车载视频的车道线和路面指示标志检测与识别方法 |
CN111968088A (zh) * | 2020-08-14 | 2020-11-20 | 西安电子科技大学 | 一种基于像素和区域分割决策融合的建筑物检测方法 |
CN111968088B (zh) * | 2020-08-14 | 2023-09-15 | 西安电子科技大学 | 一种基于像素和区域分割决策融合的建筑物检测方法 |
CN112949097B (zh) * | 2021-04-19 | 2022-09-16 | 合肥工业大学 | 一种基于深度迁移学习的轴承剩余寿命预测模型和方法 |
CN112949097A (zh) * | 2021-04-19 | 2021-06-11 | 合肥工业大学 | 一种基于深度迁移学习的轴承剩余寿命预测模型和方法 |
CN114648534A (zh) * | 2022-05-24 | 2022-06-21 | 成都理工大学 | 基于视频帧聚类的管网缺陷智能识别方法及装置、介质 |
CN116400605A (zh) * | 2023-06-08 | 2023-07-07 | 成都航空职业技术学院 | 一种机器人自动控制方法及系统 |
CN116400605B (zh) * | 2023-06-08 | 2023-08-11 | 成都航空职业技术学院 | 一种机器人自动控制方法及系统 |
CN116863429A (zh) * | 2023-07-26 | 2023-10-10 | 小米汽车科技有限公司 | 检测模型的训练方法、可行使区域的确定方法和装置 |
CN116863429B (zh) * | 2023-07-26 | 2024-05-31 | 小米汽车科技有限公司 | 检测模型的训练方法、可行使区域的确定方法和装置 |
CN117523318A (zh) * | 2023-12-26 | 2024-02-06 | 宁波微科光电股份有限公司 | 一种抗光干扰的地铁屏蔽门异物检测方法、装置及介质 |
CN117523318B (zh) * | 2023-12-26 | 2024-04-16 | 宁波微科光电股份有限公司 | 一种抗光干扰的地铁屏蔽门异物检测方法、装置及介质 |
Also Published As
Publication number | Publication date |
---|---|
CN109726627B (zh) | 2021-03-23 |
CN109726627A (zh) | 2019-05-07 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2020062433A1 (zh) | 一种神经网络模型训练及通用接地线的检测方法 | |
WO2020244653A1 (zh) | 物体识别方法及装置 | |
CN110084850B (zh) | 一种基于图像语义分割的动态场景视觉定位方法 | |
WO2020151109A1 (zh) | 基于点云带权通道特征的三维目标检测方法及系统 | |
CN111311666B (zh) | 一种融合边缘特征和深度学习的单目视觉里程计方法 | |
Kong et al. | Vanishing point detection for road detection | |
CN111612008B (zh) | 基于卷积网络的图像分割方法 | |
CN111784747B (zh) | 一种基于关键点检测和校正的车辆多目标跟踪系统及方法 | |
WO2015010451A1 (zh) | 一种从单幅图像检测道路的方法 | |
CN111291714A (zh) | 一种基于单目视觉和激光雷达融合的车辆检测方法 | |
CN110728209A (zh) | 一种姿态识别方法、装置、电子设备及存储介质 | |
CN107545263B (zh) | 一种物体检测方法及装置 | |
CN109886200B (zh) | 一种基于生成式对抗网络的无人驾驶车道线检测方法 | |
CN111696110B (zh) | 场景分割方法及系统 | |
CN110363160B (zh) | 一种多车道线识别方法及装置 | |
CN106778668A (zh) | 一种联合ransac和cnn的鲁棒的车道线检测方法 | |
CN112949493A (zh) | 一种结合语义分割和注意力机制的车道线检测方法及系统 | |
CN114429457A (zh) | 一种基于双模态融合的风机叶片缺陷智能检测方法 | |
Saleem et al. | Steering angle prediction techniques for autonomous ground vehicles: a review | |
Muthalagu et al. | Vehicle lane markings segmentation and keypoint determination using deep convolutional neural networks | |
CN115018999A (zh) | 一种多机器人协作的稠密点云地图构建方法及装置 | |
CN113420648B (zh) | 一种具有旋转适应性的目标检测方法及系统 | |
Nguyen et al. | UnfairGAN: An enhanced generative adversarial network for raindrop removal from a single image | |
CN112053385B (zh) | 基于深度强化学习的遥感视频遮挡目标跟踪方法 | |
Jin et al. | Road curvature estimation using a new lane detection method |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 18934566 Country of ref document: EP Kind code of ref document: A1 |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 18934566 Country of ref document: EP Kind code of ref document: A1 |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 18934566 Country of ref document: EP Kind code of ref document: A1 |
|
32PN | Ep: public notification in the ep bulletin as address of the adressee cannot be established |
Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205A DATED 08.11.2021) |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 18934566 Country of ref document: EP Kind code of ref document: A1 |