CN111597897B - High-speed service area parking space recognition method - Google Patents

High-speed service area parking space recognition method Download PDF

Info

Publication number
CN111597897B
CN111597897B CN202010297837.2A CN202010297837A CN111597897B CN 111597897 B CN111597897 B CN 111597897B CN 202010297837 A CN202010297837 A CN 202010297837A CN 111597897 B CN111597897 B CN 111597897B
Authority
CN
China
Prior art keywords
representing
network
layer
vehicle
loss function
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202010297837.2A
Other languages
Chinese (zh)
Other versions
CN111597897A (en
Inventor
邵奇可
卢熠
颜世航
陈一苇
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Zhejiang University of Technology ZJUT
Original Assignee
Zhejiang University of Technology ZJUT
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Zhejiang University of Technology ZJUT filed Critical Zhejiang University of Technology ZJUT
Priority to CN202010297837.2A priority Critical patent/CN111597897B/en
Publication of CN111597897A publication Critical patent/CN111597897A/en
Application granted granted Critical
Publication of CN111597897B publication Critical patent/CN111597897B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/10Terrestrial scenes
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • G06F18/2413Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches based on distances to training or reference patterns
    • G06F18/24133Distances to prototypes
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02TCLIMATE CHANGE MITIGATION TECHNOLOGIES RELATED TO TRANSPORTATION
    • Y02T10/00Road transport of goods or passengers
    • Y02T10/10Internal combustion engine [ICE] based vehicles
    • Y02T10/40Engine management systems

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • General Physics & Mathematics (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Artificial Intelligence (AREA)
  • General Engineering & Computer Science (AREA)
  • Evolutionary Computation (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • Software Systems (AREA)
  • Mathematical Physics (AREA)
  • Health & Medical Sciences (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • Computational Linguistics (AREA)
  • General Health & Medical Sciences (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Evolutionary Biology (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Multimedia (AREA)
  • Image Analysis (AREA)
  • Traffic Control Systems (AREA)
  • Management, Administration, Business Operations System, And Electronic Commerce (AREA)

Abstract

The high-speed service area parking space identification method comprises the following steps: 1) And acquiring a large number of images of high-altitude cameras and other vehicle data sets in the parking lot, calibrating the data sets according to the field management requirements, and determining a used one-stage target detection algorithm model. 2) Constructing a parameter-adaptive loss functionAnd3) And constructing a LOSS function LOSS of the one-stage target detection algorithm model. 4) Updating the weight of the one-stage target detection algorithm model by adopting a gradient descent method until the model converges; and finishing the detection of vehicles by the trained model in an actual system, and calculating the number of the residual parking spaces according to the total number of the preset on-site parking spaces and the current number of the vehicles to realize the management of the parking spaces. The focus loss function provided by the invention has the advantages that the parameter self-adaptability of the target detection model can be improved, and the accuracy of target detection is greatly improved.

Description

High-speed service area parking space recognition method
Technical Field
The invention belongs to the technical field of image recognition and computer vision, and relates to a parking space recognition method for a high-speed service area.
Background
At present, aiming at the problem of detecting a parking space in a high-speed service area, the traditional detection method mainly comprises the following steps: micro-radar detection, infrared detection, geomagnetic induction coil detection and radio frequency identification technology. The method needs to install special sensing equipment for each parking space of the parking lot in the high-speed service area, engineering cost is high, later maintenance is difficult, and the cost of manpower and material resources needed to be input is high. The parking space state is identified in real time by utilizing a security camera in the existing high-speed service area parking lot, and then parking space information of the area is counted. The existing parking lot monitoring equipment is utilized, the ground of the parking space of the parking lot is not required to be changed, and equipment maintenance and repair are easy, so that the video-based parking space detection system has good popularization value.
The video stream of the security camera is utilized to identify the parking space state, and the accuracy of an identification algorithm and the real-time requirement on the empty parking space information in an application scene are high. Therefore, the target detection algorithm based on deep learning is reasonable. The target detection algorithm based on deep learning is divided into a two-stage model and a one-stage model. Although the two-stage convolutional neural network model has better detection precision, the forward reasoning speed is lower, and the real-time requirement of a service scene cannot be met. In the traditional one-stage target detection algorithm model, the real-time performance of the algorithm is good, but the detection accuracy of the two-stage convolutional neural network model cannot be achieved. The high-speed service area parking space identification method based on the focus loss function parameter self-adaption is beneficial to improving the detection precision of the system and ensuring the real-time performance of the system to meet the requirements of application scenes.
Disclosure of Invention
The invention aims to overcome the defects in the prior art and provides a high-speed service area parking space identification method so as to improve detection accuracy and instantaneity.
The invention improves the loss function in the one-stage target detection algorithm model. The loss function is used as an objective function of the gradient descent process in the convolutional neural network, and directly influences the training result of the convolutional neural network. The quality of the result of the convolutional neural network training directly relates to the recognition precision of target detection, so that the convolutional neural network training is particularly important for the design appearance of the loss function.
In the one-stage target detection algorithm model training process, a network contains a large number of service area background objects when the target is detected by an image, and the loss value of the service area background objects is very small, but the number of the service area background objects is far more than that of the vehicle target, so that the service area background loss value with a small probability value overwhelms the target loss value of the vehicle when the loss value is calculated, the model precision is greatly reduced, and therefore a focus loss function is embedded in the one-stage target detection algorithm model to improve the training precision. The super parameters in the focus loss function need to be set according to the experience value, and the self super parameters cannot be automatically adjusted according to the predicted class probability value.
Therefore, aiming at the problem that the focus loss function needs to be manually adjusted to the super-parameters in the training process, the parameter in the training process does not have self-adaptability, the invention provides the deep learning loss function based on semi-supervised learning, and the loss function uses a weighting method to improve the super-parameters, so that the network can be adaptively adjusted to the super-parameters in the gradient descent process of the network, and the network learning efficiency is further improved.
In order to solve the technical problems, the focus loss function with self-adaptive parameters is adopted to strengthen the network training capacity, and the recognition accuracy of the system is provided.
The high-speed service area parking space identification method comprises the following steps:
step 1: constructing a high-speed service area parking lot data set M, a training data set T, a verification data set V, a labeling vehicle category number C, a training data batch size batch, a training batch number batches, a learning rate l_rate and a proportionality coefficient zeta between the training data set T and the verification data set V.
Wherein: v.u.T=M, C.epsilon.N + ,ζ∈(0,1),batches∈N + ,l_rate∈N + ,batch∈N +Representing the height and width of the image, and r represents the number of channels of the image.
Step 2: determining a target detection model to be trained at one stage, setting the depth of a convolutional neural network as L, a convolutional kernel set G of a network convolutional layer, adopting a fully-connected mode for a network output layer, a convolutional kernel set A, a network feature map set U,representing the kth feature map in the layer 1 network +.>The corresponding grid number and anchor point set M are specifically defined as follows:
wherein:respectively represent the convolution kernel, the feature map and the anchor point corresponding to the layer I networkWide, dimension. />Representing the fill size of the layer 1 network convolution kernel,/->Represents the layer I network convolution step length, f represents the excitation function of the convolution neuron, Θ represents the selected input feature, Λ epsilon N + Representing the total number of anchor points of the layer I network, and the element E N + Represents the total number of output layer nodes, phi epsilon N + Representing the total number of network feature graphs of the layer I, delta epsilon N + Representing the total number of layer i convolution kernels.
Step 3: the focus loss function of the design parameter adaptation is as follows:
wherein:
representing a jth anchor point in an ith grid on a layer I network at image t k A loss function of confidence of the vehicle sample and the parking lot background sample; similarly, let go of>Representing loss of a vehicle prediction frameLoss function (I)>Loss function representing class of vehicle, lambda being the loss function +.>Parameters. />And->The loss functions respectively representing the vehicle target and the parking lot background target are specifically as follows:
a foreground vehicle probability value representing a prediction of a jth anchor point in an ith mesh on the layer i network, and similarly,representing the corresponding parking lot background probability value. />Respectively representing the abscissa and the ordinate of the central point of the prediction frame of the jth anchor point in the ith grid on the first layer network, and the same thing is->Respectively representing the abscissa and the ordinate of the central point of the vehicle sample calibration frame; />Respectively representing the shortest Euclidean distance from the predicted frame center point of the jth anchor point in the ith grid to the frame boundary on the first layer network, and the same is ≡>Respectively representing the shortest Euclidean distance from the center point of a vehicle sample calibration frame to the boundary of the frame; />Representing predicted values of the vehicle categories predicted by the j-th anchor point in the i-th grid on the layer-I network. Similarly, let go of>Indicating the calibration status of the vehicle class>Representing a vehicle sample for prediction +.>The method is used for indicating whether the parking lot background sample is predicted or not, and specifically calculated as follows:
wherein the parameter α ε (0, 1); iou (iou) j Representing anchor point m j The overlap ratio of the anchor point frame and the vehicle calibration frame in the ith grid, miou represents the maximum overlap ratio.
Step 4: based on a stage in Step 3And (3) carrying out gradient descent training on the model by using a training set until the model converges. In the model test stage, the total number of parking spaces is set as sum epsilon N + Outputting a test sample of the current video monitoring area to perform target detection, and recording num epsilon N + Indicating the number of vehicles in the parking lot, the empty space s_num=sum-num.
The invention has the advantages that: the provided focus loss function can improve the parameter adaptability of the target detection model, improve the detection precision of the system and ensure that the real-time performance of the system meets the requirements of application scenes.
Drawings
Fig. 1 is a network configuration diagram of a convolutional neural network of the present invention.
Fig. 2 is a block diagram of a loss function in a convolutional neural network of the present invention.
Fig. 3 is a flow chart of deployment of a parking space detection algorithm based on a convolutional neural network.
Detailed Description
In order to better explain the technical scheme of the invention, the invention is further described by an embodiment example with reference to the attached drawings.
The high-speed service area parking space identification method comprises the following steps:
step 1: collecting a large number of image data shot by high-altitude cameras, constructing a high-speed service area parking lot data set M, wherein the number of the training data sets T is 10000, the number of the verification data sets V is 2000, the number of marked vehicle categories C is 5, the marked vehicle categories C are respectively a car, an off-road car, a large truck, a police car and an engineering maintenance car, the number of training data batches is 4, the number of the training batches is 1000, the learning rate l_rate is 0.001, the proportional coefficient zeta between the training data sets T and the verification data sets V is 0.25, and the high h of the image k =416,w k =416, r=3 and satisfies the consistency of the height, width, channel number settings for all images.
Step 2: determining a one-stage target detection model as Yolov3, setting a convolutional neural network depth L as 139, wherein the height, width and dimension of a convolutional kernel are setAs shown in fig. 1, the fill size of the convolution kernelDefault to 1, convolution step +.>Default to 1, the excitation function f of the convolutional neuron defaults to the leakage_relu excitation function; anchors are shared in each layer of network, and the anchor set M takes values of { (10, 13), (30,61), (156,198) }, Λ=3; the network output layer adopts a full connection mode, and the convolution kernel set A of the network output layer takes values of { (1,1,30), (1,1,30), (1,1,30) } and xi=3.
Step 3: as shown in fig. 2, a parameter-adaptive focus LOSS function LOSS is constructed, the value of the parameter α is 0.25, and the value of the parameter λ is 0.5.
Step 4: and training the model by using a gradient descent method based on a loss function of the one-stage target detection algorithm model in Step 3 until the model converges. As shown in fig. 3, the video stream of the camera placed in the parking lot is used for real-time detection, the sum of parking spaces is set to be 10, a test sample of the current video monitoring area is output for target detection, and the remaining parking spaces are calculated according to the detected number of vehicles and the sum of the parking spaces, so that the management of the parking spaces is realized.
The embodiments described in the present specification are merely examples of implementation forms of the inventive concept, and the scope of protection of the present invention should not be construed as being limited to the specific forms set forth in the embodiments, and the scope of protection of the present invention and equivalent technical means that can be conceived by those skilled in the art based on the inventive concept.

Claims (1)

1. The high-speed service area parking space identification method comprises the following steps:
step 1: constructing a high-speed service area parking lot data set M', a training data set T, a verification data set V, a labeling vehicle class number C, a training data batch size batch, a training batch number batches, a learning rate l_rate, and a proportionality coefficient zeta between the training data set T and the verification data set V;
wherein: v ∈T=M', C∈N + ,ζ∈(0,1),batches∈N + ,l_rate∈N + ,batch∈N +Representing the height and width of an image, r representing the number of channels of the image;
step 2: determining a target detection model to be trained at one stage, setting the depth of a convolutional neural network as L, a convolutional kernel set G of a network convolutional layer, adopting a fully-connected mode for a network output layer, a convolutional kernel set A, a network feature map set U,representing the kth feature map in the layer 1 network +.>The corresponding grid quantity and anchor point set M specifically comprise:
wherein:representing the height, width, dimension of the convolution kernel corresponding to the layer 1 network, < >>Representing the height, width, dimension, +.>Representing the height and width of the anchor point; />Representing the fill size of the layer 1 network convolution kernel,/->Represents the convolution step length of the layer I network, f represents the excitation function of the convolution neuron, Θ represents the selected input characteristic, L epsilon N + Representing the total number of anchor points of the layer I network, and the element E N + Represents the total number of output layer nodes, phi epsilon N + Representing the total number of network feature graphs of layer I, delta epsilon N + Representing the total number of layer i convolution kernels;
step 3: the focus loss function with adaptive design parameters specifically comprises:
wherein:
representing a jth anchor point in an ith grid on a layer I network at image t k A loss function of confidence of the vehicle sample and the parking lot background sample; similarly, let go of>Loss function representing a prediction box of a vehicle, +.>A loss function representing the class of the vehicle, lambda.epsilon.Q is the loss function +.>Parameters; />And->The loss functions respectively representing the vehicle target and the parking lot background target are specifically as follows:
a foreground vehicle probability value representing a prediction of a jth anchor point in an ith mesh on the layer i network, and similarly,representing a corresponding parking lot background probability value; />Respectively representing the abscissa and the ordinate of the central point of the prediction frame of the jth anchor point in the ith grid on the first layer network, and the same thing is->Respectively representing the abscissa and the ordinate of the central point of the vehicle sample calibration frame; />Respectively representing the shortest Euclidean distance from the center point of a calibration frame of the jth anchor point in the ith grid to the boundary of the frame on the first layer network, and the same thing is ≡>Respectively representing the shortest Euclidean distance from the central point of a vehicle sample prediction frame to the boundary of the frame; />A predicted value of the vehicle category, which represents the prediction of the jth anchor point in the ith grid on the first network; similarly, let go of>Indicating the calibration status of the vehicle class>Indicating whether a prediction is made for a vehicle sample, +.>The method is used for indicating whether the parking lot background sample is predicted or not, and specifically calculated as follows:
wherein the parameter α ε (0, 1); iou (iou) j Representing anchor point m j In the ith grid, the overlapping rate of the anchor point frame and the vehicle calibration frame, wherein miou represents the maximum overlapping rate;
step 4: training the model by using a loss function of a one-stage target detection algorithm model in Step 3 by using a gradient descent method until the model converges; in the system operation stage, extracting network characteristic values by using a first-order target detection model, determining anchor points based on a K-means clustering method, and setting the total number of parking spaces as sum epsilon N + Outputting the number num epsilon N of target detection vehicles in the current video monitoring area + Then the empty space s_num=sum-num.
CN202010297837.2A 2020-04-16 2020-04-16 High-speed service area parking space recognition method Active CN111597897B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202010297837.2A CN111597897B (en) 2020-04-16 2020-04-16 High-speed service area parking space recognition method

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202010297837.2A CN111597897B (en) 2020-04-16 2020-04-16 High-speed service area parking space recognition method

Publications (2)

Publication Number Publication Date
CN111597897A CN111597897A (en) 2020-08-28
CN111597897B true CN111597897B (en) 2023-10-24

Family

ID=72187569

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202010297837.2A Active CN111597897B (en) 2020-04-16 2020-04-16 High-speed service area parking space recognition method

Country Status (1)

Country Link
CN (1) CN111597897B (en)

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112686340B (en) * 2021-03-12 2021-07-13 成都点泽智能科技有限公司 Dense small target detection method based on deep neural network

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9916522B2 (en) * 2016-03-11 2018-03-13 Kabushiki Kaisha Toshiba Training constrained deconvolutional networks for road scene semantic segmentation
CN109902677A (en) * 2019-01-30 2019-06-18 深圳北斗通信科技有限公司 A kind of vehicle checking method based on deep learning
WO2019144575A1 (en) * 2018-01-24 2019-08-01 中山大学 Fast pedestrian detection method and device
CN110443208A (en) * 2019-08-08 2019-11-12 南京工业大学 A kind of vehicle target detection method, system and equipment based on YOLOv2

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9916522B2 (en) * 2016-03-11 2018-03-13 Kabushiki Kaisha Toshiba Training constrained deconvolutional networks for road scene semantic segmentation
WO2019144575A1 (en) * 2018-01-24 2019-08-01 中山大学 Fast pedestrian detection method and device
CN109902677A (en) * 2019-01-30 2019-06-18 深圳北斗通信科技有限公司 A kind of vehicle checking method based on deep learning
CN110443208A (en) * 2019-08-08 2019-11-12 南京工业大学 A kind of vehicle target detection method, system and equipment based on YOLOv2

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
Multi-resolution Networks for Ship Detection in Infrared Remote Sensing Images;Min Zhou et al.;《Infrared Physics & Technology》;20180831;全文 *
基于深度学习的高速服务区车位检测算法;邵奇可等;《计算机系统应用》;20190525;第28卷(第6期);全文 *

Also Published As

Publication number Publication date
CN111597897A (en) 2020-08-28

Similar Documents

Publication Publication Date Title
CN112380952B (en) Power equipment infrared image real-time detection and identification method based on artificial intelligence
CN111597902B (en) Method for monitoring motor vehicle illegal parking
CN109241817B (en) Crop image recognition method shot by unmanned aerial vehicle
CN111079640B (en) Vehicle type identification method and system based on automatic amplification sample
CN111597901A (en) Illegal billboard monitoring method
CN111709336B (en) Expressway pedestrian detection method, equipment and readable storage medium
CN110175615B (en) Model training method, domain-adaptive visual position identification method and device
CN112232151B (en) Iterative polymerization neural network high-resolution remote sensing scene classification method embedded with attention mechanism
CN111126332B (en) Frequency hopping signal classification method based on contour features
CN111723854B (en) Expressway traffic jam detection method, equipment and readable storage medium
CN112435356B (en) ETC interference signal identification method and detection system
CN112749654A (en) Deep neural network model construction method, system and device for video fog monitoring
CN113034378B (en) Method for distinguishing electric automobile from fuel automobile
CN111027347A (en) Video identification method and device and computer equipment
CN111597897B (en) High-speed service area parking space recognition method
CN111597900B (en) Illegal dog walking identification method
CN111239137B (en) Grain quality detection method based on transfer learning and adaptive deep convolution neural network
CN113139594A (en) Airborne image unmanned aerial vehicle target self-adaptive detection method
CN114299567B (en) Model training method, living body detection method, electronic device, and storage medium
CN114549909A (en) Pseudo label remote sensing image scene classification method based on self-adaptive threshold
CN112348820B (en) Remote sensing image semantic segmentation method based on depth discrimination enhancement network
CN116630828B (en) Unmanned aerial vehicle remote sensing information acquisition system and method based on terrain environment adaptation
CN108229271B (en) Method and device for interpreting remote sensing image and electronic equipment
CN110674689B (en) Vehicle re-identification method and system based on feature embedding space geometric constraint
CN116977834A (en) Method for identifying internal and external images distributed under open condition

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant