CN109522854B - Pedestrian traffic statistical method based on deep learning and multi-target tracking - Google Patents
Pedestrian traffic statistical method based on deep learning and multi-target tracking Download PDFInfo
- Publication number
- CN109522854B CN109522854B CN201811400758.9A CN201811400758A CN109522854B CN 109522854 B CN109522854 B CN 109522854B CN 201811400758 A CN201811400758 A CN 201811400758A CN 109522854 B CN109522854 B CN 109522854B
- Authority
- CN
- China
- Prior art keywords
- pedestrian
- coordinates
- current
- current pedestrian
- frame
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/50—Context or environment of the image
- G06V20/52—Surveillance or monitoring of activities, e.g. for recognising suspicious objects
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/20—Image preprocessing
- G06V10/30—Noise filtering
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/40—Extraction of image or video features
- G06V10/46—Descriptors for shape, contour or point-related descriptors, e.g. scale invariant feature transform [SIFT] or bags of words [BoW]; Salient regional features
- G06V10/462—Salient features, e.g. scale invariant feature transforms [SIFT]
Abstract
The invention relates to the technical field of image processing, and aims to provide a pedestrian flow statistical method based on deep learning and multi-target tracking. The invention mainly comprises the following steps: s1: shooting a pedestrian monitoring video and reading an image in the video; s2: setting an effective area and flow count of the image; s3: constructing a pedestrian detection model based on deep learning and training the pedestrian detection model; s4: detecting the current pedestrian to obtain the coordinate and the image block of the current pedestrian frame; s5: tracking the current pedestrian by using a multi-target tracking algorithm based on deep learning, and generating the coordinate of the current pedestrian; s6: generating a moving track of the current pedestrian; s7: judging whether the current pedestrian leaves the effective area; if yes, go to step S8, otherwise go to step S4; s8: selecting a noise threshold value and judging noise; s9: and deleting the coordinates of the current pedestrian in the continuous video frames. The invention can provide an accurate flow statistic result in an actual use scene.
Description
Technical Field
The invention relates to the technical field of image processing, in particular to a pedestrian flow statistical method based on deep learning and multi-target tracking.
Background
The popularization of the monitoring camera and the development of the image processing technology provide good conditions for intelligent monitoring. Pedestrian flow statistics is widely applied to intelligent monitoring, and can be used in various occasions such as hospitals, channels, shops and the like. The accurate flow information can help to carry out reasonable resource allocation, estimate the rent level of the shop, the operating condition and the like, and has great significance.
At present, there are many people flow rate statistical methods, one is relying on hardware sensor equipment, and the other is directly processing video. Wherein the sensor is greatly influenced by the density of the stream of people; the methods for carrying out people flow statistics on video processing mainly include the following methods: a. a face recognition based method that is affected by face occlusion and head pose; b. according to the method based on human head detection and tracking, a camera is placed at the top, requirements are placed on the installation environment, meanwhile, cap shielding can generate large influence, and a tracking algorithm is simple and easy to cause tracking errors; c. based on a head and shoulder detection and tracking method, a tracking algorithm is simple, but the pedestrian density is high, and errors are easy to occur; d. in the method, the pedestrian is judged by adopting clothes or some traditional characteristics, so that a large error is easy to generate; e. a multi-camera based approach requires multiple cameras to match count the pedestrians. In addition, the people flow counting mode basically accumulates when the target moving track crosses the auxiliary line, and due to the fact that the adopted tracking algorithm is simple, tracking is disordered when a plurality of people exist, the track is disordered, counting is affected, and the pedestrian flow counting accuracy is low.
Disclosure of Invention
The invention provides a pedestrian flow statistical method based on deep learning and multi-target tracking, which can provide accurate flow statistical results in practical use scenes.
The technical scheme adopted by the invention is as follows:
a pedestrian flow statistical method based on deep learning and multi-target tracking comprises the following steps:
s1: shooting a pedestrian monitoring video in real time, and reading images in continuous video frames of the pedestrian monitoring video;
s2: setting an effective area of images in continuous video frames, and setting a flow count with an initial value of 0;
s3: constructing a pedestrian detection model based on deep learning and training the pedestrian detection model;
s4: zooming the image with the effective area by using the trained pedestrian detection model, and then detecting the current pedestrian of the image to obtain the coordinate and the image block of the current pedestrian frame, wherein the current pedestrian frame is used for indicating the area with the current pedestrian in the image;
s5: tracking the current pedestrian in real time by using a multi-target tracking algorithm based on deep learning, and generating coordinates of the current pedestrian in a continuous video frame image;
s6: generating a moving track of the current pedestrian according to the coordinates of the current pedestrian in the images of the continuous video frames;
s7: judging whether the current pedestrian leaves the effective area in real time by utilizing a multi-target tracking algorithm according to the moving track of the current pedestrian; if yes, go to step S8, otherwise go to step S4;
s8: selecting a noise threshold value and judging the noise of the moving track of the current pedestrian; if the number of coordinates of the moving track of the current pedestrian is smaller than the noise threshold value, judging that the moving track of the current pedestrian is a noise track; if the number of coordinates of the moving track of the current pedestrian is larger than or equal to the noise threshold, judging that the moving track of the current pedestrian is an effective track, and adding 1 to the numerical value of the flow meter number;
s9: the coordinates of the current pedestrian in the consecutive video frames are deleted and then step S4 is repeated.
Preferably, in the step S3, the pedestrian detection model is trained by using a YOLOv3 network structure.
Preferably, the specific steps of step S5 are as follows:
s501: reading the coordinates and image blocks of the current pedestrian frame;
s502: constructing a pedestrian feature extraction model based on deep learning and training the pedestrian feature extraction model, and performing depth apparent feature extraction on image blocks of a current pedestrian frame by using the trained deep network pedestrian feature extraction model to generate depth apparent features of different current pedestrian frames in continuous video frames;
s503: according to the coordinates of the current pedestrian frame, the prediction coordinates of the current pedestrian and the updating coordinates of the current pedestrian are calculated by using a Kalman filter, and the distance d between the prediction coordinates of the current pedestrian and the coordinates of the current pedestrian frame in the Markov space is calculated1;
S504: calculating the cosine distance d between the apparent depth features of different current pedestrian frames in the continuous video frames obtained in step S5022;
S505: the distance d obtained in step S5031And the distance d obtained in step S5042Combining to form a fusion metric c, wherein c ═ λ d1+(1-λ)d2Lambda is 0.1;
s506: and according to the fusion metric c, performing target matching on different current pedestrian frames in the continuous video frames by using a Hungarian matching algorithm to obtain the coordinates of the current pedestrian in the continuous video frames.
Further preferably, in step S502, the deep network pedestrian feature extraction model is obtained based on residual error network training.
Further preferably, in step S506, the target matching is performed on each current pedestrian frame according to a cascade image matching algorithm.
Further preferably, the specific steps of step S6 are as follows:
s601: establishing a coordinate buffer queue of the current pedestrian movement track, wherein the size of the queue is set as m;
s602: taking the updated coordinates of the current pedestrian obtained in the step S503 as a current pedestrian tracking frame, calculating the frame center coordinates of the current pedestrian tracking frame, and adding the frame center coordinates as the current pedestrian trajectory coordinates into a current pedestrian coordinate buffer queue;
s603: after the current pedestrian tracking frame is updated, recalculating the central coordinates of the frame, and adding the recalculated central coordinates of the frame into a coordinate buffer queue;
s604: comparing the number of buffer queue coordinates with the queue size m; if the number of coordinates of the buffering queue is equal to the size m of the queue, adding the newly added center coordinates of the frame into the buffering queue, and removing the updated coordinates of the earliest pedestrian, so that the number of coordinates of the buffering queue is always equal to the size m of the queue, and the moving track of the current pedestrian is generated.
Preferably, in step S8, the noise threshold is selected as follows:
s801: based on the current pedestrian movement track, intercepting n sections of sample VIDEOs from the pedestrian monitoring VIDEO, and recording the n sections of sample VIDEOs as VIDEO1,…,VIDEOi,i=1,2,…,n;
S802: for n segments of sample VIDEO VIDEO1,…,VIDEOiManually counting the pedestrian volume, and recording the result of the pedestrian volume counting as NUM1,…,NUMi;
S803: setting a value set of a noise threshold value as { j; j is 2,3, …,20}, and n sample VIDEOs are subjected to VIDEO image processing based on steps S1 to S91,…,VIDEOiCarrying out flow statistics, and recording the obtained flow counting RESULT as RESULTij;
S804: calculating the difference between the flow counting result obtained in step S803 and the flow counting result obtained in step S802, and recording the difference as Eij=RESULTij-NUMi;
S805: calculating corresponding difference values E of different threshold valuesijWeighted average value F ofjWherein
S806: f when the comparison noise threshold takes different values jjIf F is the absolute value ofkIs the smallest in absolute value, where k is e { j; j is 2,3, …,20}, and k is the final selected noise threshold.
Further preferably, in step S801, the number n of segments of the sample video is 3 or 4, and the clipping time of the sample video is 0.5 to 2 hours.
Compared with the prior art, the invention has the beneficial effects that:
1) in the pedestrian flow statistical process, the pedestrian detection model based on deep learning is applied to detect the current pedestrian, the multi-target tracking algorithm is utilized to track the current pedestrian, the deep neural network has stronger representation capability compared with the traditional characteristic extraction, and the more accurate detection result can be obtained by using the deep neural network to detect the pedestrian; in addition, the multi-target tracking algorithm based on the deep learning uses a depth correlation measurement mode, so that effective tracking can be achieved when the current pedestrian faces shielding and the pedestrian is overlapped for a short time, the tracking track is more accurate, and compared with a simple tracking method, the tracking disorder is not easy to occur.
2) Compared with the pedestrian flow counting mode in the prior art, which is accumulated by the target moving track crossing the auxiliary line, the pedestrian flow counting method has the advantages that the moving track of the current pedestrian is generated in the effective area by dividing the effective area of the image, then the moving track of the current pedestrian is counted after noise judgment, the problems of inaccurate and incomplete tracks caused by serious dependence on the correctness of the moving track in the prior art are solved, and the statistical effect of the pedestrian flow is more accurate.
3) And the influence caused by uncertain factors in pedestrian statistics is effectively avoided by carrying out three-time denoising. Specifically, firstly, in step S2, an effective region of images in consecutive video frames is set, that is, environmental detection noise caused by uncertain factors such as an illumination environment and pedestrians in a non-attention region is removed, so as to realize the denoising effect of the first step; subsequently, step S7 utilizes a multi-target tracking algorithm to judge whether the current pedestrian leaves the effective area in real time, the tracking algorithm can remove pedestrian detection noise caused by uncertain factors such as blocking and incomplete pedestrian, and the denoising effect of the second step is realized; and finally, step S8 is to select a noise threshold and judge the noise of the moving track of the current pedestrian, the noise threshold can be effectively fitted with the real pedestrian flow, and the noise track is judged by setting the noise threshold, so that the denoising effect of the third step is realized macroscopically, the counting result of the pedestrian flow is well adjusted, and the statistical result is more accurate.
Drawings
In order to more clearly illustrate the embodiments of the present invention or the technical solutions in the prior art, the drawings used in the description of the embodiments or the prior art will be briefly described below, it is obvious that the drawings in the following description are only some embodiments of the present invention, and for those skilled in the art, other drawings can be obtained according to the drawings without creative efforts.
FIG. 1 is a block flow diagram of the present invention.
Detailed Description
The pedestrian traffic statistical method based on deep learning and multi-target tracking provided by the invention will be described in detail by way of embodiments with reference to the accompanying drawings. It should be noted that the description of the embodiments is provided to help understanding of the present invention, but the present invention is not limited thereto.
The term "and/or" herein is merely an association describing an associated object, meaning that three relationships may exist, e.g., a and/or B, may mean: a exists alone, B exists alone, and A and B exist at the same time, and the term "/and" is used herein to describe another association object relationship, which means that two relationships may exist, for example, A/and B, may mean: a alone, and both a and B alone, and further, the character "/" in this document generally means that the former and latter associated objects are in an "or" relationship.
Example (b):
a pedestrian flow statistical method based on deep learning and multi-target tracking comprises the following steps:
s1: and shooting the pedestrian monitoring video in real time, and reading images in continuous video frames of the pedestrian monitoring video. It should be noted that, in this step, the network monitoring camera may be but is not limited to be used to shoot the pedestrian monitoring video, the network monitoring camera is first placed at a position where the pedestrian can be shot, and then the camera is accessed to read the video image through a network, a local transmission, or other manners.
S2: the effective area of the images in successive video frames is set and the flow count is set to an initial value of 0. It should be noted that, setting of the effective region is to perform black filling and shielding on the non-attention region which is likely to affect the pedestrian statistics based on the observation requirement, so as to achieve the purpose of removing the noise generated by the surrounding environment.
S3: constructing a pedestrian detection model based on deep learning and training the pedestrian detection model; preferably, in the step S3, the pedestrian detection model is trained by using a YOLOv3 network structure.
S4: zooming the image with the effective area by using the trained pedestrian detection model, specifically, zooming the image into 416 × 416 pixels, then detecting the current pedestrian of the image to obtain the coordinate and the image block of the current pedestrian frame, wherein the current pedestrian frame is used for indicating the area with the current pedestrian in the image; it should be understood that the pedestrian detection model can also be implemented by fast R-CNN (fast Region-Based functional Network), ssd (single shot multi-box detector) or other target detection networks, and the model can be made to have both performance and speed by using YOLOv3 Network training.
S5: and tracking the current pedestrian in real time by using a multi-target tracking algorithm based on deep learning, and generating the coordinates of the current pedestrian in the continuous video frame images.
S6: and generating a moving track of the current pedestrian according to the coordinates of the current pedestrian in the images of the continuous video frames.
S7: judging whether the current pedestrian leaves the effective area in real time by utilizing a multi-target tracking algorithm according to the moving track of the current pedestrian; if so, the process proceeds to step S8, otherwise, the process proceeds to step S4.
S8: selecting a noise threshold value and judging the noise of the moving track of the current pedestrian; if the number of coordinates of the moving track of the current pedestrian is smaller than the noise threshold value, judging that the moving track of the current pedestrian is a noise track; and if the number of the coordinates of the moving track of the current pedestrian is greater than or equal to the noise threshold, judging that the moving track of the current pedestrian is an effective track, and adding 1 to the numerical value of the flow counting. Let j be a default value for the noise threshold, and j may be a randomly chosen smaller value, such as 6,7, …, 12.
It should be noted that, in the pedestrian flow statistics process, the factors that can affect the pedestrian flow statistics result mainly include the lighting environment in the video, the pedestrians in the non-attention area in the video, the blocked pedestrians, the incomplete pedestrians, and other uncertain factors, and in the statistics process, these factors may affect the obtaining of the coordinates and image blocks of the current pedestrian frame, thereby affecting the real-time tracking process of the current pedestrian and finally affecting the pedestrian flow statistics. In the invention, the uncertain factors influencing pedestrian flow statistics are collectively called noise, the setting of the noise threshold value can judge whether the moving track of the current pedestrian is a noise track, and the current pedestrian track judged as the noise track is removed, thus avoiding the influence of the uncertain factors on the pedestrian statistical result.
S9: the coordinates of the current pedestrian in the consecutive video frames are deleted and then step S4 is repeated.
In the pedestrian flow counting process, the pedestrian detection model based on deep learning is applied to detect the current pedestrian through the step S3, and the current pedestrian is tracked by the multi-target tracking algorithm in the step S4, so that the deep neural network has stronger representation capability compared with the traditional feature extraction, and more accurate detection results can be obtained by using the deep neural network to detect the pedestrian; in addition, in step S5, the multi-target tracking algorithm based on deep learning uses a depth-related metric method, so that effective tracking can be achieved when the current pedestrian faces occlusion and the pedestrian momentarily overlaps, which means that the tracking trajectory is more accurate, and compared with a simple tracking method, the tracking is less prone to being confused.
In the prior art, a pedestrian flow counting mode that a target moving track crosses an auxiliary line to accumulate is usually adopted, and the target moving track seriously depends on the correctness of the moving track, namely the quality of a tracking effect, but when the pedestrian flow is dense, the tracking effect is often poor, disorder can often occur, and finally the pedestrian track is inaccurate and incomplete, so that the statistical result of the pedestrian flow is seriously influenced. According to the invention, the effective area of the image is defined, the moving track of the current pedestrian is generated in the effective area, then the moving track of the current pedestrian is counted after noise judgment, so that the problem that the track is inaccurate and incomplete due to the fact that the moving track is seriously dependent on the correctness in the prior art is effectively avoided, and the statistical effect of the pedestrian flow is more accurate.
It should also be understood that the invention carries out three-time denoising, and effectively avoids the influence caused by uncertain factors in pedestrian statistics. Specifically, firstly, in step S2, an effective region of images in consecutive video frames is set, that is, environmental detection noise caused by uncertain factors such as an illumination environment and pedestrians in a non-attention region is removed, so as to realize the denoising effect of the first step; subsequently, step S7 utilizes a multi-target tracking algorithm to judge whether the current pedestrian leaves the effective area in real time, the tracking algorithm can remove pedestrian detection noise caused by uncertain factors such as blocking and incomplete pedestrian, and the denoising effect of the second step is realized; and finally, step S8 is to select a noise threshold and judge the noise of the moving track of the current pedestrian, the noise threshold can be effectively fitted with the real pedestrian flow, and the noise track is judged by setting the noise threshold, so that the third step of denoising is realized macroscopically, the counting result of the pedestrian flow is well adjusted, the statistical result is more accurate, and the statistical result of the pedestrian flow is more accurate.
Further, the specific steps of step S5 are as follows:
s501: and reading the coordinates and the image blocks of the current pedestrian frame.
S502: and constructing a pedestrian feature extraction model based on deep learning and training the pedestrian feature extraction model, and performing depth apparent feature extraction on the image block of the current pedestrian frame by using the trained deep network pedestrian feature extraction model to generate depth apparent features of different current pedestrian frames in continuous video frames. Preferably, the deep network pedestrian feature extraction model is obtained based on residual error network training; the depth appearance feature is a 128-dimensional feature vector.
S503: according to the coordinates of the current pedestrian frame, the prediction coordinates of the current pedestrian and the updating coordinates of the current pedestrian are calculated by using a Kalman filter, and the distance d between the prediction coordinates of the current pedestrian and the coordinates of the current pedestrian frame in the Markov space is calculated1. It should be noted that the kalman filter can provide a better estimate of the position of the moving object, andthe motion track is smoother.
S504: calculating the cosine distance d between the apparent depth features of different current pedestrian frames in the continuous video frames obtained in step S5022。
S505: the distance d obtained in step S5031And the distance d obtained in step S5042Combining to form a fusion metric c, wherein c ═ λ d1+(1-λ)d2And λ is 0.1.
S506: and according to the fusion metric c, performing target matching on different current pedestrian frames in the continuous video frames by using a Hungarian matching algorithm to obtain the coordinates of the current pedestrian in the continuous video frames. Preferably, in the step S506, the target matching is performed on each current pedestrian frame according to a cascade image matching algorithm, and the cascade image matching algorithm is based on geometric characteristics, and tends to have better invariance and better stability.
Further, the specific steps of step S6 are as follows:
s601: and establishing a coordinate buffer queue of the current pedestrian movement track, wherein the size of the queue is set as m.
S602: and taking the updated coordinates of the current pedestrian obtained in the step S503 as a current pedestrian tracking frame, calculating the frame center coordinates of the current pedestrian tracking frame, and adding the frame center coordinates as the current pedestrian track coordinates into a current pedestrian coordinate buffer queue.
S603: and after the current pedestrian tracking frame is updated, recalculating the central coordinates of the frame, and adding the recalculated central coordinates of the frame into a coordinate buffer queue.
S604: comparing the number of buffer queue coordinates with the queue size m; if the number of coordinates of the buffering queue is equal to the size m of the queue, adding the newly added center coordinates of the frame into the buffering queue, and removing the updated coordinates of the earliest pedestrian, so that the number of coordinates of the buffering queue is always equal to the size m of the queue, and the moving track of the current pedestrian is generated.
Further, in order to further reduce the error in the pedestrian flow statistics to obtain a more accurate pedestrian flow statistics result, in step S8, the noise threshold is selected as follows:
s801: based on the current pedestrian movement track, intercepting n sections of sample VIDEOs from the pedestrian monitoring VIDEO, and recording the n sections of sample VIDEOs as VIDEO1,…,VIDEOiI is 1,2, …, n. In this step, the number n of segments of the sample video is 3 or 4, and the clipping time of the sample video is 0.5 to 2 hours.
S802: for n segments of sample VIDEO VIDEO1,…,VIDEOiManually counting the pedestrian volume, and recording the result of the pedestrian volume counting as NUM1,…,NUMi。
S803: setting a value set of a noise threshold value as { j; j is 2,3, …,20}, and n sample VIDEOs are subjected to VIDEO image processing based on steps S1 to S91,…,VIDEOiCarrying out flow statistics, and recording the obtained flow counting RESULT as RESULTij。
S804: calculating the difference between the flow counting result obtained in step S803 and the flow counting result obtained in step S802, and recording the difference as Eij=RESULTij-NUMi。
S805: calculating corresponding difference values E of different threshold valuesijWeighted average value F ofjWherein
S806: f when the comparison noise threshold takes different values jjIf F is the absolute value ofkIs the smallest in absolute value, where k is e { j; j is 2,3, …,20}, and k is the finally selected noise threshold, i.e. the optimal noise threshold.
The above description is only an embodiment of the present invention, and not intended to limit the scope of the present invention, and all modifications of equivalent structures and equivalent processes performed by the present specification and drawings, or directly or indirectly applied to other related technical fields, are included in the scope of the present invention.
Claims (7)
1. A pedestrian flow statistical method based on deep learning and multi-target tracking is characterized in that: the method comprises the following steps:
s1: shooting a pedestrian monitoring video in real time, and reading images in continuous video frames of the pedestrian monitoring video;
s2: setting an effective area of images in continuous video frames, and setting a flow count with an initial value of 0;
s3: constructing a pedestrian detection model based on deep learning and training the pedestrian detection model;
s4: zooming the image with the effective area by using the trained pedestrian detection model, and then detecting the current pedestrian of the image to obtain the coordinate and the image block of the current pedestrian frame, wherein the current pedestrian frame is used for indicating the area with the current pedestrian in the image;
s5: tracking the current pedestrian in real time by using a multi-target tracking algorithm based on deep learning, and generating coordinates of the current pedestrian in a continuous video frame image;
s6: generating a moving track of the current pedestrian according to the coordinates of the current pedestrian in the images of the continuous video frames;
s7: judging whether the current pedestrian leaves the effective area in real time by utilizing a multi-target tracking algorithm according to the moving track of the current pedestrian; if yes, go to step S8, otherwise go to step S4;
s8: selecting a noise threshold value and judging the noise of the moving track of the current pedestrian; if the number of coordinates of the moving track of the current pedestrian is smaller than the noise threshold value, judging that the moving track of the current pedestrian is a noise track; if the number of coordinates of the moving track of the current pedestrian is larger than or equal to the noise threshold, judging that the moving track of the current pedestrian is an effective track, and adding 1 to the numerical value of the flow meter number;
s9: deleting the coordinates of the current pedestrian in the continuous video frames, and then repeating the step S4;
in step S8, the noise threshold is selected as follows:
s801: based on the current pedestrian movement track, intercepting n sections of sample VIDEOs from the pedestrian monitoring VIDEO, and recording the n sections of sample VIDEOs as VIDEO1,…,VIDEOi,i=1,2,…,n;
S802: for n segments of sample VIDEO VIDEO1,…,VIDEOiManually counting the pedestrian volume, and recording the result of the pedestrian volume counting as NUM1,…,NUMi;
S803: setting a value set of a noise threshold value as { j; j is 2,3, …,20}, and n sample VIDEOs are subjected to VIDEO image processing based on steps S1 to S91,…,VIDEOiCarrying out flow statistics, and recording the obtained flow counting RESULT as RESULTij;
S804: calculating the difference between the flow counting result obtained in step S803 and the flow counting result obtained in step S802, and recording the difference as Eij=RESULTij-NUMi;
S805: calculating corresponding difference values E of different threshold valuesijWeighted average value F ofjWherein
S806: f when the comparison noise threshold takes different values jjIf F is the absolute value ofkIs the smallest in absolute value, where k is e { j; j is 2,3, …,20}, and k is the final selected noise threshold.
2. The pedestrian traffic statistical method based on deep learning and multi-target tracking as claimed in claim 1, characterized in that: in the step S3, the pedestrian detection model is trained by using the YOLOv3 network structure.
3. The pedestrian traffic statistical method based on deep learning and multi-target tracking as claimed in claim 1, characterized in that: the specific steps of step S5 are as follows:
s501: reading the coordinates and image blocks of the current pedestrian frame;
s502: constructing a pedestrian feature extraction model based on deep learning and training the pedestrian feature extraction model, and performing depth apparent feature extraction on image blocks of a current pedestrian frame by using the trained deep network pedestrian feature extraction model to generate depth apparent features of different current pedestrian frames in continuous video frames;
s503: according to the coordinates of the current pedestrian frame, the prediction coordinates of the current pedestrian and the updating coordinates of the current pedestrian are calculated by using a Kalman filter, and the distance d between the prediction coordinates of the current pedestrian and the coordinates of the current pedestrian frame in the Markov space is calculated1;
S504: calculating the cosine distance d between the apparent depth features of different current pedestrian frames in the continuous video frames obtained in step S5022;
S505: the distance d obtained in step S5031And the distance d obtained in step S5042Combining to form a fusion metric c, wherein c ═ λ d1+(1-λ)d2Lambda is 0.1;
s506: and according to the fusion metric c, performing target matching on different current pedestrian frames in the continuous video frames by using a Hungarian matching algorithm to obtain the coordinates of the current pedestrian in the continuous video frames.
4. The pedestrian traffic statistical method based on deep learning and multi-target tracking as claimed in claim 3, characterized in that: in step S502, the deep network pedestrian feature extraction model is obtained based on residual error network training.
5. The pedestrian traffic statistical method based on deep learning and multi-target tracking as claimed in claim 3, characterized in that: in step S506, target matching is performed on each current pedestrian frame according to a cascade image matching algorithm.
6. The pedestrian traffic statistical method based on deep learning and multi-target tracking as claimed in claim 3, characterized in that: the specific steps of step S6 are as follows:
s601: establishing a coordinate buffer queue of the current pedestrian movement track, wherein the size of the queue is set as m;
s602: taking the updated coordinates of the current pedestrian obtained in the step S503 as a current pedestrian tracking frame, calculating the frame center coordinates of the current pedestrian tracking frame, and adding the frame center coordinates as the current pedestrian trajectory coordinates into a current pedestrian coordinate buffer queue;
s603: after the current pedestrian tracking frame is updated, recalculating the central coordinates of the frame, and adding the recalculated central coordinates of the frame into a coordinate buffer queue;
s604: comparing the number of buffer queue coordinates with the queue size m; if the number of coordinates of the buffering queue is equal to the size m of the queue, adding the newly added center coordinates of the frame into the buffering queue, and removing the updated coordinates of the earliest pedestrian, so that the number of coordinates of the buffering queue is always equal to the size m of the queue, and the moving track of the current pedestrian is generated.
7. The pedestrian traffic statistical method based on deep learning and multi-target tracking as claimed in claim 1, characterized in that: in step S801, the number of segments n of the sample video is 3 or 4, and the clipping time of the sample video is 0.5 to 2 hours.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201811400758.9A CN109522854B (en) | 2018-11-22 | 2018-11-22 | Pedestrian traffic statistical method based on deep learning and multi-target tracking |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201811400758.9A CN109522854B (en) | 2018-11-22 | 2018-11-22 | Pedestrian traffic statistical method based on deep learning and multi-target tracking |
Publications (2)
Publication Number | Publication Date |
---|---|
CN109522854A CN109522854A (en) | 2019-03-26 |
CN109522854B true CN109522854B (en) | 2021-05-11 |
Family
ID=65777627
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201811400758.9A Active CN109522854B (en) | 2018-11-22 | 2018-11-22 | Pedestrian traffic statistical method based on deep learning and multi-target tracking |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN109522854B (en) |
Families Citing this family (22)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10909387B2 (en) * | 2019-03-28 | 2021-02-02 | Mitsubishi Electric Research Laboratories, Inc. | Method and system for predicting dynamical flows from control inputs and limited observations |
CN110309717A (en) * | 2019-05-23 | 2019-10-08 | 南京熊猫电子股份有限公司 | A kind of pedestrian counting method based on deep neural network |
CN110346005B (en) * | 2019-06-11 | 2020-09-22 | 华南理工大学 | Coriolis mass flowmeter digital signal processing method based on deep learning |
CN110490901A (en) * | 2019-07-15 | 2019-11-22 | 武汉大学 | The pedestrian detection tracking of anti-attitudes vibration |
CN110543868A (en) * | 2019-09-09 | 2019-12-06 | 福建省趋普物联科技有限公司 | Monitoring method and system based on face recognition and head and shoulder detection |
CN110717408B (en) * | 2019-09-20 | 2022-04-12 | 台州智必安科技有限责任公司 | People flow counting method based on TOF camera |
CN110992305A (en) * | 2019-10-31 | 2020-04-10 | 中山大学 | Package counting method and system based on deep learning and multi-target tracking technology |
CN111145551A (en) * | 2020-01-03 | 2020-05-12 | 南京邮电大学 | Intersection traffic planning system based on CNN detection follows chapter rate |
CN111523472A (en) * | 2020-04-23 | 2020-08-11 | 杭州海康威视系统技术有限公司 | Active target counting method and device based on machine vision |
CN111640135A (en) * | 2020-05-25 | 2020-09-08 | 台州智必安科技有限责任公司 | TOF camera pedestrian counting method based on hardware front end |
CN112215873A (en) * | 2020-08-27 | 2021-01-12 | 国网浙江省电力有限公司电力科学研究院 | Method for tracking and positioning multiple targets in transformer substation |
CN112036367A (en) * | 2020-09-16 | 2020-12-04 | 南通天成现代农业科技有限公司 | People number detection method of YOLO convolutional neural network |
CN112183304A (en) * | 2020-09-24 | 2021-01-05 | 高新兴科技集团股份有限公司 | Off-position detection method, system and computer storage medium |
CN112613365A (en) * | 2020-12-11 | 2021-04-06 | 北京影谱科技股份有限公司 | Pedestrian detection and behavior analysis method and device and computing equipment |
CN112541440B (en) * | 2020-12-16 | 2023-10-17 | 中电海康集团有限公司 | Subway people stream network fusion method and people stream prediction method based on video pedestrian recognition |
CN112633205A (en) * | 2020-12-28 | 2021-04-09 | 北京眼神智能科技有限公司 | Pedestrian tracking method and device based on head and shoulder detection, electronic equipment and storage medium |
CN113763418B (en) * | 2021-03-02 | 2024-02-02 | 华南理工大学 | Multi-target tracking method based on head and shoulder detection |
CN113158813A (en) * | 2021-03-26 | 2021-07-23 | 精英数智科技股份有限公司 | Real-time statistical method and device for flow target |
CN113158897A (en) * | 2021-04-21 | 2021-07-23 | 新疆大学 | Pedestrian detection system based on embedded YOLOv3 algorithm |
CN113221808A (en) * | 2021-05-26 | 2021-08-06 | 新疆爱华盈通信息技术有限公司 | Dinner plate counting statistical method and device based on image recognition |
CN115082862A (en) * | 2022-07-07 | 2022-09-20 | 南京杰迈视讯科技有限公司 | High-precision pedestrian flow statistical method based on monocular camera |
CN116167625B (en) * | 2023-04-25 | 2023-08-18 | 湖南工商大学 | Trampling risk assessment method based on deep learning |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN101231755A (en) * | 2007-01-25 | 2008-07-30 | 上海遥薇实业有限公司 | Moving target tracking and quantity statistics method |
CN101872414A (en) * | 2010-02-10 | 2010-10-27 | 杭州海康威视软件有限公司 | People flow rate statistical method and system capable of removing false targets |
CN106600631A (en) * | 2016-11-30 | 2017-04-26 | 郑州金惠计算机系统工程有限公司 | Multiple target tracking-based passenger flow statistics method |
CN108021848A (en) * | 2016-11-03 | 2018-05-11 | 浙江宇视科技有限公司 | Passenger flow volume statistical method and device |
-
2018
- 2018-11-22 CN CN201811400758.9A patent/CN109522854B/en active Active
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN101231755A (en) * | 2007-01-25 | 2008-07-30 | 上海遥薇实业有限公司 | Moving target tracking and quantity statistics method |
CN101872414A (en) * | 2010-02-10 | 2010-10-27 | 杭州海康威视软件有限公司 | People flow rate statistical method and system capable of removing false targets |
CN108021848A (en) * | 2016-11-03 | 2018-05-11 | 浙江宇视科技有限公司 | Passenger flow volume statistical method and device |
CN106600631A (en) * | 2016-11-30 | 2017-04-26 | 郑州金惠计算机系统工程有限公司 | Multiple target tracking-based passenger flow statistics method |
Non-Patent Citations (2)
Title |
---|
SIMPLE ONLINE AND REALTIME TRACKING WITH A DEEP ASSOCIATION METRIC;Nicolai Wojke et al.;《2017 IEEE International Conference on Image Processing (ICIP) 》;20171231;第3645-3649页 * |
基于卷积神经网络的人流量统计;张雅俊 等;《重庆邮电大学学报(自然科学版)》;20170430;第29卷(第2期);第265-271页 * |
Also Published As
Publication number | Publication date |
---|---|
CN109522854A (en) | 2019-03-26 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN109522854B (en) | Pedestrian traffic statistical method based on deep learning and multi-target tracking | |
CN104751491B (en) | A kind of crowd's tracking and people flow rate statistical method and device | |
US10735694B2 (en) | System and method for activity monitoring using video data | |
TWI448977B (en) | Method and apparatus for video analytics based object counting | |
CN106778712B (en) | Multi-target detection and tracking method | |
CN104978567B (en) | Vehicle checking method based on scene classification | |
CN108229256B (en) | Road construction detection method and device | |
CN112561951B (en) | Motion and brightness detection method based on frame difference absolute error and SAD | |
Wei et al. | City-scale vehicle tracking and traffic flow estimation using low frame-rate traffic cameras | |
CN106570449B (en) | A kind of flow of the people defined based on region and popularity detection method and detection system | |
Hu et al. | A novel approach for crowd video monitoring of subway platforms | |
Fradi et al. | Spatio-temporal crowd density model in a human detection and tracking framework | |
CN110889347B (en) | Density traffic flow counting method and system based on space-time counting characteristics | |
CN109583361A (en) | The scene video text tracking method minimized based on energy | |
Almomani et al. | Segtrack: A novel tracking system with improved object segmentation | |
Cheng | Highway traffic flow estimation for surveillance scenes damaged by rain | |
Tursun et al. | A video based real-time vehicle counting system using optimized virtual loop method | |
CN107067411B (en) | Mean-shift tracking method combined with dense features | |
Makawana et al. | Moving vehicle detection and speed measurement in video sequence | |
Guo et al. | Vehicle detection and tracking based on optical field | |
CN110830734B (en) | Abrupt change and gradual change lens switching identification method and system | |
Kuplyakov et al. | A distributed tracking algorithm for counting people in video by head detection | |
CN114399532A (en) | Camera position and posture determining method and device | |
Yang et al. | Crowd activity change point detection in videos via graph stream mining | |
Pan et al. | A novel vehicle flow detection algorithm based on motion saliency for traffic surveillance system |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |