CN108830166B - Real-time bus passenger flow volume statistical method - Google Patents

Real-time bus passenger flow volume statistical method Download PDF

Info

Publication number
CN108830166B
CN108830166B CN201810505385.5A CN201810505385A CN108830166B CN 108830166 B CN108830166 B CN 108830166B CN 201810505385 A CN201810505385 A CN 201810505385A CN 108830166 B CN108830166 B CN 108830166B
Authority
CN
China
Prior art keywords
characteristic
gradient
amplitude
pyramid
image
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Expired - Fee Related
Application number
CN201810505385.5A
Other languages
Chinese (zh)
Other versions
CN108830166A (en
Inventor
章国泰
聂凯
张立斌
高珊华
王红广
王鹏
张东旭
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Tianjin Tongka Intelligent Network Technology Co ltd
Original Assignee
Tianjin Tongka Intelligent Network Technology Co ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Tianjin Tongka Intelligent Network Technology Co ltd filed Critical Tianjin Tongka Intelligent Network Technology Co ltd
Priority to CN201810505385.5A priority Critical patent/CN108830166B/en
Publication of CN108830166A publication Critical patent/CN108830166A/en
Application granted granted Critical
Publication of CN108830166B publication Critical patent/CN108830166B/en
Expired - Fee Related legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/50Context or environment of the image
    • G06V20/52Surveillance or monitoring of activities, e.g. for recognising suspicious objects
    • G06V20/53Recognition of crowd images, e.g. recognition of crowd congestion
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/44Local feature extraction by analysis of parts of the pattern, e.g. by detecting edges, contours, loops, corners, strokes or intersections; Connectivity analysis, e.g. of connected components
    • G06V10/443Local feature extraction by analysis of parts of the pattern, e.g. by detecting edges, contours, loops, corners, strokes or intersections; Connectivity analysis, e.g. of connected components by matching or filtering
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/50Extraction of image or video features by performing operations within image blocks; by using histograms, e.g. histogram of oriented gradients [HoG]; by summing image-intensity values; Projection analysis
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/56Extraction of image or video features relating to colour
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/40Scenes; Scene-specific elements in video content
    • G06V20/41Higher-level, semantic clustering, classification or understanding of video scenes, e.g. detection, labelling or Markovian modelling of sport events or news items
    • G06V20/42Higher-level, semantic clustering, classification or understanding of video scenes, e.g. detection, labelling or Markovian modelling of sport events or news items of sport video content

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Multimedia (AREA)
  • Theoretical Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Computational Linguistics (AREA)
  • Software Systems (AREA)
  • Image Analysis (AREA)
  • Image Processing (AREA)

Abstract

The invention relates to a real-time bus passenger flow rate statistical method. The method aims to detect and track the human head in the video image. The video data of the method is passenger flow video data shot from the top areas of the front door and the rear door of the bus by using a camera, and the statistical method comprises the following steps: extracting characteristics; training a model; head detection; head tracking; compared with the prior art, the invention has the advantages that: the detection method has real-time performance and can efficiently and accurately detect the passenger flow.

Description

Real-time bus passenger flow volume statistical method
The technical field is as follows:
the invention relates to the technical field of image processing in pattern recognition, and further relates to a real-time bus passenger flow volume statistical method.
Background art:
many developers have used infrared devices and pressure sensors to make passenger flow statistics. Since the error between the detection result and the actual passenger flow is large, passenger flow counting methods based on infrared devices and pressure sensors have not been used until now and are gradually abandoned. In recent years, with the continuous development of the pattern recognition field and the GPU parallel computing field, the computer vision direction has been rapidly developed. In the field of image processing, passenger flow statistics is an important application, and is a new field and direction in intelligent video monitoring at present.
In recent years, passenger flow statistics methods are mainly based on three main categories: the three methods are all defect-based on detection methods of feature points, human body segmentation and tracking and deep learning. The accuracy of the first two detection methods needs to be improved. Although the third type of detection method has high accuracy, its real-time performance and high hardware cost cannot meet the standards for popularization and use.
The invention content is as follows:
the invention aims to provide a method for detecting human heads in video images. Such a method should be able to detect the passenger flow efficiently and accurately. The specific technical scheme is as follows:
the video data of the method is passenger flow video data shot from the top areas of the front door and the rear door of the bus by using a camera, and the statistical method comprises the following steps:
step 1: model training:
step 1.1: converting real image video frames collected by a camera into a BGRA format; in the video frame, calibrating the head position of the passenger and regarding the head position as a positive sample training image set of a training detection model, and regarding the area without the head of the passenger in the video frame as a negative sample training image set of the training detection model; carrying out image scale normalization processing on the positive sample training image set and the negative sample training image set to form training input images with the same size;
step 1.2: extracting features of the image: calculating a gradient amplitude characteristic and a related direction value of the LUV characteristic; according to the related gradient direction value, quantifying the gradient amplitude characteristic and extracting the histogram characteristic of the gradient amplitude characteristic; then, connecting the LUV characteristic, the gradient amplitude characteristic and the histogram characteristic in series to form a final image characteristic;
step 1.3: constructing a characteristic pyramid: calculating a characteristic scaling coefficient of each layer of pyramid according to the number of layers of the set characteristic pyramid; then, multiplying the image characteristics by the characteristic scaling coefficient to form a characteristic pyramid;
step 1.4: constructing a training model: sending the generated characteristic pyramid into a decision tree algorithm for training until the training accuracy reaches the requirement, and ending iterative training;
step 2: head detection:
step 2.1: acquiring a video frame to be detected by using a camera;
step 2.2: calculating LUV characteristics of a video frame to be detected, gradient amplitude characteristics of the LUV characteristics and related direction values; according to the related gradient direction value, quantifying the gradient amplitude characteristic and extracting the histogram characteristic of the gradient amplitude characteristic; then, connecting the LUV characteristic, the gradient amplitude characteristic and the histogram characteristic in series to form a final image characteristic;
step 2.3: calculating a characteristic scaling coefficient of each layer of pyramid according to the number of layers of the set characteristic pyramid; then, multiplying the image characteristics by the characteristic scaling coefficient to form a characteristic pyramid;
step 2.4: inputting the feature pyramid of the video frame to be detected obtained in the step 2.3 into the decision tree classifier trained in the step 1.4, and multiplying the feature scaling coefficient of each pyramid layer by the set size of the detection window and the detection moving step length to obtain the size of the detection window and the detection moving step length for scanning each pyramid layer; scanning on the corresponding pyramid layer by using the corresponding detection window and the moving step length; judging whether the windows are the head regions or the non-head regions of passengers according to the judgment results of the decision tree on each layer of pyramid and the window scores, scaling the windows which are determined to be the heads to the original size according to the characteristic scaling coefficient, and removing overlapped windows; finally, recording the head position of the passenger in a result file;
and step 3: tracking the head:
tracking the detection window by using a related kernel filter method, forming a track, and if the track crosses a specified limit, indicating that the passenger finishes the action of getting on or off the vehicle;
and 4, step 4: and (3) passenger flow volume counting:
if the passenger has finished getting on or off, the algorithm will update the data with the addition of 1 to the passenger flow, otherwise, the passenger flow will remain unchanged.
As one of the preferable schemes, the specific process of step 1.2 is as follows: mapping the RGB pixel values of the image into a three-dimensional stereo space X, Y, Z, using this space to calculate the corresponding LUV feature values; then, respectively carrying out convolution on L, U, V features in the horizontal direction and the vertical direction, solving the square values of all feature points in the two directions, adding the square values of all feature points in the two directions, taking a square root, selecting a maximum value as a gradient amplitude feature, and selecting an arctangent value of the convolution values in the horizontal direction and the vertical direction as a gradient direction value; finally, according to the related gradient direction value, quantifying the gradient amplitude characteristic and extracting the histogram characteristic of the gradient amplitude characteristic; the LUV feature, gradient magnitude feature, and histogram feature are then concatenated to form the final image feature.
In the above preferred embodiment, the calculation method for calculating the gradient magnitude characteristic and the correlation direction value of the LUV characteristic is as follows:
Gx(x,y)=I(x+1,y)-I(x-1,y)
Gy(x,y)=I(x,y+1)-I(x,y-1)
Figure BDA0001671300400000031
α(x,y)=tan-1(Gy(x,y)/Gx(x,y))
Gx(x, y) represents the horizontal gradient magnitude of the input image at the pixel point (x, y), Gy(x, y) represents the vertical gradient amplitude of the input image at the pixel point (x, y), G (x, y) is the gradient amplitude of the pixel point (x, y), alpha (x, y) is the gradient direction of the pixel point (x, y), and I (x, y) represents the L, U, V characteristic value of the point.
As a second preferred embodiment, the method for calculating the final image feature in step 1.3 is as follows:
Fi=μi*F,Ficharacteristic of the i-th pyramid of the method, μiAnd F is the coefficient of the ith layer of pyramid characteristics of the method, and F is the formed final image characteristics.
As a third preferred scheme, in step 2, a parallel computing method is applied, on the GPU, feature extraction is performed on the current video frame by using the heterogeneous programming language OpenCL, and the passenger head detection of the previous frame is completed on the CPU, and the two are performed simultaneously, so as to achieve the purpose of parallel processing.
As a fourth preferred scheme, in step 3, a correlation filter is trained by the correlated kernel filter according to the information of the previous and subsequent frames, and correlation calculation is performed with the newly input frame, and the obtained confidence map is the predicted tracking result; the point or block with the highest score is the tracking result.
As the fifth preferred scheme, the method also comprises a step 5, and the process is as follows:
and 5: storing the coordinates and width and height of the passenger head detected by each frame of image into a file of a detection result, and storing each frame of image; when the bus stops running, the method can output the final passenger flow.
Compared with the prior art, the invention has the advantages that:
the detection method has real-time performance and can efficiently and accurately detect the passenger flow.
In the embodiment, the head detection thread is implemented on the CPU, and the feature extraction thread is implemented on the GPU, so that the head detection thread and the feature extraction thread can simultaneously operate, realize parallelization processing and save calculation time.
Description of the drawings:
fig. 1 is a flow chart of the bus passenger flow rate statistical method of the patent.
FIG. 2 is a schematic diagram of a process of extracting features by GPU parallel computing.
FIG. 3 is a schematic flow chart of creating a feature pyramid.
FIG. 4 is a schematic diagram of algorithmic two-thread processing; in the figure, a DP thread represents a head detection thread; the FC thread represents a feature extraction thread. Since the head detection thread is implemented on the CPU and the feature extraction thread is implemented on the GPU. Therefore, the two can be operated simultaneously, thereby realizing parallelization processing and saving calculation time.
The specific implementation mode is as follows:
example (b):
a real-time bus passenger flow volume statistical method is characterized in that video data of the method are passenger flow video data shot from top areas of a front door and a rear door of a bus by using cameras, and the statistical method comprises the following steps:
step 1: model training:
step 1.1: converting real image video frames collected by a camera into a BGRA format; in the video frame, calibrating the head position of the passenger and regarding the head position as a positive sample training image set of a training detection model, and regarding the area without the head of the passenger in the video frame as a negative sample training image set of the training detection model; carrying out image scale normalization processing on the positive sample training image set and the negative sample training image set to form training input images with the same size;
step 1.2: extracting features of the image:
step 1.2.1: LUV features were extracted on RGB images: mapping the RGB pixel values of the image into a three-dimensional stereo space X, Y, Z, using this space to calculate the corresponding LUV feature values;
step 1.2.2: carrying out convolution of triangular filtering on the LUV characteristics;
step 1.2.3: respectively calculating the gradient amplitude and the gradient direction of L, U, V, and selecting the maximum gradient amplitude of the three as a final amplitude result; calculating the amplitude direction according to the gradient amplitude, and outputting the amplitude direction if the amplitude direction is greater than or equal to zero; if the amplitude direction is smaller than zero, outputting the amplitude direction plus one hundred eighty degrees;
step 1.2.4: performing convolution of triangular filtering on the gradient amplitude characteristics;
step 1.2.5: quantizing the gradient amplitude characteristics and the gradient direction to generate a final gradient direction histogram;
step 1.2.6: scaling and connecting the LUV characteristic, the gradient amplitude characteristic and the gradient direction histogram characteristic in series;
in the above scheme, the gradient magnitude characteristic calculation method for calculating the LUV characteristic is as follows:
Gx(x,y)=I(x+1,y)-I(x-1,y)
Gy(x,y)=I(x,y+1)-I(x,y-1)
Figure BDA0001671300400000061
α(x,y)=tan-1(Gy(x,y)/Gx(x,y))
Gx(x, y) represents the horizontal gradient magnitude of the input image at the pixel point (x, y), Gy(x, y) represents the gradient amplitude of the input image in the vertical direction at the pixel point (x, y), G (x, y) is the gradient amplitude of the pixel point (x, y), alpha (x, y) is the gradient direction of the pixel point (x, y), and I (x, y) represents the characteristic value of L, U, V of the point;
step 1.3: constructing a characteristic pyramid: calculating a characteristic scaling coefficient of each layer of pyramid according to the number of layers of the set characteristic pyramid; then, multiplying the image characteristics by the characteristic scaling coefficient to form a characteristic pyramid; the final image feature calculation method is as follows:
Fi=μi*F,Ficharacteristic of the i-th pyramid of the method, μiThe coefficient of the ith layer pyramid feature of the method is shown, and F is the formed final image feature;
step 1.4: constructing a training model: sending the generated characteristic pyramid into a decision tree algorithm for training until the training accuracy reaches the requirement, and ending iterative training;
step 2: head detection: by using a parallel computing method, on a GPU, feature extraction is carried out on a current video frame by using a heterogeneous programming language OpenCL, passenger head detection of the previous frame is completed on a CPU, and the two are carried out simultaneously, so that the purpose of parallel processing is achieved; the specific process is as follows:
step 2.1: acquiring a video frame to be detected by using a camera;
step 2.2: extracting LUV characteristics from the video frame:
step 2.2.1: LUV features were extracted on RGB images: mapping the RGB pixel values of the image into a three-dimensional stereo space X, Y, Z, using this space to calculate the corresponding LUV feature values;
step 2.2.2: carrying out convolution of triangular filtering on the LUV characteristics;
step 2.2.3: respectively calculating the gradient amplitude and the gradient direction of L, U, V, and selecting the maximum gradient amplitude of the three as a final amplitude result; calculating the amplitude direction according to the gradient amplitude, and outputting the amplitude direction if the amplitude direction is greater than or equal to zero; if the amplitude direction is smaller than zero, outputting the amplitude direction plus one hundred eighty degrees;
step 2.2.4: performing convolution of triangular filtering on the gradient amplitude characteristics;
step 2.2.5: quantizing the gradient amplitude characteristics and the gradient direction to generate a final gradient direction histogram;
step 2.2.6: scaling and connecting the LUV characteristic, the gradient amplitude characteristic and the gradient direction histogram characteristic in series;
in the above scheme, the gradient magnitude characteristic calculation method for calculating the LUV characteristic is as follows:
Gx(x,y)=I(x+1,y)-I(x-1,y)
Gy(X,y)=I(x,y+1)-I(x,y-1)
Figure BDA0001671300400000071
α(x,y)=tan-1(Gy(x,y)/Gx(x,y))
Gx(x, y) represents the horizontal gradient magnitude of the input image at the pixel point (x, y), Gy(x, y) represents the gradient amplitude of the input image in the vertical direction at the pixel point (x, y), G (x, y) is the gradient amplitude of the pixel point (x, y), alpha (x, y) is the gradient direction of the pixel point (x, y), and I (x, y) represents the characteristic value of L, U, V of the point;
step 2.3: calculating a characteristic scaling coefficient of each layer of pyramid according to the number of layers of the set characteristic pyramid; then, multiplying the image characteristics by the characteristic scaling coefficient to form a characteristic pyramid; the final image feature calculation method is as follows:
i=μi*F,Ficharacteristic of the i-th pyramid of the method, μiThe coefficient of the ith layer pyramid feature of the method is shown, and F is the formed final image feature;
step 2.4: inputting the feature pyramid of the video frame to be detected obtained in the step 2.3 into the decision tree classifier trained in the step 1.4, and multiplying the feature scaling coefficient of each pyramid layer by the set size of the detection window and the detection moving step length to obtain the size of the detection window and the detection moving step length for scanning each pyramid layer; scanning on the corresponding pyramid layer by using the corresponding detection window and the moving step length; judging whether the windows are the head regions or the non-head regions of passengers according to the judgment results of the decision tree on each layer of pyramid and the window scores, scaling the windows which are determined to be the heads to the original size according to the characteristic scaling coefficient, and removing overlapped windows; finally, recording the head position of the passenger in a result file;
and step 3: a tracking head stage:
tracking the detection window by using a related kernel filter method, forming a track, and if the track crosses a specified limit, indicating that the passenger finishes the action of getting on or off the vehicle; training a relevant filter by the relevant kernel filter according to the information of the previous and next frames, and carrying out relevance calculation on the relevant filter and the newly input frame to obtain a confidence image which is a predicted tracking result; the point or block with the highest score is the tracking result;
and 4, step 4: and (3) passenger flow counting:
if the passenger finishes the boarding action, the algorithm updates the data of the added 1 to the passenger flow; otherwise, the passenger flow volume can be kept unchanged;
and 5: storing the coordinates and width and height of the passenger head detected by each frame of image into a file of a detection result, and storing each frame of image; when the bus stops running, the method can output the final passenger flow.

Claims (5)

1. A real-time bus passenger flow volume statistical method is characterized in that video data of the method are passenger flow video data shot from top areas of a front door and a rear door of a bus by using cameras, and the statistical method comprises the following steps:
step 1: model training:
step 1.1: converting real image video frames collected by a camera into a BGRA format; in the video frame, calibrating the head position of the passenger and regarding the head position as a positive sample training image set of a training detection model, and regarding the area without the head of the passenger in the video frame as a negative sample training image set of the training detection model; carrying out image scale normalization processing on the positive sample training image set and the negative sample training image set to form training input images with the same size;
step 1.2: extracting features of the image: calculating a gradient amplitude characteristic and a related direction value of the LUV characteristic; according to the related gradient direction value, quantifying the gradient amplitude characteristic and extracting the histogram characteristic of the gradient amplitude characteristic; then, connecting the LUV characteristic, the gradient amplitude characteristic and the histogram characteristic in series to form a final image characteristic; the specific process is as follows:
step 1.2.1: extracting LUV characteristics on the RGB image;
step 1.2.2: carrying out convolution of triangular filtering on the LUV characteristics;
step 1.2.3: respectively calculating the gradient amplitude and the gradient direction of L, U, V, and selecting the maximum gradient amplitude of the three as a final amplitude result; calculating the amplitude direction according to the gradient amplitude, and outputting the amplitude direction if the amplitude direction is greater than or equal to zero; if the amplitude direction is smaller than zero, outputting the amplitude direction plus one hundred eighty degrees;
step 1.2.4: performing convolution of triangular filtering on the gradient amplitude characteristics;
step 1.2.5: quantizing the gradient amplitude characteristics and the gradient direction to generate a final gradient direction histogram;
step 1.2.6: scaling and connecting the LUV characteristic, the gradient amplitude characteristic and the gradient direction histogram characteristic in series;
the step 2.2 comprises the following processes:
step 2.2.1: extracting LUV characteristics on the RGB image;
step 2.2.2: carrying out convolution of triangular filtering on the LUV characteristics;
step 2.2.3: respectively calculating the gradient amplitude and the gradient direction of L, U, V, and selecting the maximum gradient amplitude of the three as a final amplitude result; calculating the amplitude direction according to the gradient amplitude, and outputting the amplitude direction if the amplitude direction is greater than or equal to zero; if the amplitude direction is smaller than zero, outputting the amplitude direction plus one hundred eighty degrees;
step 2.2.4: performing convolution of triangular filtering on the gradient amplitude characteristics;
step 2.2.5: quantizing the gradient amplitude characteristics and the gradient direction to generate a final gradient direction histogram;
step 2.2.6: scaling and connecting the LUV characteristic, the gradient amplitude characteristic and the gradient direction histogram characteristic in series;
the calculation method for calculating the gradient amplitude characteristic and the correlation direction value of the LUV characteristic is as follows:
Gx(x,y)=I(x+1,y)-I(x-1,y)
Gy(x,y)=I(x,y+1)-I(x,y-1)
Figure FDA0002992417100000021
α(x,y)=tan-1(Gy(x,y)/Gx(x,y))
Gx(x, y) represents the horizontal gradient magnitude of the input image at the pixel point (x, y), Gy(x, y) represents the gradient amplitude of the input image in the vertical direction at the pixel point (x, y), G (x, y) is the gradient amplitude of the pixel point (x, y), alpha (x, y) is the gradient direction of the pixel point (x, y), and I (x, y) represents the characteristic value of L, U, V of the point;
step 1.3: constructing a characteristic pyramid: calculating a characteristic scaling coefficient of each layer of pyramid according to the number of layers of the set characteristic pyramid; then, multiplying the image characteristics by the characteristic scaling coefficient to form a characteristic pyramid;
step 1.4: constructing a training model: sending the generated characteristic pyramid into a decision tree algorithm for training until the training accuracy reaches the requirement, and ending iterative training;
step 2: head detection:
step 2.1: acquiring a video frame to be detected by using a camera;
step 2.2: calculating LUV characteristics of a video frame to be detected, gradient amplitude characteristics of the LUV characteristics and related direction values; according to the related gradient direction value, quantifying the gradient amplitude characteristic and extracting the histogram characteristic of the gradient amplitude characteristic; then, connecting the LUV characteristic, the gradient amplitude characteristic and the histogram characteristic in series to form a final image characteristic;
step 2.3: calculating a characteristic scaling coefficient of each layer of pyramid according to the number of layers of the set characteristic pyramid; then, multiplying the image characteristics by the characteristic scaling coefficient to form a characteristic pyramid;
step 2.4: inputting the feature pyramid of the video frame to be detected obtained in the step 2.3 into the decision tree classifier trained in the step 1.4, and multiplying the feature scaling coefficient of each pyramid layer by the set size of the detection window and the detection moving step length to obtain the size of the detection window and the detection moving step length for scanning each pyramid layer; scanning on the corresponding pyramid layer by using the corresponding detection window and the moving step length; judging whether the windows are the head regions or the non-head regions of passengers according to the judgment results of the decision tree on each layer of pyramid and the window scores, scaling the windows which are determined to be the heads to the original size according to the characteristic scaling coefficient, and removing overlapped windows; finally, recording the head position of the passenger in a result file;
and step 3: tracking the head:
tracking the detection window by using a related kernel filter method, forming a track, and if the track crosses a specified limit, indicating that the passenger finishes the action of getting on or off the vehicle;
and 4, step 4: and (3) passenger flow volume counting:
if the passenger has finished getting on or off, the algorithm will update the data with the addition of 1 to the passenger flow, otherwise, the passenger flow will remain unchanged.
2. The method for real-time statistics of the passenger flow volume of the bus according to claim 1, wherein the method for calculating the final image features in the step 1.3 is as follows:
Fi=μiF,Ficharacteristic of the i-th pyramid of the method, μiAnd F is the coefficient of the ith layer of pyramid characteristics of the method, and F is the formed final image characteristics.
3. The method for real-time bus passenger flow statistics according to any one of claims 1-2, wherein in step 2, a parallel computing method is applied, and on the GPU, feature extraction is performed on a current video frame by using a heterogeneous programming language (OpenCL), and passenger head detection of a previous frame is completed on the CPU, and the two are performed simultaneously, so as to achieve the purpose of parallel processing.
4. The method for real-time statistics of bus passenger flow according to any one of claims 1-2, characterized in that in step 3, a correlation kernel filter is trained according to the information of the previous and next frames, and correlation calculation is performed with the newly inputted frame, and the obtained confidence map is the predicted tracking result; the point or block with the highest score is the tracking result.
5. The method for real-time bus passenger flow statistics according to any one of claims 1-2, characterized by further comprising a step 5, wherein the process is as follows:
and 5: storing the coordinates and width and height of the passenger head detected by each frame of image into a file of a detection result, and storing each frame of image; when the bus stops running, the method can output the final passenger flow.
CN201810505385.5A 2018-05-24 2018-05-24 Real-time bus passenger flow volume statistical method Expired - Fee Related CN108830166B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201810505385.5A CN108830166B (en) 2018-05-24 2018-05-24 Real-time bus passenger flow volume statistical method

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201810505385.5A CN108830166B (en) 2018-05-24 2018-05-24 Real-time bus passenger flow volume statistical method

Publications (2)

Publication Number Publication Date
CN108830166A CN108830166A (en) 2018-11-16
CN108830166B true CN108830166B (en) 2021-06-29

Family

ID=64148514

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201810505385.5A Expired - Fee Related CN108830166B (en) 2018-05-24 2018-05-24 Real-time bus passenger flow volume statistical method

Country Status (1)

Country Link
CN (1) CN108830166B (en)

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP7331769B2 (en) * 2020-04-30 2023-08-23 トヨタ自動車株式会社 Position estimation system and position estimation method

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103985182A (en) * 2014-05-30 2014-08-13 长安大学 Automatic public transport passenger flow counting method and system
CN206400573U (en) * 2016-12-27 2017-08-11 清华大学苏州汽车研究院(吴江) A kind of counting passenger flow of buses system based on ToF cameras
CN107239762A (en) * 2017-06-06 2017-10-10 电子科技大学 Patronage statistical method in a kind of bus of view-based access control model

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
TWI448990B (en) * 2012-09-07 2014-08-11 Univ Nat Chiao Tung Real-time people counting system using layer scanning method

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103985182A (en) * 2014-05-30 2014-08-13 长安大学 Automatic public transport passenger flow counting method and system
CN206400573U (en) * 2016-12-27 2017-08-11 清华大学苏州汽车研究院(吴江) A kind of counting passenger flow of buses system based on ToF cameras
CN107239762A (en) * 2017-06-06 2017-10-10 电子科技大学 Patronage statistical method in a kind of bus of view-based access control model

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
"基于卷积神经网络的人脸检测与特征点标定算法研究";王维;《中国优秀硕士学位论文全文数据库(信息科技辑)》;20180415(第4期);全文 *
"监控场景中人数统计算法的研究与应用";马海军;《中国优秀硕士学位论文全文数据库(信息科技辑)》;20161015(第10期);第2、3章 *

Also Published As

Publication number Publication date
CN108830166A (en) 2018-11-16

Similar Documents

Publication Publication Date Title
CN110059558B (en) Orchard obstacle real-time detection method based on improved SSD network
WO2021238019A1 (en) Real-time traffic flow detection system and method based on ghost convolutional feature fusion neural network
US10445602B2 (en) Apparatus and method for recognizing traffic signs
CN106648078B (en) Multi-mode interaction method and system applied to intelligent robot
CN110826389B (en) Gait recognition method based on attention 3D frequency convolution neural network
CN110956082B (en) Face key point detection method and detection system based on deep learning
WO2015131468A1 (en) Method and system for estimating fingerprint pose
CN112861785B (en) Instance segmentation and image restoration-based pedestrian re-identification method with shielding function
CN112784712B (en) Missing child early warning implementation method and device based on real-time monitoring
WO2023151237A1 (en) Face pose estimation method and apparatus, electronic device, and storage medium
CN111915583A (en) Vehicle and pedestrian detection method based on vehicle-mounted thermal infrared imager in complex scene
CN113963333B (en) Traffic sign board detection method based on improved YOLOF model
CN112784494B (en) Training method of false positive recognition model, target recognition method and device
CN108830166B (en) Real-time bus passenger flow volume statistical method
CN113112479A (en) Progressive target detection method and device based on key block extraction
CN117152625A (en) Remote sensing small target identification method, system, equipment and medium based on CoordConv and Yolov5
CN108985216B (en) Pedestrian head detection method based on multivariate logistic regression feature fusion
CN106909936B (en) Vehicle detection method based on double-vehicle deformable component model
CN114067359B (en) Pedestrian detection method integrating human body key points and visible part attention characteristics
CN113658223B (en) Multi-row person detection and tracking method and system based on deep learning
CN113111850B (en) Human body key point detection method, device and system based on region-of-interest transformation
CN112801020B (en) Pedestrian re-identification method and system based on background graying
CN115937736A (en) Small target detection method based on attention and context awareness
CN114639013A (en) Remote sensing image airplane target detection and identification method based on improved Orient RCNN model
CN114038011A (en) Method for detecting abnormal behaviors of human body in indoor scene

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant
CF01 Termination of patent right due to non-payment of annual fee
CF01 Termination of patent right due to non-payment of annual fee

Granted publication date: 20210629