CN112598709A - Pedestrian movement speed intelligent sensing method based on video stream - Google Patents

Pedestrian movement speed intelligent sensing method based on video stream Download PDF

Info

Publication number
CN112598709A
CN112598709A CN202011559927.0A CN202011559927A CN112598709A CN 112598709 A CN112598709 A CN 112598709A CN 202011559927 A CN202011559927 A CN 202011559927A CN 112598709 A CN112598709 A CN 112598709A
Authority
CN
China
Prior art keywords
pedestrian
coordinate system
frame
camera
video stream
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202011559927.0A
Other languages
Chinese (zh)
Other versions
CN112598709B (en
Inventor
寄珊珊
李特
孟启炜
朱世强
顾建军
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Zhejiang Lab
Original Assignee
Zhejiang Lab
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Zhejiang Lab filed Critical Zhejiang Lab
Priority to CN202011559927.0A priority Critical patent/CN112598709B/en
Publication of CN112598709A publication Critical patent/CN112598709A/en
Application granted granted Critical
Publication of CN112598709B publication Critical patent/CN112598709B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/20Analysis of motion
    • G06T7/246Analysis of motion using feature-based methods, e.g. the tracking of corners or segments
    • G06T7/248Analysis of motion using feature-based methods, e.g. the tracking of corners or segments involving reference images or patches
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/70Determining position or orientation of objects or cameras
    • G06T7/73Determining position or orientation of objects or cameras using feature-based methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/80Analysis of captured images to determine intrinsic or extrinsic camera parameters, i.e. camera calibration
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/103Static body considered as a whole, e.g. static pedestrian or occupant recognition
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10016Video; Image sequence
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30196Human being; Person
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30204Marker
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30232Surveillance

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Multimedia (AREA)
  • Human Computer Interaction (AREA)
  • Image Analysis (AREA)

Abstract

The invention discloses a pedestrian movement speed intelligent sensing method based on video streaming, which comprises the following steps: acquiring internal parameters and external parameters of a camera by adopting a camera calibration method, and establishing a coordinate conversion model from an image coordinate system to a space coordinate system; acquiring a real-time video stream under a monitoring camera, and carrying out pedestrian target detection and tracking to obtain a pedestrian ID and a pedestrian detection frame coarse positioning result under a corresponding image coordinate system; performing component segmentation on the obtained pedestrian target by using a human body component segmentation model to obtain a pedestrian foot point fine positioning result under an image coordinate system; the method comprises the steps of utilizing a coordinate conversion model to realize position information resolving of pedestrian foot points from an image coordinate system to a space coordinate system in the real world, calculating the moving speed of pedestrians by combining video stream frame interval time, and realizing the intelligent sensing method of the moving speed of the pedestrians based on the video stream of the camera.

Description

Pedestrian movement speed intelligent sensing method based on video stream
Technical Field
The invention relates to the technical field of machine vision and deep learning, in particular to a pedestrian movement speed intelligent sensing method based on video streaming.
Background
The pedestrian monitoring system has the advantages that intelligent sensing such as detection, tracking, space position positioning and real-time movement speed calculation is carried out on pedestrians under the monitoring camera, and the pedestrian monitoring system is of great significance in helping to establish intelligent security and free manpower.
The traditional method for acquiring the motion information of the target pedestrian by means of an external sensor or an external object worn on the pedestrian has poor user experience, adopts an intelligent algorithm based on vision to realize non-contact pedestrian spatial information perception and motion speed calculation, is limited by the precision of a model converted from an image coordinate system to a spatial coordinate system in terms of precision, and is particularly easily influenced by the precision of a pedestrian detection frame.
Disclosure of Invention
Aiming at the defects of the prior art, the invention provides a pedestrian movement speed intelligent sensing method based on video streaming.
The purpose of the invention is realized by the following technical scheme: a pedestrian movement speed intelligent sensing method based on video streaming comprises the following steps:
acquiring a calibration picture under a camera, calibrating the camera, acquiring internal parameters and external parameters of a monitoring camera, and establishing a coordinate conversion model from an image coordinate system to a world coordinate system;
acquiring a real-time video stream of a monitoring camera according to a streaming media protocol, detecting and tracking pedestrians by adopting a pedestrian detection model and a multi-target tracking model, outputting a pedestrian ID and a detection frame, and obtaining a pedestrian coarse positioning result under an image coordinate system;
step three, acquiring a pedestrian picture according to the pedestrian detection frame in the step two, inputting a trained human body part segmentation model for segmenting the pedestrian part, and acquiring a pedestrian foot point fine positioning result under an image coordinate system;
step four, according to pixel information of the foot points of the pedestrians in an image coordinate system, coordinate conversion calculation is carried out by combining the coordinate conversion model established in the step one, and space coordinate information of the pedestrians taking the ground plane as a two-dimensional plane XY axis and the ground perpendicular to the ground as a Z axis is obtained; and combining the space coordinate information with a pedestrian tracking algorithm to obtain the space position moving distance of the target pedestrian under the interval frame, and then combining the interval frame time to calculate the pedestrian movement speed.
Furthermore, the camera calibration method comprises the steps of obtaining internal parameters and external parameters of the camera, wherein the internal parameters are obtained by a checkerboard calibration method and comprise an internal parameter matrix MCAnd a distortion coefficient dist, wherein external parameters are calculated by adopting an ArUcormarker two-dimensional code calibration cloth and combining with internal parameters of the monitoring camera by utilizing a PNP method, and the external parameters comprise a rotation matrix R and a translation matrix T;
further, the third step comprises the following substeps:
(3.1) according to the pedestrian coarse positioning result obtained in the step two, cutting all pedestrian pictures from the original image frame of the real-time video stream of the monitoring camera for batch processing, inputting a trained human body part segmentation model for pedestrian segmentation, and outputting a pedestrian part segmentation result;
(3.2) judging whether the video frame image contains the pedestrian foot points according to the pedestrian segmentation result, and if not, not carrying out coordinate conversion calculation; if the pedestrian foot points are contained, the pedestrian foot points are converted into pixel coordinates of the original video image frame through scale conversion, then the lower frame of the pedestrian detection frame is corrected according to the lower boundary result of the pedestrian foot point pixels, and the middle point of the connecting line of the lower left corner and the lower right corner of the corrected pedestrian detection frame is output and used as a pedestrian foot point fine positioning output value.
Further, the training process of the human body part segmentation model comprises the following steps: the PPSS dataset is used as a training set, a human body part segmentation model is input, training is carried out in a random gradient descent mode, the initial learning rate is set to be 0.03, the learning rate is updated through a linear attenuation learning strategy, and when the training times or the loss function reaches a threshold value, training of the human body part segmentation model is completed.
Further, the pedestrian movement speed
Figure BDA0002860053520000021
The calculation process is as follows:
Figure BDA0002860053520000031
wherein the content of the first and second substances,
Figure BDA0002860053520000032
represents the abscissa of the pedestrian with the ID of i in the ground two-dimensional coordinate system at the time of the N-k frame,
Figure BDA0002860053520000033
represents the ordinate of the pedestrian with ID i in the ground two-dimensional coordinate system at the time of the N-k frame,
Figure BDA0002860053520000034
represents the abscissa of the pedestrian with the ID of i in the ground two-dimensional coordinate system at the time of the Nth frame,
Figure BDA0002860053520000035
and f is the video stream frame frequency, and k is the interval frame number.
Compared with the prior art, the invention has the following beneficial effects: adopting a non-contact intelligent sensing method, utilizing a coordinate conversion model to realize space position information resolving and calculating the moving speed of the pedestrian by combining the video stream frame interval time; in the conversion from an image coordinate system to a space coordinate system, the model precision is limited by the foot point position information of the pedestrian, the obtained detection frame only adopting the pedestrian detection model is relatively rough, and under the condition that the pedestrian is shielded, the lower boundary of the pedestrian given by the pedestrian detection frame is not the lower boundary of the foot point position of the real pedestrian, so that a large error is easily caused; on the basis of a detection frame output by pedestrian detection, a pedestrian component segmentation network is adopted to segment human body components, so that pedestrian foot point judgment can be carried out, and more accurate pixel-level pedestrian foot point positions in an image coordinate system can be obtained; therefore, a more accurate pedestrian space position can be obtained, the calculation accuracy of subsequent pedestrian movement speed perception is correspondingly improved to some extent, and the accuracy can reach centimeter level.
Drawings
Fig. 1 is a flow chart of a pedestrian movement speed intelligent sensing method based on video streaming.
Detailed Description
Fig. 1 is a flow chart of a pedestrian movement speed intelligent sensing method based on video streaming, and the pedestrian movement speed intelligent sensing method includes the following steps:
acquiring a calibration picture under a camera, calibrating the camera, acquiring internal parameters and external parameters of a monitoring camera, and establishing a coordinate conversion model from an image coordinate system to a world coordinate system;
manufacturing a checkerboard calibration plate with inner angular points of 11 x 8, and collecting about 20 checkerboard calibration pictures at different positions, at different angles and in different postures; detecting the angular points of the chessboard pattern calibration picture, searching sub-pixel angular points on the basis, removing the pictures with the detected angular points less than 88, calibrating the camera according to the Zhang calibration method, and obtaining a camera internal reference matrix MCAnd a distortion coefficient dist, performing precision judgment by adopting a back projection error, performing distortion correction according to the distortion coefficient, and outputting the internal parameters of the corrected and optimized camera;
the method comprises the steps of generating two-dimensional code markers for external parameter calibration by adopting an ArUco two-dimensional code dictionary, manufacturing calibration cloth, wherein the number of the two-dimensional code markers on the calibration cloth is 5 × 5, the length of each marker is 50cm, the interval between the markers is also 50cm, and the larger size of the marker is selected to meet the requirement that a monitoring camera can identify the two-dimensional code mark points at the far position in a monitoring vision field. The method comprises the steps of placing a two-dimensional code calibration cloth on the ground flatly, collecting a picture, detecting two-dimensional code mark points in the calibrated picture, obtaining pixel coordinates of the mark points when the number of the detected two-dimensional code mark points is more than or equal to 4, carrying out PNP calculation by combining internal parameters of a camera, calculating a camera rotation matrix R and a translation matrix T, and obtaining external parameters of the camera.
According to the field range of the monitoring camera deployed in the indoor monitoring scene,establishing a three-dimensional world coordinate system which takes a ground plane as a two-dimensional plane X axis and a two-dimensional plane Y axis and takes a vertical direction to the ground as a Z axis, and selecting a proper reference point as a coordinate origin. And establishing a conversion relation between an image coordinate system and a world coordinate system by combining internal parameters and external parameters of the camera:
Figure BDA0002860053520000041
from P-1P ═ E can be obtained:
Figure BDA0002860053520000042
thereby obtaining a conversion formula from a two-dimensional image coordinate system to a three-dimensional world coordinate system:
Figure BDA0002860053520000043
wherein Z isCDenotes camera coordinates, [ u, v ]]Representing image pixel coordinates, [ X ]W,YW,ZW]Representing world coordinates of a pedestrian target, R representing a surveillance camera rotation matrix, T representing a surveillance camera translation matrix, MCRepresenting a surveillance camera internal reference matrix. For unknown parameter ZCThe solution is as follows:
Figure BDA0002860053520000051
order to
Figure BDA0002860053520000052
R-1·T=P2From
Figure BDA0002860053520000053
The third behavior 1 may result in: zC·P1[2]=Zw·P2[2]Thereby obtaining ZC=(Zw+P2[2])/P1[2];
All the parameters from the two-dimensional image coordinate system to the three-dimensional world coordinate system are obtained, and the coordinates [ u, v ] of the pedestrian foot point in the image coordinate system are obtained]Substituting and making the coordinate axis perpendicular to the ground in the three-dimensional world coordinate system as ZwAnd (5) obtaining the coordinate output result of the pedestrian under the space coordinate system when the coordinate is 0.
Step two, acquiring a real-time video stream of the monitoring camera according to a streaming media protocol, wherein the frame rate is 25fps, and the resolution is 1920 x 1080; adopting a pedestrian detection model and a multi-target tracking model to detect and track pedestrians, outputting a pedestrian ID and a detection frame, and obtaining a pedestrian coarse positioning result under an image coordinate system;
the pedestrian detection model adopts a YoloV3 algorithm, and a backbone network is Darknet 53; the size of the input picture after preprocessing is 416, the number of detection frames output by the network is [ (13 × 13) + (26 × 26) + (52 × 52) ] × 3-10647, then the detection frames output by the non-maximum suppression NMS algorithm are subjected to post-processing, and frame screening is carried out according to the set threshold and the confidence coefficient to obtain the output of pedestrian detection;
the multi-target tracking model adopts a Deepsort-based multi-target tracking algorithm to continuously track the pedestrians, firstly, a Kalman filter is adopted to estimate the motion state of the pedestrians, and motion information association is carried out according to the Markov distance between the detected frame position of the current frame and the position predicted by the Kalman filter; appearance information association firstly adopts a pedestrian ReID feature extraction network based on Resnet50 to extract appearance features, the feature dimension is 512, then similarity calculation is carried out by adopting Euclidean distance, whether association is successful or not is judged through a set threshold value, and finally linear weighting is carried out by combining two associations of motion features and appearance features to serve as a final association result; and if the tracked target continuously exceeds 3 frames and is not matched with the detection frame, judging whether the target is deleted or not by setting the maximum survival period, thereby realizing continuous tracking of multiple pedestrians under continuous frame video images and obtaining the ID of the pedestrian.
Step three, acquiring a pedestrian picture according to the pedestrian detection frame in the step two, inputting a trained human body part segmentation model for segmenting the pedestrian part, and acquiring a pedestrian foot point fine positioning result under an image coordinate system;
the human body part segmentation model adopts a semantic segmentation network DeepLab V3+, a light-weight network MobileNet V2 is adopted as a backbone network, the segmentation model comprises an encoder module and a decoder module, the encoder module performs feature extraction through the series connection of a plurality of layers of convolutional layers, and then multi-scale features are obtained through a void space pyramid pooling module; the decoder module performs two times of upsampling by using the bilinear difference value, wherein the first time of upsampling fuses bottom layer features with the same resolution ratio features in the decoder module network, and the feature pixels restored to the original input size can be output by the segmentation network after two times of upsampling. The training process of the human body part segmentation model comprises the following steps: the method comprises the steps of inputting a human body part segmentation model by using a PPSS dataset as a training set, training by using a random gradient descent mode, setting an initial learning rate to be 0.03, updating the learning rate by using a linear attenuation learning strategy, finishing training of the human body part segmentation model after the training times or the Loss function reaches a threshold value, and adopting a Softmax Loss function as a Loss function of the human body part segmentation model by a person skilled in the art.
The third step specifically comprises the following substeps:
(3.1) according to the pedestrian rough positioning result in the second step, cutting all pedestrian pictures from the original image frame of the real-time video stream of the monitoring camera to perform batch processing, inputting N C H W into a trained human body part segmentation model to perform pedestrian segmentation, and outputting a pedestrian part segmentation result, wherein N is the number of pedestrian targets in a single frame image, C is the number of channels of the input image, C is 3, H is the image height, H is 160, W is the image width, W is 80, obtaining a semantic feature map after dimensionality reduction through a feature coding unit, then recovering the feature of the input resolution of the original image after two times of upsampling through a feature decoding unit, and the value set of feature pixel values is {0,1,2,3,4,5,6,7}, which respectively represent the human body parts: hair, face, upper body, arms, lower body, legs, feet, and background;
(3.2) after the human body part segmentation model is processed, a fine-grained semantic segmentation result of a human body pixel level can be further obtained on the basis of the pedestrian detection frame, so that the pedestrian detection frame is further corrected; judging whether the video frame image contains the pedestrian foot points according to the pedestrian segmentation result, if not, judging that the current pedestrian does not have the foot point pedestrian foot points according to the characteristic value 6(6 marks the pixel value of the pedestrian foot points), and not carrying out coordinate conversion calculation; it should be noted that the situation that the foot points do not exist may be due to the blocking phenomenon between pedestrians, and when the target reappears, we can perform continuous tracking according to the pedestrian tracking algorithm and perform subsequent calculation of the moving speed of the pedestrian according to the recorded time interval and the moving distance. If the pedestrian foot points are contained, the pedestrian foot points are converted into pixel coordinates of the original video image frame through scale conversion, then the lower frame of the pedestrian detection frame is corrected according to the lower boundary result of the pedestrian foot point pixels, and the middle point of the connecting line of the lower left corner and the lower right corner of the corrected pedestrian detection frame is output and used as a pedestrian foot point fine positioning output value.
Step four, according to pixel information of the foot points of the pedestrians in an image coordinate system, coordinate conversion calculation is carried out by combining the coordinate conversion model established in the step one, and space coordinate information of the pedestrians taking the ground plane as a two-dimensional plane XY axis and the ground perpendicular to the ground as a Z axis is obtained; and combining the space coordinate information with a pedestrian tracking algorithm to obtain the space position moving distance of the target pedestrian under the interval frame, and then combining the interval frame time to calculate the pedestrian movement speed.
(4.1) two-dimensional image coordinates [ u, v ] of pedestrian foot points in the image coordinate system determined according to the pedestrian detection frame and the pedestrian component segmentation network]And combining the image coordinate system established in the step one with a conversion model from the world coordinate system to obtain:
Figure BDA0002860053520000071
corresponding parameters are brought in, and the coordinates [ X, Y, Z ] of the foot points of the pedestrians under the world coordinate system can be obtained];
(4.2) in combination with a pedestrian tracking algorithm,the pedestrian ID and the corresponding pedestrian space coordinate information under continuous interval frames can be obtained; ideally, the height of the foot point of the pedestrian on the ground is 0, that is, Z is 0, so that the XY-axis coordinates [ X, Y ] of the foot point of the pedestrian on the two-dimensional plane of the ground plane can be obtained]Suppose for a continuously tracked pedestrian that its identity ID is i, its coordinates in the ground two-dimensional coordinates at frame N-k are
Figure BDA0002860053520000081
The coordinates of the ground two-dimensional coordinate system at the Nth frame are
Figure BDA0002860053520000082
The video stream frame rate is f, and k is the interval frame number; the moving distance of the pedestrian in the nth frame of the current frame on the two-dimensional plane can be calculated by the euclidean distance, and the moving speed of the pedestrian can be obtained by combining the time of the interval frame:
Figure BDA0002860053520000083
and sending the pedestrian target ID and the pedestrian movement speed to the front end for visual display through the message middleware, thereby finishing intelligent perception of the pedestrian movement speed based on the video stream.
In summary, the video stream is acquired by the aid of the camera deployed at the monitoring point, the real-time movement speed of the target pedestrian is acquired by the aid of a non-contact visual perception algorithm, behaviors such as static pedestrians and excessive pedestrian movement speed are judged, and effective early warning and customized service is provided for a pedestrian intelligent perception system in a monitoring scene. The method realizes the coordinate conversion from an image coordinate system to a space coordinate system in the real world through a camera calibration method, and further foot point judgment and fine positioning are carried out by adopting a human body part segmentation model on the basis of a pedestrian detection frame because the conversion precision is limited by the positioning result of the foot points of pedestrians in the image coordinate system, so that the intelligent sensing of the pedestrian movement speed based on the camera video stream is more accurate.

Claims (5)

1. A pedestrian movement speed intelligent sensing method based on video streaming is characterized by comprising the following steps:
acquiring a calibration picture under a camera, calibrating the camera, acquiring internal parameters and external parameters of a monitoring camera, and establishing a coordinate conversion model from an image coordinate system to a world coordinate system;
acquiring a real-time video stream of a monitoring camera according to a streaming media protocol, detecting and tracking pedestrians by adopting a pedestrian detection model and a multi-target tracking model, outputting a pedestrian ID and a detection frame, and obtaining a pedestrian coarse positioning result under an image coordinate system;
step three, acquiring a pedestrian picture according to the pedestrian detection frame in the step two, inputting a trained human body part segmentation model for segmenting the pedestrian part, and acquiring a pedestrian foot point fine positioning result under an image coordinate system;
step four, according to pixel information of the foot points of the pedestrians in an image coordinate system, coordinate conversion calculation is carried out by combining the coordinate conversion model established in the step one, and space coordinate information of the pedestrians taking the ground plane as a two-dimensional plane XY axis and the ground perpendicular to the ground as a Z axis is obtained; and combining the space coordinate information with a pedestrian tracking algorithm to obtain the space position moving distance of the target pedestrian under the interval frame, and then combining the interval frame time to calculate the pedestrian movement speed.
2. The method according to claim 1, wherein the camera calibration method comprises obtaining internal parameters and external parameters of the camera, the internal parameters are obtained by a checkerboard calibration method, and the internal parameters comprise an internal parameter matrix MCAnd distortion coefficients dist, wherein external parameters are calculated by adopting ArUcoMarker two-dimensional code calibration cloth and combining with internal parameters of the monitoring camera by utilizing a PNP method, and the external parameters comprise a rotation matrix R and a translation matrix T.
3. The video stream-based pedestrian motion speed intelligent sensing method according to claim 1, wherein the third step comprises the following sub-steps:
(3.1) according to the pedestrian coarse positioning result obtained in the step two, cutting all pedestrian pictures from the original image frame of the real-time video stream of the monitoring camera for batch processing, inputting a trained human body part segmentation model for pedestrian segmentation, and outputting a pedestrian part segmentation result;
(3.2) judging whether the video frame image contains the pedestrian foot points according to the pedestrian segmentation result, and if not, not carrying out coordinate conversion calculation; if the pedestrian foot points are contained, the pedestrian foot points are converted into pixel coordinates of the original video image frame through scale conversion, then the lower frame of the pedestrian detection frame is corrected according to the lower boundary result of the pedestrian foot point pixels, and the middle point of the connecting line of the lower left corner and the lower right corner of the corrected pedestrian detection frame is output and used as a pedestrian foot point fine positioning output value.
4. The pedestrian motion speed intelligent sensing method based on the video stream as claimed in claim 1, wherein the training process of the human body part segmentation model is as follows: the PPSS dataset is used as a training set, a human body part segmentation model is input, training is carried out in a random gradient descent mode, the initial learning rate is set to be 0.03, the learning rate is updated through a linear attenuation learning strategy, and when the training times or the loss function reaches a threshold value, training of the human body part segmentation model is completed.
5. The video stream-based pedestrian motion speed intelligent sensing method according to claim 1, wherein the pedestrian motion speed is
Figure FDA0002860053510000021
The calculation process is as follows:
Figure FDA0002860053510000022
wherein the content of the first and second substances,
Figure FDA0002860053510000023
represents the abscissa of the pedestrian with the ID of i in the ground two-dimensional coordinate system at the time of the N-k frame,
Figure FDA0002860053510000024
represents the ordinate of the pedestrian with ID i in the ground two-dimensional coordinate system at the time of the N-k frame,
Figure FDA0002860053510000025
represents the abscissa of the pedestrian with the ID of i in the ground two-dimensional coordinate system at the time of the Nth frame,
Figure FDA0002860053510000026
and f is the video stream frame frequency, and k is the interval frame number.
CN202011559927.0A 2020-12-25 2020-12-25 Pedestrian movement speed intelligent sensing method based on video stream Active CN112598709B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202011559927.0A CN112598709B (en) 2020-12-25 2020-12-25 Pedestrian movement speed intelligent sensing method based on video stream

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202011559927.0A CN112598709B (en) 2020-12-25 2020-12-25 Pedestrian movement speed intelligent sensing method based on video stream

Publications (2)

Publication Number Publication Date
CN112598709A true CN112598709A (en) 2021-04-02
CN112598709B CN112598709B (en) 2022-11-01

Family

ID=75201948

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202011559927.0A Active CN112598709B (en) 2020-12-25 2020-12-25 Pedestrian movement speed intelligent sensing method based on video stream

Country Status (1)

Country Link
CN (1) CN112598709B (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113658221A (en) * 2021-07-28 2021-11-16 同济大学 Monocular camera-based AGV pedestrian following method
CN113840228A (en) * 2021-08-25 2021-12-24 北京航空航天大学杭州创新研究院 Pedestrian indoor positioning method based on positioning matching

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20120237086A1 (en) * 2009-12-03 2012-09-20 National Institute Of Advanced Industrial Science And Technology Moving body positioning device
US20140348382A1 (en) * 2013-05-22 2014-11-27 Hitachi, Ltd. People counting device and people trajectory analysis device
US20180075593A1 (en) * 2016-09-15 2018-03-15 Qualcomm Incorporated Automatic scene calibration method for video analytics
US20200005490A1 (en) * 2016-09-28 2020-01-02 Chung Ang University Academic Cooperation Foundation Normalized metadata generation device, object occlusion detection device and method

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20120237086A1 (en) * 2009-12-03 2012-09-20 National Institute Of Advanced Industrial Science And Technology Moving body positioning device
US20140348382A1 (en) * 2013-05-22 2014-11-27 Hitachi, Ltd. People counting device and people trajectory analysis device
US20180075593A1 (en) * 2016-09-15 2018-03-15 Qualcomm Incorporated Automatic scene calibration method for video analytics
US20200005490A1 (en) * 2016-09-28 2020-01-02 Chung Ang University Academic Cooperation Foundation Normalized metadata generation device, object occlusion detection device and method

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
CAI, H.,ET.AL: "Multi-scale body-part mask guided attention for person reidentification", 《2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION》 *
梅鸿翔: "一种基于图像坐标检测的目标定位方法", 《计算机与数字工程》 *
陈成等: "单目视频人体三维运动高效恢复", 《计算机辅助设计与图形学学报》 *

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113658221A (en) * 2021-07-28 2021-11-16 同济大学 Monocular camera-based AGV pedestrian following method
CN113658221B (en) * 2021-07-28 2024-04-26 同济大学 AGV pedestrian following method based on monocular camera
CN113840228A (en) * 2021-08-25 2021-12-24 北京航空航天大学杭州创新研究院 Pedestrian indoor positioning method based on positioning matching
CN113840228B (en) * 2021-08-25 2024-04-02 北京航空航天大学杭州创新研究院 Pedestrian indoor positioning method based on positioning matching

Also Published As

Publication number Publication date
CN112598709B (en) 2022-11-01

Similar Documents

Publication Publication Date Title
CN111462200B (en) Cross-video pedestrian positioning and tracking method, system and equipment
CN106846359B (en) Moving target rapid detection method based on video sequence
CN110569704B (en) Multi-strategy self-adaptive lane line detection method based on stereoscopic vision
JP6095018B2 (en) Detection and tracking of moving objects
Sidla et al. Pedestrian detection and tracking for counting applications in crowded situations
CN109934848B (en) Method for accurately positioning moving object based on deep learning
CN108629946B (en) Human body falling detection method based on RGBD sensor
CN107481315A (en) A kind of monocular vision three-dimensional environment method for reconstructing based on Harris SIFT BRIEF algorithms
CN104598883B (en) Target knows method for distinguishing again in a kind of multiple-camera monitoring network
CN112801074B (en) Depth map estimation method based on traffic camera
CN112598709B (en) Pedestrian movement speed intelligent sensing method based on video stream
CN111027432B (en) Gait feature-based visual following robot method
CN104200492B (en) Video object automatic detection tracking of taking photo by plane based on profile constraints
CN110941996A (en) Target and track augmented reality method and system based on generation of countermeasure network
CN110516639B (en) Real-time figure three-dimensional position calculation method based on video stream natural scene
CN107481267A (en) A kind of shooting projection interactive system and method based on binocular vision
CN115376034A (en) Motion video acquisition and editing method and device based on human body three-dimensional posture space-time correlation action recognition
CN116309686A (en) Video positioning and speed measuring method, device and equipment for swimmers and storage medium
KR100574227B1 (en) Apparatus and method for separating object motion from camera motion
CN107045630B (en) RGBD-based pedestrian detection and identity recognition method and system
CN115880643B (en) Social distance monitoring method and device based on target detection algorithm
CN108986162B (en) Dish and background segmentation method based on inertial measurement unit and visual information
Vigus et al. Video object tracking using region split and merge and a Kalman filter tracking algorithm
CN114170317B (en) Swimming pool drowning prevention head position judging method and device and computer equipment
CN112395985B (en) Ground unmanned vehicle vision road detection method based on unmanned aerial vehicle image

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant