CN112598709B - Pedestrian movement speed intelligent sensing method based on video stream - Google Patents

Pedestrian movement speed intelligent sensing method based on video stream Download PDF

Info

Publication number
CN112598709B
CN112598709B CN202011559927.0A CN202011559927A CN112598709B CN 112598709 B CN112598709 B CN 112598709B CN 202011559927 A CN202011559927 A CN 202011559927A CN 112598709 B CN112598709 B CN 112598709B
Authority
CN
China
Prior art keywords
pedestrian
coordinate system
frame
camera
video stream
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202011559927.0A
Other languages
Chinese (zh)
Other versions
CN112598709A (en
Inventor
寄珊珊
李特
孟启炜
朱世强
顾建军
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Zhejiang Lab
Original Assignee
Zhejiang Lab
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Zhejiang Lab filed Critical Zhejiang Lab
Priority to CN202011559927.0A priority Critical patent/CN112598709B/en
Publication of CN112598709A publication Critical patent/CN112598709A/en
Application granted granted Critical
Publication of CN112598709B publication Critical patent/CN112598709B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/20Analysis of motion
    • G06T7/246Analysis of motion using feature-based methods, e.g. the tracking of corners or segments
    • G06T7/248Analysis of motion using feature-based methods, e.g. the tracking of corners or segments involving reference images or patches
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/70Determining position or orientation of objects or cameras
    • G06T7/73Determining position or orientation of objects or cameras using feature-based methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/80Analysis of captured images to determine intrinsic or extrinsic camera parameters, i.e. camera calibration
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/103Static body considered as a whole, e.g. static pedestrian or occupant recognition
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10016Video; Image sequence
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30196Human being; Person
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30204Marker
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30232Surveillance

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Multimedia (AREA)
  • Human Computer Interaction (AREA)
  • Image Analysis (AREA)

Abstract

The invention discloses a pedestrian movement speed intelligent sensing method based on video streaming, which comprises the following steps: acquiring internal parameters and external parameters of a camera by adopting a camera calibration method, and establishing a coordinate conversion model from an image coordinate system to a space coordinate system; acquiring a real-time video stream under a monitoring camera, and carrying out pedestrian target detection and tracking to obtain a pedestrian ID and a pedestrian detection frame coarse positioning result under a corresponding image coordinate system; performing component segmentation on the obtained pedestrian target by using a human body component segmentation model to obtain a pedestrian foot point fine positioning result under an image coordinate system; the method comprises the steps of utilizing a coordinate conversion model to realize position information resolving of pedestrian foot points from an image coordinate system to a space coordinate system in the real world, calculating the moving speed of pedestrians by combining video stream frame interval time, and realizing the intelligent sensing method of the moving speed of the pedestrians based on the video stream of the camera.

Description

Pedestrian movement speed intelligent sensing method based on video stream
Technical Field
The invention relates to the technical field of machine vision and deep learning, in particular to a pedestrian movement speed intelligent sensing method based on video streaming.
Background
The pedestrian detection and tracking system has the advantages that intelligent sensing such as detection, tracking, spatial position positioning and real-time movement speed calculation is carried out on pedestrians under the monitoring camera, and the pedestrian detection and tracking system is of great significance to help to establish intelligent security and manpower liberation.
The traditional method for acquiring the motion information of the target pedestrian by means of an external sensor or an external object worn on the pedestrian has poor user experience, adopts an intelligent algorithm based on vision to realize non-contact pedestrian spatial information perception and motion speed calculation, is limited by the precision of a model converted from an image coordinate system to a spatial coordinate system in terms of precision, and is particularly easily influenced by the precision of a pedestrian detection frame.
Disclosure of Invention
Aiming at the defects of the prior art, the invention provides a pedestrian movement speed intelligent perception method based on video streaming.
The purpose of the invention is realized by the following technical scheme: a pedestrian movement speed intelligent sensing method based on video streaming comprises the following steps:
acquiring a calibration picture under a camera, calibrating the camera, acquiring internal parameters and external parameters of a monitoring camera, and establishing a coordinate conversion model from an image coordinate system to a world coordinate system;
step two, acquiring a real-time video stream of a monitoring camera according to a streaming media protocol, detecting and tracking pedestrians by adopting a pedestrian detection model and a multi-target tracking model, outputting a pedestrian ID and a detection frame, and obtaining a pedestrian coarse positioning result under an image coordinate system;
step three, acquiring a pedestrian picture according to the pedestrian detection frame in the step two, inputting a trained human body part segmentation model for segmenting the pedestrian part, and acquiring a pedestrian foot point fine positioning result under an image coordinate system;
step four, according to pixel information of the foot points of the pedestrians in an image coordinate system, coordinate conversion calculation is carried out by combining the coordinate conversion model established in the step one, and space coordinate information of the pedestrians taking the ground plane as a two-dimensional plane XY axis and the ground perpendicular to the ground as a Z axis is obtained; and combining the space coordinate information with a pedestrian tracking algorithm to obtain the space position moving distance of the target pedestrian under the interval frame, and then combining the interval frame time to calculate the pedestrian movement speed.
Furthermore, the camera calibration method comprises the steps of obtaining internal parameters and external parameters of the camera, wherein the internal parameters are checked by a checkerboardObtaining by a calibration method, including an internal parameter matrix MCAnd a distortion coefficient dist, wherein external parameters are calculated by adopting an ArUcormarker two-dimensional code calibration cloth and combining with internal parameters of the monitoring camera by utilizing a PNP method, and the external parameters comprise a rotation matrix R and a translation matrix T;
further, the third step comprises the following substeps:
(3.1) according to the pedestrian coarse positioning result obtained in the step two, cutting all pedestrian pictures from the original image frame of the real-time video stream of the monitoring camera for batch processing, inputting a trained human body part segmentation model for pedestrian segmentation, and outputting a pedestrian part segmentation result;
(3.2) judging whether the video frame image contains the pedestrian foot points according to the pedestrian segmentation result, and if not, not carrying out coordinate conversion calculation; if the pedestrian foot points are contained, the pedestrian foot points are converted into pixel coordinates of the original video image frame through scale conversion, then the lower frame of the pedestrian detection frame is corrected according to the lower boundary result of the pedestrian foot point pixels, and the middle point of the connecting line of the lower left corner and the lower right corner of the corrected pedestrian detection frame is output and used as a pedestrian foot point fine positioning output value.
Further, the training process of the human body part segmentation model comprises the following steps: the PPSS dataset is used as a training set, a human body part segmentation model is input, training is carried out in a random gradient descent mode, the initial learning rate is set to be 0.03, the learning rate is updated through a linear attenuation learning strategy, and when the training times or the loss function reaches a threshold value, training of the human body part segmentation model is completed.
Further, the pedestrian moving speed
Figure BDA0002860053520000021
The calculation process is as follows:
Figure BDA0002860053520000031
wherein, the first and the second end of the pipe are connected with each other,
Figure BDA0002860053520000032
represents the abscissa of the pedestrian with the ID of i in the ground two-dimensional coordinate system at the time of the N-k frame,
Figure BDA0002860053520000033
represents the ordinate of the pedestrian with ID i in the ground two-dimensional coordinate system at the time of the N-k frame,
Figure BDA0002860053520000034
represents the abscissa of the pedestrian with the ID of i in the ground two-dimensional coordinate system at the time of the Nth frame,
Figure BDA0002860053520000035
and f is the video stream frame frequency, and k is the interval frame number.
Compared with the prior art, the invention has the following beneficial effects: adopting a non-contact intelligent sensing method, utilizing a coordinate conversion model to realize space position information resolving and calculating the moving speed of the pedestrian by combining the video stream frame interval time; in the conversion from an image coordinate system to a space coordinate system, the accuracy of the model is limited by the information of the positions of the foot points of the pedestrians, the detection frame obtained by only adopting the pedestrian detection model is relatively rough, and under the condition that the pedestrians are shielded, the lower boundary of the pedestrian given by the pedestrian detection frame is not the lower boundary of the positions of the foot points of the real pedestrians, so that a large error is easily caused; on the basis of a detection frame output by pedestrian detection, a pedestrian component segmentation network is adopted to segment human body components, so that pedestrian foot point judgment can be carried out, and more accurate pixel-level pedestrian foot point positions in an image coordinate system can be obtained; therefore, a more accurate pedestrian space position can be obtained, the calculation accuracy of subsequent pedestrian movement speed perception is correspondingly improved to some extent, and the accuracy can reach centimeter level.
Drawings
Fig. 1 is a flow chart of a pedestrian movement speed intelligent sensing method based on video streaming.
Detailed Description
Fig. 1 is a flow chart of a pedestrian movement speed intelligent sensing method based on video streaming, and the pedestrian movement speed intelligent sensing method includes the following steps:
acquiring a calibration picture under a camera, calibrating the camera, acquiring internal parameters and external parameters of a monitoring camera, and establishing a coordinate conversion model from an image coordinate system to a world coordinate system;
manufacturing a checkerboard calibration plate with inner angular points of 11-8, and collecting about 20 checkerboard calibration pictures at different positions, different angles and different postures; detecting the angular points of the chessboard pattern calibration picture, searching sub-pixel angular points on the basis, removing the pictures with the detected angular points less than 88, calibrating the camera according to the Zhang calibration method, and obtaining a camera internal reference matrix MCAnd a distortion coefficient dist, performing precision judgment by adopting a back projection error, performing distortion correction according to the distortion coefficient, and outputting the internal parameters of the corrected and optimized camera;
the method comprises the steps of generating two-dimensional code markers for external reference calibration by adopting an ArUco two-dimensional code dictionary, manufacturing calibration cloth, wherein the number of the two-dimensional code markers on the calibration cloth is 5 × 5, the length of each marker is 50cm, the interval between the markers is also 50cm, and selecting a larger marker size to meet the requirement that a monitoring camera can identify two-dimensional code mark points far away in a monitoring view field. The method comprises the steps of placing a two-dimensional code calibration cloth on the ground flatly, collecting a picture, detecting two-dimensional code mark points in the calibrated picture, obtaining pixel coordinates of the mark points when the number of the detected two-dimensional code mark points is more than or equal to 4, carrying out PNP calculation by combining internal parameters of a camera, calculating a camera rotation matrix R and a translation matrix T, and obtaining external parameters of the camera.
According to the field range of a monitoring camera deployed in an indoor monitoring scene, a three-dimensional world coordinate system taking a ground plane as a two-dimensional plane X axis and a two-dimensional plane Y axis and taking a position perpendicular to the ground as a Z axis is established, and a proper reference point is selected as a coordinate origin. And establishing a conversion relation between an image coordinate system and a world coordinate system by combining internal parameters and external parameters of the camera:
Figure BDA0002860053520000041
from P-1P = E can yield:
Figure BDA0002860053520000042
thereby obtaining a conversion formula from a two-dimensional image coordinate system to a three-dimensional world coordinate system:
Figure BDA0002860053520000043
wherein Z isCDenotes camera coordinates, [ u, v ]]Representing image pixel coordinates, [ X ]W,YW,ZW]Representing the world coordinates of the pedestrian target, R representing a surveillance camera rotation matrix, T representing a surveillance camera translation matrix, MCAnd representing a monitoring camera internal reference matrix. For unknown parameter ZCThe solution is as follows:
Figure BDA0002860053520000051
order to
Figure BDA0002860053520000052
R-1·T=P2From
Figure BDA0002860053520000053
The third behavior 1 may result in: zC·P1[2]=Zw·P2[2]Thereby obtaining ZC=(Zw+P2[2])/P1[2];
All the parameters from the two-dimensional image coordinate system to the three-dimensional world coordinate system are obtained, and the coordinates [ u, v ] of the pedestrian foot point in the image coordinate system are obtained]Substituting and making the coordinate axis perpendicular to the ground in the three-dimensional world coordinate system as ZwAnd =0, the coordinate output result of the pedestrian under the space coordinate system can be obtained.
Step two, acquiring a real-time video stream of the monitoring camera according to a streaming media protocol, wherein the frame rate is 25fps, and the resolution is 1920 × 1080; adopting a pedestrian detection model and a multi-target tracking model to detect and track pedestrians, outputting a pedestrian ID and a detection frame, and obtaining a pedestrian coarse positioning result under an image coordinate system;
the pedestrian detection model adopts a YoloV3 algorithm, and a backbone network is Darknet53; the size of the input picture after preprocessing is 416 × 416, the number of detection frames output by the network is [ (13 × 13) + (26 × 26) + (52 × 52) ] + 3=10647, then the detection frames output by the non-maximum suppression NMS algorithm are subjected to post-processing, and frame screening is performed according to the set threshold and confidence coefficient to obtain the output of pedestrian detection;
the multi-target tracking model adopts a Deepsort-based multi-target tracking algorithm to continuously track the pedestrian, firstly, a Kalman filter is adopted to estimate the motion state of the pedestrian, and the relevance of motion information is carried out according to the Markov distance between the position of the detected frame of the current frame and the position predicted by the Kalman filter; appearance information association is carried out by firstly adopting a pedestrian ReID feature extraction network based on Resnet50 to extract appearance features, wherein the feature dimension is 512, then similarity calculation is carried out by adopting Euclidean distance, whether association succeeds or not is judged through a set threshold value, and finally linear weighting is carried out by combining two associations of motion features and appearance features to serve as a final association result; and if the Kalman prediction result of the new target can be matched with the detection result in 3 continuous frames, adding the tracker of the target into a tracking list, distributing a new ID, and if the tracked target continuously exceeds 3 frames and is not matched with the detection frame, judging whether the target is deleted or not by setting a maximum survival period, thereby realizing continuous tracking of multiple pedestrians under continuous frame video images and obtaining the pedestrian ID.
Step three, acquiring a pedestrian picture according to the pedestrian detection frame in the step two, inputting a trained human body part segmentation model to segment a pedestrian part, and acquiring a pedestrian foot point fine positioning result under an image coordinate system;
the human body part segmentation model adopts a semantic segmentation network DeepLab V3+ and a light-weight network MobileNet V2 as a main network, the segmentation model comprises an encoder module and a decoder module, the encoder module performs feature extraction through the series connection of a plurality of layers of convolution layers, and then multi-scale features are obtained through a void space pyramid pooling module; the decoder module performs two times of upsampling by using the bilinear difference value, wherein the first time of upsampling fuses bottom layer features with the same resolution ratio features in the decoder module network, and the feature pixels restored to the original input size can be output by the segmentation network after two times of upsampling. The training process of the human body part segmentation model comprises the following steps: the method comprises the steps of inputting a human body part segmentation model by using a PPSS dataset as a training set, training by using a random gradient descent mode, setting an initial learning rate to be 0.03, updating the learning rate by using a linear attenuation learning strategy, finishing training of the human body part segmentation model after the training times or the Loss function reaches a threshold value, and adopting a Softmax Loss function as a Loss function of the human body part segmentation model by a person skilled in the art.
The third step specifically comprises the following substeps:
(3.1) according to the pedestrian rough positioning result in the second step, cutting all pedestrian pictures from the original image frame of the real-time video stream of the monitoring camera to perform batch processing, inputting N C H W into a trained human body part segmentation model to perform pedestrian segmentation, and outputting a pedestrian part segmentation result, wherein N is the number of pedestrian targets in a single frame image, C is the number of channels of the input image, C =3, H is the image height, H =160, W is the image width, W =80, a semantic feature map after dimensionality reduction is obtained through a feature coding unit, then the feature is recovered to the feature of the input resolution of the original image after two times of upsampling through a feature decoding unit, the value set of the feature pixel value is {0,1,2,3,4,5,6,7}, and respectively represents a human body part: hair, face, upper body, arms, lower body, legs, feet, and background;
(3.2) after the human body part segmentation model is processed, a human body pixel level fine-grained semantic segmentation result can be further obtained on the basis of the pedestrian detection frame, so that the pedestrian detection frame is further corrected; judging whether the video frame image contains the pedestrian foot points according to the pedestrian segmentation result, if not, judging that the current pedestrian does not have the foot point pedestrian foot points according to the characteristic value 6 (6 marks the pixel value of the pedestrian foot points), and not carrying out coordinate conversion calculation; it should be noted that the situation that the foot points do not exist may be due to the blocking phenomenon between pedestrians, and when the target reappears, we can perform continuous tracking according to the pedestrian tracking algorithm and perform subsequent calculation of the moving speed of the pedestrian according to the recorded time interval and the moving distance. If the pedestrian foot points are contained, the pedestrian foot points are converted into pixel coordinates of the original video image frame through scale conversion, then the lower frame of the pedestrian detection frame is corrected according to the lower boundary result of the pedestrian foot point pixels, and the middle point of the connecting line of the lower left corner and the lower right corner of the corrected pedestrian detection frame is output and used as a pedestrian foot point fine positioning output value.
Step four, according to pixel information of the foot points of the pedestrians in an image coordinate system, coordinate conversion calculation is carried out by combining the coordinate conversion model established in the step one, and space coordinate information of the pedestrians taking the ground plane as a two-dimensional plane XY axis and the ground perpendicular to the ground as a Z axis is obtained; and combining the space coordinate information with a pedestrian tracking algorithm to obtain the space position moving distance of the target pedestrian under the interval frame, and then combining the interval frame time to calculate the pedestrian movement speed.
(4.1) two-dimensional image coordinates [ u, v ] of pedestrian foot points in the image coordinate system determined according to the pedestrian detection frame and the pedestrian component segmentation network]Combining the conversion model from the image coordinate system to the world coordinate system established in the step one to obtain:
Figure BDA0002860053520000071
corresponding parameters are brought in, and the coordinates [ X, Y, Z ] of the foot points of the pedestrians under the world coordinate system can be obtained];
(4.2) combining with a pedestrian tracking algorithm, obtaining the pedestrian ID and corresponding pedestrian space coordinate information under continuous interval frames; ideally, the height of the foot point of the pedestrian on the ground is 0, that is, Z =0, so that the XY-axis coordinates [ X, Y ] of the foot point of the pedestrian on the two-dimensional plane of the ground plane can be obtained]Suppose for a continuously tracked pedestrian that its identity ID is i, its coordinates in the ground two-dimensional coordinates at frame N-k are
Figure BDA0002860053520000081
Terrestrial two at Nth frameThe coordinates of the dimensional coordinate system are
Figure BDA0002860053520000082
The video stream frame rate is f, and k is the interval frame number; the moving distance of the pedestrian in the nth frame of the current frame on the two-dimensional plane can be calculated by the euclidean distance, and the moving speed of the pedestrian can be obtained by combining the time of the interval frame:
Figure BDA0002860053520000083
and sending the pedestrian target ID and the pedestrian movement speed to the front end for visual display through the message middleware, thereby finishing intelligent perception of the pedestrian movement speed based on the video stream.
To sum up, this application utilizes the camera of deployment at the control point to acquire the video stream, adopts contactless visual perception algorithm to acquire the real-time velocity of motion of target pedestrian, judges to behaviors such as static pedestrian and pedestrian's velocity of motion overquick, provides effectual early warning and customization service for pedestrian's intelligence perception system under the control scene. The method realizes the coordinate conversion from an image coordinate system to a space coordinate system in the real world through a camera calibration method, and further foot point judgment and fine positioning are carried out by adopting a human body part segmentation model on the basis of a pedestrian detection frame because the conversion precision is limited by the positioning result of the foot points of pedestrians in the image coordinate system, so that the intelligent sensing of the pedestrian movement speed based on the camera video stream is more accurate.

Claims (4)

1. A pedestrian movement speed intelligent sensing method based on video streaming is characterized by comprising the following steps:
acquiring a calibration picture under a camera, calibrating the camera, acquiring internal parameters and external parameters of a monitoring camera, and establishing a coordinate conversion model from an image coordinate system to a world coordinate system;
acquiring a real-time video stream of a monitoring camera according to a streaming media protocol, detecting and tracking pedestrians by adopting a pedestrian detection model and a multi-target tracking model, outputting a pedestrian ID and a detection frame, and obtaining a pedestrian coarse positioning result under an image coordinate system;
step three, acquiring a pedestrian picture according to the pedestrian detection frame in the step two, inputting a trained human body part segmentation model for segmenting the pedestrian part, and acquiring a pedestrian foot point fine positioning result under an image coordinate system;
the third step comprises the following substeps:
(3.1) according to the pedestrian coarse positioning result obtained in the step two, cutting all pedestrian pictures from the original image frame of the real-time video stream of the monitoring camera for batch processing, inputting a trained human body part segmentation model for pedestrian segmentation, and outputting a pedestrian part segmentation result;
(3.2) judging whether the video frame image contains the pedestrian foot points according to the pedestrian segmentation result, and if not, not carrying out coordinate conversion calculation; if the pedestrian foot points are contained, converting the pixel coordinates into pixel coordinates of the original video image frame through scale conversion, correcting a lower border of the pedestrian detection frame according to a lower border result of the pixel of the pedestrian foot points, and outputting a middle point of a connecting line of a lower left corner and a lower right corner of the corrected pedestrian detection frame as a pedestrian foot point fine positioning output value;
step four, according to pixel information of the pedestrian foot points in an image coordinate system, coordinate conversion calculation is carried out by combining the coordinate conversion model established in the step one, and space coordinate information of the pedestrian in a three-dimensional world coordinate system with the ground plane as a two-dimensional plane XY axis and a coordinate axis perpendicular to the ground as a Z axis is obtained; and combining the space coordinate information with a pedestrian tracking algorithm to obtain the moving distance of the space position of the target pedestrian under the interval frame, and then combining the interval frame time to calculate the moving speed of the pedestrian.
2. The method as claimed in claim 1, wherein the camera calibration method comprises obtaining internal parameters and external parameters of the camera, the internal parameters are obtained by a checkerboard calibration method, and the internal parameters comprise an internal parameter matrix
Figure 84495DEST_PATH_IMAGE001
And distortion coefficient dist, and ArUcoMar is adopted as external parameterAnd resolving the internal parameters of the monitoring camera by using a PNP method in combination with the marker two-dimensional code calibration cloth, wherein the PNP method comprises a rotation matrix R and a translation matrix T.
3. The pedestrian motion speed intelligent perception method based on the video stream as claimed in claim 1, wherein the training process of the human body part segmentation model is as follows: the PPSS dataset is used as a training set, a human body part segmentation model is input, training is carried out in a random gradient descent mode, the initial learning rate is set to be 0.03, the learning rate is updated through a linear attenuation learning strategy, and when the training times or the loss function reaches a threshold value, training of the human body part segmentation model is completed.
4. The video stream-based pedestrian motion speed intelligent sensing method according to claim 1, wherein the pedestrian motion speed is
Figure 234854DEST_PATH_IMAGE002
The calculation process is as follows:
Figure DEST_PATH_IMAGE003
wherein the content of the first and second substances,
Figure 800089DEST_PATH_IMAGE004
represents the abscissa of the pedestrian with the ID of i in the ground two-dimensional coordinate system at the time of the N-k frame,
Figure 326886DEST_PATH_IMAGE005
represents the ordinate of the pedestrian with the ID being i in the ground two-dimensional coordinate system at the time of the N-k frame,
Figure 323661DEST_PATH_IMAGE006
represents the abscissa of the pedestrian with the ID of i in the ground two-dimensional coordinate system at the time of the Nth frame,
Figure 520287DEST_PATH_IMAGE007
and f is the video stream frame frequency, and k is the interval frame number.
CN202011559927.0A 2020-12-25 2020-12-25 Pedestrian movement speed intelligent sensing method based on video stream Active CN112598709B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202011559927.0A CN112598709B (en) 2020-12-25 2020-12-25 Pedestrian movement speed intelligent sensing method based on video stream

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202011559927.0A CN112598709B (en) 2020-12-25 2020-12-25 Pedestrian movement speed intelligent sensing method based on video stream

Publications (2)

Publication Number Publication Date
CN112598709A CN112598709A (en) 2021-04-02
CN112598709B true CN112598709B (en) 2022-11-01

Family

ID=75201948

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202011559927.0A Active CN112598709B (en) 2020-12-25 2020-12-25 Pedestrian movement speed intelligent sensing method based on video stream

Country Status (1)

Country Link
CN (1) CN112598709B (en)

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113658221B (en) * 2021-07-28 2024-04-26 同济大学 AGV pedestrian following method based on monocular camera
CN113840228B (en) * 2021-08-25 2024-04-02 北京航空航天大学杭州创新研究院 Pedestrian indoor positioning method based on positioning matching

Family Cites Families (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPWO2011068184A1 (en) * 2009-12-03 2013-04-18 独立行政法人産業技術総合研究所 Mobile positioning device
JP6276519B2 (en) * 2013-05-22 2018-02-07 株式会社 日立産業制御ソリューションズ Person counting device and human flow line analyzing device
US10372970B2 (en) * 2016-09-15 2019-08-06 Qualcomm Incorporated Automatic scene calibration method for video analytics
WO2018062647A1 (en) * 2016-09-28 2018-04-05 중앙대학교 산학협력단 Normalized-metadata generation apparatus, object occlusion detection apparatus, and methods thereof

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
Multi-scale body-part mask guided attention for person reidentification;Cai, H.,et.al;《2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition》;20191231;全文 *
一种基于图像坐标检测的目标定位方法;梅鸿翔;《计算机与数字工程》;20160320(第03期);全文 *
单目视频人体三维运动高效恢复;陈成等;《计算机辅助设计与图形学学报》;20090815(第08期);全文 *

Also Published As

Publication number Publication date
CN112598709A (en) 2021-04-02

Similar Documents

Publication Publication Date Title
CN111462200B (en) Cross-video pedestrian positioning and tracking method, system and equipment
JP6095018B2 (en) Detection and tracking of moving objects
CN106846359B (en) Moving target rapid detection method based on video sequence
Sidla et al. Pedestrian detection and tracking for counting applications in crowded situations
CN104204721B (en) Single camera distance estimations
Zhou et al. Seamless fusion of LiDAR and aerial imagery for building extraction
CN107481315A (en) A kind of monocular vision three-dimensional environment method for reconstructing based on Harris SIFT BRIEF algorithms
CN115439424A (en) Intelligent detection method for aerial video image of unmanned aerial vehicle
CN110458877B (en) Navigation method based on bionic vision for fusing infrared and visible light information
CN112598709B (en) Pedestrian movement speed intelligent sensing method based on video stream
CN110941996A (en) Target and track augmented reality method and system based on generation of countermeasure network
CN104200492B (en) Video object automatic detection tracking of taking photo by plane based on profile constraints
CN114973028B (en) Aerial video image real-time change detection method and system
CN111247526B (en) Method and system for tracking position and direction of target object moving on two-dimensional plane
CN114022910A (en) Swimming pool drowning prevention supervision method and device, computer equipment and storage medium
CN110458862A (en) A kind of motion target tracking method blocked under background
CN115376034A (en) Motion video acquisition and editing method and device based on human body three-dimensional posture space-time correlation action recognition
CN116309686A (en) Video positioning and speed measuring method, device and equipment for swimmers and storage medium
CN116978009A (en) Dynamic object filtering method based on 4D millimeter wave radar
KR100574227B1 (en) Apparatus and method for separating object motion from camera motion
CN111531546B (en) Robot pose estimation method, device, equipment and storage medium
CN108986162B (en) Dish and background segmentation method based on inertial measurement unit and visual information
Kröhnert Automatic waterline extraction from smartphone images
CN113920254B (en) Monocular RGB (Red Green blue) -based indoor three-dimensional reconstruction method and system thereof
CN103345762B (en) Bayes's visual tracking method based on manifold learning

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant