WO2022126668A1 - Method for pedestrian identification in public places and human flow statistics system - Google Patents

Method for pedestrian identification in public places and human flow statistics system Download PDF

Info

Publication number
WO2022126668A1
WO2022126668A1 PCT/CN2020/137803 CN2020137803W WO2022126668A1 WO 2022126668 A1 WO2022126668 A1 WO 2022126668A1 CN 2020137803 W CN2020137803 W CN 2020137803W WO 2022126668 A1 WO2022126668 A1 WO 2022126668A1
Authority
WO
WIPO (PCT)
Prior art keywords
pedestrian
feature
historical
current
bounding box
Prior art date
Application number
PCT/CN2020/137803
Other languages
French (fr)
Chinese (zh)
Inventor
舒元昊
张一杨
马小雯
刘倚剑
Original Assignee
中电海康集团有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 中电海康集团有限公司 filed Critical 中电海康集团有限公司
Publication of WO2022126668A1 publication Critical patent/WO2022126668A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/50Context or environment of the image
    • G06V20/52Surveillance or monitoring of activities, e.g. for recognising suspicious objects
    • G06V20/53Recognition of crowd images, e.g. recognition of crowd congestion
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/80Analysis of captured images to determine intrinsic or extrinsic camera parameters, i.e. camera calibration
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/44Local feature extraction by analysis of parts of the pattern, e.g. by detecting edges, contours, loops, corners, strokes or intersections; Connectivity analysis, e.g. of connected components
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/40Scenes; Scene-specific elements in video content
    • G06V20/46Extracting features or characteristics from the video content, e.g. video fingerprints, representative shots or key frames
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/103Static body considered as a whole, e.g. static pedestrian or occupant recognition
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10016Video; Image sequence
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20081Training; Learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30196Human being; Person
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30244Camera pose

Definitions

  • the invention belongs to the technical field of computer vision, and in particular relates to a pedestrian identification method and a people flow statistics system in a public place.
  • People flow statistics involve pedestrian identification, pedestrian stay time in the statistical area, and entry and exit trajectories.
  • the commonly used statistical methods include base station-based statistical methods, such as Bluetooth base stations, 4G base stations, etc., but the positioning accuracy of the above methods is not accurate enough;
  • the statistical methods of optical imaging equipment such as the statistical methods of infrared arrays and millimeter-wave radars, the above methods have relatively high positioning accuracy, but cannot accurately identify pedestrians, which is likely to cause repeated statistics; there are statistical methods based on optical imaging equipment, such as cameras, positioning It has high accuracy and can accurately identify pedestrians, but there is the problem of pedestrians being occluded. Some of them are based on the statistical method of pedestrian re-identification, and there are also repeated statistical problems caused by the discrepancy between the pedestrian movement pattern and the filtering predicted trajectory.
  • the purpose of the present invention is to provide a pedestrian identification method and a people flow statistics system in a public place, which can accurately identify pedestrians and have a high accuracy rate of people flow statistics.
  • the technical scheme adopted by the present invention is:
  • a pedestrian identification method in a public place comprising:
  • Step 1 Obtain an optical image, detect pedestrians in the optical image, and output the three-dimensional bounding box of the pedestrian and the corresponding timestamp;
  • Step 2 Obtain pedestrian features based on optical images and 3D bounding boxes, including:
  • Step 2.1 extract the human body shape and features of the pedestrians in the optical image as the pedestrian appearance features of each pedestrian, and save them in the historical feature database;
  • Step 2.2 based on the current three-dimensional bounding box of the pedestrian and the three-dimensional bounding box distributed according to the time series in the historical feature library, extract the three-dimensional motion feature of each pedestrian, and save it in the historical feature library;
  • Step 2.3 based on the pedestrian 3D motion feature and the pedestrian 3D motion feature within the specified time in the historical feature library, predict the pedestrian 3D motion feature at the next moment, and save it in the historical feature library;
  • Step 3 Perform pedestrian recognition based on the pedestrian features in the historical feature library, including:
  • Step 3.1 Calculate the apparent feature distance one by one based on the current pedestrian apparent features and the historical pedestrian apparent features of each pedestrian in the historical feature database. If the apparent feature distance is greater than the apparent threshold, determine the current pedestrian apparent feature and history The pedestrian apparent features in the feature library belong to the same pedestrian, and the current apparent feature distance is determined as the pedestrian's apparent feature distance;
  • Step 3.2 Calculate the spatial feature distances one by one based on the current three-dimensional motion features of pedestrians and the three-dimensional motion features of pedestrians at the next moment predicted from the previous moment of each pedestrian in the historical feature database. If the spatial feature distance is greater than the spatial threshold, determine the current The three-dimensional motion feature of the pedestrian and the three-dimensional motion feature of the pedestrian at the next moment predicted in the historical feature database belong to the same pedestrian, and the current spatial feature distance is determined as the spatial feature distance of the pedestrian;
  • Step 3.3 Based on the current three-dimensional motion features of pedestrians, apparent feature distances, spatial feature distances, and the three-dimensional motion features of each pedestrian history in the historical feature database, determine whether it conforms to the motion pattern of the same pedestrian, and output the motion pattern matching degree as the Pedestrian motion pattern matching degree;
  • Step 3.4 Perform weighted calculation on the apparent feature distance, spatial feature distance, and motion pattern matching degree belonging to the same pedestrian, and obtain the matching result between the pedestrian in the current three-dimensional bounding box and the pedestrian in the historical feature database, and the matching result includes a successful matching or a matching result. If the matching fails, the pedestrian information obtained by the matching is also included when the matching is successful;
  • Step 4 Mark the pedestrian status of the pedestrian corresponding to the three-dimensional bounding box according to the current matching results and the historical matching results as initial matching successful, lost, successful re-matching after loss, successful continuous matching or out of the camera range.
  • each optional method can be independently implemented for the above-mentioned overall solution.
  • the combination can also be a combination between multiple optional ways.
  • the pedestrian in the optical image is detected, and a three-dimensional bounding box of the pedestrian is output, including:
  • the camera used to obtain the optical image is calibrated, and the mapping relationship between the pixel in the optical image and the distance of the camera is obtained;
  • the 3D bounding box of the pedestrian is obtained.
  • the three-dimensional motion feature of each pedestrian is extracted based on the current three-dimensional bounding box of the pedestrian and the three-dimensional bounding box distributed according to the time series in the historical feature library, including:
  • Step 2.2.1. Extraction of direction vector extract the movement direction of pedestrians in the horizontal direction and the movement direction in the vertical direction through the current 3D bounding box and the historical 3D bounding box;
  • Step 2.2.2. Movement speed extraction extract the movement speed of the person in the horizontal direction and the movement speed in the vertical direction through the current 3D bounding box and the historical 3D bounding box;
  • Step 2.2.3 Relative position extraction: According to the mapping relationship obtained after the camera is calibrated, the coordinates of the pedestrian in the three-dimensional coordinate system centered on the camera are output based on the current three-dimensional bounding box and the historical three-dimensional bounding box;
  • Step 2.2.4 Use the direction vector, motion speed and relative position extracted in steps 2.2.1 to 2.2.3 as the three-dimensional motion feature of the pedestrian.
  • the pedestrian state of the pedestrian corresponding to the three-dimensional bounding box is marked according to the current matching result and the historical matching result as initial matching success, loss, rematching success after loss, continuous matching success or out of the camera range, including:
  • the state of the pedestrian is marked as walking out of the camera range, and M ⁇ N.
  • new pedestrian information is allocated to the pedestrian in the historical feature database, and the pedestrian characteristic of the pedestrian is associated with the newly allocated pedestrian information.
  • the present invention also provides a people flow statistics system, and the people flow statistics system includes:
  • the pedestrian detection module is used to obtain optical images, detect pedestrians in the optical images, and output the three-dimensional bounding boxes of pedestrians and corresponding timestamps;
  • the feature extraction module is used to obtain pedestrian features based on optical images and 3D bounding boxes.
  • the specific steps are as follows:
  • the pedestrian recognition module is used for pedestrian recognition based on the pedestrian features in the historical feature database, and the specific steps are as follows:
  • the spatial feature distances one by one based on the current three-dimensional motion features of pedestrians and the three-dimensional motion features of pedestrians at the next moment predicted from the previous moment of each pedestrian in the historical feature database. If the spatial feature distance is greater than the spatial threshold, the current pedestrian is judged The three-dimensional motion feature and the three-dimensional motion feature of the pedestrian at the next moment predicted at the previous moment in the historical feature library belong to the same pedestrian, and the current spatial feature distance is determined as the spatial feature distance of the pedestrian;
  • the pedestrian marking module is used to mark the pedestrian status of the pedestrian corresponding to the three-dimensional bounding box according to the current matching results and the historical matching results as initial matching success, loss, rematching success after loss, continuous matching success or out of the camera range;
  • the people flow statistics module is used to count the flow of people within the statistical range corresponding to the optical image within the preset time according to the pedestrian state.
  • the pedestrian in the optical image is detected, and the three-dimensional bounding box of the pedestrian is output, and the following operations are performed:
  • the camera used to obtain the optical image is calibrated, and the mapping relationship between the pixel in the optical image and the distance of the camera is obtained;
  • the 3D bounding box of the pedestrian is obtained.
  • the three-dimensional bounding box based on the current three-dimensional bounding box of pedestrians and the three-dimensional bounding box distributed according to time series in the historical feature library, extract the three-dimensional motion feature of pedestrians of each pedestrian, and perform the following operations:
  • Direction vector extraction Through the current 3D bounding box and the historical 3D bounding box, the pedestrian's movement direction in the horizontal direction and the movement direction in the vertical direction are extracted;
  • Movement speed extraction extract the movement speed of the person in the horizontal direction and the movement speed in the vertical direction through the current 3D bounding box and the historical 3D bounding box;
  • Relative position extraction According to the mapping relationship obtained after the camera is calibrated, the coordinates of the pedestrian in the three-dimensional coordinate system centered on the camera are output based on the current three-dimensional bounding box and the historical three-dimensional bounding box;
  • Feature integration The extracted direction vector, motion speed and relative position are used as pedestrian 3D motion features.
  • the pedestrian states of the pedestrians corresponding to the three-dimensional bounding box are marked as successful initial matching, lost, successful re-matching after loss, successful continuous matching, or out of the camera range, and the following operations are performed. :
  • the state of the pedestrian is marked as walking out of the camera range, and M ⁇ N.
  • new pedestrian information is allocated to the pedestrian in the historical feature database, and the pedestrian characteristic of the pedestrian is associated with the newly allocated pedestrian information.
  • the present invention provides a pedestrian identification method in a public place, which comprehensively considers pedestrian appearance features, three-dimensional motion features and motion patterns, accurately identifies pedestrians, obtains the time and position of pedestrians entering and exiting the statistical range, and the movement trajectory within the statistical range, and based on the This method proposes a people flow statistics system, which can accurately count the flow of people entering and leaving the statistical range in unit time.
  • Fig. 1 is the flow chart of the pedestrian identification method in public places of the present invention
  • Fig. 2 is the flow chart of the present invention outputting the three-dimensional bounding box of pedestrian
  • FIG. 3 is a flow chart of the present invention for acquiring pedestrian features based on an optical image and a three-dimensional bounding box;
  • FIG. 4 is a schematic diagram of the present invention extracting motion features by human body structure in a right-handed coordinate system
  • Fig. 5 is the flow chart that the present invention carries out pedestrian recognition based on the pedestrian characteristic in the historical characteristic database
  • FIG. 6 is a flow chart of the present invention for marking pedestrian status
  • FIG. 7 is a structural diagram of the people flow statistics system of the present invention.
  • a pedestrian identification method in a public place which can accurately identify pedestrians and can be used for urban planning based on pedestrian identification statistics, business strategy adjustment based on pedestrian flow statistics in shopping malls, and subway shift adjustment based on subway station traffic statistics.
  • the method for identifying pedestrians in public places in this embodiment includes the following steps:
  • Step 1 Obtain an optical image, detect pedestrians in the optical image, and output a three-dimensional bounding box of the pedestrian and a corresponding timestamp.
  • the optical image is acquired by the camera, and the obtained time stamp is the time when the optical image is captured by the camera. It is easy to understand that the acquisition of the optical image may be based on any image acquisition device, and this embodiment takes a camera as an example for description.
  • this embodiment includes the following steps when forming the 3D bounding box, as shown in Figure 2:
  • the camera used to obtain the optical image is calibrated, and the mapping relationship between the pixel and the camera distance in the optical image is obtained; the pedestrian in the optical image is detected, and the two-dimensional bounding box (BBox) of the pedestrian in the optical image is obtained; based on the two-dimensional The bounding box and the mapping relationship get the pedestrian's three-dimensional bounding box (3D Bounding Box, 3D BBox).
  • the mapping relationship between the pedestrian pixel and the camera distance in the optical image is obtained by calibrating the camera, and the corresponding depth information is obtained based on the mapping relationship.
  • the depth information reflects the actual distance between the pedestrian and the camera, including the movement of the pedestrian. It is convenient to extract the three-dimensional motion features of pedestrians based on the depth information.
  • a monocular fixed-focus camera is used to shoot a video
  • a cube with a side length of 1 meter is used to calibrate the mapping relationship.
  • Each face of the cube is evenly divided into 100 black and white grids, and the camera shooting range is Statistical range.
  • the calibration is a conventional technical means used by the camera, and the specific steps of the calibration are not limited in this embodiment, and the depth information is obtained based on the mapping relationship of the camera calibration, which is the preferred method provided in this embodiment, but is not limited to the only means. , for example, the depth information can be superimposed by the combination of the camera and the depth camera.
  • pedestrians in the optical image are identified based on the pedestrian detection method and a two-dimensional bounding box is output.
  • the pedestrian detection method used is a conventional method in image recognition, for example, the recognition based on Yolo extension obtained by training on the pedestrian dataset is used. network.
  • the 3D bounding box is obtained, the 2D bounding box and the mapping relationship are input into the 3D estimation method, and the 3D bounding box is output.
  • the three-dimensional estimation method used in this embodiment is a monocular depth estimation method based on optical flow, which can output the inverse depth, and the depth information can be obtained by calculating the inverse depth.
  • the depth information has different error coefficients in different ranges from the camera.
  • an error matrix is used.
  • Pedestrian recognition based on the 3D bounding box can effectively overcome the pedestrian occlusion problem. Since the human body structure conforms to the geometric constraints, the partially occluded 2D bounding box can be restored to a complete 3D bounding box, and its spatial error is within the allowable range.
  • Step 2 Obtain pedestrian features based on the optical image and the 3D bounding box, as shown in Figure 3, including:
  • Step 2.1 extract the human body shape and features of the pedestrians in the optical image as the pedestrian appearance features of each pedestrian, and save them in the historical feature database.
  • this embodiment adopts the pedestrian apparent feature extraction method, mainly extracting the human body traits and features that can be observed visually, and marking the pedestrian apparent feature as F for the convenience of distinction appearance .
  • the embedding structure (embedding, embedding) of the recognition network based on Yolo extension is used as the method for extracting the apparent feature of pedestrians.
  • Step 2.2 based on the current three-dimensional bounding box of the pedestrian and the three-dimensional bounding box distributed according to the time series in the historical feature library, extract the three-dimensional motion feature of each pedestrian, and save it in the historical feature library.
  • the three-dimensional motion feature of pedestrians is the feature of the position change of pedestrians in three-dimensional space.
  • F displacement is an important feature for data association in time series.
  • a right-handed coordinate system is established, and the three-dimensional bounding box is divided into three parts: head, upper body, and lower body according to the human body structure, and the three-dimensional bounding box is input to the pedestrian three-dimensional motion feature extraction method, which mainly extracts people in three-dimensional space.
  • the position change feature in the output is the three-dimensional motion feature of the pedestrian.
  • the three-dimensional motion feature extraction method consists of the following parts:
  • Step 2.2.1. Extraction of direction vector Extract the moving direction of pedestrians in the horizontal direction and the moving direction in the vertical direction through the current three-dimensional bounding box and the historical three-dimensional bounding box.
  • the direction vector of the pedestrian can be obtained based on the position change of the 3D bounding box based on the time distribution.
  • the direction vector of the pedestrian can be obtained based on the position change of the 3D bounding box based on the time distribution.
  • only the first two bounding boxes can be used to determine the direction vector, or multiple pairs of 3D bounding boxes can be used, and the mean, median or other values of the multiple direction vectors can be taken as the final direction vector.
  • Step 2.2.2. Movement speed extraction extract the movement speed of the person in the horizontal direction and the movement speed in the vertical direction through the current 3D bounding box and the historical 3D bounding box.
  • the 3D bounding box based on the time distribution obtains the movement speed of the pedestrian according to the time difference and the position difference of the corresponding 3D bounding box, and multiple 3D bounding boxes can only use the first two bounding boxes to calculate the movement speed, It is also possible to use multiple pairs of three-dimensional bounding boxes, and take the mean, median or other values of multiple motion velocities as the final motion velocity.
  • Step 2.2.3 Relative position extraction: According to the mapping relationship obtained after the camera is calibrated, the coordinates of the pedestrian in the three-dimensional coordinate system centered on the camera are output based on the current three-dimensional bounding box and the historical three-dimensional bounding box.
  • each three-dimensional bounding box is equivalent to a point, and the coordinates of the point are obtained as the coordinates of the pedestrian.
  • the point can be the center point of the three-dimensional bounding box, a vertex or any point.
  • Step 2.2.4 Use the direction vector, motion speed and relative position extracted in steps 2.2.1 to 2.2.3 as the three-dimensional motion feature of the pedestrian.
  • the current 3D bounding box is the 3D bounding box of a new pedestrian entering the statistical range, and the above matching fails to obtain the corresponding historical 3D bounding box, set the direction vector and movement speed of the new pedestrian as default values (for example, The direction vector is none and the transport speed is 0), and the coordinates are the coordinates of the current three-dimensional bounding box as the three-dimensional motion feature of the pedestrian.
  • Step 2.3 Predict the three-dimensional motion feature of the pedestrian at the next moment based on the three-dimensional motion feature of the pedestrian and the three-dimensional motion feature of the pedestrian in the historical feature library at a specified time, and save the three-dimensional motion feature of the pedestrian in the historical feature library.
  • the three-dimensional motion feature of the pedestrian at the next moment is marked as F predicted , which is predicted by the trajectory prediction algorithm and represents the movement trend of the pedestrian in space and time. Using this feature can further solve the problem of target loss caused by the pedestrian being occluded; in this embodiment Use Kalman filtering to predict the three-dimensional motion characteristics of pedestrians at the next moment.
  • Step 3 Perform pedestrian recognition based on the pedestrian features in the historical feature database, as shown in Figure 5, including:
  • Step 3.1 Calculate the apparent feature distance one by one based on the current pedestrian apparent features and the historical pedestrian apparent features of each pedestrian in the historical feature database. If the apparent feature distance is greater than the apparent threshold, determine the current pedestrian apparent feature and history The pedestrian apparent features in the feature library belong to the same pedestrian, and the apparent feature distance is determined as the pedestrian's apparent feature distance.
  • the pedestrian apparent feature matching method is used to calculate the current pedestrian apparent feature Pedestrian appearance features in historical feature database To judge whether they belong to the same pedestrian, for example, using the weighted Mahalanobis distance and cosine distance to calculate the apparent feature distance, the coefficients are 0.02 and 0.98 respectively.
  • the historical pedestrian appearance feature in this embodiment mainly adopts the pedestrian appearance feature of the previous moment.
  • a list of apparently similar pedestrians can be established, each pedestrian corresponds to a list of apparently similar pedestrians, and the table is further distinguished based on the list. View feature distance.
  • the current moment has the apparent characteristics of two persons, A and B, where A and A ⁇ t-1 ⁇ (the pedestrian apparent characteristics of the previous moment), B and A ⁇ t-1 ⁇ are also very similar, even
  • a and A ⁇ t-1 ⁇ the pedestrian apparent characteristics of the previous moment
  • B and A ⁇ t-1 ⁇ are also very similar, even
  • the similarity of multiple time periods in the apparently similar pedestrian lists of B and A is high, but the similarity of multiple time periods in the apparently similar pedestrian lists of B and B is higher, then it can be judged that B is B.
  • this search method is time-consuming, it is generally used in scenarios with high requirements for pedestrian recognition.
  • Step 3.2 Calculate the spatial feature distances one by one based on the current three-dimensional motion features of pedestrians and the three-dimensional motion features of pedestrians at the next moment predicted from the previous moment of each pedestrian in the historical feature database. If the spatial feature distance is greater than the spatial threshold, determine the current The three-dimensional motion feature of the pedestrian and the three-dimensional motion feature of the pedestrian at the next moment predicted in the historical feature database belong to the same pedestrian, and the current spatial feature distance is determined as the spatial feature distance of the pedestrian.
  • the three-dimensional motion feature of the pedestrian at the next moment predicted at the previous moment is the predicted three-dimensional motion feature of the current pedestrian.
  • Matching the predicted and the actual current three-dimensional motion feature of the pedestrian can be used as the basis for judging whether they belong to the same pedestrian.
  • the Hungarian algorithm is used as the pedestrian three-dimensional motion feature matching method for calculation to determine whether they belong to the same pedestrian.
  • a list of spatially similar pedestrians can be established.
  • Step 3.3 Based on the current three-dimensional motion features of pedestrians, apparent feature distances, spatial feature distances, and the three-dimensional motion features of each pedestrian history in the historical feature database, determine whether it conforms to the motion pattern of the same pedestrian, and output the motion pattern matching degree as the Pedestrian motion pattern matching degree.
  • this embodiment focuses on the change speed of pedestrians in time series and the movement logic in spatial position.
  • the movement logic includes but is not limited to common behaviors such as turning back, staying in place, trotting, and squatting.
  • the change speed of the target within 3 seconds and the movement logic of the target in the shooting space are mainly investigated. Since the cameras acquire optical images based on preset intervals, pedestrians with reasonable speed changes can be judged as the same pedestrian.
  • the matching degree can be directly output based on the pre-trained neural network, or it can be directly judged by predicting the preset matching rules.
  • the former is relatively flexible in judgment, but needs to be trained based on a large number of samples.
  • the latter can be directly generated and used, and is easy to add, delete and modify, but the flexibility is relatively low, and an appropriate method can be selected according to actual needs.
  • a matching rule (specific probability values omitted) is established as shown in Table 1, which represents the probability of converting the behavior pattern of the previous stage to the corresponding behavior pattern of the current stage.
  • the motion pattern matching degree is performed based on Table 1, the historical three-dimensional motion features of pedestrians corresponding to the apparent feature distances are taken, and the three-dimensional historical three-dimensional motion features of pedestrians corresponding to the spatial feature distances are taken. If the pedestrian corresponding to the motion feature is not the same pedestrian, this match will be discarded. If it is the same pedestrian, the behavior pattern and current behavior pattern of the pedestrian in the previous stage will be judged according to the obtained historical pedestrian 3D motion feature and the current pedestrian 3D motion feature. The behavior pattern of the stage can be looked up in the table to obtain the probability value as the motion pattern matching degree.
  • the behavior pattern of a stage is determined by at least two pedestrian three-dimensional motion features. Since a pedestrian three-dimensional motion feature has a direction vector, motion speed and coordinates, the current stage can be determined by the change of the two direction vectors. Going forward, turning back or turning, combined with the coordinate change can further distinguish whether it is going forward or staying, and further distinguish whether it is going forward or accelerating forward combined with the movement speed.
  • the above table is a preferred matching rule adopted in this embodiment, which can be further optimized in actual use, for example, the turning is refined into a left turn or a right turn, etc., and the probability values in the table can also be calculated according to the actual use. The probability is updated to improve the pedestrian recognition rate.
  • Step 3.4 Perform weighted calculation on the apparent feature distance, spatial feature distance, and motion pattern matching degree belonging to the same pedestrian (ie, input it into the weighting calculator), and obtain the matching result between the pedestrian in the current three-dimensional bounding box and the pedestrian in the historical feature database.
  • the matching result includes matching success or matching failure, and when the matching is successful, it also includes the pedestrian information obtained by the matching.
  • the weights of apparent feature distance, spatial feature distance, and motion pattern matching degree are 0.6, 0.2, and 0.2, respectively. Since apparent feature is the most intuitive feature to distinguish different pedestrians, the apparent feature distance is set in this embodiment. has the highest weight. Of course, in actual use, the weight can be adjusted, such as increasing the weight of the motion pattern matching degree, so as to avoid misjudgment caused by two people whose apparent characteristics are too similar.
  • the matching failure in the final matching result indicates that there is no historical record for the characteristics of the current pedestrian, that is, the pedestrian is a pedestrian who has newly entered the statistical range; and the matching success indicates that the current pedestrian's characteristics have historical records, so the matching pedestrian information is output to correlate. New and historical features of the same pedestrian.
  • Pedestrian information can be a unique identifier (eg, ID value), spatial location, time, and the like.
  • Step 4 Mark the pedestrian status of the pedestrian corresponding to the three-dimensional bounding box according to the current matching results and the historical matching results as initial matching successful, lost, successful re-matching after loss, successful continuous matching or out of the camera range.
  • a specific matching method provided in this embodiment may be:
  • the pedestrian feature is successfully extracted (that is, the recognition is successful at the current moment), but the matching result is a matching failure (that is, there is no historical record), the status of the current pedestrian is marked as the initial matching success;
  • M consecutive times for example, 50 consecutive times (10 seconds*5 times/sec)
  • the current recognition fails and the previous moment also fails, or the current recognition fails and the continuous recognition fails
  • the number of times is not greater than the threshold
  • the pedestrian marked as lost is successfully re-matched in the current matching result (for example, the current moment is successfully identified and there is a historical record, but the previous match failed), then the status of the pedestrian is updated to be lost and the re-match is successful;
  • N consecutive times for example, 150 consecutive times (10 seconds*15 times/sec)
  • the current recognition fails and the number of consecutive recognition failures is greater than the threshold
  • mark the status of the pedestrian is out of the camera range, and M ⁇ N.
  • a people flow statistics system comprising:
  • the pedestrian detection module is used to obtain optical images, detect pedestrians in the optical images, and output the three-dimensional bounding boxes of pedestrians and corresponding timestamps;
  • the feature extraction module is used to obtain pedestrian features based on optical images and 3D bounding boxes.
  • the specific steps are as follows:
  • the pedestrian recognition module is used for pedestrian recognition based on the pedestrian features in the historical feature database, and the specific steps are as follows:
  • the spatial feature distances one by one based on the current three-dimensional motion features of pedestrians and the three-dimensional motion features of pedestrians at the next moment predicted from the previous moment of each pedestrian in the historical feature database. If the spatial feature distance is greater than the spatial threshold, the current pedestrian is judged The three-dimensional motion feature and the three-dimensional motion feature of the pedestrian at the next moment predicted at the previous moment in the historical feature library belong to the same pedestrian, and the current spatial feature distance is determined as the spatial feature distance of the pedestrian;
  • the pedestrian marking module is used to mark the pedestrian status of the pedestrian corresponding to the three-dimensional bounding box according to the current matching results and the historical matching results as initial matching success, loss, rematching success after loss, continuous matching success or out of the camera range;
  • the people flow statistics module is used to count the flow of people within the statistical range corresponding to the optical image within the preset time according to the pedestrian state.
  • the detection of pedestrians in the optical image, outputting a three-dimensional bounding box of the pedestrians performs the following operations:
  • the camera used to obtain the optical image is calibrated, and the mapping relationship between the pixel in the optical image and the distance of the camera is obtained;
  • the 3D bounding box of the pedestrian is obtained.
  • the pedestrian detection module has a camera calibration function and a parameter management function.
  • the camera calibration device can be used as a part independent of the people flow statistics system in this embodiment, and the calibrated external parameters of the equipment and The mapping relationship may be sent to the parameter management module of the people flow statistics system in this embodiment.
  • the people flow statistics system further includes a video acquisition module, the video acquisition module is connected with the video acquisition device of the peripheral equipment, and after acquiring the real-time video within the statistical range, each The optical picture of the frame is sent to the pedestrian detection module.
  • the three-dimensional motion feature of each pedestrian is extracted based on the current three-dimensional bounding box of the pedestrian and the three-dimensional bounding box distributed according to the time series in the historical feature library, and the following operations are performed:
  • Direction vector extraction Through the current 3D bounding box and the historical 3D bounding box, the pedestrian's movement direction in the horizontal direction and the movement direction in the vertical direction are extracted;
  • Movement speed extraction Extract the movement speed of the person in the horizontal direction and the movement speed in the vertical direction through the current 3D bounding box and the historical 3D bounding box;
  • Relative position extraction According to the mapping relationship obtained after the camera is calibrated, the coordinates of the pedestrian in the three-dimensional coordinate system centered on the camera are output based on the current three-dimensional bounding box and the historical three-dimensional bounding box;
  • Feature integration The extracted direction vector, motion speed and relative position are used as pedestrian 3D motion features.
  • the pedestrian state of the pedestrian corresponding to the three-dimensional bounding box is marked according to the current matching result and the historical matching result as successful initial matching, lost, successful re-matching after loss, successful continuous matching, or out of camera range , do the following:
  • the state of the pedestrian is marked as walking out of the camera range, and M ⁇ N.
  • new pedestrian information is allocated to the pedestrian in the historical feature database, and the pedestrian characteristic of the pedestrian is associated with the newly allocated pedestrian information.

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Multimedia (AREA)
  • Biophysics (AREA)
  • General Health & Medical Sciences (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Artificial Intelligence (AREA)
  • Biomedical Technology (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Computational Linguistics (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Computation (AREA)
  • Health & Medical Sciences (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • General Engineering & Computer Science (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Human Computer Interaction (AREA)
  • Image Analysis (AREA)

Abstract

A method for pedestrian identification in public places and a human flow statistics system. The method comprises: acquiring an optical image, detecting a pedestrian in the optical image, and outputting a three-dimensional bounding box and a corresponding time stamp for the pedestrian; acquiring pedestrian features on the basis of the optical image and the three-dimensional bounding box; performing pedestrian identification on the basis of the pedestrian features in a historical feature library; according to a matching result at said time and a historical matching result, labeling the pedestrian status of the pedestrian that corresponds to the three-dimensional bounding box as an initial matching success, failure, re-matching success after failure, continuous matching success, or moved out of a camera range. By taking into consideration the apparent features, three-dimensional motion features and motion mode of a pedestrian, the method accurately identifies the pedestrian and acquires the time and position at which the pedestrian enters or exits a statistics range and the movement trajectory within the statistics range. On the basis of the method, a human flow statistics system is provided to accurately perform statistics on the people flow amount exiting or entering the statistics range within a unit of time.

Description

一种公共场所行人识别方法及人流统计系统A method for identifying pedestrians in public places and a people flow statistics system 技术领域technical field
本发明属于计算机视觉技术领域,具体涉及一种公共场所行人识别方法及人流统计系统。The invention belongs to the technical field of computer vision, and in particular relates to a pedestrian identification method and a people flow statistics system in a public place.
背景技术Background technique
人流统计涉及到行人识别、行人在统计区域内的停留时间与出入轨迹,目前常用的统计方法有基于基站的统计方法,如蓝牙基站、4G基站等,但以上方法定位精度不够准确;有基于非光学成像设备的统计方法,如红外阵列、毫米波雷达的统计方法,以上方法定位精度相对较高,但无法准确识别行人,容易造成重复统计;有基于光学成像设备的统计方法,如摄像头,定位精度较高,能够准确识别行人,但是存在行人被遮挡的问题,部分基于行人重识别的统计方法,还存在行人运动模式与滤波预测轨迹不符导致的重复统计问题。People flow statistics involve pedestrian identification, pedestrian stay time in the statistical area, and entry and exit trajectories. Currently, the commonly used statistical methods include base station-based statistical methods, such as Bluetooth base stations, 4G base stations, etc., but the positioning accuracy of the above methods is not accurate enough; The statistical methods of optical imaging equipment, such as the statistical methods of infrared arrays and millimeter-wave radars, the above methods have relatively high positioning accuracy, but cannot accurately identify pedestrians, which is likely to cause repeated statistics; there are statistical methods based on optical imaging equipment, such as cameras, positioning It has high accuracy and can accurately identify pedestrians, but there is the problem of pedestrians being occluded. Some of them are based on the statistical method of pedestrian re-identification, and there are also repeated statistical problems caused by the discrepancy between the pedestrian movement pattern and the filtering predicted trajectory.
发明内容SUMMARY OF THE INVENTION
本发明的目的在于提供一种公共场所行人识别方法及人流统计系统,准确识别行人,人流量统计准确率高。The purpose of the present invention is to provide a pedestrian identification method and a people flow statistics system in a public place, which can accurately identify pedestrians and have a high accuracy rate of people flow statistics.
为实现上述目的,本发明所采取的技术方案为:To achieve the above object, the technical scheme adopted by the present invention is:
一种公共场所行人识别方法,所述公共场所行人识别方法,包括:A pedestrian identification method in a public place, the method for pedestrian identification in a public place, comprising:
步骤1、获取光学图像,检测光学图像中的行人,输出行人的三维包围盒及对应的时间戳;Step 1. Obtain an optical image, detect pedestrians in the optical image, and output the three-dimensional bounding box of the pedestrian and the corresponding timestamp;
步骤2、基于光学图像和三维包围盒获取行人特征,包括:Step 2. Obtain pedestrian features based on optical images and 3D bounding boxes, including:
步骤2.1、提取光学图像中行人的人体形状和特征作为每个行人的行人表观特征,并保存至历史特征库中;Step 2.1, extract the human body shape and features of the pedestrians in the optical image as the pedestrian appearance features of each pedestrian, and save them in the historical feature database;
步骤2.2、基于行人当前的三维包围盒以及历史特征库中按照时间序列分布的三维包围盒,提取每个行人的行人三维运动特征,并保存至历史特征库中;Step 2.2, based on the current three-dimensional bounding box of the pedestrian and the three-dimensional bounding box distributed according to the time series in the historical feature library, extract the three-dimensional motion feature of each pedestrian, and save it in the historical feature library;
步骤2.3、基于行人三维运动特征及历史特征库中指定时间内的行人三维运动特征预测下一时刻的行人三维运动特征,并保存至历史特征库中;Step 2.3, based on the pedestrian 3D motion feature and the pedestrian 3D motion feature within the specified time in the historical feature library, predict the pedestrian 3D motion feature at the next moment, and save it in the historical feature library;
步骤3、基于历史特征库中的行人特征进行行人识别,包括:Step 3. Perform pedestrian recognition based on the pedestrian features in the historical feature library, including:
步骤3.1、基于当前的行人表观特征和历史特征库中每个行人历史的行人表观特征逐一计算表观特征距离,若表观特征距离大于表观阈值则判断当前的行人表观特征与历史特征库中的行人表观特征属于同一行人,确定当前的表观特征距离作为该行人的表观特征距离;Step 3.1. Calculate the apparent feature distance one by one based on the current pedestrian apparent features and the historical pedestrian apparent features of each pedestrian in the historical feature database. If the apparent feature distance is greater than the apparent threshold, determine the current pedestrian apparent feature and history The pedestrian apparent features in the feature library belong to the same pedestrian, and the current apparent feature distance is determined as the pedestrian's apparent feature distance;
步骤3.2、基于当前的行人三维运动特征以及历史特征库中每个行人的上一时刻预测得到的下一时刻的行人三维运动特征逐一计算空间特征距离,若空间特征距离大于空间阈值则判断当前的行人三维运动特征与历史特征库中的上一时刻预测得到的下一时刻的行人三维运动特征属于同一行人,确定当前的空间特征距离作为该行人的空间特征距离;Step 3.2. Calculate the spatial feature distances one by one based on the current three-dimensional motion features of pedestrians and the three-dimensional motion features of pedestrians at the next moment predicted from the previous moment of each pedestrian in the historical feature database. If the spatial feature distance is greater than the spatial threshold, determine the current The three-dimensional motion feature of the pedestrian and the three-dimensional motion feature of the pedestrian at the next moment predicted in the historical feature database belong to the same pedestrian, and the current spatial feature distance is determined as the spatial feature distance of the pedestrian;
步骤3.3、基于当前的行人三维运动特征、表观特征距离、空间特征距离,以及历史特征库中每个行人历史的行人三维运动特征判断是否符合同一行人的运动模式,输出运动模式匹配度作为该行人的运动模式匹配度;Step 3.3. Based on the current three-dimensional motion features of pedestrians, apparent feature distances, spatial feature distances, and the three-dimensional motion features of each pedestrian history in the historical feature database, determine whether it conforms to the motion pattern of the same pedestrian, and output the motion pattern matching degree as the Pedestrian motion pattern matching degree;
步骤3.4、将属于同一行人的表观特征距离、空间特征距离、运动模式匹配度进行加权计算,得到当前三维包围盒内行人与历史特征库中行人的匹配结果,所述匹配结果包括匹配成功或匹配失败,匹配成功时还包括匹配得到的行人信息;Step 3.4: Perform weighted calculation on the apparent feature distance, spatial feature distance, and motion pattern matching degree belonging to the same pedestrian, and obtain the matching result between the pedestrian in the current three-dimensional bounding box and the pedestrian in the historical feature database, and the matching result includes a successful matching or a matching result. If the matching fails, the pedestrian information obtained by the matching is also included when the matching is successful;
步骤4、根据本次的匹配结果以及历史的匹配结果标记三维包围盒所对应行人的行人状态为初次匹配成功、丢失、丢失后重新匹配成功、连续匹配成功或走出摄像范围。Step 4: Mark the pedestrian status of the pedestrian corresponding to the three-dimensional bounding box according to the current matching results and the historical matching results as initial matching successful, lost, successful re-matching after loss, successful continuous matching or out of the camera range.
以下还提供了若干可选方式,但并不作为对上述总体方案的额外限定,仅仅是进一步的增补或优选,在没有技术或逻辑矛盾的前提下,各可选方式可单独针对上述总体方案进行组合,还可以是多个可选方式之间进行组合。Several optional methods are also provided below, which are not intended to be additional limitations on the above-mentioned overall solution, but are merely further additions or optimizations. On the premise of no technical or logical contradiction, each optional method can be independently implemented for the above-mentioned overall solution. The combination can also be a combination between multiple optional ways.
作为优选,所述检测光学图像中的行人,输出行人的三维包围盒,包括:Preferably, the pedestrian in the optical image is detected, and a three-dimensional bounding box of the pedestrian is output, including:
对用于获取光学图像的摄像头进行标定,得到光学图像中像素与摄像头距离的映射关系;The camera used to obtain the optical image is calibrated, and the mapping relationship between the pixel in the optical image and the distance of the camera is obtained;
检测光学图像中的行人,获取光学图像中行人的二维包围框;Detect pedestrians in optical images, and obtain two-dimensional bounding boxes of pedestrians in optical images;
基于二维包围框和映射关系得到行人的三维包围盒。Based on the 2D bounding box and the mapping relationship, the 3D bounding box of the pedestrian is obtained.
作为优选,所述基于行人当前的三维包围盒以及历史特征库中按照时间序列分布的三维包围盒,提取每个行人的行人三维运动特征,包括:Preferably, the three-dimensional motion feature of each pedestrian is extracted based on the current three-dimensional bounding box of the pedestrian and the three-dimensional bounding box distributed according to the time series in the historical feature library, including:
步骤2.2.1、方向矢量提取:通过当前三维包围盒及历史的三维包围盒,提取行人在水平方向的运动方向及垂直方向的运动方向;Step 2.2.1. Extraction of direction vector: extract the movement direction of pedestrians in the horizontal direction and the movement direction in the vertical direction through the current 3D bounding box and the historical 3D bounding box;
步骤2.2.2、运动速度提取:通过当前三维包围盒及历史的三维包围盒,提取人在水平方向的运动速度及垂直方向的运动速度;Step 2.2.2. Movement speed extraction: extract the movement speed of the person in the horizontal direction and the movement speed in the vertical direction through the current 3D bounding box and the historical 3D bounding box;
步骤2.2.3、相对位置提取:根据摄像头标定后得到的映射关系,基于当前三维包围盒及历史的三维包围盒输出行人在以摄像头为中心的三维坐标系中的坐标;Step 2.2.3. Relative position extraction: According to the mapping relationship obtained after the camera is calibrated, the coordinates of the pedestrian in the three-dimensional coordinate system centered on the camera are output based on the current three-dimensional bounding box and the historical three-dimensional bounding box;
步骤2.2.4、将步骤2.2.1~2.2.3中提取的方向矢量、运动速度和相对位置作为行人三维运动特征。Step 2.2.4. Use the direction vector, motion speed and relative position extracted in steps 2.2.1 to 2.2.3 as the three-dimensional motion feature of the pedestrian.
作为优选,所述根据本次的匹配结果以及历史的匹配结果标记三维包围盒所对应行人的行人状态为初次匹配成功、丢失、丢失后重新匹配成功、连续匹配成功或走出摄像范围,包括:Preferably, the pedestrian state of the pedestrian corresponding to the three-dimensional bounding box is marked according to the current matching result and the historical matching result as initial matching success, loss, rematching success after loss, continuous matching success or out of the camera range, including:
若成功提取行人特征,但匹配结果为匹配失败,则标记当前行人的状态为初次匹配成功;If the pedestrian feature is successfully extracted, but the matching result is that the matching fails, the status of the current pedestrian is marked as the initial matching success;
若连续M次未匹配到历史匹配结果中的同一行人,则标记该行人的状态为丢失;If the same pedestrian in the historical matching results is not matched for M consecutive times, the status of the pedestrian is marked as lost;
若被标记为丢失的行人本次匹配结果中重新匹配成功,则更新该行人的状态为丢失后重新匹配成功;If the pedestrian marked as lost is successfully re-matched in the current matching result, the status of the pedestrian is updated to be lost and the re-match is successful;
若连续L次匹配到历史匹配结果中的同一行人,则更新该行人的状态为连续匹配成功;If the same pedestrian in the historical matching result is matched for L consecutive times, the status of the pedestrian is updated to indicate that the continuous matching is successful;
若连续N次未匹配到历史匹配结果中的同一行人,则标记该行人的状态为走出摄像范围,且M<N。If the same pedestrian in the historical matching result is not matched for N consecutive times, the state of the pedestrian is marked as walking out of the camera range, and M<N.
作为优选,若当前行人的状态标记为初次匹配成功,则在历史特征库中为该行人分配新的行人信息,并将该行人的行人特征与新分配的行人信息关联。Preferably, if the status of the current pedestrian is marked as successful for the first time, new pedestrian information is allocated to the pedestrian in the historical feature database, and the pedestrian characteristic of the pedestrian is associated with the newly allocated pedestrian information.
本发明还提供一种人流统计系统,所述人流统计系统,包括:The present invention also provides a people flow statistics system, and the people flow statistics system includes:
行人检测模块,用于获取光学图像,检测光学图像中的行人,输出行人的三维包围盒及对应的时间戳;The pedestrian detection module is used to obtain optical images, detect pedestrians in the optical images, and output the three-dimensional bounding boxes of pedestrians and corresponding timestamps;
特征提取模块,用于基于光学图像和三维包围盒获取行人特征,具体执行以下步骤:The feature extraction module is used to obtain pedestrian features based on optical images and 3D bounding boxes. The specific steps are as follows:
a、提取光学图像中行人的人体形状和特征作为每个行人的行人表观特征,并保存至历史特征库中;a. Extract the human body shape and features of the pedestrians in the optical image as the pedestrian appearance features of each pedestrian, and save them in the historical feature database;
b、基于行人当前的三维包围盒以及历史特征库中按照时间序列分布的三维包围盒,提取每个行人的行人三维运动特征,并保存至历史特征库中;b. Based on the current 3D bounding box of the pedestrian and the 3D bounding box distributed according to the time series in the historical feature library, extract the three-dimensional motion feature of each pedestrian and save it in the historical feature library;
c、基于行人三维运动特征及历史特征库中指定时间内的行人三维运动特征预测下一时刻的行人三维运动特征,并保存至历史特征库中;c. Predict the 3D motion feature of the pedestrian at the next moment based on the 3D motion feature of the pedestrian and the 3D motion feature of the pedestrian within the specified time in the historical feature library, and save it in the historical feature library;
行人识别模块,用于基于历史特征库中的行人特征进行行人识别,具体执行以下步骤:The pedestrian recognition module is used for pedestrian recognition based on the pedestrian features in the historical feature database, and the specific steps are as follows:
a、基于当前的行人表观特征和历史特征库中每个行人历史的行人表观特征逐一计算表观特征距离,若表观特征距离大于表观阈值则判断当前的行人表观特征与历史特征库中的行人表观特征属于同一行人,确定当前的表观特征距离作为该行人的表观特征距离;a. Calculate the apparent feature distance one by one based on the current pedestrian apparent feature and the historical feature of each pedestrian in the historical feature database. If the apparent feature distance is greater than the apparent threshold, then judge the current pedestrian apparent feature and historical feature The pedestrian apparent features in the library belong to the same pedestrian, and the current apparent feature distance is determined as the pedestrian's apparent feature distance;
b、基于当前的行人三维运动特征以及历史特征库中每个行人的上一时刻预测得到的下一时刻的行人三维运动特征逐一计算空间特征距离,若空间特征距离大于空间阈值则判断当前的行人三维运动特征与历史特征库中的上一时刻预测得到的下一时刻的行人三维运动特征属于同一行人,确定当前的空间特征距离作为该行人的空间特征距离;b. Calculate the spatial feature distances one by one based on the current three-dimensional motion features of pedestrians and the three-dimensional motion features of pedestrians at the next moment predicted from the previous moment of each pedestrian in the historical feature database. If the spatial feature distance is greater than the spatial threshold, the current pedestrian is judged The three-dimensional motion feature and the three-dimensional motion feature of the pedestrian at the next moment predicted at the previous moment in the historical feature library belong to the same pedestrian, and the current spatial feature distance is determined as the spatial feature distance of the pedestrian;
c、基于当前的行人三维运动特征、表观特征距离、空间特征距离,以及历史特征库中每个行人历史的行人三维运动特征判断是否符合同一行人的运动模式,输出运动模式匹配度作为该行人的运动模式匹配度;c. Judging whether it conforms to the motion pattern of the same pedestrian based on the current three-dimensional motion features, apparent feature distance, spatial feature distance, and the three-dimensional motion features of each pedestrian in the historical feature database, and output the motion pattern matching degree as the pedestrian Motion pattern matching degree;
d、将属于同一行人的表观特征距离、空间特征距离、运动模式匹配度进行加权计算,得到当前三维包围盒内行人与历史特征库中行人的匹配结果,所述匹配结果包括匹配成功或匹配失败,匹配成功时还包括匹配得到的行人信息;d. Perform weighted calculation on the apparent feature distance, spatial feature distance, and motion pattern matching degree belonging to the same pedestrian, and obtain the matching result between the pedestrian in the current three-dimensional bounding box and the pedestrian in the historical feature database, and the matching result includes matching success or matching. If the match fails, the pedestrian information obtained by the match is also included when the match is successful;
行人标记模块,用于根据本次的匹配结果以及历史的匹配结果标记三维包围盒所对应行人的行人状态为初次匹配成功、丢失、丢失后重新匹配成功、连续匹配成功或走出摄像范围;The pedestrian marking module is used to mark the pedestrian status of the pedestrian corresponding to the three-dimensional bounding box according to the current matching results and the historical matching results as initial matching success, loss, rematching success after loss, continuous matching success or out of the camera range;
人流统计模块,用于根据行人状态统计预设时间内光学图像所对应的统计范围内的人流量。The people flow statistics module is used to count the flow of people within the statistical range corresponding to the optical image within the preset time according to the pedestrian state.
作为优选,所述检测光学图像中的行人,输出行人的三维包围盒,执行如下操作:Preferably, the pedestrian in the optical image is detected, and the three-dimensional bounding box of the pedestrian is output, and the following operations are performed:
对用于获取光学图像的摄像头进行标定,得到光学图像中像素与摄像头距离的映射关系;The camera used to obtain the optical image is calibrated, and the mapping relationship between the pixel in the optical image and the distance of the camera is obtained;
检测光学图像中的行人,获取光学图像中行人的二维包围框;Detect pedestrians in optical images, and obtain two-dimensional bounding boxes of pedestrians in optical images;
基于二维包围框和映射关系得到行人的三维包围盒。Based on the 2D bounding box and the mapping relationship, the 3D bounding box of the pedestrian is obtained.
作为优选,所述基于行人当前的三维包围盒以及历史特征库中按照时间序 列分布的三维包围盒,提取每个行人的行人三维运动特征,执行如下操作:As preferably, the three-dimensional bounding box based on the current three-dimensional bounding box of pedestrians and the three-dimensional bounding box distributed according to time series in the historical feature library, extract the three-dimensional motion feature of pedestrians of each pedestrian, and perform the following operations:
方向矢量提取:通过当前三维包围盒及历史的三维包围盒,提取行人在水平方向的运动方向及垂直方向的运动方向;Direction vector extraction: Through the current 3D bounding box and the historical 3D bounding box, the pedestrian's movement direction in the horizontal direction and the movement direction in the vertical direction are extracted;
运动速度提取:通过当前三维包围盒及历史的三维包围盒,提取人在水平方向的运动速度及垂直方向的运动速度;Movement speed extraction: extract the movement speed of the person in the horizontal direction and the movement speed in the vertical direction through the current 3D bounding box and the historical 3D bounding box;
相对位置提取:根据摄像头标定后得到的映射关系,基于当前三维包围盒及历史的三维包围盒输出行人在以摄像头为中心的三维坐标系中的坐标;Relative position extraction: According to the mapping relationship obtained after the camera is calibrated, the coordinates of the pedestrian in the three-dimensional coordinate system centered on the camera are output based on the current three-dimensional bounding box and the historical three-dimensional bounding box;
特征整合:将提取的方向矢量、运动速度和相对位置作为行人三维运动特征。Feature integration: The extracted direction vector, motion speed and relative position are used as pedestrian 3D motion features.
作为优选,所述根据本次的匹配结果以及历史的匹配结果标记三维包围盒所对应行人的行人状态为初次匹配成功、丢失、丢失后重新匹配成功、连续匹配成功或走出摄像范围,执行如下操作:Preferably, according to the current matching results and the historical matching results, the pedestrian states of the pedestrians corresponding to the three-dimensional bounding box are marked as successful initial matching, lost, successful re-matching after loss, successful continuous matching, or out of the camera range, and the following operations are performed. :
若成功提取行人特征,但匹配结果为匹配失败,则标记当前行人的状态为初次匹配成功;If the pedestrian feature is successfully extracted, but the matching result is that the matching fails, the status of the current pedestrian is marked as the initial matching success;
若连续M次未匹配到历史匹配结果中的同一行人,则标记该行人的状态为丢失;If the same pedestrian in the historical matching results is not matched for M consecutive times, the status of the pedestrian is marked as lost;
若被标记为丢失的行人本次匹配结果中重新匹配成功,则更新该行人的状态为丢失后重新匹配成功;If the pedestrian marked as lost is successfully re-matched in the current matching result, the status of the pedestrian is updated to be lost and the re-match is successful;
若连续L次匹配到历史匹配结果中的同一行人,则更新该行人的状态为连续匹配成功;If the same pedestrian in the historical matching result is matched for L consecutive times, the status of the pedestrian is updated to indicate that the continuous matching is successful;
若连续N次未匹配到历史匹配结果中的同一行人,则标记该行人的状态为走出摄像范围,且M<N。If the same pedestrian in the historical matching result is not matched for N consecutive times, the state of the pedestrian is marked as walking out of the camera range, and M<N.
作为优选,若当前行人的状态标记为初次匹配成功,则在历史特征库中为该行人分配新的行人信息,并将该行人的行人特征与新分配的行人信息关联。Preferably, if the status of the current pedestrian is marked as successful for the first time, new pedestrian information is allocated to the pedestrian in the historical feature database, and the pedestrian characteristic of the pedestrian is associated with the newly allocated pedestrian information.
本发明提供的一种公共场所行人识别方法,综合考虑行人表观特征、三维运动特征及运动模式,准确识别行人,获取行人进出统计范围的时间、位置及在统计范围内的移动轨迹,并基于该方法提出一种人流统计系统,能够准确地统计出单位时间内出入统计范围的人流量。The present invention provides a pedestrian identification method in a public place, which comprehensively considers pedestrian appearance features, three-dimensional motion features and motion patterns, accurately identifies pedestrians, obtains the time and position of pedestrians entering and exiting the statistical range, and the movement trajectory within the statistical range, and based on the This method proposes a people flow statistics system, which can accurately count the flow of people entering and leaving the statistical range in unit time.
附图说明Description of drawings
图1为本发明公共场所行人识别方法的流程图;Fig. 1 is the flow chart of the pedestrian identification method in public places of the present invention;
图2为本发明输出行人的三维包围盒的流程图;Fig. 2 is the flow chart of the present invention outputting the three-dimensional bounding box of pedestrian;
图3为本发明基于光学图像和三维包围盒获取行人特征的流程图;3 is a flow chart of the present invention for acquiring pedestrian features based on an optical image and a three-dimensional bounding box;
图4为本发明在右手坐标系下按人体结构提取运动特征的示意图;4 is a schematic diagram of the present invention extracting motion features by human body structure in a right-handed coordinate system;
图5为本发明基于历史特征库中的行人特征进行行人识别的流程图;Fig. 5 is the flow chart that the present invention carries out pedestrian recognition based on the pedestrian characteristic in the historical characteristic database;
图6为本发明进行行人状态标记的流程图;6 is a flow chart of the present invention for marking pedestrian status;
图7为本发明人流统计系统的结构图。FIG. 7 is a structural diagram of the people flow statistics system of the present invention.
具体实施方式Detailed ways
下面将结合本发明实施例中的附图,对本发明实施例中的技术方案进行清楚、完整地描述,显然,所描述的实施例仅是本发明一部分实施例,而不是全部的实施例。基于本发明中的实施例,本领域普通技术人员在没有做出创造性劳动前提下所获得的所有其他实施例,都属于本发明保护的范围。The technical solutions in the embodiments of the present invention will be clearly and completely described below with reference to the accompanying drawings in the embodiments of the present invention. Obviously, the described embodiments are only a part of the embodiments of the present invention, but not all of the embodiments. Based on the embodiments of the present invention, all other embodiments obtained by those of ordinary skill in the art without creative efforts shall fall within the protection scope of the present invention.
除非另有定义,本文所使用的所有的技术和科学术语与属于本发明的技术领域的技术人员通常理解的含义相同。本文中在本发明的说明书中所使用的术语只是为了描述具体的实施例的目的,不是在于限制本发明。Unless otherwise defined, all technical and scientific terms used herein have the same meaning as commonly understood by one of ordinary skill in the art to which this invention belongs. The terms used herein in the description of the present invention are for the purpose of describing specific embodiments only, and are not intended to limit the present invention.
其中一个实施例中,提供一种公共场所行人识别方法,对行人识别准确,可用于基于行人识别统计人流进行城市规划,商场人流统计进行营业策略调整、地铁站人流统计进行地铁班次调整等场景。In one embodiment, a pedestrian identification method in a public place is provided, which can accurately identify pedestrians and can be used for urban planning based on pedestrian identification statistics, business strategy adjustment based on pedestrian flow statistics in shopping malls, and subway shift adjustment based on subway station traffic statistics.
如图1所示,本实施例中的公共场所行人识别方法,包括以下步骤:As shown in Figure 1, the method for identifying pedestrians in public places in this embodiment includes the following steps:
步骤1、获取光学图像,检测光学图像中的行人,输出行人的三维包围盒及对应的时间戳。Step 1. Obtain an optical image, detect pedestrians in the optical image, and output a three-dimensional bounding box of the pedestrian and a corresponding timestamp.
本实施例基于摄像头获取光学图像,得到的时间戳即为摄像头拍摄光学图像时的时间。容易理解的是,获取光学图像可以是基于任意图像采集设备,本实施例以摄像头为例进行说明。In this embodiment, the optical image is acquired by the camera, and the obtained time stamp is the time when the optical image is captured by the camera. It is easy to understand that the acquisition of the optical image may be based on any image acquisition device, and this embodiment takes a camera as an example for description.
由于三维包围盒是带有深度信息的,而光学图像中不含有深度信息,因此本实施例在形成三维包围盒时包括以下步骤,如图2所示:Since the 3D bounding box has depth information, and the optical image does not contain depth information, this embodiment includes the following steps when forming the 3D bounding box, as shown in Figure 2:
对用于获取光学图像的摄像头进行标定,得到光学图像中像素与摄像头距离的映射关系;检测光学图像中的行人,获取光学图像中行人的二维包围框(Bounding Box,BBox);基于二维包围框和映射关系得到行人的三维包围盒(3D Bounding Box,3D BBox)。The camera used to obtain the optical image is calibrated, and the mapping relationship between the pixel and the camera distance in the optical image is obtained; the pedestrian in the optical image is detected, and the two-dimensional bounding box (BBox) of the pedestrian in the optical image is obtained; based on the two-dimensional The bounding box and the mapping relationship get the pedestrian's three-dimensional bounding box (3D Bounding Box, 3D BBox).
本实施例通过对摄像头的标定得到光学图像中行人像素与摄像头距离的映 射关系,基于该映射关系得到对应的深度信息,该深度信息体现了行人与摄像头之间的实际距离,包含了行人的移动变化,便于后续基于深度信息提取出行人的三维运动特征。In this embodiment, the mapping relationship between the pedestrian pixel and the camera distance in the optical image is obtained by calibrating the camera, and the corresponding depth information is obtained based on the mapping relationship. The depth information reflects the actual distance between the pedestrian and the camera, including the movement of the pedestrian. It is convenient to extract the three-dimensional motion features of pedestrians based on the depth information.
在本实施例中,使用单目定焦摄像头拍摄视频,使用各边长为1米的立方体对映射关系进行标定,该立方体每一个面平均划分成100个黑白相间的网格,摄像头拍摄范围为统计范围。In this embodiment, a monocular fixed-focus camera is used to shoot a video, and a cube with a side length of 1 meter is used to calibrate the mapping relationship. Each face of the cube is evenly divided into 100 black and white grids, and the camera shooting range is Statistical range.
需要说明的是,标定为摄像头使用的常规技术手段,本实施例中不限制标定具体步骤,并且基于摄像头标定的映射关系得到深度信息,为本实施例提供的优选方法,但不限制为唯一手段,例如可以通过摄像机和深度相机的结合叠加深度信息等。It should be noted that the calibration is a conventional technical means used by the camera, and the specific steps of the calibration are not limited in this embodiment, and the depth information is obtained based on the mapping relationship of the camera calibration, which is the preferred method provided in this embodiment, but is not limited to the only means. , for example, the depth information can be superimposed by the combination of the camera and the depth camera.
本实施例中基于行人检测方法识别光学图像中的行人并输出二维包围框,所使用的行人检测方法为图像识别中的常规方法,例如使用在行人数据集上训练得到的基于Yolo拓展的识别网络。而在得到三维包围盒时,将二维包围框和映射关系输入三维估算方法,输出三维包围盒。In this embodiment, pedestrians in the optical image are identified based on the pedestrian detection method and a two-dimensional bounding box is output. The pedestrian detection method used is a conventional method in image recognition, for example, the recognition based on Yolo extension obtained by training on the pedestrian dataset is used. network. When the 3D bounding box is obtained, the 2D bounding box and the mapping relationship are input into the 3D estimation method, and the 3D bounding box is output.
本实施例中使用的三维估算方法为基于光流的单目深度估计方法,该方法可以输出逆深度,由逆深度可计算得到深度信息。该深度信息在距离摄像头不同范围内误差系数不同,在本实施例中,使用了一个误差矩阵。The three-dimensional estimation method used in this embodiment is a monocular depth estimation method based on optical flow, which can output the inverse depth, and the depth information can be obtained by calculating the inverse depth. The depth information has different error coefficients in different ranges from the camera. In this embodiment, an error matrix is used.
基于三维包围盒进行行人识别可有效克服行人遮挡问题,由于人的身体结构符合几何约束,被部分遮挡的二维包围框可以恢复成完整的三维包围盒,其空间误差在容许范围内。Pedestrian recognition based on the 3D bounding box can effectively overcome the pedestrian occlusion problem. Since the human body structure conforms to the geometric constraints, the partially occluded 2D bounding box can be restored to a complete 3D bounding box, and its spatial error is within the allowable range.
步骤2、基于光学图像和三维包围盒获取行人特征,如图3所示,包括:Step 2. Obtain pedestrian features based on the optical image and the 3D bounding box, as shown in Figure 3, including:
步骤2.1、提取光学图像中行人的人体形状和特征作为每个行人的行人表观特征,并保存至历史特征库中。Step 2.1, extract the human body shape and features of the pedestrians in the optical image as the pedestrian appearance features of each pedestrian, and save them in the historical feature database.
由于人体形状和特征是区别不同行人的重要特征,所以本实施例采用行人表观特征提取方法,主要提取可以通过视觉观察到的人体性状和特征,为便于区分,并标记行人表观特征为F appearanceSince the shape and features of the human body are important features to distinguish different pedestrians, this embodiment adopts the pedestrian apparent feature extraction method, mainly extracting the human body traits and features that can be observed visually, and marking the pedestrian apparent feature as F for the convenience of distinction appearance .
本实施例中使用基于Yolo拓展的识别网络的嵌入结构(embedding,嵌入)作为行人表观特征提取方法。In this embodiment, the embedding structure (embedding, embedding) of the recognition network based on Yolo extension is used as the method for extracting the apparent feature of pedestrians.
步骤2.2、基于行人当前的三维包围盒以及历史特征库中按照时间序列分布的三维包围盒,提取每个行人的行人三维运动特征,并保存至历史特征库中。Step 2.2, based on the current three-dimensional bounding box of the pedestrian and the three-dimensional bounding box distributed according to the time series in the historical feature library, extract the three-dimensional motion feature of each pedestrian, and save it in the historical feature library.
行人三维运动特征为行人在三维空间中的位置变化特征,为便于区分,标 记为F displacement,是在时间序列上进行数据关联的重要特征。 The three-dimensional motion feature of pedestrians is the feature of the position change of pedestrians in three-dimensional space. For the convenience of distinction, it is marked as F displacement , which is an important feature for data association in time series.
如图4所示,本实施例中建立右手坐标系,依据人体结构将三维包围盒分为头、上身、下身三部分,输入三维包围盒到行人三维运动特征提取方法,主要提取人在三维空间中的位置变化特征,输出行人三维运动特征。As shown in Figure 4, in this embodiment, a right-handed coordinate system is established, and the three-dimensional bounding box is divided into three parts: head, upper body, and lower body according to the human body structure, and the three-dimensional bounding box is input to the pedestrian three-dimensional motion feature extraction method, which mainly extracts people in three-dimensional space. The position change feature in the output is the three-dimensional motion feature of the pedestrian.
在本实施例中,三维运动特征提取方法由以下部分组成:In this embodiment, the three-dimensional motion feature extraction method consists of the following parts:
步骤2.2.1、方向矢量提取:通过当前三维包围盒及历史的三维包围盒,提取行人在水平方向的运动方向及垂直方向的运动方向。Step 2.2.1. Extraction of direction vector: Extract the moving direction of pedestrians in the horizontal direction and the moving direction in the vertical direction through the current three-dimensional bounding box and the historical three-dimensional bounding box.
由于三维包围盒具有对应的时间戳,因此基于时间分布的三维包围盒的位置变化能够得到行人的方向矢量。多个三维包围盒可以仅使用首位两个包围盒进行方向矢量确定,也可以使用多对三维包围盒,并取多个方向矢量的均值、中值或其他值作为最终确定的方向矢量。Since the 3D bounding box has a corresponding timestamp, the direction vector of the pedestrian can be obtained based on the position change of the 3D bounding box based on the time distribution. For multiple 3D bounding boxes, only the first two bounding boxes can be used to determine the direction vector, or multiple pairs of 3D bounding boxes can be used, and the mean, median or other values of the multiple direction vectors can be taken as the final direction vector.
步骤2.2.2、运动速度提取:通过当前三维包围盒及历史的三维包围盒,提取人在水平方向的运动速度及垂直方向的运动速度。Step 2.2.2. Movement speed extraction: extract the movement speed of the person in the horizontal direction and the movement speed in the vertical direction through the current 3D bounding box and the historical 3D bounding box.
与方向矢量提取类似,基于时间分布的三维包围盒,根据时间差以及对应的三维包围盒的位置差得到行人的运动速度,并且多个三维包围盒可以仅使用首位两个包围盒进行运动速度计算,也可以使用多对三维包围盒,并取多个运动速度的均值、中值或其他值作为最终确定的运动速度。Similar to the direction vector extraction, the 3D bounding box based on the time distribution obtains the movement speed of the pedestrian according to the time difference and the position difference of the corresponding 3D bounding box, and multiple 3D bounding boxes can only use the first two bounding boxes to calculate the movement speed, It is also possible to use multiple pairs of three-dimensional bounding boxes, and take the mean, median or other values of multiple motion velocities as the final motion velocity.
步骤2.2.3、相对位置提取:根据摄像头标定后得到的映射关系,基于当前三维包围盒及历史的三维包围盒输出行人在以摄像头为中心的三维坐标系中的坐标。Step 2.2.3. Relative position extraction: According to the mapping relationship obtained after the camera is calibrated, the coordinates of the pedestrian in the three-dimensional coordinate system centered on the camera are output based on the current three-dimensional bounding box and the historical three-dimensional bounding box.
在确定坐标时,将每个三维包围盒等价为一个点,求取该点的坐标作为行人的坐标,该点可以是三维包围盒的中心点、某一顶点或任意一点。When determining the coordinates, each three-dimensional bounding box is equivalent to a point, and the coordinates of the point are obtained as the coordinates of the pedestrian. The point can be the center point of the three-dimensional bounding box, a vertex or any point.
步骤2.2.4、将步骤2.2.1~2.2.3中提取的方向矢量、运动速度和相对位置作为行人三维运动特征。Step 2.2.4. Use the direction vector, motion speed and relative position extracted in steps 2.2.1 to 2.2.3 as the three-dimensional motion feature of the pedestrian.
由于历史特征库中通常存在多个行人的三维包围盒,因此在进行行人三维运动特征提取时,首先对当前和历史的三维包围盒进行特征匹配(例如采用匈牙利匹配算法),取未被使用且匹配度最高的历史的三维包围盒进行行人三维运动特征的提取。Since there are usually multiple 3D bounding boxes of pedestrians in the historical feature database, when extracting the 3D motion features of pedestrians, first perform feature matching on the current and historical 3D bounding boxes (for example, using the Hungarian matching algorithm), and take the unused and The historical 3D bounding box with the highest matching degree is used to extract the 3D motion features of pedestrians.
若当前的三维包围盒为新进入统计范围的新的行人的三维包围盒,则上述匹配未能得到相应的历史的三维包围盒,则设置新的行人的方向矢量、运动速度为默认值(例如方向矢量为无、运输速度为0),坐标为当前三维包围盒的坐 标作为该行人的行人三维运动特征。If the current 3D bounding box is the 3D bounding box of a new pedestrian entering the statistical range, and the above matching fails to obtain the corresponding historical 3D bounding box, set the direction vector and movement speed of the new pedestrian as default values (for example, The direction vector is none and the transport speed is 0), and the coordinates are the coordinates of the current three-dimensional bounding box as the three-dimensional motion feature of the pedestrian.
步骤2.3、基于行人三维运动特征及历史特征库中指定时间内的行人三维运动特征预测下一时刻的行人三维运动特征,并保存至历史特征库中。Step 2.3: Predict the three-dimensional motion feature of the pedestrian at the next moment based on the three-dimensional motion feature of the pedestrian and the three-dimensional motion feature of the pedestrian in the historical feature library at a specified time, and save the three-dimensional motion feature of the pedestrian in the historical feature library.
下一时刻的行人三维运动特征标记为F predicted,由轨迹预测算法预测得到,表现的是行人在时空上的运动趋势,使用该特征能够进一步解决行人被遮挡导致的目标丢失问题;本实施例中使用卡尔曼滤波预测行人下一时刻的行人三维运动特征。 The three-dimensional motion feature of the pedestrian at the next moment is marked as F predicted , which is predicted by the trajectory prediction algorithm and represents the movement trend of the pedestrian in space and time. Using this feature can further solve the problem of target loss caused by the pedestrian being occluded; in this embodiment Use Kalman filtering to predict the three-dimensional motion characteristics of pedestrians at the next moment.
与行人三维运动特征提取类似,由于历史特征库中通常存在多个行人的行人三维运动特征,因此在进行行人三维运动特征预测时,首先对当前和历史的行人三维运动特征进行特征匹配(例如采用匈牙利匹配算法),取未被使用且匹配度最高的历史的行人三维运动特征进行行人三维运动特征的预测。Similar to the extraction of pedestrian 3D motion features, since there are usually multiple pedestrian 3D motion features in the historical feature database, when predicting pedestrian 3D motion features, feature matching is first performed on the current and historical pedestrian 3D motion features (for example, using Hungarian matching algorithm), take the historical pedestrian 3D motion features that are not used and have the highest matching degree to predict pedestrian 3D motion features.
若当前的行人三维运动特征为新进入统计范围的新行人的行人三维运动特征,则上述匹配未能得到相应的历史的行人三维运动特征,则直接基于当前的行人三维运动特征进行预测。步骤3、基于历史特征库中的行人特征进行行人识别,如图5所示,包括:If the current pedestrian 3D motion feature is the pedestrian 3D motion feature of a new pedestrian entering the statistical range, and the above matching fails to obtain the corresponding historical pedestrian 3D motion feature, the prediction is made directly based on the current pedestrian 3D motion feature. Step 3. Perform pedestrian recognition based on the pedestrian features in the historical feature database, as shown in Figure 5, including:
步骤3.1、基于当前的行人表观特征和历史特征库中每个行人历史的行人表观特征逐一计算表观特征距离,若表观特征距离大于表观阈值则判断当前的行人表观特征与历史特征库中的行人表观特征属于同一行人,确定该表观特征距离作为该行人的表观特征距离。Step 3.1. Calculate the apparent feature distance one by one based on the current pedestrian apparent features and the historical pedestrian apparent features of each pedestrian in the historical feature database. If the apparent feature distance is greater than the apparent threshold, determine the current pedestrian apparent feature and history The pedestrian apparent features in the feature library belong to the same pedestrian, and the apparent feature distance is determined as the pedestrian's apparent feature distance.
本实施例中使用行人表观特征匹配方法,通过计算当前行人表观特征
Figure PCTCN2020137803-appb-000001
与历史特征库中行人表观特征
Figure PCTCN2020137803-appb-000002
的特征距离,判断是否属于同一行人,例如使用马氏距离和余弦距离的加权计算表观特征距离,系数分别为0.02和0.98。本实施例中的历史的行人表观特征主要取用上一时刻的行人表观特征。为了提高匹配结果,在另一个实施例中,若判断到属于同一行人的行人表观特征,则可以建立表观相似行人列表,每个行人对应一个表观相似行人列表,基于该列表进一步区分表观特征距离。
In this embodiment, the pedestrian apparent feature matching method is used to calculate the current pedestrian apparent feature
Figure PCTCN2020137803-appb-000001
Pedestrian appearance features in historical feature database
Figure PCTCN2020137803-appb-000002
To judge whether they belong to the same pedestrian, for example, using the weighted Mahalanobis distance and cosine distance to calculate the apparent feature distance, the coefficients are 0.02 and 0.98 respectively. The historical pedestrian appearance feature in this embodiment mainly adopts the pedestrian appearance feature of the previous moment. In order to improve the matching result, in another embodiment, if the apparent characteristics of pedestrians belonging to the same pedestrian are determined, a list of apparently similar pedestrians can be established, each pedestrian corresponds to a list of apparently similar pedestrians, and the table is further distinguished based on the list. View feature distance.
例如当前时刻具有A和B两个人的表观特征,其中A和A{t-1}(上一时刻的行人表观特征),B和A{t-1}的相似度也很高,甚至B和A的表观相似行人列表中多个时间段的相似度都很高,但是B和B的表观相似行人列表中多个时间段的相似度更高,那么可以判断B就是B。但是由于该搜索方法会比较耗时,因此一般在行人识别要求较高的场景下使用。For example, the current moment has the apparent characteristics of two persons, A and B, where A and A{t-1} (the pedestrian apparent characteristics of the previous moment), B and A{t-1} are also very similar, even The similarity of multiple time periods in the apparently similar pedestrian lists of B and A is high, but the similarity of multiple time periods in the apparently similar pedestrian lists of B and B is higher, then it can be judged that B is B. However, since this search method is time-consuming, it is generally used in scenarios with high requirements for pedestrian recognition.
步骤3.2、基于当前的行人三维运动特征以及历史特征库中每个行人的上一时刻预测得到的下一时刻的行人三维运动特征逐一计算空间特征距离,若空间特征距离大于空间阈值则判断当前的行人三维运动特征与历史特征库中的上一时刻预测得到的下一时刻的行人三维运动特征属于同一行人,确定当前的空间特征距离作为该行人的空间特征距离。Step 3.2. Calculate the spatial feature distances one by one based on the current three-dimensional motion features of pedestrians and the three-dimensional motion features of pedestrians at the next moment predicted from the previous moment of each pedestrian in the historical feature database. If the spatial feature distance is greater than the spatial threshold, determine the current The three-dimensional motion feature of the pedestrian and the three-dimensional motion feature of the pedestrian at the next moment predicted in the historical feature database belong to the same pedestrian, and the current spatial feature distance is determined as the spatial feature distance of the pedestrian.
上一时刻预测得到的下一时刻的行人三维运动特征也就是预测得到的当前的行人三维运动特征,将预测的和实际的当前的行人三维运动特征进行匹配,可以作为是否属于同一行人的判断依据之一,因为同一行人的三维运动特征变化不会过大,因此该匹配具有参考性。在本实施例中使用匈牙利算法作为行人三维运动特征匹配方法进行计算,判断是否属于同一行人。The three-dimensional motion feature of the pedestrian at the next moment predicted at the previous moment is the predicted three-dimensional motion feature of the current pedestrian. Matching the predicted and the actual current three-dimensional motion feature of the pedestrian can be used as the basis for judging whether they belong to the same pedestrian. One, because the three-dimensional motion characteristics of the same pedestrian will not change too much, so the matching is for reference. In this embodiment, the Hungarian algorithm is used as the pedestrian three-dimensional motion feature matching method for calculation to determine whether they belong to the same pedestrian.
与表观相似行人列表类似,在另一个实施例中,若判断到属于同一行人的行人三维运动特征,则可以建立空间相似行人列表。Similar to the list of apparently similar pedestrians, in another embodiment, if three-dimensional motion characteristics of pedestrians belonging to the same pedestrian are determined, a list of spatially similar pedestrians can be established.
步骤3.3、基于当前的行人三维运动特征、表观特征距离、空间特征距离,以及历史特征库中每个行人历史的行人三维运动特征判断是否符合同一行人的运动模式,输出运动模式匹配度作为该行人的运动模式匹配度。Step 3.3. Based on the current three-dimensional motion features of pedestrians, apparent feature distances, spatial feature distances, and the three-dimensional motion features of each pedestrian history in the historical feature database, determine whether it conforms to the motion pattern of the same pedestrian, and output the motion pattern matching degree as the Pedestrian motion pattern matching degree.
针对行人的运动模式,本实施例着重考虑行人在时间序列上的变化速度和空间位置上的移动逻辑,移动逻辑包括但不限于折返、原地停留、小跑、下蹲等常见行为。同时考虑到行人的常规运动,本实施例中重点考察目标在3秒内的变化速度和目标在拍摄空间内的移动逻辑。由于摄像机基于预设间隔获取光学图像,因此可以将速度变化合理的行人判断为同一行人。For the movement pattern of pedestrians, this embodiment focuses on the change speed of pedestrians in time series and the movement logic in spatial position. The movement logic includes but is not limited to common behaviors such as turning back, staying in place, trotting, and squatting. At the same time, considering the regular movement of pedestrians, in this embodiment, the change speed of the target within 3 seconds and the movement logic of the target in the shooting space are mainly investigated. Since the cameras acquire optical images based on preset intervals, pedestrians with reasonable speed changes can be judged as the same pedestrian.
将当前行人三维运动特征与历史特征库中的三维运动特征、表观特征距离、空间特征距离输入运动模式匹配方法,判断行人行为是否符合公共场所行人常见运动模式,若运动模式匹配度小于运动阈值则认为属于同一行人,建立运动模式相似行人列表。Input the current pedestrian 3D motion feature and the 3D motion feature, apparent feature distance, and spatial feature distance in the historical feature database into the motion pattern matching method to determine whether the pedestrian behavior conforms to the common motion patterns of pedestrians in public places. If the motion pattern matching degree is less than the motion threshold It is considered that they belong to the same pedestrian, and a list of pedestrians with similar movement patterns is established.
计算行人的运动模式匹配度可基于预训练的神经网络直接输出匹配度,也可以预计预设的匹配规则进行直接判断。前者判断相对灵活,但需要基于大量样本进行神经网络的训练,后者可以直接生成使用,并且便于添加、删除和修改,但是灵活度相对较低,可根据实际需求选择合适的方式。To calculate the matching degree of the pedestrian's motion pattern, the matching degree can be directly output based on the pre-trained neural network, or it can be directly judged by predicting the preset matching rules. The former is relatively flexible in judgment, but needs to be trained based on a large number of samples. The latter can be directly generated and used, and is easy to add, delete and modify, but the flexibility is relatively low, and an appropriate method can be selected according to actual needs.
在一个实施例中,根据实际观察及统计,建立了匹配规则(具体概率值略)如表1所示,该表格表示上一阶段的行为模式向当前阶段对应行为模式转换的概率。In one embodiment, based on actual observations and statistics, a matching rule (specific probability values omitted) is established as shown in Table 1, which represents the probability of converting the behavior pattern of the previous stage to the corresponding behavior pattern of the current stage.
表1上一阶段的行为模式向当前阶段对应行为模式转换的概率Table 1 Probability of transition from the behavior pattern of the previous stage to the corresponding behavior pattern of the current stage
Figure PCTCN2020137803-appb-000003
Figure PCTCN2020137803-appb-000003
基于表1进行运动模式匹配度时,取表观特征距离所对应的行人的历史行人三维运动特征,取空间特征距离所对应的行人的历史行人三维运动特征,若所取的两个历史行人三维运动特征所对应的行人不是同一个行人,则放弃本次匹配,若是同一个行人,则根据取到的历史行人三维运动特征和当前的行人三维运动特征判断该行人上一阶段的行为模式和当前阶段的行为模式,即可查表得到概率值作为运动模式匹配度。When the motion pattern matching degree is performed based on Table 1, the historical three-dimensional motion features of pedestrians corresponding to the apparent feature distances are taken, and the three-dimensional historical three-dimensional motion features of pedestrians corresponding to the spatial feature distances are taken. If the pedestrian corresponding to the motion feature is not the same pedestrian, this match will be discarded. If it is the same pedestrian, the behavior pattern and current behavior pattern of the pedestrian in the previous stage will be judged according to the obtained historical pedestrian 3D motion feature and the current pedestrian 3D motion feature. The behavior pattern of the stage can be looked up in the table to obtain the probability value as the motion pattern matching degree.
需要说明的是,一个阶段的行为模式至少由两个行人三维运动特征进行确定,由于一个行人三维运动特征带有方向矢量、运动速度和坐标,因此通过两个方向矢量的变化可以确定当前阶段为前进、折返或转弯,并结合坐标变化可以进一步区分是前进还是停留,结合运动速度进一步区分为前进还是加速前进。It should be noted that the behavior pattern of a stage is determined by at least two pedestrian three-dimensional motion features. Since a pedestrian three-dimensional motion feature has a direction vector, motion speed and coordinates, the current stage can be determined by the change of the two direction vectors. Going forward, turning back or turning, combined with the coordinate change can further distinguish whether it is going forward or staying, and further distinguish whether it is going forward or accelerating forward combined with the movement speed.
当然上述表格为本实施例采用的一种优选的匹配规则,在实际使用时可进一步优化,例如细化转弯为左转弯或右转弯等,并且表格中的概率值也可以根据实际使用时统计的概率进行更新,以提高行人识别率。Of course, the above table is a preferred matching rule adopted in this embodiment, which can be further optimized in actual use, for example, the turning is refined into a left turn or a right turn, etc., and the probability values in the table can also be calculated according to the actual use. The probability is updated to improve the pedestrian recognition rate.
步骤3.4、将属于同一行人的表观特征距离、空间特征距离、运动模式匹配度进行加权计算(即输入至加权计算器中),得到当前三维包围盒内行人与历史特征库中行人的匹配结果,所述匹配结果包括匹配成功或匹配失败,匹配成功时还包括匹配得到的行人信息。Step 3.4. Perform weighted calculation on the apparent feature distance, spatial feature distance, and motion pattern matching degree belonging to the same pedestrian (ie, input it into the weighting calculator), and obtain the matching result between the pedestrian in the current three-dimensional bounding box and the pedestrian in the historical feature database. , the matching result includes matching success or matching failure, and when the matching is successful, it also includes the pedestrian information obtained by the matching.
本实施例中表观特征距离、空间特征距离、运动模式匹配度的权重分别为0.6、0.2、0.2,由于表观特征是区分不同行人最为直观的特征,因此本实施例中设置表观特征距离具有最高权重。当然在实际使用时,可以进行权重调整,例提升运动模式匹配度的权重,以避免因表观特征过于相似的两个人造成误判。In this embodiment, the weights of apparent feature distance, spatial feature distance, and motion pattern matching degree are 0.6, 0.2, and 0.2, respectively. Since apparent feature is the most intuitive feature to distinguish different pedestrians, the apparent feature distance is set in this embodiment. has the highest weight. Of course, in actual use, the weight can be adjusted, such as increasing the weight of the motion pattern matching degree, so as to avoid misjudgment caused by two people whose apparent characteristics are too similar.
最终匹配结果中的匹配失败表示当前行人的特征不存在历史记录,即该行人为新进入统计范围的行人;而匹配成功表示当前行人的特征存在历史记录, 因此输出匹配得到的行人信息,以关联同一行人的新特征和历史特征。行人信息可以是唯一标识符(例如ID值)、空间位置、时间等。The matching failure in the final matching result indicates that there is no historical record for the characteristics of the current pedestrian, that is, the pedestrian is a pedestrian who has newly entered the statistical range; and the matching success indicates that the current pedestrian's characteristics have historical records, so the matching pedestrian information is output to correlate. New and historical features of the same pedestrian. Pedestrian information can be a unique identifier (eg, ID value), spatial location, time, and the like.
步骤4、根据本次的匹配结果以及历史的匹配结果标记三维包围盒所对应行人的行人状态为初次匹配成功、丢失、丢失后重新匹配成功、连续匹配成功或走出摄像范围。Step 4: Mark the pedestrian status of the pedestrian corresponding to the three-dimensional bounding box according to the current matching results and the historical matching results as initial matching successful, lost, successful re-matching after loss, successful continuous matching or out of the camera range.
如图6所示本实施例中提供一种具体的匹配方法可以是:As shown in FIG. 6, a specific matching method provided in this embodiment may be:
若成功提取行人特征(即当前时刻识别成功),但匹配结果为匹配失败(即不存在历史记录),则标记当前行人的状态为初次匹配成功;If the pedestrian feature is successfully extracted (that is, the recognition is successful at the current moment), but the matching result is a matching failure (that is, there is no historical record), the status of the current pedestrian is marked as the initial matching success;
若连续M次(例如连续50次(10秒*5次/秒))未匹配到历史匹配结果中的同一行人(例如当前识别失败且上一时刻也识别失败,或者当前识别失败且连续识别失败的次数不大于阈值),则标记该行人的状态为丢失;If M consecutive times (for example, 50 consecutive times (10 seconds*5 times/sec)) do not match the same pedestrian in the historical matching results (for example, the current recognition fails and the previous moment also fails, or the current recognition fails and the continuous recognition fails The number of times is not greater than the threshold), the status of the pedestrian is marked as lost;
若被标记为丢失的行人本次匹配结果中重新匹配成功(例如当前时刻识别成功,存在历史记录,但上一时刻匹配失败),则更新该行人的状态为丢失后重新匹配成功;If the pedestrian marked as lost is successfully re-matched in the current matching result (for example, the current moment is successfully identified and there is a historical record, but the previous match failed), then the status of the pedestrian is updated to be lost and the re-match is successful;
若连续L次(例如连续50次(10秒*5次/秒))匹配到历史匹配结果中的同一行人(例如当前时刻识别成功,存在历史记录,且上一时刻也是匹配成功),则更新该行人的状态为连续匹配成功;If the same pedestrian in the historical matching result is matched for L consecutive times (for example, 50 consecutive times (10 seconds*5 times/second)) (for example, the current moment is successfully recognized, there is a historical record, and the previous moment was also successfully matched), update The status of the pedestrian is continuous matching success;
若连续N次(例如连续150次(10秒*15次/秒))未匹配到历史匹配结果中的同一行人(例如当前识别失败且连续识别失败的次数大于阈值),则标记该行人的状态为走出摄像范围,且M<N。If the same pedestrian in the historical matching results is not matched for N consecutive times (for example, 150 consecutive times (10 seconds*15 times/sec)) (for example, the current recognition fails and the number of consecutive recognition failures is greater than the threshold), mark the status of the pedestrian is out of the camera range, and M<N.
若当前行人的状态标记为初次匹配成功,则在历史特征库中为该行人分配新的行人信息,并将该行人的行人特征与新分配的行人信息关联,在下一时刻即可作为历史数据用于对该行人进行识别跟踪。If the status of the current pedestrian is marked as successful for the first time, assign new pedestrian information to the pedestrian in the historical feature database, and associate the pedestrian characteristics of the pedestrian with the newly allocated pedestrian information, which can be used as historical data at the next moment. to identify and track the pedestrian.
如图7所示,在另一个实施中,提供一种人流统计系统,包括:As shown in Figure 7, in another implementation, a people flow statistics system is provided, comprising:
行人检测模块,用于获取光学图像,检测光学图像中的行人,输出行人的三维包围盒及对应的时间戳;The pedestrian detection module is used to obtain optical images, detect pedestrians in the optical images, and output the three-dimensional bounding boxes of pedestrians and corresponding timestamps;
特征提取模块,用于基于光学图像和三维包围盒获取行人特征,具体执行以下步骤:The feature extraction module is used to obtain pedestrian features based on optical images and 3D bounding boxes. The specific steps are as follows:
a、提取光学图像中行人的人体形状和特征作为每个行人的行人表观特征,并保存至历史特征库中;a. Extract the human body shape and features of the pedestrians in the optical image as the pedestrian appearance features of each pedestrian, and save them in the historical feature database;
b、基于行人当前的三维包围盒以及历史特征库中按照时间序列分布的三维 包围盒,提取每个行人的行人三维运动特征,并保存至历史特征库中;b. Based on the current 3D bounding box of the pedestrian and the 3D bounding box distributed according to the time series in the historical feature library, extract the three-dimensional motion feature of each pedestrian and save it in the historical feature library;
c、基于行人三维运动特征及历史特征库中指定时间内的行人三维运动特征预测下一时刻的行人三维运动特征,并保存至历史特征库中;c. Predict the 3D motion feature of the pedestrian at the next moment based on the 3D motion feature of the pedestrian and the 3D motion feature of the pedestrian within the specified time in the historical feature library, and save it in the historical feature library;
行人识别模块,用于基于历史特征库中的行人特征进行行人识别,具体执行以下步骤:The pedestrian recognition module is used for pedestrian recognition based on the pedestrian features in the historical feature database, and the specific steps are as follows:
a、基于当前的行人表观特征和历史特征库中每个行人历史的行人表观特征逐一计算表观特征距离,若表观特征距离大于表观阈值则判断当前的行人表观特征与历史特征库中的行人表观特征属于同一行人,确定当前的表观特征距离作为该行人的表观特征距离;a. Calculate the apparent feature distance one by one based on the current pedestrian apparent feature and the historical feature of each pedestrian in the historical feature database. If the apparent feature distance is greater than the apparent threshold, then judge the current pedestrian apparent feature and historical feature The pedestrian apparent features in the library belong to the same pedestrian, and the current apparent feature distance is determined as the pedestrian's apparent feature distance;
b、基于当前的行人三维运动特征以及历史特征库中每个行人的上一时刻预测得到的下一时刻的行人三维运动特征逐一计算空间特征距离,若空间特征距离大于空间阈值则判断当前的行人三维运动特征与历史特征库中的上一时刻预测得到的下一时刻的行人三维运动特征属于同一行人,确定当前的空间特征距离作为该行人的空间特征距离;b. Calculate the spatial feature distances one by one based on the current three-dimensional motion features of pedestrians and the three-dimensional motion features of pedestrians at the next moment predicted from the previous moment of each pedestrian in the historical feature database. If the spatial feature distance is greater than the spatial threshold, the current pedestrian is judged The three-dimensional motion feature and the three-dimensional motion feature of the pedestrian at the next moment predicted at the previous moment in the historical feature library belong to the same pedestrian, and the current spatial feature distance is determined as the spatial feature distance of the pedestrian;
c、基于当前的行人三维运动特征、表观特征距离、空间特征距离,以及历史特征库中每个行人历史的行人三维运动特征判断是否符合同一行人的运动模式,输出运动模式匹配度作为该行人的运动模式匹配度;c. Judging whether it conforms to the motion pattern of the same pedestrian based on the current three-dimensional motion features, apparent feature distance, spatial feature distance, and the three-dimensional motion features of each pedestrian in the historical feature database, and output the motion pattern matching degree as the pedestrian Motion pattern matching degree;
d、将属于同一行人的表观特征距离、空间特征距离、运动模式匹配度进行加权计算,得到当前三维包围盒内行人与历史特征库中行人的匹配结果,所述匹配结果包括匹配成功或匹配失败,匹配成功时还包括匹配得到的行人信息;d. Perform weighted calculation on the apparent feature distance, spatial feature distance, and motion pattern matching degree belonging to the same pedestrian, and obtain the matching result between the pedestrian in the current three-dimensional bounding box and the pedestrian in the historical feature database, and the matching result includes matching success or matching. If the match fails, the pedestrian information obtained by the match is also included when the match is successful;
行人标记模块,用于根据本次的匹配结果以及历史的匹配结果标记三维包围盒所对应行人的行人状态为初次匹配成功、丢失、丢失后重新匹配成功、连续匹配成功或走出摄像范围;The pedestrian marking module is used to mark the pedestrian status of the pedestrian corresponding to the three-dimensional bounding box according to the current matching results and the historical matching results as initial matching success, loss, rematching success after loss, continuous matching success or out of the camera range;
人流统计模块,用于根据行人状态统计预设时间内光学图像所对应的统计范围内的人流量。The people flow statistics module is used to count the flow of people within the statistical range corresponding to the optical image within the preset time according to the pedestrian state.
关于人流统计系统中的具体限定,参见上述对于公共场所行人识别方法的具体限定,在此不再进行赘述。For the specific limitations in the people flow statistics system, please refer to the above-mentioned specific limitations on the pedestrian identification method in public places, which will not be repeated here.
在一个优选实施例中,所述检测光学图像中的行人,输出行人的三维包围盒,执行如下操作:In a preferred embodiment, the detection of pedestrians in the optical image, outputting a three-dimensional bounding box of the pedestrians, performs the following operations:
对用于获取光学图像的摄像头进行标定,得到光学图像中像素与摄像头距离的映射关系;The camera used to obtain the optical image is calibrated, and the mapping relationship between the pixel in the optical image and the distance of the camera is obtained;
检测光学图像中的行人,获取光学图像中行人的二维包围框;Detect pedestrians in optical images, and obtain two-dimensional bounding boxes of pedestrians in optical images;
基于二维包围框和映射关系得到行人的三维包围盒。Based on the 2D bounding box and the mapping relationship, the 3D bounding box of the pedestrian is obtained.
本实施例中行人检测模块带有摄像头标定功能以及参数管理功能,在其他实施例中,摄像头标定装置可作为独立于本实施例的人流统计系统之外的部分,将标定后的设备外参以及映射关系发送至本实施例的人流统计系统的参数管理模块即可。In this embodiment, the pedestrian detection module has a camera calibration function and a parameter management function. In other embodiments, the camera calibration device can be used as a part independent of the people flow statistics system in this embodiment, and the calibrated external parameters of the equipment and The mapping relationship may be sent to the parameter management module of the people flow statistics system in this embodiment.
需要说明的是,本实施例基于光学图像进程人流统计,即人流统计系统还包括视频采集模块,该视频采集模块与外设的视频采集装置连接,获取统计范围内的实时视频后,将每一帧的光学图片发送至行人检测模块。It should be noted that this embodiment is based on optical image process people flow statistics, that is, the people flow statistics system further includes a video acquisition module, the video acquisition module is connected with the video acquisition device of the peripheral equipment, and after acquiring the real-time video within the statistical range, each The optical picture of the frame is sent to the pedestrian detection module.
在另一个实施例中,所述基于行人当前的三维包围盒以及历史特征库中按照时间序列分布的三维包围盒,提取每个行人的行人三维运动特征,执行如下操作:In another embodiment, the three-dimensional motion feature of each pedestrian is extracted based on the current three-dimensional bounding box of the pedestrian and the three-dimensional bounding box distributed according to the time series in the historical feature library, and the following operations are performed:
方向矢量提取:通过当前三维包围盒及历史的三维包围盒,提取行人在水平方向的运动方向及垂直方向的运动方向;Direction vector extraction: Through the current 3D bounding box and the historical 3D bounding box, the pedestrian's movement direction in the horizontal direction and the movement direction in the vertical direction are extracted;
运动速度提取:通过当前三维包围盒及历史的三维包围盒,提取人在水平方向的运动速度及垂直方向的运动速度;Movement speed extraction: Extract the movement speed of the person in the horizontal direction and the movement speed in the vertical direction through the current 3D bounding box and the historical 3D bounding box;
相对位置提取:根据摄像头标定后得到的映射关系,基于当前三维包围盒及历史的三维包围盒输出行人在以摄像头为中心的三维坐标系中的坐标;Relative position extraction: According to the mapping relationship obtained after the camera is calibrated, the coordinates of the pedestrian in the three-dimensional coordinate system centered on the camera are output based on the current three-dimensional bounding box and the historical three-dimensional bounding box;
特征整合:将提取的方向矢量、运动速度和相对位置作为行人三维运动特征。Feature integration: The extracted direction vector, motion speed and relative position are used as pedestrian 3D motion features.
在另一个实施例中,所述根据本次的匹配结果以及历史的匹配结果标记三维包围盒所对应行人的行人状态为初次匹配成功、丢失、丢失后重新匹配成功、连续匹配成功或走出摄像范围,执行如下操作:In another embodiment, the pedestrian state of the pedestrian corresponding to the three-dimensional bounding box is marked according to the current matching result and the historical matching result as successful initial matching, lost, successful re-matching after loss, successful continuous matching, or out of camera range , do the following:
若成功提取行人特征,但匹配结果为匹配失败,则标记当前行人的状态为初次匹配成功;If the pedestrian feature is successfully extracted, but the matching result is that the matching fails, the status of the current pedestrian is marked as the initial matching success;
若连续M次未匹配到历史匹配结果中的同一行人,则标记该行人的状态为丢失;If the same pedestrian in the historical matching results is not matched for M consecutive times, the status of the pedestrian is marked as lost;
若被标记为丢失的行人本次匹配结果中重新匹配成功,则更新该行人的状态为丢失后重新匹配成功;If the pedestrian marked as lost is successfully re-matched in the current matching result, the status of the pedestrian is updated to be lost and the re-match is successful;
若连续L次匹配到历史匹配结果中的同一行人,则更新该行人的状态为连续匹配成功;If the same pedestrian in the historical matching result is matched for L consecutive times, the status of the pedestrian is updated to indicate that the continuous matching is successful;
若连续N次未匹配到历史匹配结果中的同一行人,则标记该行人的状态为走出摄像范围,且M<N。If the same pedestrian in the historical matching result is not matched for N consecutive times, the state of the pedestrian is marked as walking out of the camera range, and M<N.
在另一个实施例中,若当前行人的状态标记为初次匹配成功,则在历史特征库中为该行人分配新的行人信息,并将该行人的行人特征与新分配的行人信息关联。In another embodiment, if the status of the current pedestrian is marked as successful for the first match, new pedestrian information is allocated to the pedestrian in the historical feature database, and the pedestrian characteristic of the pedestrian is associated with the newly allocated pedestrian information.
以上所述实施例的各技术特征可以进行任意的组合,为使描述简洁,未对上述实施例中的各技术特征所有可能的组合都进行描述,然而,只要这些技术特征的组合不存在矛盾,都应当认为是本说明书记载的范围。The technical features of the above-described embodiments can be combined arbitrarily. For the sake of brevity, all possible combinations of the technical features in the above-described embodiments are not described. However, as long as there is no contradiction between the combinations of these technical features, All should be regarded as the scope described in this specification.
以上所述实施例仅表达了本发明的几种实施方式,其描述较为具体和详细,但并不能因此而理解为对发明专利范围的限制。应当指出的是,对于本领域的普通技术人员来说,在不脱离本发明构思的前提下,还可以做出若干变形和改进,这些都属于本发明的保护范围。因此,本发明专利的保护范围应以所附权利要求为准。The above-mentioned embodiments only represent several embodiments of the present invention, and the descriptions thereof are more specific and detailed, but should not be construed as a limitation on the scope of the invention patent. It should be pointed out that for those of ordinary skill in the art, without departing from the concept of the present invention, several modifications and improvements can also be made, which all belong to the protection scope of the present invention. Therefore, the protection scope of the patent of the present invention should be subject to the appended claims.

Claims (10)

  1. 一种公共场所行人识别方法,其特征在于,所述公共场所行人识别方法,包括:A method for pedestrian identification in a public place, characterized in that the method for pedestrian identification in a public place comprises:
    步骤1、获取光学图像,检测光学图像中的行人,输出行人的三维包围盒及对应的时间戳;Step 1. Obtain an optical image, detect pedestrians in the optical image, and output the three-dimensional bounding box of the pedestrian and the corresponding timestamp;
    步骤2、基于光学图像和三维包围盒获取行人特征,包括:Step 2. Obtain pedestrian features based on optical images and 3D bounding boxes, including:
    步骤2.1、提取光学图像中行人的人体形状和特征作为每个行人的行人表观特征,并保存至历史特征库中;Step 2.1, extract the human body shape and features of the pedestrians in the optical image as the pedestrian appearance features of each pedestrian, and save them in the historical feature database;
    步骤2.2、基于行人当前的三维包围盒以及历史特征库中按照时间序列分布的三维包围盒,提取每个行人的行人三维运动特征,并保存至历史特征库中;Step 2.2, based on the current three-dimensional bounding box of the pedestrian and the three-dimensional bounding box distributed according to the time series in the historical feature library, extract the three-dimensional motion feature of each pedestrian, and save it in the historical feature library;
    步骤2.3、基于行人三维运动特征及历史特征库中指定时间内的行人三维运动特征预测下一时刻的行人三维运动特征,并保存至历史特征库中;Step 2.3, based on the pedestrian 3D motion feature and the pedestrian 3D motion feature within the specified time in the historical feature library, predict the pedestrian 3D motion feature at the next moment, and save it in the historical feature library;
    步骤3、基于历史特征库中的行人特征进行行人识别,包括:Step 3. Perform pedestrian recognition based on the pedestrian features in the historical feature library, including:
    步骤3.1、基于当前的行人表观特征和历史特征库中每个行人历史的行人表观特征逐一计算表观特征距离,若表观特征距离大于表观阈值则判断当前的行人表观特征与历史特征库中的行人表观特征属于同一行人,确定当前的表观特征距离作为该行人的表观特征距离;Step 3.1. Calculate the apparent feature distance one by one based on the current pedestrian apparent features and the historical pedestrian apparent features of each pedestrian in the historical feature database. If the apparent feature distance is greater than the apparent threshold, determine the current pedestrian apparent feature and history The pedestrian apparent features in the feature library belong to the same pedestrian, and the current apparent feature distance is determined as the pedestrian's apparent feature distance;
    步骤3.2、基于当前的行人三维运动特征以及历史特征库中每个行人的上一时刻预测得到的下一时刻的行人三维运动特征逐一计算空间特征距离,若空间特征距离大于空间阈值则判断当前的行人三维运动特征与历史特征库中的上一时刻预测得到的下一时刻的行人三维运动特征属于同一行人,确定当前的空间特征距离作为该行人的空间特征距离;Step 3.2. Calculate the spatial feature distances one by one based on the current three-dimensional motion features of pedestrians and the three-dimensional motion features of pedestrians at the next moment predicted from the previous moment of each pedestrian in the historical feature database. If the spatial feature distance is greater than the spatial threshold, determine the current The three-dimensional motion feature of the pedestrian and the three-dimensional motion feature of the pedestrian at the next moment predicted in the historical feature database belong to the same pedestrian, and the current spatial feature distance is determined as the spatial feature distance of the pedestrian;
    步骤3.3、基于当前的行人三维运动特征、表观特征距离、空间特征距离,以及历史特征库中每个行人历史的行人三维运动特征判断是否符合同一行人的运动模式,输出运动模式匹配度作为该行人的运动模式匹配度;Step 3.3. Based on the current three-dimensional motion features of pedestrians, apparent feature distances, spatial feature distances, and the three-dimensional motion features of each pedestrian history in the historical feature database, determine whether it conforms to the motion pattern of the same pedestrian, and output the motion pattern matching degree as the Pedestrian motion pattern matching degree;
    步骤3.4、将属于同一行人的表观特征距离、空间特征距离、运动模式匹配度进行加权计算,得到当前三维包围盒内行人与历史特征库中行人的匹配结果,所述匹配结果包括匹配成功或匹配失败,匹配成功时还包括匹配得到的行人信息;Step 3.4: Perform weighted calculation on the apparent feature distance, spatial feature distance, and motion pattern matching degree belonging to the same pedestrian, and obtain the matching result between the pedestrian in the current three-dimensional bounding box and the pedestrian in the historical feature database, and the matching result includes a successful matching or a matching result. If the matching fails, the pedestrian information obtained by the matching is also included when the matching is successful;
    步骤4、根据本次的匹配结果以及历史的匹配结果标记三维包围盒所对应行人的行人状态为初次匹配成功、丢失、丢失后重新匹配成功、连续匹配成功或 走出摄像范围。Step 4. Mark the pedestrian status of the pedestrian corresponding to the three-dimensional bounding box according to the current matching results and the historical matching results as initial matching success, loss, re-matching success after loss, continuous matching success or out of the camera range.
  2. 如权利要求1所述的公共场所行人识别方法,其特征在于,所述检测光学图像中的行人,输出行人的三维包围盒,包括:The method for pedestrian identification in a public place according to claim 1, wherein the detecting a pedestrian in an optical image and outputting a three-dimensional bounding box of the pedestrian comprises:
    对用于获取光学图像的摄像头进行标定,得到光学图像中像素与摄像头距离的映射关系;The camera used to obtain the optical image is calibrated, and the mapping relationship between the pixel in the optical image and the distance of the camera is obtained;
    检测光学图像中的行人,获取光学图像中行人的二维包围框;Detect pedestrians in optical images, and obtain two-dimensional bounding boxes of pedestrians in optical images;
    基于二维包围框和映射关系得到行人的三维包围盒。Based on the 2D bounding box and the mapping relationship, the 3D bounding box of the pedestrian is obtained.
  3. 如权利要求2所述的公共场所行人识别方法,其特征在于,所述基于行人当前的三维包围盒以及历史特征库中按照时间序列分布的三维包围盒,提取每个行人的行人三维运动特征,包括:The method for pedestrian identification in a public place according to claim 2, wherein the three-dimensional motion feature of each pedestrian is extracted based on the current three-dimensional bounding box of the pedestrian and the three-dimensional bounding box distributed according to time series in the historical feature library, include:
    步骤2.2.1、方向矢量提取:通过当前三维包围盒及历史的三维包围盒,提取行人在水平方向的运动方向及垂直方向的运动方向;Step 2.2.1. Extraction of direction vector: extract the movement direction of pedestrians in the horizontal direction and the movement direction in the vertical direction through the current 3D bounding box and the historical 3D bounding box;
    步骤2.2.2、运动速度提取:通过当前三维包围盒及历史的三维包围盒,提取人在水平方向的运动速度及垂直方向的运动速度;Step 2.2.2. Movement speed extraction: extract the movement speed of the person in the horizontal direction and the movement speed in the vertical direction through the current 3D bounding box and the historical 3D bounding box;
    步骤2.2.3、相对位置提取:根据摄像头标定后得到的映射关系,基于当前三维包围盒及历史的三维包围盒输出行人在以摄像头为中心的三维坐标系中的坐标;Step 2.2.3. Relative position extraction: According to the mapping relationship obtained after the camera is calibrated, the coordinates of the pedestrian in the three-dimensional coordinate system centered on the camera are output based on the current three-dimensional bounding box and the historical three-dimensional bounding box;
    步骤2.2.4、将步骤2.2.1~2.2.3中提取的方向矢量、运动速度和相对位置作为行人三维运动特征。Step 2.2.4. Use the direction vector, motion speed and relative position extracted in steps 2.2.1 to 2.2.3 as the three-dimensional motion feature of the pedestrian.
  4. 如权利要求1所述的公共场所行人识别方法,其特征在于,所述根据本次的匹配结果以及历史的匹配结果标记三维包围盒所对应行人的行人状态为初次匹配成功、丢失、丢失后重新匹配成功、连续匹配成功或走出摄像范围,包括:The method for pedestrian identification in a public place according to claim 1, wherein the pedestrian state of the pedestrian corresponding to the three-dimensional bounding box is marked according to the current matching result and the historical matching results as successful initial matching, loss, and re-setting after loss. Successful match, continuous match, or out of camera range, including:
    若成功提取行人特征,但匹配结果为匹配失败,则标记当前行人的状态为初次匹配成功;If the pedestrian feature is successfully extracted, but the matching result is that the matching fails, the status of the current pedestrian is marked as the initial matching success;
    若连续M次未匹配到历史匹配结果中的同一行人,则标记该行人的状态为丢失;If the same pedestrian in the historical matching results is not matched for M consecutive times, the status of the pedestrian is marked as lost;
    若被标记为丢失的行人本次匹配结果中重新匹配成功,则更新该行人的状态为丢失后重新匹配成功;If the pedestrian marked as lost is successfully re-matched in the current matching result, the status of the pedestrian is updated to be lost and the re-match is successful;
    若连续L次匹配到历史匹配结果中的同一行人,则更新该行人的状态为连续匹配成功;If the same pedestrian in the historical matching result is matched for L consecutive times, the status of the pedestrian is updated to indicate that the continuous matching is successful;
    若连续N次未匹配到历史匹配结果中的同一行人,则标记该行人的状态为走出摄像范围,且M<N。If the same pedestrian in the historical matching result is not matched for N consecutive times, the state of the pedestrian is marked as walking out of the camera range, and M<N.
  5. 如权利要求1所述的公共场所行人识别方法,其特征在于,若当前行人的状态标记为初次匹配成功,则在历史特征库中为该行人分配新的行人信息,并将该行人的行人特征与新分配的行人信息关联。The pedestrian identification method in a public place according to claim 1, wherein if the status of the current pedestrian is marked as successful for the first time, new pedestrian information is allocated to the pedestrian in the historical feature database, and the pedestrian feature of the pedestrian is assigned to the pedestrian. Associated with the newly assigned pedestrian information.
  6. 一种人流统计系统,其特征在于,所述人流统计系统,包括:A people flow statistics system, characterized in that the people flow statistics system includes:
    行人检测模块,用于获取光学图像,检测光学图像中的行人,输出行人的三维包围盒及对应的时间戳;The pedestrian detection module is used to obtain optical images, detect pedestrians in the optical images, and output the three-dimensional bounding boxes of pedestrians and corresponding timestamps;
    特征提取模块,用于基于光学图像和三维包围盒获取行人特征,具体执行以下步骤:The feature extraction module is used to obtain pedestrian features based on optical images and 3D bounding boxes. The specific steps are as follows:
    a、提取光学图像中行人的人体形状和特征作为每个行人的行人表观特征,并保存至历史特征库中;a. Extract the human body shape and features of the pedestrians in the optical image as the pedestrian appearance features of each pedestrian, and save them in the historical feature database;
    b、基于行人当前的三维包围盒以及历史特征库中按照时间序列分布的三维包围盒,提取每个行人的行人三维运动特征,并保存至历史特征库中;b. Based on the current 3D bounding box of the pedestrian and the 3D bounding box distributed according to the time series in the historical feature library, extract the three-dimensional motion feature of each pedestrian and save it in the historical feature library;
    c、基于行人三维运动特征及历史特征库中指定时间内的行人三维运动特征预测下一时刻的行人三维运动特征,并保存至历史特征库中;c. Predict the 3D motion feature of the pedestrian at the next moment based on the 3D motion feature of the pedestrian and the 3D motion feature of the pedestrian within the specified time in the historical feature library, and save it in the historical feature library;
    行人识别模块,用于基于历史特征库中的行人特征进行行人识别,具体执行以下步骤:The pedestrian recognition module is used for pedestrian recognition based on the pedestrian features in the historical feature database, and the specific steps are as follows:
    a、基于当前的行人表观特征和历史特征库中每个行人历史的行人表观特征逐一计算表观特征距离,若表观特征距离大于表观阈值则判断当前的行人表观特征与历史特征库中的行人表观特征属于同一行人,确定当前的表观特征距离作为该行人的表观特征距离;a. Calculate the apparent feature distance one by one based on the current pedestrian apparent feature and the historical feature of each pedestrian in the historical feature database. If the apparent feature distance is greater than the apparent threshold, then judge the current pedestrian apparent feature and historical feature The pedestrian apparent features in the library belong to the same pedestrian, and the current apparent feature distance is determined as the pedestrian's apparent feature distance;
    b、基于当前的行人三维运动特征以及历史特征库中每个行人的上一时刻预测得到的下一时刻的行人三维运动特征逐一计算空间特征距离,若空间特征距离大于空间阈值则判断当前的行人三维运动特征与历史特征库中的上一时刻预测得到的下一时刻的行人三维运动特征属于同一行人,确定当前的空间特征距离作为该行人的空间特征距离;b. Calculate the spatial feature distances one by one based on the current three-dimensional motion features of pedestrians and the three-dimensional motion features of pedestrians at the next moment predicted from the previous moment of each pedestrian in the historical feature database. If the spatial feature distance is greater than the spatial threshold, the current pedestrian is judged The three-dimensional motion feature and the three-dimensional motion feature of the pedestrian at the next moment predicted at the previous moment in the historical feature library belong to the same pedestrian, and the current spatial feature distance is determined as the spatial feature distance of the pedestrian;
    c、基于当前的行人三维运动特征、表观特征距离、空间特征距离,以及历史特征库中每个行人历史的行人三维运动特征判断是否符合同一行人的运动模式,输出运动模式匹配度作为该行人的运动模式匹配度;c. Judging whether it conforms to the motion pattern of the same pedestrian based on the current three-dimensional motion features, apparent feature distance, spatial feature distance, and the three-dimensional motion features of each pedestrian in the historical feature database, and output the motion pattern matching degree as the pedestrian Motion pattern matching degree;
    d、将属于同一行人的表观特征距离、空间特征距离、运动模式匹配度进行 加权计算,得到当前三维包围盒内行人与历史特征库中行人的匹配结果,所述匹配结果包括匹配成功或匹配失败,匹配成功时还包括匹配得到的行人信息;d. Perform weighted calculation on the apparent feature distance, spatial feature distance, and motion pattern matching degree belonging to the same pedestrian, and obtain the matching result between the pedestrian in the current three-dimensional bounding box and the pedestrian in the historical feature database, and the matching result includes matching success or matching. If the match fails, the pedestrian information obtained by the match is also included when the match is successful;
    行人标记模块,用于根据本次的匹配结果以及历史的匹配结果标记三维包围盒所对应行人的行人状态为初次匹配成功、丢失、丢失后重新匹配成功、连续匹配成功或走出摄像范围;The pedestrian marking module is used to mark the pedestrian status of the pedestrian corresponding to the three-dimensional bounding box according to the current matching results and the historical matching results as initial matching success, loss, rematching success after loss, continuous matching success or out of the camera range;
    人流统计模块,用于根据行人状态统计预设时间内光学图像所对应的统计范围内的人流量。The people flow statistics module is used to count the flow of people within the statistical range corresponding to the optical image within the preset time according to the pedestrian state.
  7. 如权利要求6所述的人流统计系统,其特征在于,所述检测光学图像中的行人,输出行人的三维包围盒,执行如下操作:The people flow statistics system according to claim 6, characterized in that, the pedestrian in the optical image is detected, the three-dimensional bounding box of the pedestrian is output, and the following operations are performed:
    对用于获取光学图像的摄像头进行标定,得到光学图像中像素与摄像头距离的映射关系;The camera used to obtain the optical image is calibrated, and the mapping relationship between the pixel in the optical image and the distance of the camera is obtained;
    检测光学图像中的行人,获取光学图像中行人的二维包围框;Detect pedestrians in optical images, and obtain two-dimensional bounding boxes of pedestrians in optical images;
    基于二维包围框和映射关系得到行人的三维包围盒。Based on the 2D bounding box and the mapping relationship, the 3D bounding box of the pedestrian is obtained.
  8. 如权利要求7所述的人流统计系统,其特征在于,所述基于行人当前的三维包围盒以及历史特征库中按照时间序列分布的三维包围盒,提取每个行人的行人三维运动特征,执行如下操作:The people flow statistics system according to claim 7, wherein the three-dimensional motion feature of each pedestrian is extracted based on the current three-dimensional bounding box of the pedestrian and the three-dimensional bounding box distributed according to the time series in the historical feature library, and the execution is as follows operate:
    方向矢量提取:通过当前三维包围盒及历史的三维包围盒,提取行人在水平方向的运动方向及垂直方向的运动方向;Direction vector extraction: Through the current 3D bounding box and the historical 3D bounding box, the pedestrian's movement direction in the horizontal direction and the movement direction in the vertical direction are extracted;
    运动速度提取:通过当前三维包围盒及历史的三维包围盒,提取人在水平方向的运动速度及垂直方向的运动速度;Movement speed extraction: Extract the movement speed of the person in the horizontal direction and the movement speed in the vertical direction through the current 3D bounding box and the historical 3D bounding box;
    相对位置提取:根据摄像头标定后得到的映射关系,基于当前三维包围盒及历史的三维包围盒输出行人在以摄像头为中心的三维坐标系中的坐标;Relative position extraction: According to the mapping relationship obtained after the camera is calibrated, the coordinates of the pedestrian in the three-dimensional coordinate system centered on the camera are output based on the current three-dimensional bounding box and the historical three-dimensional bounding box;
    特征整合:将提取的方向矢量、运动速度和相对位置作为行人三维运动特征。Feature integration: The extracted direction vector, motion speed and relative position are used as pedestrian 3D motion features.
  9. 如权利要求6所述的人流统计系统,其特征在于,所述根据本次的匹配结果以及历史的匹配结果标记三维包围盒所对应行人的行人状态为初次匹配成功、丢失、丢失后重新匹配成功、连续匹配成功或走出摄像范围,执行如下操作:The people flow statistics system according to claim 6, wherein the pedestrian state of the pedestrian corresponding to the three-dimensional bounding box is marked according to the current matching result and the historical matching result as successful initial matching, lost, and successful re-matching after loss , continuous matching successfully or out of the camera range, perform the following operations:
    若成功提取行人特征,但匹配结果为匹配失败,则标记当前行人的状态为初次匹配成功;If the pedestrian feature is successfully extracted, but the matching result is that the matching fails, the status of the current pedestrian is marked as the initial matching success;
    若连续M次未匹配到历史匹配结果中的同一行人,则标记该行人的状态为 丢失;If the same pedestrian in the historical matching result is not matched for M consecutive times, the state of the pedestrian is marked as lost;
    若被标记为丢失的行人本次匹配结果中重新匹配成功,则更新该行人的状态为丢失后重新匹配成功;If the pedestrian marked as lost is successfully re-matched in the current matching result, the status of the pedestrian is updated to be lost and the re-match is successful;
    若连续L次匹配到历史匹配结果中的同一行人,则更新该行人的状态为连续匹配成功;If the same pedestrian in the historical matching result is matched for L consecutive times, the status of the pedestrian is updated to indicate that the continuous matching is successful;
    若连续N次未匹配到历史匹配结果中的同一行人,则标记该行人的状态为走出摄像范围,且M<N。If the same pedestrian in the historical matching result is not matched for N consecutive times, the state of the pedestrian is marked as walking out of the camera range, and M<N.
  10. 如权利要求6所述的人流统计系统,其特征在于,若当前行人的状态标记为初次匹配成功,则在历史特征库中为该行人分配新的行人信息,并将该行人的行人特征与新分配的行人信息关联。The people flow statistics system according to claim 6, wherein if the current state of the pedestrian is marked as successful for the first time, new pedestrian information is allocated to the pedestrian in the historical feature database, and the pedestrian feature of the pedestrian is compared with the new pedestrian information. Assigned pedestrian information association.
PCT/CN2020/137803 2020-12-15 2020-12-19 Method for pedestrian identification in public places and human flow statistics system WO2022126668A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN202011477711.X 2020-12-15
CN202011477711.XA CN112446355B (en) 2020-12-15 2020-12-15 Pedestrian recognition method and people stream statistics system in public place

Publications (1)

Publication Number Publication Date
WO2022126668A1 true WO2022126668A1 (en) 2022-06-23

Family

ID=74739432

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2020/137803 WO2022126668A1 (en) 2020-12-15 2020-12-19 Method for pedestrian identification in public places and human flow statistics system

Country Status (2)

Country Link
CN (1) CN112446355B (en)
WO (1) WO2022126668A1 (en)

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113011329B (en) * 2021-03-19 2024-03-12 陕西科技大学 Multi-scale feature pyramid network-based and dense crowd counting method

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106650562A (en) * 2016-06-14 2017-05-10 西安电子科技大学 Online continuous human behavior identification method based on Kinect
CN109829476A (en) * 2018-12-27 2019-05-31 青岛中科慧畅信息科技有限公司 End-to-end three-dimension object detection method based on YOLO
CN111709974A (en) * 2020-06-22 2020-09-25 苏宁云计算有限公司 Human body tracking method and device based on RGB-D image
US10839203B1 (en) * 2016-12-27 2020-11-17 Amazon Technologies, Inc. Recognizing and tracking poses using digital imagery captured from multiple fields of view
CN111968235A (en) * 2020-07-08 2020-11-20 杭州易现先进科技有限公司 Object attitude estimation method, device and system and computer equipment

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110490901A (en) * 2019-07-15 2019-11-22 武汉大学 The pedestrian detection tracking of anti-attitudes vibration

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106650562A (en) * 2016-06-14 2017-05-10 西安电子科技大学 Online continuous human behavior identification method based on Kinect
US10839203B1 (en) * 2016-12-27 2020-11-17 Amazon Technologies, Inc. Recognizing and tracking poses using digital imagery captured from multiple fields of view
CN109829476A (en) * 2018-12-27 2019-05-31 青岛中科慧畅信息科技有限公司 End-to-end three-dimension object detection method based on YOLO
CN111709974A (en) * 2020-06-22 2020-09-25 苏宁云计算有限公司 Human body tracking method and device based on RGB-D image
CN111968235A (en) * 2020-07-08 2020-11-20 杭州易现先进科技有限公司 Object attitude estimation method, device and system and computer equipment

Also Published As

Publication number Publication date
CN112446355B (en) 2023-10-17
CN112446355A (en) 2021-03-05

Similar Documents

Publication Publication Date Title
US10417503B2 (en) Image processing apparatus and image processing method
US10909695B2 (en) System and process for detecting, tracking and counting human objects of interest
US9646212B2 (en) Methods, devices and systems for detecting objects in a video
CN108416250B (en) People counting method and device
LU102028B1 (en) Multiple view multiple target tracking method and system based on distributed camera network
CN109344690B (en) People counting method based on depth camera
EP2345999A1 (en) Method for automatic detection and tracking of multiple objects
CN110874583A (en) Passenger flow statistics method and device, storage medium and electronic equipment
WO2022227761A1 (en) Target tracking method and apparatus, electronic device, and storage medium
JP6503079B2 (en) Specific person detection system, specific person detection method and detection device
US20220366570A1 (en) Object tracking device and object tracking method
CN111666860A (en) Vehicle track tracking method integrating license plate information and vehicle characteristics
CN112613668A (en) Scenic spot dangerous area management and control method based on artificial intelligence
WO2022078134A1 (en) People traffic analysis method and system, electronic device, and readable storage medium
WO2022126668A1 (en) Method for pedestrian identification in public places and human flow statistics system
CN111260696A (en) Method for edge-end-oriented pedestrian tracking and accurate people counting
CN113536891B (en) Pedestrian traffic statistical method, storage medium and electronic equipment
CN113763418B (en) Multi-target tracking method based on head and shoulder detection
CN115588149A (en) Cross-camera multi-target cascade matching method based on matching priority
JP5472976B2 (en) Object detection device, object detection method, program, and recording medium
CN115620379B (en) Mall person searching method based on computer vision
Virgona et al. Socially constrained tracking in crowded environments using shoulder pose estimates
WO2023184197A1 (en) Target tracking method and apparatus, system, and storage medium
CN117237879B (en) Track tracking method and system
Mei et al. Human localization and identification by combining visual detection and UWB

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 20965681

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 20965681

Country of ref document: EP

Kind code of ref document: A1

122 Ep: pct application non-entry in european phase

Ref document number: 20965681

Country of ref document: EP

Kind code of ref document: A1