CN111260696B - Method for edge-end-oriented pedestrian tracking and accurate people counting - Google Patents

Method for edge-end-oriented pedestrian tracking and accurate people counting Download PDF

Info

Publication number
CN111260696B
CN111260696B CN202010068506.1A CN202010068506A CN111260696B CN 111260696 B CN111260696 B CN 111260696B CN 202010068506 A CN202010068506 A CN 202010068506A CN 111260696 B CN111260696 B CN 111260696B
Authority
CN
China
Prior art keywords
pedestrian
pedestrians
tracking
list
counting
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202010068506.1A
Other languages
Chinese (zh)
Other versions
CN111260696A (en
Inventor
黄樟钦
盛梦雪
张硕
李洪亮
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing University of Technology
Original Assignee
Beijing University of Technology
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing University of Technology filed Critical Beijing University of Technology
Priority to CN202010068506.1A priority Critical patent/CN111260696B/en
Publication of CN111260696A publication Critical patent/CN111260696A/en
Priority to PCT/CN2021/072538 priority patent/WO2021147826A1/en
Application granted granted Critical
Publication of CN111260696B publication Critical patent/CN111260696B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/20Analysis of motion
    • G06T7/277Analysis of motion involving stochastic approaches, e.g. using Kalman filters
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/22Matching criteria, e.g. proximity measures
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/20Analysis of motion
    • G06T7/246Analysis of motion using feature-based methods, e.g. the tracking of corners or segments
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/60Analysis of geometric attributes
    • G06T7/62Analysis of geometric attributes of area, perimeter, diameter or volume
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10016Video; Image sequence
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30196Human being; Person
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30242Counting objects in image

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • General Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • Multimedia (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Geometry (AREA)
  • Artificial Intelligence (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Evolutionary Biology (AREA)
  • Evolutionary Computation (AREA)
  • General Engineering & Computer Science (AREA)
  • Image Analysis (AREA)
  • Traffic Control Systems (AREA)

Abstract

An edge-end-oriented pedestrian tracking and people number accurate counting method relates to the field of pedestrian tracking and people number counting based on artificial intelligence components in an edge calculation scene. The method initially establishes and maintains a pedestrian dynamic tracking list in real time. And acquiring pedestrian result information from the artificial intelligence component at the edge end, and establishing a pedestrian static detection list on the basis of the pedestrian result information. And accurately matching the pedestrian static detection list with the pedestrian dynamic tracking list. Cost and performance limitations at the edge end make the edge end computing power limited, leading to failure of pedestrian exact matching. At the moment, the method carries out intelligent fuzzy matching on the pedestrian according to the dynamic tracking information of the pedestrian. And updating a pedestrian dynamic tracking list in real time according to the matched result, accurately counting the number of the pedestrians in the counting area, and if the accurate counting fails, using a delay counting strategy for the pedestrians by the method.

Description

Method for edge-end-oriented pedestrian tracking and accurate people counting
Technical Field
The invention relates to the field of pedestrian tracking and people counting based on artificial intelligence components in an edge calculation scene.
Background
With the general improvement of safety consciousness of people, monitoring cameras are distributed all over streets and alleys, and the integration of technologies such as image processing, machine vision and the like enables a computer to replace human brain to analyze information in videos, judge the conditions in the videos and enable monitoring to be intelligent. In recent years, edge computing has emerged, and embedded systems become the most important processing platform of edge computing due to the characteristics of low power consumption and single function, so that intelligent monitoring technology can be widely applied to actual life.
The efficiency of the embedded special chip is far higher than that of a PC (personal computer), the cost and the power consumption are greatly lower than those of a CPU (central processing unit), and a plurality of scholars design the special chip for embedded image processing and design methods for pedestrian tracking and people counting in a targeted manner. The pedestrian tracking strategy comprises the steps of carrying out similarity matching by utilizing the characteristic information of color, posture and the like of pedestrians analyzed by a special chip or judging the overlapping rate of front and rear frame detection frames. People counting is mainly triggered by crossing detection lines.
Artificial intelligence technology is now rapidly falling to the ground, and is also expanding towards the edge and terminal areas. In recent years, each large designer specially designs an artificial intelligence component for an artificial intelligence algorithm, however, the requirement of the artificial intelligence component on the computing capacity of a deployment platform is high, the processing capacity of an edge computing platform may limit the processing performance of the artificial intelligence component, and meanwhile, each large designer also tends to reduce the processing performance of the artificial intelligence component to meet the balance of cost and selling price in practical application.
Disclosure of Invention
In view of the above, the objective of the algorithm is to design a general pedestrian tracking and people counting method based on pedestrian result information obtained by intelligently analyzing a real-time video and related information thereof by an artificial intelligence processing component, so that when the algorithm is applied to an edge computing scene, the pedestrian tracking and people counting can still be accurately performed under the resource limitations of insufficient processing capability of the artificial intelligence processing component, limited computing capability of an edge end processing platform, product cost pricing and the like.
A method for pedestrian tracking and people number accurate statistics facing to an edge end is characterized by comprising the following steps:
step S1: and establishing and maintaining a pedestrian dynamic tracking list in real time for dynamically recording pedestrian information in a tracking area.
Step S2: and acquiring pedestrian result information of the edge artificial intelligence processing component.
And step S3: and extracting important pedestrian result information to establish a pedestrian static detection list.
And step S4: and matching the pedestrians in the static pedestrian detection list and the dynamic pedestrian tracking list.
Step S5: and updating the pedestrian dynamic tracking list according to the matching result.
Step S6: and accurately counting the number of the pedestrians in the counting area according to the dynamic pedestrian tracking list.
The information recorded in the pedestrian dynamic tracking list in the step S1 should include the tracking id, direction, moving radius, and frame loss information of the pedestrian.
The information recorded in the static pedestrian detection list in step S3 should include the tracking id, position coordinate, and time information of the pedestrian.
The algorithm is further characterized by: the resource limitation of the edge end, including the self processing capacity limitation of the artificial intelligent processing part, the insufficient computing capacity of the edge end processing platform, the product cost limitation and the like, can cause the intelligent analysis process of the real-time video to have the frame loss problem, thereby causing the deviation of the pedestrian result information. Therefore, the matching process in the step S4 is firstly carried out with accurate matching, and if the accurate matching fails, the pedestrian with the failed accurate matching is carried out with intelligent fuzzy matching.
Step S41: and (4) precise matching.
And judging whether the pedestrians in the two lists are the same person or not according to the tracking id in the static detection list and the dynamic tracking list of the pedestrians, if so, judging that the pedestrians in the two lists are the same person, and if not, judging that the pedestrians in the two lists are the same person, so that the accurate matching is successful, otherwise, the accurate matching is failed.
Step S42: and carrying out intelligent fuzzy matching on the pedestrians with the accurate matching failure.
For the pedestrians which do not have the frame loss problem and have definite directions, the position range which is possibly generated in the next frame is predicted in the motion direction of the pedestrians, and the pedestrians which have the largest similarity with the prediction result and exceed the matching threshold MATCH _ THRESHOID in the static detection list are selected to be matched with the pedestrians. If MATCH _ THRESHOID is percentage, then MATCH _ THRESHOID is more than or equal to 0 and less than or equal to 100 percent, and the matching condition is more severe as the MATCH _ THRESHOID approaches 100 percent, and generally the matching condition is set to be more than or equal to 80 percent.
And for pedestrians which do not have the frame loss problem but have ambiguous directions, bidirectionally predicting the position range of the pedestrians which may appear in the next frame, and selecting the pedestrians which have the largest similarity with the prediction result and exceed MATCH _ THRESHOID in the static detection list to be matched with the pedestrians. After matching is successful, the direction of the pedestrian can be judged according to the moving condition of the position coordinate.
For the pedestrian with frame loss problem, calculating the number n of the frame loss, and expanding the prediction range by a multiple of alpha x n, wherein alpha is an expansion coefficient, and the matching range is a part of a circle with a moving radius as a radius under the general condition, and alpha = pi n 2 . When n is less than or equal to LOST _ THRESHOID, selecting static detection list and predictionThe pedestrian with the largest similarity and exceeding MATCH _ threshold is matched with the result. LOST _ THRESHOID is a frame loss threshold, and LOST _ THRESHOID =3 can be generally set.
And for the pedestrians with changed directions, the pedestrians with changed directions can not be successfully matched through the steps, the position range of the pedestrians with changed directions possibly appearing in the next frame is reversely predicted, and the pedestrians with the maximum similarity to the prediction result and exceeding MATCH _ THRESHOID in the static detection list are selected to be matched with the pedestrians. If the matching is successful, recording the direction of the sudden change.
The step S5 of updating the pedestrian dynamic tracking list includes:
and recording frame loss information when the pedestrians which are not successfully matched in the dynamic tracking list are not matched.
And (4) marking the pedestrians which are not successfully matched in the static detection list as new pedestrians, judging the direction and adding the new pedestrians into the dynamic tracking list.
And (4) walking out of the tracking area or deleting the pedestrians with the frame loss number larger than LOST _ THRESHOID, and deleting the dynamic tracking list.
The algorithm is further characterized by: the direction is used as an intelligent fuzzy matching index, so that the accuracy of people counting is improved better. And judging the direction of the pedestrian in the static detection list according to the initial position coordinate when the pedestrian is judged to be a new person. If the initial position coordinates of the pedestrian are located at the upper part of the tracking area, the initial position coordinates are generally regarded as the upper 1/3 part of the tracking area, and the direction is judged to be entering. If the initial position coordinate of the pedestrian is located at the lower part of the tracking area, the initial position coordinate is generally regarded as the lower 1/3 part of the tracking area, and the direction is judged to be out. If the initial position coordinates of the pedestrian are located in other parts of the tracking area, it is considered that the direction cannot be judged, and the direction can be confirmed with delay according to the processing method of the pedestrian whose frame loss problem does not occur but the direction is unknown in step S42, so as to make a correct judgment on the direction.
The accurate people counting process in the step S6 comprises the following steps:
step S61: and establishing and maintaining a pedestrian dynamic statistical column, and using the table for accurate statistics of the number of people.
Step S62: the accurate statistics of the number of the pedestrians in the statistical area is specifically as follows:
if the pedestrian direction is known and enters the statistical area for the first time, the total number of people and the total number of people in the direction are recorded.
If the direction of the pedestrian is unknown and the pedestrian enters the counting area for the first time, only the total number of people is recorded, and an intelligent delay counting strategy is adopted for the total number of people in the direction.
If the direction of the pedestrians changes suddenly and enters the statistical area for the first time, the total number of people is increased, and the counting with the direction is not carried out.
Step S63: when the pedestrian is counted normally and leaves the statistical area or is in the statistical area but the number of dropped frames is greater than LOST _ THRESHOID, it is deleted from the pedestrian dynamic statistical list.
The accurate counting of the number of the pedestrians cannot be carried out when the dynamic tracking information of the pedestrians is ambiguous, and the algorithm carries out intelligent delay counting processing on the pedestrians, and specifically comprises the following steps:
if the pedestrian enters the statistical area but the direction is not determined, counting is not carried out until the direction is determined through the pedestrian tracking process.
If the pedestrian is not counted normally but leaves the statistical area, the pedestrian is not deleted until the direction is determined through the pedestrian tracking process and the pedestrian is counted normally.
If the pedestrian is not counted normally but the number of frame loss is larger than LOST _ THRESHOID, the pedestrian is deleted. Considering the actual situation of the set statistical region and LOST _ THRESHOID, if the pedestrian is considered not to be detected again in the statistical region, the pedestrian is counted in the total number, otherwise, the pedestrian is ignored.
In the algorithm, except for an intelligent delay counting strategy, counting is only carried out when a pedestrian enters a statistical area for the first time, and when the pedestrian stays in the statistical area, the counting is not repeated although the pedestrian stays in a dynamic statistical list of the pedestrian.
The algorithm of claim 1, further characterized by: a method for delineating a range of positions where a pedestrian is likely to be present in a next frame. The scribing process is detailed as follows:
one of the most rough methods is to define the prediction range within a full circle with the center of mass of the pedestrian as the center and the radius of movement as the radius.
Further, the method uses the direction as an important matching index, and the limitation of the direction helps us to reduce the prediction range from a full circle to a semicircle in the moving direction. The semicircle is specifically: taking the center of mass of the pedestrian as an end point, making a ray parallel to the moving direction, dividing a whole circle into semicircles by a straight line which passes through the end point and is perpendicular to the ray, and taking the semicircle containing the ray.
Furthermore, the time difference between the two video frames before and after being analyzed intelligently is always within 1 second, and the moving direction of the pedestrian in 1 second can not be changed greatly, specifically, the angle of the pedestrian to the left or right in the moving direction can not be greater than 45 degrees, so that the prediction range is reduced to the sector of the semi-circle with the radius and the included angle of the ray being less than 45 degrees and > -45 degrees.
The algorithm has the advantages that:
(1) The method is universal for various detection models such as face detection, pedestrian detection, head and shoulder detection and the like, the accurate matching conditions can be changed according to actual conditions, unidirectional or bidirectional people counting can be carried out, and the method has good expandability and universality.
(2) The intelligent fuzzy matching method has the advantages that the problem that the accuracy of people counting is reduced due to errors of pedestrian result information caused by the problems of limited processing capacity of an artificial intelligent processing part, insufficient computing capacity of an edge end processing platform, product cost pricing and other resource limitations is considered, and the intelligent fuzzy matching is added on the basis of accurate matching, so that the pedestrian tracking and the people counting are more accurate.
Description of the drawings:
FIG. 1 Algorithm flow chart
FIG. 2 is a diagram illustrating a rough prediction range
FIG. 3 is a schematic view of the prediction range of the tape direction
FIG. 4 is a schematic diagram of the prediction range with angles
FIG. 5 is a diagram of prediction range with matching distance coefficients
FIG. 6 is a diagram illustrating bi-directional match prediction range
FIG. 7 is a diagram illustrating the prediction range of reverse matching
The specific implementation mode is as follows:
the technical solutions of the present invention will be clearly and completely described below with reference to the embodiments of the present invention, and the examples are only for explaining the present invention and do not limit the application of the present invention.
In this embodiment, the pedestrian result information provided by the artificial intelligence component to us includes the tracking id, the position coordinates, and the current frame number Seq of the pedestrian curr Face feature vectors, etc. The performance of the artificial intelligence component is good, the processing rate is 12 frames per second, but the preprocessing capability of the edge end processing platform for video frames is limited, and only 3 frames per second can be achieved, so that the analysis is performed at intervals of fixed frame number frames (which are called as detection frames ofps), in the embodiment, frame =7, the same video frames are used as input data of the artificial intelligence component among groups, which inevitably causes the processing performance of the artificial intelligence component to be reduced, and the pedestrian tracking id is inaccurate.
The invention provides a method for edge-end-oriented pedestrian tracking and accurate people number counting, which comprises the following specific steps of:
step S1: establishing and maintaining a pedestrian dynamic tracking list and a pedestrian dynamic statistical list in real time, wherein the pedestrian dynamic tracking list is used for dynamically recording pedestrian information in a tracking area, and adding pedestrians into the pedestrian dynamic statistical list when the pedestrians enter the statistical area, and the two lists comprise the following 4 characteristics:
1) Tracking id: included in the pedestrian result information provided by the artificial intelligence component.
2) Moving direction d: if the initial position coordinates of the pedestrian are located in the upper 1/3 part of the tracking area, we consider the pedestrian as the entering direction, and make it carry the entering direction attribute, note d = in. If the initial position coordinates of the pedestrian are located in the lower 1/3 part of the tracking area, we consider the pedestrian as an outgoing direction, and make it carry out the directional attribute, note d = out. If the initial position coordinates of the pedestrian are located in other parts of the tracking area, the direction cannot be judged, the direction attribute is not carried, and the direction is confirmed in a delayed mode.
3) Moving radius r: defined as the pedestrian starting from entering the tracking area until the presentThe average distance moved by each detection frame. Let the initial centroid coordinate of the pedestrian entering the tracking area be (x) orgi ,y orgi ) The coordinates of the centroid of the current frame are (x) curr ,y curr ) The initial frame number is Seq orgi The current frame number is Seq curr Then, then
Figure 1
4) Frame number lost Seq lost : seq when no frame is lost lost = -1, if the pedestrian is not successfully matched with any pedestrian in the pedestrian static detection list after one round of algorithm, recording frame loss, seq lost =Seq curr
Step S2: and acquiring pedestrian result information of the current video frame from the artificial intelligence component, establishing a pedestrian static detection list on the basis, and matching pedestrians in the static detection list with pedestrians in the dynamic tracking list.
The basic principle of the algorithm matching is as follows: firstly, carrying out accurate matching, namely, if the tracking id is the same, the accurate matching is successful, otherwise, according to an intelligent fuzzy matching method, if the pedestrian direction in the dynamic tracking list is known, selecting the pedestrian which is positioned in the moving direction of the dynamic tracking list, is in the prediction range and is closest to the dynamic tracking list to be matched with the pedestrian.
The prediction horizon of the algorithm we make the following analysis:
one of the coarsest approaches may be to define the prediction range as being the pedestrian's centroid (x) curr ,y curr ) As the center of the circle, the circle with the radius of movement r as the radius, as shown in fig. 2. The basic matching principle of the algorithm can help us to narrow the prediction range from fig. 2 to fig. 3 to be within a semicircle in the moving direction by adding the limit of the moving direction d. However, the algorithm can also perform bidirectional people counting, so that the situation that two pedestrians walking in the same direction get closer before passing by the shoulder (as shown by two pedestrians id2 and id1 in fig. 4) to cause matching error may occur. Therefore, it is not enough to reduce the prediction range to fig. 3, and we think that two pedestrians will avoid the collision as much as possible, so it is considered that the pedestrians are confronted with each otherWhen the walking is too close, the walking and the walking are mutually avoided to generate an angle as shown in the graph 4, the time difference between the two video frames before and after being analyzed by the intelligent analysis is certainly within 1 second, the moving direction of the pedestrian does not change greatly within 1 second, and particularly, the angle of the pedestrian deviating to the left or the right in the moving direction is not larger than 45 degrees, so that the prediction range is further reduced to the graph 4, namely, the radius is in a sector with the included angle of less than 45 degrees and > -45 degrees with the ray. In conclusion, the algorithm considers that the part outside the shadow of FIG. 4 does not belong to the prediction range (so in FIG. 4, we consider id3 to be a better intelligent fuzzy matching object with id 1).
The algorithm uses a matching table for analysis, with the transverse representing pedestrians in a dynamic tracking list and the longitudinal representing pedestrians in a static detection list. The values that may appear in the table and their meanings are given in the following table:
numerical value Means of
-2 Out of prediction range
-1 Has been matched
0 Tracking id identity
distance (other value) Relative distance between two pedestrian centroids
The form filling process is as follows:
step S21: if the two tracking ids are the same, the artificial intelligence component analyzes to obtain that the two persons are the same person, the accurate matching is finished, and 0 is filled in
Step S22: if the pedestrian in the static detection list is not within the predicted range of the pedestrian in the dynamic tracking list, fill-2.
Step S23: otherwise, distance fill is calculated.
And step S3: sequentially extracting minimum distances other than-2 from the table min If the following relationships are satisfied simultaneously:
1) The pedestrians of the corresponding dynamic tracking list and the pedestrians of the static detection list are not matched.
2) If MIN _ COEFFICIENT is the minimum COEFFICIENT of the matching distance and MAX _ COEFFICIENT is the maximum COEFFICIENT of the matching distance, MIN _ COEFFICIENT multiplied by r is not more than distance min MAX _ COEFFIIENT x r, the prediction horizon is further refined in this step, as shown in FIG. 5. The MIN _ COEFFICENT and MAX _ COEFFICENT can be selected to be the optimal values within the range of 0-1 according to actual conditions. If MATCH _ threshold =80%, MIN _ coeffient =0.9 and MAX _ coeffient =1.1 may be set.
We consider the two pedestrians as successfully matched and set the correspondence to-1, indicating that the two pedestrians have been matched.
Because the frame loss problem caused by the performance reduction of the artificial intelligence component influences the accuracy of pedestrian tracking and accurate people counting, the intelligent fuzzy matching is carried out on the pedestrians which are not successfully matched after the three steps. Specifically, as shown in step S4:
the delay confirmation strategy of the moving direction d in the step S1 is specifically:
the pedestrian with uncertain direction in the dynamic tracking list is subjected to two-way range matching, namely, the prediction range is expanded from the graph 5 to the graph 6, the pedestrian with the nearest distance in the range is selected to be matched with the pedestrian, and after the matching is successful, the moving direction of the pedestrian can be confirmed according to the moving direction of the center of mass of the pedestrian.
Frame loss compensation strategy:
by calculating we canObtaining the ofps of a plurality of detection frames lost by the pedestrian according to the calculation formula
Figure DEST_PATH_IMAGE002
Accordingly we also enlarge the matching radius of movement r by a factor of n. Namely, MINCEOEFFICIENT × r × n < distance min < MAX _ COEFFICIENT × r × n. The pedestrian is deleted from the pedestrian dynamic tracking list when n > LOST _ THRESHOID. If framestap =7, lost _threshold =3.
Sudden change of the pedestrian direction:
if the pedestrians who are not successfully matched still exist after the judgment, the pedestrians are considered whether the moving direction is changed or not, so that the pedestrians are subjected to reverse matching as shown in the figure 7, and if the matching is successful, the pedestrians are marked to have sudden changes of the direction. When the number of people is accurately counted, the total number of people is calculated, but the direction-carrying counting is not carried out.
Step S5: and updating the pedestrian dynamic tracking list according to the accurate matching and intelligent fuzzy matching results.
The pedestrians which are not successfully matched in the dynamic tracking list are recorded with Seq lost =Seq curr
And (4) recording the pedestrians which are not successfully matched in the static detection list as new pedestrians, judging the moving direction d, and adding the new pedestrians into the dynamic tracking list.
And (4) walking out of the tracking area or deleting the pedestrians with the frame loss number larger than LOST _ THRESHOID, and deleting the dynamic tracking list.
Step S6: the accurate statistics of the number of the pedestrians in the statistical area is specifically as follows:
step S61: if the pedestrians enter the statistical area for the first time and d = in, the total number of people is increased, and the number of people in the direction of the in is increased.
If the pedestrians enter the statistical area for the first time and d = out, the number of the pedestrians in the total area is increased, and the number of the pedestrians in the out direction is increased.
If the pedestrians enter the statistical area for the first time and the direction is unknown, only the total number of people is recorded, and an intelligent delay counting strategy is adopted for the total number of people in the direction.
If the pedestrians enter the statistical area for the first time and carry the direction mutation attribute, the total number of the pedestrians is increased, and the pedestrians are not counted in the direction.
Step S62: when the pedestrian is counted normally and leaves the statistical area or the number of frames LOST in the statistical area is greater than LOST _ THRESHOID, the pedestrian is deleted from the dynamic statistical list.
The intelligent delay counting strategy in step S61 is specifically:
if the pedestrian enters the statistical area but the direction is not determined, counting is not carried out until the direction is determined through the pedestrian tracking process.
If the pedestrians do not count normally but walk out of the statistical area, the pedestrians are not deleted until the direction is determined and the pedestrians count normally in the tracking process.
If the pedestrian is not counted normally but the number of frame loss is larger than LOST _ THRESHOID, the pedestrian is deleted. Considering the actual situation of the statistical region and LOST _ THRESHOID set, if the pedestrian is not detected again in the statistical region, the pedestrian is counted in the total number, otherwise, the pedestrian is ignored.
In the algorithm, except for a delay counting strategy, counting is only carried out when a pedestrian enters a statistical area for the first time, and when the pedestrian stays in a dynamic statistical area, although the pedestrian stays in a statistical list, the counting is not repeated.

Claims (5)

1. A method for tracking pedestrians facing to an edge end and accurately counting the number of the pedestrians is characterized by comprising the following steps of:
step S1: establishing and maintaining a pedestrian dynamic tracking list in real time for dynamically recording pedestrian information in a tracking area;
step S2: acquiring pedestrian result information of an edge artificial intelligence processing component;
and step S3: extracting important pedestrian result information and establishing a pedestrian static detection list;
and step S4: matching the pedestrians in the static pedestrian detection list and the dynamic pedestrian tracking list;
step S5: updating a pedestrian dynamic tracking list according to the matching result;
step S6: accurately counting the number of pedestrians in the counting area according to the dynamic pedestrian tracking list;
the information recorded by the pedestrian dynamic tracking list in the step S1 comprises the tracking id, direction, moving radius and frame loss information of the pedestrian;
in the step S3, the information recorded by the pedestrian static detection list comprises the tracking id, the position coordinate and the time information of the pedestrian;
therefore, the matching process in the step S4 is firstly carried out with accurate matching, and if the accurate matching fails, the pedestrian with the failed accurate matching is carried out with intelligent fuzzy matching;
step S41: precise matching;
whether the pedestrians in the two lists are the same person can be judged according to the tracking id in the static detection list and the dynamic tracking list of the pedestrians, if the tracking id is the same person, the pedestrians are the same person, the accurate matching is successful, and if not, the accurate matching is failed;
step S42: carrying out intelligent fuzzy matching on the pedestrians with the accurate matching failure;
for the pedestrians which do not have the frame loss problem and have definite directions, predicting the position range which possibly appears in the next frame in the motion direction of the pedestrians, and selecting the pedestrians which have the largest similarity with the prediction result and exceed the matching threshold MATCH _ THRESHOID in the static detection list to be matched with the pedestrians; if MATCH _ THRESHOID is percentage, then MATCH _ THRESHOID is more than or equal to 0 and less than or equal to 100 percent, and MATCH _ THRESHOID is set to be more than or equal to 80 percent;
for pedestrians who do not have the frame loss problem but have ambiguous directions, performing bidirectional prediction on the possible position range of the pedestrians in the next frame, and selecting the pedestrian which has the largest similarity with the prediction result and exceeds MATCH _ THRESHOID in the static detection list to MATCH with the pedestrian; judging the direction of the pedestrian according to the moving condition of the position coordinate of the pedestrian after successful matching;
for the pedestrian with frame loss problem, calculating the number n of the frame loss, expanding the prediction range by alpha multiplied by n, wherein alpha is an expansion coefficient, the matching range is a part of a circle with the moving radius as the radius, and alpha = pi n 2 (ii) a When n is less than or equal to LOST _ THRESHOID, selecting the pedestrian with the maximum similarity to the prediction result and exceeding MATCH _ THRESHOID in the static detection list to be matched with the pedestrian; LOST _ THRESHOID is a frame loss threshold, and LOST _ THRESHOID =3;
for the pedestrians with changed directions, the pedestrians with changed directions can not be successfully matched through the steps, the position range of the pedestrians with changed directions possibly appearing in the next frame is reversely predicted, and the pedestrians with the largest similarity to the prediction result and exceeding MATCH _ THRESHOID in the static detection list are selected to be matched with the pedestrians; if the matching is successful, recording the direction of the terminal to generate mutation;
the step S5 of updating the pedestrian dynamic tracking list includes:
if the pedestrian is not successfully matched in the dynamic tracking list, recording frame loss information;
the pedestrians which are not successfully matched in the static detection list are marked as new people, the direction is judged, and a dynamic tracking list is added;
and (4) walking out of the tracking area or deleting the pedestrians with the frame loss number larger than LOST _ THRESHOID, and deleting the dynamic tracking list.
2. The method of claim 1, wherein: judging the direction of the pedestrian in the static detection list according to the initial position coordinate when the pedestrian is judged to be a new person; if the initial position coordinate of the pedestrian is positioned at the upper part of the tracking area, the initial position coordinate of the pedestrian is regarded as the upper 1/3 part of the tracking area, and the direction is judged to be entering; if the initial position coordinate of the pedestrian is positioned at the lower part of the tracking area, the initial position coordinate of the pedestrian is regarded as the lower 1/3 part of the tracking area, and the direction is judged to be out; if the initial position coordinates of the pedestrian are located in other parts of the tracking area, it is considered that the direction cannot be judged, and the direction can be confirmed with delay according to the processing method of the pedestrian whose frame loss problem does not occur but the direction is unknown in step S42, so as to make a correct judgment on the direction.
3. The method of claim 1, wherein: the accurate people counting process in the step S6 comprises the following steps:
step S61: establishing and maintaining a pedestrian dynamic statistical column, wherein a table is used for accurately counting the number of people;
step S62: the accurate statistics of the number of the pedestrians in the statistical area is specifically as follows:
if the direction of the pedestrians is known and the pedestrians enter the statistical area for the first time, recording the total number of the pedestrians and the total number of the pedestrians in the direction;
if the direction of the pedestrian is unknown and the pedestrian enters the counting area for the first time, only the total number of people is recorded, and an intelligent delay counting strategy is adopted for the total number of people in the direction;
if the pedestrian direction changes suddenly and enters the statistical area for the first time, the total number of people is increased, and the counting with the direction is not carried out;
step S63: when the pedestrian is counted normally and leaves the statistical area or is in the statistical area but the number of dropped frames is greater than LOST _ THRESHOID, it is deleted from the pedestrian dynamic statistical list.
4. The method of claim 1, wherein: the accurate statistics of the number of the pedestrians cannot be carried out when the dynamic tracking information of the pedestrians is not clear, and the algorithm carries out intelligent delay counting processing on the pedestrians, and specifically comprises the following steps:
if the pedestrian enters the statistical area but the direction is not determined, counting is not carried out until the direction is determined through the pedestrian tracking process;
if the pedestrian is not counted normally but leaves the statistical area, the pedestrian is not deleted until the direction is determined through the pedestrian tracking process and the pedestrian is counted normally;
if the pedestrian is not counted normally but the frame loss number is larger than LOST _ THRESHOID, deleting the pedestrian; considering the set statistical region and the actual situation of LOST _ THRESHOID, if the pedestrian is not detected again in the statistical region, the pedestrian is counted into the total number, otherwise, the pedestrian is ignored;
in the algorithm, except for an intelligent delay counting strategy, counting is only carried out when a pedestrian enters a statistical area for the first time, and when the pedestrian stays in the statistical area, the counting is not repeated although the pedestrian stays in a dynamic statistical list of the pedestrian.
5. The method of claim 1, wherein: the method for defining the position range of the pedestrian possibly appearing in the next frame specifically comprises the following steps:
defining the prediction range in a whole circle with the center of mass of the pedestrian as the center of a circle and the moving radius as the radius;
the direction limitation reduces the prediction range from a full circle to a semicircle in the moving direction of the prediction range; the semicircle is specifically: taking the center of mass of the pedestrian as an end point, making rays parallel to the moving direction, dividing a whole circle into semicircles by a straight line which passes through the end point and is perpendicular to the rays, and taking the semicircle containing the rays;
the time difference between the two video frames is always within 1 second, and the angle of the moving direction to the left or right is not more than 45 degrees, so the prediction range is reduced to the sector of which the radius of the semicircle and the ray form an angle of <45 degrees and > -45 degrees.
CN202010068506.1A 2020-01-21 2020-01-21 Method for edge-end-oriented pedestrian tracking and accurate people counting Active CN111260696B (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
CN202010068506.1A CN111260696B (en) 2020-01-21 2020-01-21 Method for edge-end-oriented pedestrian tracking and accurate people counting
PCT/CN2021/072538 WO2021147826A1 (en) 2020-01-21 2021-01-18 Edge end-oriented pedestrian tracking and accurate people counting method

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202010068506.1A CN111260696B (en) 2020-01-21 2020-01-21 Method for edge-end-oriented pedestrian tracking and accurate people counting

Publications (2)

Publication Number Publication Date
CN111260696A CN111260696A (en) 2020-06-09
CN111260696B true CN111260696B (en) 2023-04-18

Family

ID=70954344

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202010068506.1A Active CN111260696B (en) 2020-01-21 2020-01-21 Method for edge-end-oriented pedestrian tracking and accurate people counting

Country Status (2)

Country Link
CN (1) CN111260696B (en)
WO (1) WO2021147826A1 (en)

Families Citing this family (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111260696B (en) * 2020-01-21 2023-04-18 北京工业大学 Method for edge-end-oriented pedestrian tracking and accurate people counting
CN113361456B (en) * 2021-06-28 2024-05-07 北京影谱科技股份有限公司 Face recognition method and system
CN115546192B (en) * 2022-11-03 2023-03-21 中国平安财产保险股份有限公司 Livestock quantity identification method, device, equipment and storage medium

Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103049787A (en) * 2011-10-11 2013-04-17 汉王科技股份有限公司 People counting method and system based on head and shoulder features

Family Cites Families (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101231755B (en) * 2007-01-25 2013-03-06 上海遥薇(集团)有限公司 Moving target tracking and quantity statistics method
CN103854292B (en) * 2012-11-28 2018-02-23 中兴通讯股份有限公司 A kind of number and the computational methods and device in crowd movement direction
CN106778570B (en) * 2016-12-05 2018-08-24 清华大学深圳研究生院 A kind of pedestrian detection and tracking in real time
CN109754409A (en) * 2017-11-06 2019-05-14 北京航天长峰科技工业集团有限公司 A kind of monitor video pedestrian target matched jamming System and method for
CN109255802B (en) * 2018-08-21 2023-08-18 平安科技(深圳)有限公司 Pedestrian tracking method, device, computer equipment and storage medium
CN110443210B (en) * 2019-08-08 2021-11-26 北京百度网讯科技有限公司 Pedestrian tracking method and device and terminal
CN111260696B (en) * 2020-01-21 2023-04-18 北京工业大学 Method for edge-end-oriented pedestrian tracking and accurate people counting

Patent Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103049787A (en) * 2011-10-11 2013-04-17 汉王科技股份有限公司 People counting method and system based on head and shoulder features

Also Published As

Publication number Publication date
CN111260696A (en) 2020-06-09
WO2021147826A1 (en) 2021-07-29

Similar Documents

Publication Publication Date Title
CN111260696B (en) Method for edge-end-oriented pedestrian tracking and accurate people counting
CN107657226B (en) People number estimation method based on deep learning
Atev et al. Learning traffic patterns at intersections by spectral clustering of motion trajectories
CN111488804A (en) Labor insurance product wearing condition detection and identity identification method based on deep learning
CN108647649B (en) Method for detecting abnormal behaviors in video
CN107909044B (en) People counting method combining convolutional neural network and track prediction
CN107545256B (en) Camera network pedestrian re-identification method combining space-time and network consistency
CN101344922A (en) Human face detection method and device
CN113963445A (en) Pedestrian falling action recognition method and device based on attitude estimation
CN107315994B (en) Clustering method based on Spectral Clustering space trajectory
CN111126459A (en) Method and device for identifying fine granularity of vehicle
CN110232331B (en) Online face clustering method and system
CN113763427B (en) Multi-target tracking method based on coarse-to-fine shielding processing
CN107944628A (en) A kind of accumulation mode under road network environment finds method and system
CN106570471A (en) Scale adaptive multi-attitude face tracking method based on compressive tracking algorithm
CN115423735A (en) Passenger flow volume statistical method and system
Chen et al. A precise information extraction algorithm for lane lines
CN112347983B (en) Lane line detection processing method, lane line detection processing device, computer equipment and storage medium
CN106023262A (en) Crowd flowing main direction estimating method and device
Xie et al. A multi-object tracking system for surveillance video analysis
WO2020232697A1 (en) Online face clustering method and system
CN114742280B (en) Road condition prediction method and corresponding model training method, device, equipment and medium
Rituerto et al. Label propagation in videos indoors with an incremental non-parametric model update
CN114937060A (en) Monocular pedestrian indoor positioning prediction method guided by map meaning
WO2022126668A1 (en) Method for pedestrian identification in public places and human flow statistics system

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant