WO2019006633A1 - Fuzzy logic based video multi-target tracking method and device - Google Patents

Fuzzy logic based video multi-target tracking method and device Download PDF

Info

Publication number
WO2019006633A1
WO2019006633A1 PCT/CN2017/091575 CN2017091575W WO2019006633A1 WO 2019006633 A1 WO2019006633 A1 WO 2019006633A1 CN 2017091575 W CN2017091575 W CN 2017091575W WO 2019006633 A1 WO2019006633 A1 WO 2019006633A1
Authority
WO
WIPO (PCT)
Prior art keywords
result
prediction result
trajectory
observation
prediction
Prior art date
Application number
PCT/CN2017/091575
Other languages
French (fr)
Chinese (zh)
Inventor
李良群
湛西羊
罗升
刘宗香
谢维信
Original Assignee
深圳大学
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 深圳大学 filed Critical 深圳大学
Priority to PCT/CN2017/091575 priority Critical patent/WO2019006633A1/en
Publication of WO2019006633A1 publication Critical patent/WO2019006633A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/20Analysis of motion
    • G06T7/246Analysis of motion using feature-based methods, e.g. the tracking of corners or segments
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/20Analysis of motion
    • G06T7/254Analysis of motion involving subtraction of images
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/20Analysis of motion
    • G06T7/269Analysis of motion using gradient-based methods

Definitions

  • the present invention relates to the field of target tracking, and in particular to a video multi-target tracking method and apparatus based on fuzzy logic.
  • Video multi-target tracking technology is an important research branch in the field of computer vision. It is related to many frontier disciplines, such as image processing, pattern recognition, artificial intelligence, automatic control and computer integration. In intelligent video surveillance, human-computer interaction, robot vision Navigation, virtual reality, medical diagnosis, traffic control and surveillance have very important practical value.
  • the invention provides a video multi-target tracking method and device based on fuzzy logic, which can effectively improve the correct association between multi-objectives and observations, and accurately correct multiple targets under the conditions of apparent similarity, frequent interaction, occlusion and background interference. Tracking, while significantly reducing the number of target tag changes in multi-target tracking, has strong robustness and accuracy.
  • a technical solution adopted by the present invention is to provide a video multi-target tracking method based on fuzzy logic, which comprises: performing online target motion detection on a current video frame, and detecting a possible moving object as an observation result; Data correlation between the observation result and the prediction result of the target, wherein the prediction result is obtained by predicting at least the trajectory of the target of the previous video frame; and the prediction result and the observation on the unassociated Performing trajectory management, comprising: acquiring a termination trajectory segment by using the prediction result not associated with the prediction result, and acquiring a new trajectory segment by using the observation result that is not associated, the termination trajectory segment and the new The track segment is tracked.
  • a technical solution adopted by the present invention is to provide a device for video multi-target tracking based on fuzzy logic, comprising: a processor, configured to perform a current video frame acquired from the camera Online target motion detection, detecting the obtained possible moving object as an observation result; performing data association on the observation result and the prediction result of the target, wherein the prediction result is obtained by predicting at least the trajectory of the target of the previous video frame Performing trajectory management on the prediction result that is not associated with the observation result, including acquiring the termination trajectory segment by using the prediction result that is not associated with the prediction result, and acquiring the observation result by using the unrelated association New track a segment, performing trajectory association on the terminating trajectory segment and the new trajectory segment.
  • the invention has the beneficial effects of providing a video multi-target tracking method and device based on fuzzy logic, which performs data association by the observation result in the current video frame and the prediction result of the target, and the observation result and the prediction result on the unassociated Trajectory management can effectively improve the correct correlation between multi-objectives and observations, and accurately track multiple targets under the conditions of apparent similarity, frequent interaction, occlusion and background interference, which has strong robustness and accuracy.
  • FIG. 1 is a schematic flow chart of a first embodiment of a video multi-target tracking method based on fuzzy logic
  • FIG. 2 is a schematic flow chart of a second embodiment of a video multi-target tracking method based on fuzzy logic according to the present invention
  • FIG. 3 is a schematic diagram of occlusion between prediction results of different targets of the present invention.
  • FIG. 4 is a schematic flow chart of a third embodiment of a video multi-target tracking method based on fuzzy logic according to the present invention.
  • FIG. 5 is a schematic flowchart of an embodiment of step S233 in FIG. 4;
  • FIG. 6 is a schematic flow chart of a fourth embodiment of a video multi-target tracking method based on fuzzy logic according to the present invention.
  • step S23b in FIG. 6 is a schematic flow chart of an embodiment of step S23b in FIG. 6;
  • FIG. 8 is a schematic structural diagram of multi-feature clue fusion according to the present invention.
  • Figure 9 is a fuzzy input variable f k (i, j) of the present invention, Schematic diagram of the membership function
  • FIG. 10 is a schematic diagram of a membership function of the output fuzzy variable ⁇ M of the present invention.
  • FIG. 11 is a schematic flowchart diagram of a fifth embodiment of a video multi-target tracking method based on fuzzy logic according to the present invention.
  • step S31 in FIG. 11 is a schematic flow chart of an embodiment of step S31 in FIG. 11;
  • step S33 in FIG. 11 is a schematic flow chart of an embodiment of step S33 in FIG. 11;
  • 15 is a schematic diagram of a position of the present invention for acquiring a loss prediction point
  • 16 is a schematic structural diagram of a first embodiment of a video multi-target tracking apparatus based on fuzzy logic according to the present invention
  • 17 is a schematic structural diagram of a second embodiment of a video multi-target tracking apparatus based on fuzzy logic according to the present invention.
  • a schematic flowchart of a first embodiment of a video multi-target tracking method based on fuzzy logic the method includes the following steps:
  • S1 Perform online target motion detection on the current video frame, and detect possible motion objects as Observations.
  • the online target motion detection can use motion detection algorithms such as frame difference method, optical flow method, background subtraction method, and mixed Gaussian background model.
  • the invention mainly adopts a mixed Gaussian background model to perform motion detection on the current video frame to find pixels belonging to the foreground of the motion, supplemented by median filtering and simple morphological processing, and finally obtain possible moving objects in the current video frame.
  • Observed object An observation object is an image block in the current video frame. Generally, the shape of the observation object is a rectangle.
  • Targets include reliable targets for stable tracking and temporary targets for unstable tracking.
  • the target state in this step that is, whether each target is marked as a reliable target or a temporary target, is determined by the trajectory management of the previous video frame.
  • the temporary target includes a new target established by the observation that the previous video frame is a candidate result that is not associated and is not a successful match, and a target whose consecutively associated successful number of frames is less than or equal to the first frame number threshold and has not been deleted.
  • a reliable target includes a target whose number of consecutively successful frames is greater than the first frame number threshold and has not been deleted.
  • the prediction result of the target is obtained by predicting at least the trajectory of the target of the previous video frame.
  • the data association method in step S2 can deal with the data association problem of high-frequency occlusion occurring in a short period of time and multi-target tracking under a large number of false observation conditions, but in the case of long-term occlusion and missed detection, some
  • the target state is not updated for a long time, and the target motion trajectory is difficult to maintain, and the target trajectory is broken, that is, the same target has multiple motion trajectories.
  • the corresponding new target trajectory needs to be initialized, and if the target leaves the scene, the corresponding target trajectory is also deleted.
  • the fuzzy membership degree is established by introducing the feature similarity measure of the target trajectory, and the fuzzy synthesis function is used to calculate the trajectory.
  • the comprehensive similarity between the segments is then used to achieve the correlation of the same target trajectory with the maximum comprehensive similarity and threshold discriminant principle, and predictively fill the missing points between the trajectory segments of the same target, and finally obtain a complete continuous target trajectory.
  • the data association is performed by the observation result in the current video frame and the prediction result of the target, and the trajectory management is performed on the uncorrelated observation result and the prediction result, thereby effectively improving the multi-view.
  • the correct correlation between the target and the observation accurate tracking of multiple targets under the conditions of apparent similarity, frequent interaction, occlusion and background interference, with strong robustness and accuracy
  • FIG. 2 is a schematic flowchart diagram of a second embodiment of a video multi-target tracking method based on fuzzy logic according to the present invention, and a second embodiment of the present invention is a further extension of step S2 in the first embodiment.
  • the method includes the following steps:
  • the prediction result of the target in this step is obtained by predicting at least the trajectory of the target of the previous video frame.
  • an occlusion calculation is performed between the prediction results of all the targets in the current frame video to determine whether occlusion occurs between the prediction results of all the targets in the current frame video.
  • FIG. 3 is a schematic diagram of occlusion between prediction results of different targets of the present invention.
  • the tracking icon shapes of the prediction result A and the prediction result B are all rectangular, and there is overlap between the two, and the parameter of the prediction result A is expressed as: [x, y, w, d], wherein x, y represents the coordinates of the rectangular frame, w represents the width of the rectangular frame, d represents the height of the rectangular frame, and the parameter of the prediction result B is expressed as: [x', y', w', h'], where x', y' represents the coordinates of the rectangular frame, w' represents the width of the rectangular frame, h' represents the height of the rectangular frame, and the shaded portion between the predicted result A and the predicted result B is expressed as: [x o , y o , w o , h o ] And its overlapping parts are expressed as:
  • the area of the overlap between the prediction result A and the prediction result B is expressed as w o *h o . If the above w o , h o does not satisfy w o >0 or h o >0, the two tracking rectangles do not form an overlapping rectangle, that is, the overlapping rectangle area is 0.
  • the occlusion degree between the two is defined as:
  • s( ⁇ ) represents the area of the area
  • the occlusion degree satisfies 0 ⁇ ⁇ (A, B) ⁇ 1.
  • ⁇ (A, B) is greater than 0, it means that occlusion occurs between the prediction result A and the prediction result B.
  • the ambiguity determination is performed on the prediction results of all the targets in the current video frame scene, and the overlap ratio ⁇ ij of the tracking rectangle between the different target prediction results of the current video frame is calculated according to the formula (15).
  • the occlusion degree between the prediction results is determined, and it is judged whether the occlusion degree of each prediction result and other prediction results is smaller than the first occlusion determination threshold ⁇ over .
  • the first occlusion determination threshold ⁇ over satisfies ⁇ over ⁇ [0, 1]. If ⁇ ij is smaller than the first occlusion determination threshold ⁇ over , occlusion is considered to occur between the prediction results. If ⁇ over is equal to 0, it indicates that no occlusion occurs between the prediction results.
  • the first data is associated with the observation result in the current video frame.
  • An occlusion occurs between the predicted result and other predicted results, and a second data association is performed.
  • the first data association is different from the first data association, and the second data association is more complex than the first data association.
  • occlusion and non-occlusion are respectively performed between the prediction results of the target, and data association between the prediction result and the observation result is performed. It can accurately track multiple targets under the conditions of apparent similarity, frequent interaction, occlusion and background interference, and has strong robustness and accuracy.
  • FIG. 4 is a third embodiment of a video multi-target tracking method based on fuzzy logic according to the present invention, which is a further extension of S23 in the second embodiment of the video multi-target tracking method based on the fuzzy logic of the present invention.
  • the invention discloses a video multi-target tracking method based on fuzzy logic. The same steps of the second embodiment are not described herein again.
  • This embodiment includes:
  • step S23 further includes the following sub-steps:
  • the second similarity measure is used to measure the predicted result and the distance between the observed result.
  • the second similarity measure includes: a spatial distance feature similarity measure and an appearance feature similarity measure.
  • the spatial distance feature is one of the features that can more effectively match the observation and prediction results of the target.
  • the spatial distance feature similarity measure f D ( ⁇ ) between the observation d and the prediction result o is defined as:
  • the appearance feature similarity measure f S ( ⁇ ) between the observation result d and the prediction result o is defined as:
  • the multiplicative fusion is used to fuse the spatial distance feature similarity measure and the appearance feature similarity measure to obtain the correlation between the observation result and the prediction result, and is defined as:
  • the correlation cost matrix between the observation result and the prediction result is obtained according to the degree of association, and is defined as:
  • step S233 further includes the following sub-steps:
  • S2332 Determine whether the maximum value is the maximum value in the row and column, and meet the greater than the first threshold.
  • the spatial distance feature similarity measure and the appearance feature similarity measure between the observation result and the prediction result are fused to obtain an associated cost matrix of the two.
  • the optimization solution can find the correct correlation observations and prediction results.
  • FIG. 6 is a fourth embodiment of a video multi-target tracking method based on fuzzy logic according to the present invention, which is a further expansion of S23 in a second embodiment of a video multi-target tracking method based on fuzzy logic according to the present invention. exhibition.
  • Step S23 further includes the following sub-steps:
  • S23a Calculate a third similarity measure between the observation and the prediction.
  • the color feature has better ability to resist the deformation of the target, but it lacks the description of the spatial structure of the prediction result, and is sensitive to illumination, and the edge feature can well describe the edge of the human body.
  • the variation and the small amount of offset are insensitive, so the color and edge features have complementary characteristics, so the present invention uses these two kinds of information fusion to establish the appearance characteristics of the prediction result.
  • the distance between the observation result and the prediction result is measured by a third similarity measure, and the third similarity measure includes an appearance feature similarity measure, a geometric feature similarity measure, and a motion feature similarity measure. And spatial distance feature similarity measures.
  • the appearance feature similarity measure f A ( ⁇ ) between the observation result d and the prediction result o is defined as:
  • H c ( ⁇ ) is a color histogram feature weighted by the background of the current video frame image
  • H g ( ⁇ ) is a block gradient direction histogram feature
  • Is a variance constant Is the variance constant.
  • the motion feature similarity measure f M ( ⁇ ) between the observation d and the prediction result o is defined as:
  • the spatial distance feature similarity measure f D ( ⁇ ) between the observation d and the prediction result o is defined as:
  • the appearance feature similarity measure f S ( ⁇ ) between the observation d and the prediction result o is defined as:
  • the target model and the candidate model corresponding to the appearance feature similarity measure and the geometric feature similarity measure are respectively defined as: the target model:
  • Bhattacharyya coefficient which is defined as:
  • the motion model of the predicted result of the target is described by the coordinates and velocity of its centroid.
  • the motion of the video target motion is not very large.
  • the tracking rectangle (x, y, w, h) of the prediction result of each target establishes a motion state parameter model based on position, size, and velocity. Then define the state variable X k of the Kalman filter as:
  • x and y respectively represent the horizontal and vertical coordinates of the centroid of the tracking rectangle of the observation result of the kth frame, and respectively represent the velocity of the centroid of the tracking rectangle of the observation result of the kth frame in the x-axis and the y-axis direction.
  • S23b The fuzzy inference system model is used to calculate the weight value of each feature similarity measure in the third similarity measure.
  • the fuzzy inference system in the present invention mainly includes four basic elements: fuzzification of input variables, establishment of fuzzy rule base, fuzzy inference engine, and defuzzifier (fuzzy innovation precise output).
  • the input of the fuzzy inference system is defined by the similarity measure of each feature, and the adaptive weighting coefficient of each feature is obtained by inference.
  • step S23b further includes the following substeps:
  • FIG. 8 is a schematic structural diagram of multi-feature clue fusion according to the present invention.
  • the motion feature similarity measure is used as the first fuzzy input variable, and the similarity measure of the remaining three features is the second fuzzy input variable, and the calculation of the similarity measure mean of the other three features can be defined as:
  • the accuracy of the output variable is affected by the number of fuzzy sets.
  • FIG. 9 is a fuzzy input variable f k (i, j) according to the present invention. Schematic diagram of the membership function.
  • Input variable f k (i,j) and The fuzzification is performed by using five language fuzzy sets ⁇ ZE, SP, MP, LP, VP ⁇ , and the membership functions are ⁇ 0, ZE (i, j), ⁇ 0, SP (i, j), ⁇ 0, respectively.
  • MP (i, j), ⁇ 0, LP (i, j), and ⁇ 0, VP (i, j) indicate that the five fuzzy sets represent zero, positive, medium, positive, and very large, respectively. .
  • FIG. 10 is a schematic diagram of the membership function of the output fuzzy variable ⁇ M according to the present invention.
  • the output fuzzy variable ⁇ M contains five fuzzy sets: ⁇ ZE, SP, MP, LP, VP, EP ⁇ , EP represents the maximum fuzzy set, and its membership function is ⁇ 1, ZE (i, j), ⁇ 1, SP (i, j), ⁇ 1, MP (i, j), ⁇ 1, LP (i, j), ⁇ 1, VP (i, j), and ⁇ 1, EP (i, j).
  • S23b3 obtains the weight value of each feature similarity measure in the third similarity measure by using the inference rule of the fuzzy inference system.
  • the fuzzy inference rule can be as follows:
  • rule 1 is taken as an example to give a detailed reasoning process:
  • the fuzzy set corresponding to the fuzzy input variable f M (i, j) is ZE, and the corresponding fuzzy membership degree can be obtained by using the value of f M (i, j) according to the fuzzy membership function shown in FIG. 9 . value
  • the same method can be used to find fuzzy input variables Corresponding fuzzy membership value
  • means take small.
  • rule 1 the corresponding fuzzy output is EP, then the output of rule 1 can be calculated by:
  • the fuzzy rule m corresponds to the centroid of the output fuzzy set.
  • the fuzzy inference system is constructed for different features, and the weight value coefficients ⁇ A , ⁇ S and ⁇ D of the geometric shape feature, the motion direction feature and the spatial distance can be obtained respectively.
  • ⁇ k ⁇ k ⁇ A, M, S, D ⁇ is the fusion coefficient of each feature similarity measure and satisfies f k (i,j) k ⁇ A,M,S,D ⁇ are observations And forecast results A measure of similarity between each feature.
  • the greedy algorithm is used to achieve the correct correlation between the prediction result and the observation result, so that the correlation between the prediction result and the observation result further includes:
  • the environment adaptively assigns different weight values to different feature information, obtains the weighted sum of the multi-attribute features, constitutes the associated cost matrix of the prediction result of the frame target and the observation result, and then uses the greedy algorithm to optimize the solution allocation, which can effectively improve the multi-objective The correct association with the observations.
  • FIG. 11 is a schematic flowchart of a fifth embodiment of a video multi-target tracking method based on the fuzzy logic, which is a further extension of step S3 in the first embodiment of the video multi-target fuzzy data association method of the present invention.
  • Embodiments further include:
  • the fuzzy logic data association method can deal with the data association problem of high-frequency occlusion and multi-target tracking under a large number of false observations in a short period of time.
  • some target states are not long.
  • the target motion trajectory is difficult to maintain, and there will be a situation where the target trajectory is broken, that is, the same target has multiple motion trajectories.
  • the corresponding new target trajectory needs to be initialized, and if the target leaves the scene, the corresponding target trajectory is also deleted.
  • step S31 further includes the following sub-steps:
  • the target detector will inevitably produce some false observations under the condition of maintaining high detection rate, which will not be associated with any existing targets. These false observations may also be incorrectly initialized to new targets.
  • the new target initialization module uses the observations in the continuous T init frame to determine whether there is an overlap of the rectangular frame areas and the same size, and the area overlap ratio of the rectangular frame between the observation results is defined as:
  • the observation values at time t and time t+1 are respectively, and area( ⁇ ) indicates the area of the observation result.
  • Representation of observations versus The overlap area, h is the height value of the observation rectangle.
  • ⁇ ⁇ and ⁇ r represent the overlap rate threshold and the size similarity threshold, respectively.
  • the area overlap ratio and the size similarity of the observations in the continuous init frame are greater than the set threshold, that is, when the init is greater than or equal to T init , it is converted into a valid track, that is, a new track segment is started. And add it to the target tracking sequence. Therefore, the method can effectively eliminate false observations generated by the target detector, thereby reducing the erroneous target trajectory start.
  • the target termination trajectory may be a trajectory segment or a complete target trajectory
  • the last position of the termination trajectory is used to determine whether the trajectory is disconnected or leaves the scene in the scene. If the last position of the end track is within the scene, its trajectory is the end track segment. Meanwhile, when the start frame of the target track segment is the current time, it indicates that the new track segment is a new observed temporary track.
  • the set of termination trajectory segments is defined as:
  • the set of new track fragments is defined as: Where n a and n b respectively represent the number of the end track segment set and the new track segment set.
  • the first similarity measure includes an appearance similarity measure, a shape similarity measure, and a motion similarity measure, wherein the appearance similarity measure is defined as:
  • H g ( ⁇ ) represents the direction gradient histogram feature, Is a variance constant
  • the shape similarity measure is defined as:
  • h i represents the height of the end track segment T i in the image
  • h j represents the height of the new track segment T j in the image
  • the motion similarity measure is defined as:
  • G( ⁇ ) represents a Gaussian distribution
  • is the variance of the Gaussian distribution
  • ⁇ t is the first observed frame interval of the last trajectory segment T i and the new trajectory segment T j is observed.
  • v i is the end position and velocity of the terminating track segment T i , respectively.
  • v j is the starting position and velocity of the new track segment.
  • Figure 13 is a motion similarity measure for the end track segment and the new track segment in the occlusion case. It is assumed that the error between the position of the predicted result and the position of the actual observation satisfies the Gaussian distribution, that is, the smaller the distance between the predicted position of the terminated trajectory segment and the actual position of the new trajectory segment, the similarity between the two trajectory segments The bigger (for example versus The closer the distance is, The value is larger).
  • ⁇ gap is the associated time interval threshold, a time frame indicating that the trajectory segment T i is broken, Indicates the time frame at which the new track segment T j starts.
  • the present application uses a fuzzy comprehensive model based fuzzy model to measure the degree of matching between the ending trajectory segment and the new trajectory segment, which is defined as:
  • indicates that the matching degree takes a minimum value
  • indicates that the matching degree takes a maximum value
  • the associated cost matrix between the terminating trajectory segment and the new trajectory segment is defined as:
  • the prerequisites for the two track segments to be associated are:
  • the time has continuity, that is, there is no overlapping area in the corresponding time frame interval, ie
  • the reasonable correlation time interval threshold ⁇ gap can be set to associate the trajectories that may be associated within a relatively small range, which can improve the time efficiency of the algorithm. Excludes some trace segments that cannot be successfully associated.
  • the terminating trajectory segment T i is associated with the new trajectory segment T j* , and the new trajectory segment T j* is no longer associated with other terminating trajectory segments T i , otherwise it is not associated with the trajectory segment, where ⁇ is a threshold parameter, and 0 ⁇ ⁇ ⁇ 1.
  • the above correlation method can be used to associate two broken trajectories together, but the two trajectory segments There is often a lack of detection point information for lost frames. Therefore, the item The target does not form a complete continuous trajectory, and it also needs to predict the filling of the gap between them.
  • step S33 includes the following sub-steps:
  • S331 Perform bidirectional prediction on the missing track segment between the associated ending track segment and the new track segment to obtain position information of the predicted point.
  • T f in front of the two tracks off a track segment, the track segment is terminated
  • T b is the back of a track segment, i.e. the new track segment.
  • the position of the two-way continuous predicted target within the disconnection time interval is determined by the end position, the new starting position, and the speed information of the two trajectories where the same target is broken.
  • the process of acquiring the position information of the predicted point is as shown in FIG.
  • p f represents a specific location when using the track segment T f for the target where forward prediction
  • p b represents a specific location when using the track segment T b of the reverse prediction target
  • t f T f represents a forward prediction for the current
  • the number of frames, t b represents the current number of frames when T b is used for backward prediction.
  • step 2) until t f ⁇ t b , and finally obtain the position information of the missing points between the two track segments.
  • the averaging method is used to obtain the width and height of the rectangular frame of the predicted point, which is:
  • h k and w k represent the height and width of the rectangular frame of the detection point in the kth frame.
  • the height and width of the rectangle representing the end of the track segment Tf The height and width of the rectangular box representing the head of the track segment T b .
  • S333 Fill the missing track segment according to the position information of the predicted point and the rectangular frame information.
  • the prediction result and the observation result of the already associated target are filtered and predicted by a filter to obtain an actual track point and a prediction result in the current video frame of the target
  • the filter used in the present application may include, but is not limited to, a Kalman filter. Further, extrapolation prediction is performed on the prediction result without the associated target, and the prediction result is obtained, so as to accurately track the multi-target. And the prediction result of the target is used for data association in the next frame of the video frame.
  • the missing points between the broken trajectories of the same target are predicted and filled, and a complete continuous target trajectory is formed, which can effectively solve the problems of smoothing and predicting the target trajectory, terminating the target trajectory, and starting the new target trajectory.
  • FIG. 16 is a schematic structural diagram of a first embodiment of a video multi-target tracking apparatus based on fuzzy logic, including:
  • the detecting module 11 is configured to perform online target motion detection on the current video frame, and detect the obtained possible moving object as an observation result.
  • the association module 12 is configured to perform data association between the observation result and the prediction result of the target, wherein the prediction result is obtained by predicting at least the trajectory of the target of the previous video frame.
  • the trajectory management module 13 is configured to perform trajectory management on the unrelated prediction results and the observation results, including acquiring the ending trajectory segments by using the unrelated prediction results, and acquiring the new trajectory segments by using the unrelated observation results, Terminate the track segment and the new track segment for track association.
  • FIG. 17 is a schematic structural diagram of a second embodiment of a video multi-target tracking apparatus based on fuzzy logic according to the present invention, including: a processor 110 and a camera 120.
  • the camera 120 can be a local camera, the processor 110 is connected to the camera 120 through a bus; the camera 120 can also be a remote camera, and the processor 110 is connected to the camera 120 via a local area network or the Internet.
  • the processor 110 controls the operation of the fuzzy logic based video multi-target tracking device, which may also be referred to as a CPU (Central Processing Unit).
  • Processor 110 may be an integrated circuit chip with signal processing capabilities.
  • the processor 110 can also be a general purpose processor, a digital signal processor (DSP), an application specific integrated circuit (ASIC), an off-the-shelf programmable gate array (FPGA) or other programmable logic device, a discrete gate or transistor logic device, and discrete hardware components.
  • the general purpose processor may be a microprocessor or the processor or any conventional processor or the like.
  • the video multi-target tracking device based on the fuzzy logic may further include a memory (not shown) for storing instructions and data necessary for the operation of the processor 110, and may also store video data captured by the transmitter 120.
  • the processor 110 is configured to perform online target motion detection on the current video frame acquired from the camera 120, and detect the obtained possible moving object as an observation result; perform data association on the observation result and the target prediction result, wherein the prediction result is at least the previous one.
  • the trajectory of the target of the video frame is predicted; the trajectory management is performed on the unconstrained prediction result and the observation result, including the use of the unrelated prediction result to obtain the termination trajectory fragment and the use of the unrelated observation result to acquire the new trajectory
  • each part included in the video multi-target tracking device based on the fuzzy logic of the present invention can refer to the present
  • the description of the video multi-target tracking method based on the fuzzy logic in the corresponding embodiments is not described herein.
  • the present invention provides a video multi-target tracking method and apparatus based on fuzzy logic, which performs data association by observing results in a current video frame and prediction results of a target, and is not associated.
  • the observation and prediction results on the trajectory management can effectively improve the correct correlation between multi-objectives and observations, and accurately track multiple targets under the conditions of apparent similarity, frequent interaction, occlusion and background interference. Robustness and accuracy.

Abstract

The present invention discloses a fuzzy logic based video multi-target tracking method and device, the method comprising: performing online target motion detection on a current video frame, using a possible moving object obtained by detection as an observation result; creating a data association between the observation result and a prediction result of a target, wherein the prediction result is obtained by at least using a target trajectory in a previous video frame for prediction; performing trajectory management on unassociated prediction results and observation results, wherein the trajectory management comprises using an unassociated prediction result to obtain a terminating trajectory segment and using an unassociated observation result to obtain a new trajectory segment; and creating a trajectory association between the terminating trajectory segment and the new trajectory segment. The method of the present invention can be used to effectively increase the number of correct associations between multiple targets and observation results, thus significantly reducing the number of variations in target tags during tracking of multiple target objects, resulting in improved robustness and accuracy.

Description

基于模糊逻辑的视频多目标跟踪方法及装置Video multi-target tracking method and device based on fuzzy logic 【技术领域】[Technical Field]
本发明涉及目标跟踪领域,特别是涉及一种基于模糊逻辑的视频多目标跟踪方法及装置。The present invention relates to the field of target tracking, and in particular to a video multi-target tracking method and apparatus based on fuzzy logic.
【背景技术】【Background technique】
视频多目标跟踪技术作为计算机视觉领域的一个重要研究分支,其关系到很多前沿学科,如图像处理、模式识别、人工智能、自动控制以及计算机于一体,在智能视频监控、人机交互、机器人视觉导航、虚拟现实、医学诊断、交通管制及监控等领域有着非常重要的实用价值。Video multi-target tracking technology is an important research branch in the field of computer vision. It is related to many frontier disciplines, such as image processing, pattern recognition, artificial intelligence, automatic control and computer integration. In intelligent video surveillance, human-computer interaction, robot vision Navigation, virtual reality, medical diagnosis, traffic control and surveillance have very important practical value.
但针对复杂背景环境下的视频目标,开发出一种鲁棒高效的多目标跟踪算法仍存在诸多困难,如目标的相互遮挡、目标的数目以及虚假观测等。这些情况在实际的行人目标跟踪中具有很强的随意性和不确定性,用传统概率方法并不能很好的进行建模。However, for a video target in a complex background environment, there are still many difficulties in developing a robust and efficient multi-target tracking algorithm, such as mutual occlusion of targets, number of targets, and false observations. These situations have strong arbitrariness and uncertainty in actual pedestrian target tracking, and traditional probabilistic methods are not well modeled.
【发明内容】[Summary of the Invention]
本发明提供一种基于模糊逻辑的视频多目标跟踪方法及装置,能够有效提高多目标与观测之间的正确关联,对表观相似、频繁交互、遮挡以及背景干扰等情况下的多目标进行准确跟踪,同时能大幅减少多目标跟踪中目标标签变化的数量,具有较强的鲁棒性和准确性。The invention provides a video multi-target tracking method and device based on fuzzy logic, which can effectively improve the correct association between multi-objectives and observations, and accurately correct multiple targets under the conditions of apparent similarity, frequent interaction, occlusion and background interference. Tracking, while significantly reducing the number of target tag changes in multi-target tracking, has strong robustness and accuracy.
为了解决上述技术问题,本发明采用的一个技术方案是:提供一种基于模糊逻辑的视频多目标跟踪方法,包括:对当前视频帧进行在线目标运动检测,检测得到的可能运动对象作为观测结果;对所述观测结果和目标的预测结果进行数据关联,其中所述预测结果是至少利用前一视频帧的目标的轨迹进行预测而得到的;对未被关联上的所述预测结果及所述观测结果进行轨迹管理,包括利用所述未被关联上所述预测结果获取终止轨迹片段以及利用所述未被关联上的所述观测结果获取新的轨迹片段,对所述终止轨迹片段及所述新的轨迹片段进行轨迹关联。In order to solve the above technical problem, a technical solution adopted by the present invention is to provide a video multi-target tracking method based on fuzzy logic, which comprises: performing online target motion detection on a current video frame, and detecting a possible moving object as an observation result; Data correlation between the observation result and the prediction result of the target, wherein the prediction result is obtained by predicting at least the trajectory of the target of the previous video frame; and the prediction result and the observation on the unassociated Performing trajectory management, comprising: acquiring a termination trajectory segment by using the prediction result not associated with the prediction result, and acquiring a new trajectory segment by using the observation result that is not associated, the termination trajectory segment and the new The track segment is tracked.
为了解决上述技术问题,本发明采用的一个技术方案是:提供一种基于模糊逻辑的视频多目标跟踪的装置,包括:处理器,所述处理器用于对从所述摄像机获取的当前视频帧进行在线目标运动检测,检测得到的可能运动对象作为观测结果;对所述观测结果和目标的预测结果进行数据关联,其中所述预测结果是至少利用前一视频帧的目标的轨迹进行预测而得到的;对未被关联上的所述预测结果及所述观测结果进行轨迹管理,包括利用所述未被关联上所述预测结果获取终止轨迹片段以及利用所述未被关联上的所述观测结果获取新的轨迹 片段,对所述终止轨迹片段及所述新的轨迹片段进行轨迹关联。In order to solve the above technical problem, a technical solution adopted by the present invention is to provide a device for video multi-target tracking based on fuzzy logic, comprising: a processor, configured to perform a current video frame acquired from the camera Online target motion detection, detecting the obtained possible moving object as an observation result; performing data association on the observation result and the prediction result of the target, wherein the prediction result is obtained by predicting at least the trajectory of the target of the previous video frame Performing trajectory management on the prediction result that is not associated with the observation result, including acquiring the termination trajectory segment by using the prediction result that is not associated with the prediction result, and acquiring the observation result by using the unrelated association New track a segment, performing trajectory association on the terminating trajectory segment and the new trajectory segment.
本发明的有益效果是:提供一种基于模糊逻辑的视频多目标跟踪方法及装置,通过当前视频帧中的观测结果和目标的预测结果进行数据关联,并对未关联上的观测结果和预测结果进行轨迹管理,能有效提高多目标与观测之间的正确关联,对表观相似、频繁交互、遮挡以及背景干扰等情况下的多目标进行准确跟踪,具有较强的鲁棒性和准确性。The invention has the beneficial effects of providing a video multi-target tracking method and device based on fuzzy logic, which performs data association by the observation result in the current video frame and the prediction result of the target, and the observation result and the prediction result on the unassociated Trajectory management can effectively improve the correct correlation between multi-objectives and observations, and accurately track multiple targets under the conditions of apparent similarity, frequent interaction, occlusion and background interference, which has strong robustness and accuracy.
【附图说明】[Description of the Drawings]
图1是基于模糊逻辑的视频多目标跟踪方法第一实施例的流程示意图;1 is a schematic flow chart of a first embodiment of a video multi-target tracking method based on fuzzy logic;
图2是本发明基于模糊逻辑的视频多目标跟踪方法第二实施例的流程示意图;2 is a schematic flow chart of a second embodiment of a video multi-target tracking method based on fuzzy logic according to the present invention;
图3是本发明不同目标的预测结果之间遮挡示意图;3 is a schematic diagram of occlusion between prediction results of different targets of the present invention;
图4是本发明基于模糊逻辑的视频多目标跟踪方法第三实施例的流程示意图;4 is a schematic flow chart of a third embodiment of a video multi-target tracking method based on fuzzy logic according to the present invention;
图5是图4中步骤S233一实施方式的流程示意图;FIG. 5 is a schematic flowchart of an embodiment of step S233 in FIG. 4;
图6是本发明基于模糊逻辑的视频多目标跟踪方法第四实施例的流程示意图;6 is a schematic flow chart of a fourth embodiment of a video multi-target tracking method based on fuzzy logic according to the present invention;
图7是图6中步骤S23b一实施方式的流程示意图;7 is a schematic flow chart of an embodiment of step S23b in FIG. 6;
图8是本发明多特征线索融合的结构示意图;8 is a schematic structural diagram of multi-feature clue fusion according to the present invention;
图9是本发明模糊输入变量fk(i,j)、
Figure PCTCN2017091575-appb-000001
的隶属度函数示意图;
Figure 9 is a fuzzy input variable f k (i, j) of the present invention,
Figure PCTCN2017091575-appb-000001
Schematic diagram of the membership function;
图10是本发明输出模糊变量αM的隶属度函数示意图;10 is a schematic diagram of a membership function of the output fuzzy variable α M of the present invention;
图11是本发明基于模糊逻辑的视频多目标跟踪方法第五实施例的流程示意图;11 is a schematic flowchart diagram of a fifth embodiment of a video multi-target tracking method based on fuzzy logic according to the present invention;
图12是图11中步骤S31一实施方式的流程示意图;12 is a schematic flow chart of an embodiment of step S31 in FIG. 11;
图13是本发明遮挡情况下终止轨迹片段和新的轨迹片段的运动相似性度量;13 is a motion similarity metric of a terminating trajectory segment and a new trajectory segment in the occlusion case of the present invention;
图14是图11中步骤S33一实施方式的流程示意图;14 is a schematic flow chart of an embodiment of step S33 in FIG. 11;
图15是本发明获取丢失预测点的位置示意图;15 is a schematic diagram of a position of the present invention for acquiring a loss prediction point;
图16是本发明基于模糊逻辑的视频多目标跟踪装置第一实施方式的结构示意图;16 is a schematic structural diagram of a first embodiment of a video multi-target tracking apparatus based on fuzzy logic according to the present invention;
图17是本发明基于模糊逻辑的视频多目标跟踪装置第二实施方式的结构示意图。17 is a schematic structural diagram of a second embodiment of a video multi-target tracking apparatus based on fuzzy logic according to the present invention.
【具体实施方式】【Detailed ways】
如图1所示,本发明基于模糊逻辑的视频多目标跟踪方法第一实施例的流程示意图,该方法包括如下步骤:As shown in FIG. 1 , a schematic flowchart of a first embodiment of a video multi-target tracking method based on fuzzy logic, the method includes the following steps:
S1:对当前视频帧进行在线目标运动检测,检测得到的可能运动对象作为 观测结果。S1: Perform online target motion detection on the current video frame, and detect possible motion objects as Observations.
在线目标运动检测可以使用帧差法、光流法、背景减除法、混合高斯背景模型等运动检测算法。本发明主要采用混合高斯背景模型,对当前视频帧进行运动检测,以从中找出属于运动前景的像素,辅以中值滤波和简单的形态学处理,最终得到当前视频帧中的可能运动对象作为观测对象。一个观测对象是当前视频帧中的一个图像块,一般而言,观测对象的形状为矩形。The online target motion detection can use motion detection algorithms such as frame difference method, optical flow method, background subtraction method, and mixed Gaussian background model. The invention mainly adopts a mixed Gaussian background model to perform motion detection on the current video frame to find pixels belonging to the foreground of the motion, supplemented by median filtering and simple morphological processing, and finally obtain possible moving objects in the current video frame. Observed object. An observation object is an image block in the current video frame. Generally, the shape of the observation object is a rectangle.
采用混合高斯背景模型对运动目标进行检测,得到检测目标集合为z={z1,...,zr},由于检测出的目标的预测结果并没有身份ID标识,并不能判断观测结果与前一帧目标的预测结果的对应情况。为此,还必须以检测结果z={z1,...,zr}作为当前观测信息,对目标的预测结果以及观测结果作进一步的关联判断。The mixed Gaussian background model is used to detect the moving target, and the set of detection targets is z={z 1 ,...,z r }. Since the predicted result of the detected target does not have the identity ID, the observation result cannot be judged. The corresponding situation of the prediction result of the previous frame target. To this end, the detection result z={z 1 ,..., z r } must also be used as the current observation information to make further correlation judgment between the prediction result of the target and the observation result.
S2,对观测结果和目标的预测结果进行数据关联。S2, data association between the observation result and the prediction result of the target.
由于在视频多目标跟踪中的目标绝大多数为非刚体,其运动具有一定的随机性,且在实际复杂场景中经常会有光照变化、目标遮挡、相似物干扰等因素,都有可能引起目标跟踪的不确定性。目标包括稳定跟踪的可靠目标及不稳定跟踪的临时目标。本步骤中的目标状态,即每个目标被标记为可靠目标还是临时目标,是由前一视频帧的轨迹管理决定的。临时目标包括在前一视频帧为未被关联且不是匹配成功的候选结果的观测结果建立的新的目标,以及连续关联成功的帧数小于或者等于第一帧数阈值且未被删除的目标。可靠目标包括连续关联成功的帧数大于第一帧数阈值且未被删除的目标。目标的预测结果是至少利用前一视频帧的目标的轨迹进行预测而得到的。Since most of the targets in video multi-target tracking are non-rigid, their motions have certain randomness, and in actual complex scenes, there are often lighting changes, target occlusion, similar object interference, etc., which may cause targets. Tracking uncertainty. Targets include reliable targets for stable tracking and temporary targets for unstable tracking. The target state in this step, that is, whether each target is marked as a reliable target or a temporary target, is determined by the trajectory management of the previous video frame. The temporary target includes a new target established by the observation that the previous video frame is a candidate result that is not associated and is not a successful match, and a target whose consecutively associated successful number of frames is less than or equal to the first frame number threshold and has not been deleted. A reliable target includes a target whose number of consecutively successful frames is greater than the first frame number threshold and has not been deleted. The prediction result of the target is obtained by predicting at least the trajectory of the target of the previous video frame.
S3,对未被关联上的预测结果及观测结果进行轨迹管理,包括利用未被关联上预测结果获取终止轨迹片段以及利用未被关联上的观测结果获取新的轨迹片段,对终止轨迹片段及新的轨迹片段进行轨迹关联。S3, performing trajectory management on the unrelated prediction results and observation results, including acquiring the trajectory segments by using the unrelated prediction results and acquiring new trajectory segments by using the unrelated observations, and terminating the trajectory segments and new The track segment is tracked.
具体而言,步骤S2中的数据关联方法能够处理在短时期内出现的高频率遮挡和大量虚假观测条件下的多目标跟踪的数据关联问题,然而在长时间的遮挡以及漏检情况下,一些目标状态长时间得不到更新,其目标运动轨迹很难维持,会出现目标轨迹断批的情况,即同一个目标拥有多条运动轨迹。同时,当新目标进入场景时,需要初始化相应的新的目标轨迹,如果目标离开场景时,也要删除相应的目标轨迹。Specifically, the data association method in step S2 can deal with the data association problem of high-frequency occlusion occurring in a short period of time and multi-target tracking under a large number of false observation conditions, but in the case of long-term occlusion and missed detection, some The target state is not updated for a long time, and the target motion trajectory is difficult to maintain, and the target trajectory is broken, that is, the same target has multiple motion trajectories. At the same time, when the new target enters the scene, the corresponding new target trajectory needs to be initialized, and if the target leaves the scene, the corresponding target trajectory is also deleted.
本申请通过分别利用未被关联上预测结果及未被关联上的观测结果获取终止轨迹片段以及新的轨迹片段,通过引入目标轨迹的特征相似性度量建立模糊隶属度,利用模糊综合函数来计算轨迹片段之间的综合相似度,然后采用最大综合相似度和阈值判别原则来实现同一目标轨迹的关联,并对同一目标的轨迹片段间的缺失点进行预测填充,最后得到一条完整连续的目标轨迹。In this application, by using the unrelated prediction results and the unrelated observations to obtain the termination trajectory segments and the new trajectory segments, the fuzzy membership degree is established by introducing the feature similarity measure of the target trajectory, and the fuzzy synthesis function is used to calculate the trajectory. The comprehensive similarity between the segments is then used to achieve the correlation of the same target trajectory with the maximum comprehensive similarity and threshold discriminant principle, and predictively fill the missing points between the trajectory segments of the same target, and finally obtain a complete continuous target trajectory.
上述实施方式中,通过当前视频帧中的观测结果和目标的预测结果进行数据关联,并对未关联上的观测结果和预测结果进行轨迹管理,能有效提高多目 标与观测之间的正确关联,对表观相似、频繁交互、遮挡以及背景干扰等情况下的多目标进行准确跟踪,具有较强的鲁棒性和准确性In the above embodiment, the data association is performed by the observation result in the current video frame and the prediction result of the target, and the trajectory management is performed on the uncorrelated observation result and the prediction result, thereby effectively improving the multi-view. The correct correlation between the target and the observation, accurate tracking of multiple targets under the conditions of apparent similarity, frequent interaction, occlusion and background interference, with strong robustness and accuracy
请参阅图2,图2为本发明基于模糊逻辑的视频多目标跟踪方法第二实施例的流程示意图,且本发明的第二实施例是在第一实施例中步骤S2的进一步拓展,所述方法包括如下步骤:Referring to FIG. 2, FIG. 2 is a schematic flowchart diagram of a second embodiment of a video multi-target tracking method based on fuzzy logic according to the present invention, and a second embodiment of the present invention is a further extension of step S2 in the first embodiment. The method includes the following steps:
S21,计算当前视频帧中不同目标的预测结果之间的遮挡度。S21. Calculate an occlusion degree between prediction results of different targets in the current video frame.
本步骤中的目标的预测结果是至少利用前一视频帧的目标的轨迹进行预测而得到的。首先对于当前帧视频中的所有目标的预测结果之间进行遮挡度计算,以判定当前帧视频中的所有目标的预测结果之间是否发生遮挡。The prediction result of the target in this step is obtained by predicting at least the trajectory of the target of the previous video frame. First, an occlusion calculation is performed between the prediction results of all the targets in the current frame video to determine whether occlusion occurs between the prediction results of all the targets in the current frame video.
请进一步参阅图3,图3为本发明不同目标的预测结果之间遮挡示意图。在当前视频帧中,预测结果A和预测结果B的跟踪图标形状均为矩形,且二者之间有重叠,且预测结果A的参数表述为:[x,y,w,d],其中,x,y表示矩形框的坐标,w表示矩形框宽度,d表示矩形框的高度,且预测结果B的参数表述为:[x′,y′,w′,h′],其中,x′,y′表示矩形框的坐标,w′表示矩形框宽度,h′表示矩形框的高度,预测结果A和预测结果B之间阴影部分表示为:[xo,yo,wo,ho],且其重叠部分表示为:Please refer to FIG. 3 for further reference. FIG. 3 is a schematic diagram of occlusion between prediction results of different targets of the present invention. In the current video frame, the tracking icon shapes of the prediction result A and the prediction result B are all rectangular, and there is overlap between the two, and the parameter of the prediction result A is expressed as: [x, y, w, d], wherein x, y represents the coordinates of the rectangular frame, w represents the width of the rectangular frame, d represents the height of the rectangular frame, and the parameter of the prediction result B is expressed as: [x', y', w', h'], where x', y' represents the coordinates of the rectangular frame, w' represents the width of the rectangular frame, h' represents the height of the rectangular frame, and the shaded portion between the predicted result A and the predicted result B is expressed as: [x o , y o , w o , h o ] And its overlapping parts are expressed as:
xo=max(x,x′)x o =max(x,x')
yo=max(y,y′)y o =max(y,y')
wo=min(x+w,x′+w′)-xo w o =min(x+w,x'+w')-x o
ho=min(y+h,y′+h′)-yo  (14)h o =min(y+h,y'+h')-y o (14)
由此可知,预测结果A和预测结果B之间重叠部分的面积表示为wo*ho。若上述的wo、ho不满足wo>0或者ho>0,则两跟踪矩形框之间不构成重叠矩形,也即重叠矩形面积为0。From this, it can be seen that the area of the overlap between the prediction result A and the prediction result B is expressed as w o *h o . If the above w o , h o does not satisfy w o >0 or h o >0, the two tracking rectangles do not form an overlapping rectangle, that is, the overlapping rectangle area is 0.
假设预测结果A与预测结果B发生如图2所示的遮挡,且两跟踪矩形框之间重叠的阴影部分表示遮挡区域,则定义二者之间的遮挡度为:Assuming that the prediction result A and the prediction result B are occluded as shown in FIG. 2, and the shaded portion between the two tracking rectangles represents the occlusion area, the occlusion degree between the two is defined as:
Figure PCTCN2017091575-appb-000002
Figure PCTCN2017091575-appb-000002
其中,s(·)表示区域面积,且遮挡度满足0≤ω(A,B)≤1。当ω(A,B)大于0,则说明预测结果A与预测结果B之间发生遮挡。且进一步由分别代表预测结果A与预测结果B的两跟踪矩形框底部的纵向图像坐标值yA与yB可知,若yA>yB,则说明预测结果B被预测结果A遮挡,反之,则说明预测结果A被预测结果B遮挡。Where s(·) represents the area of the area, and the occlusion degree satisfies 0 ≤ ω (A, B) ≤ 1. When ω(A, B) is greater than 0, it means that occlusion occurs between the prediction result A and the prediction result B. And further, the longitudinal image coordinate values y A and y B at the bottom of the two tracking rectangles respectively representing the prediction result A and the prediction result B, if y A > y B , the prediction result B is blocked by the prediction result A, and vice versa. Then, the prediction result A is occluded by the prediction result B.
S22:根据遮挡度分别判断每一预测结果与其他预测结果之间是否发生遮挡。S22: Determine whether occlusion occurs between each prediction result and other prediction results according to the occlusion degree.
本步骤中,对于当前视频帧场景中的全部目标的预测结果进行遮挡度判断,并按照式(15)计算当前视频帧不同目标预测结果之间的跟踪矩形框的重叠率ωij (不同目标即预测结果之间的遮挡度),并判断每一预测结果与其他预测结果的遮挡度是否小于第一遮挡判定阈值τover。其中,第一遮挡判定阈值τover满足τover∈[0,1]。若ωij小于第一遮挡判定阈值τover则认为预测结果之间发生遮挡,若τover等于0,则表明预测结果之间未发生遮挡。In this step, the ambiguity determination is performed on the prediction results of all the targets in the current video frame scene, and the overlap ratio ω ij of the tracking rectangle between the different target prediction results of the current video frame is calculated according to the formula (15). The occlusion degree between the prediction results is determined, and it is judged whether the occlusion degree of each prediction result and other prediction results is smaller than the first occlusion determination threshold τ over . The first occlusion determination threshold τ over satisfies τ over ∈ [0, 1]. If ω ij is smaller than the first occlusion determination threshold τ over , occlusion is considered to occur between the prediction results. If τ over is equal to 0, it indicates that no occlusion occurs between the prediction results.
S23:若预测结果与任何其他预测结果之间均未发生遮挡,则对预测结果和观测结果进行第一数据关联;若预测结果与其他预测结果之间发生遮挡,则对预测结果和观测结果进行第二数据关联。S23: if there is no occlusion between the prediction result and any other prediction results, the first data correlation is performed on the prediction result and the observation result; if occlusion occurs between the prediction result and other prediction results, the prediction result and the observation result are performed The second data association.
对当前视频帧中所有目标的预测结果进行遮挡度判定后,对预测结果与任何其他预测结果之间均未发生遮挡的,将其与当前视频帧中的观测结果进行第一数据关联。对预测结果与其他预测结果之间发生遮挡,进行第二数据关联。其中,第一数据关联和第一数据关联不同,且第二数据关联比第一数据关联复杂。After occlusion determination is performed on the prediction result of all the targets in the current video frame, if no occlusion occurs between the prediction result and any other prediction results, the first data is associated with the observation result in the current video frame. An occlusion occurs between the predicted result and other predicted results, and a second data association is performed. The first data association is different from the first data association, and the second data association is more complex than the first data association.
上述实施方式中,首先通过判定当前视频帧中所有目标的预测结果之间是否发生遮挡,在分别对目标的预测结果之间发生遮挡和不遮挡的情况,进行预测结果和观测结果间的数据关联,能够对表观相似、频繁交互、遮挡以及背景干扰等情况下的多目标进行准确跟踪,具有较强的鲁棒性和准确性。In the above embodiment, firstly, by determining whether occlusion occurs between prediction results of all targets in the current video frame, occlusion and non-occlusion are respectively performed between the prediction results of the target, and data association between the prediction result and the observation result is performed. It can accurately track multiple targets under the conditions of apparent similarity, frequent interaction, occlusion and background interference, and has strong robustness and accuracy.
请参阅图4,图4为本发明基于模糊逻辑的视频多目标跟踪方法第三实施例,是在本发明基于模糊逻辑的视频多目标跟踪方法第二实施例中S23的进一步扩展,因此与本发明基于模糊逻辑的视频多目标跟踪方法第二实施例相同的步骤在此不再赘述。本实施例包括:Referring to FIG. 4, FIG. 4 is a third embodiment of a video multi-target tracking method based on fuzzy logic according to the present invention, which is a further extension of S23 in the second embodiment of the video multi-target tracking method based on the fuzzy logic of the present invention. The invention discloses a video multi-target tracking method based on fuzzy logic. The same steps of the second embodiment are not described herein again. This embodiment includes:
参阅图4,步骤S23进一步包括如下子步骤:Referring to FIG. 4, step S23 further includes the following sub-steps:
S231,计算观测结果和预测结果之间的第二相似性度量。S231. Calculate a second similarity measure between the observation result and the prediction result.
采用第二相似性度量对预测结果和与观测结果间的距离进行度量。其中,第二相似性度量包括:空间距离特征相似性度量以及外观特征相似性度量。通常,目标在相邻帧图像之间的位置不会发生较大变化,因此,空间距离特征是能够较为有效地匹配目标的观测结果与预测结果的特征之一。在具体实施例中,观测结果d与预测结果o之间的空间距离特征相似性度量fD(·)定义为:The second similarity measure is used to measure the predicted result and the distance between the observed result. The second similarity measure includes: a spatial distance feature similarity measure and an appearance feature similarity measure. Generally, the position of the target between adjacent frame images does not change greatly. Therefore, the spatial distance feature is one of the features that can more effectively match the observation and prediction results of the target. In a specific embodiment, the spatial distance feature similarity measure f D (·) between the observation d and the prediction result o is defined as:
Figure PCTCN2017091575-appb-000003
Figure PCTCN2017091575-appb-000003
其中,||·||2为二范数,(xo,yo)为预测结果o在当前视频帧中的中心坐标,(xd,yd)为观测结果d在当前视频帧中的中心坐标,ho为预测结果o在当前视频帧中的高度,
Figure PCTCN2017091575-appb-000004
为方差常量。
Where ||·|| 2 is a two-norm, (x o , y o ) is the central coordinate of the prediction result o in the current video frame, and (x d , y d ) is the observation result d in the current video frame. Center coordinates, h o is the height of the prediction result o in the current video frame,
Figure PCTCN2017091575-appb-000004
Is the variance constant.
进一步,观测结果d与预测结果o之间的外观特征相似性度量fS(·)定义为: Further, the appearance feature similarity measure f S (·) between the observation result d and the prediction result o is defined as:
Figure PCTCN2017091575-appb-000005
Figure PCTCN2017091575-appb-000005
其中,hd为观测结果d在当前视频帧中的高度,
Figure PCTCN2017091575-appb-000006
为方差常量。
Where h d is the height of the observation d in the current video frame,
Figure PCTCN2017091575-appb-000006
Is the variance constant.
S232,利用第一相似性度量计算观测结果和预测结果之间的关联代价矩阵。S232. Calculate an associative cost matrix between the observation result and the prediction result by using the first similarity measure.
采用乘性融合对空间距离特征相似性度量以及外观特征相似性度量进行融合,以得到观测结果和预测结果之间的关联度,且定义为:The multiplicative fusion is used to fuse the spatial distance feature similarity measure and the appearance feature similarity measure to obtain the correlation between the observation result and the prediction result, and is defined as:
sij=fD(o,d)×fs(o,d)  (3)s ij =f D (o,d)×f s (o,d) (3)
根据关联度得到观测结果和预测结果之间的关联代价矩阵,且定义为:The correlation cost matrix between the observation result and the prediction result is obtained according to the degree of association, and is defined as:
S=[sij]n×l  (4)S=[s ij ] n×l (4)
其中,i=1,2,…n,j=1,2,…,l。Where i=1, 2,...n, j=1, 2,...,l.
S233,采用贪婪算法对关联代价矩阵进行优化求解,找出关联的观测结果和预测结果。S233, using the greedy algorithm to optimize the correlation cost matrix to find the associated observations and prediction results.
采用贪婪算法实现预测结果与观测结果之间的正确的关联,从而得到预测结果与观测结果之间的关联对。请参阅图5,该步骤S233进一步包括如下子步骤:The greedy algorithm is used to achieve the correct correlation between the prediction result and the observation result, so as to obtain the correlation between the prediction result and the observation result. Referring to FIG. 5, the step S233 further includes the following sub-steps:
S2331,找出关联代价矩阵S中未被标记的所有元素中的最大值。S2331, find the maximum value among all the elements in the associated cost matrix S that are not marked.
找出关联代价矩阵S中未被标记的所有元素中的最大值Spq=max([Sij]n*l),其中,p=1,2,3......n,q=1,2,3......l,并标记该关联代价矩阵S中该最大值spq所在的第p行及第q列的所有元素。Find the maximum value Spq=max([Sij]n*l) of all elements in the associated cost matrix S that are not marked, where p=1, 2, 3...n, q=1, 2, 3...l, and mark all elements of the pth row and the qth column where the maximum value s pq is located in the associated cost matrix S.
S2332,判断最大值是否为所在行列中的最大值,且满足大于第一阈值。S2332: Determine whether the maximum value is the maximum value in the row and column, and meet the greater than the first threshold.
判断该最大值spq大是否为所在行和所在列中的最大值,即是否满足:spq≥{spj}j=1,2,…l、spq≥{siq}i=1,2,…,n。进一步判断该最大值spq是否大于第一阈值λ1,即预测结果p与观测结果q的关联概率是否大于第一阈值λ1,且该第一阈值的满足λ1∈[0.6,0.9]。It is determined whether the maximum value s pq is the maximum value in the row and the column in which it is located, that is, whether: s pq ≥{s pj } j=1,2,...l , s pq ≥{s iq } i=1, 2,...,n . It is further determined whether the maximum value s pq is greater than the first threshold λ 1 , that is, whether the correlation probability of the prediction result p and the observation result q is greater than the first threshold λ 1 , and the first threshold satisfies λ 1 ∈ [0.6, 0.9].
S2333,若大于,则观测结果和预测结果正确关联。S2333, if it is greater, the observation result is correctly correlated with the prediction result.
该最大值spq满足上述判定条件,则认为预测结果p与观测结果q和之间正确关联,则将该关联对记录进已关联预测结果和观测结果的集合中。循环执行上述步骤直至关联代价矩阵S中的所有行或所有列均被标记。When the maximum value s pq satisfies the above-described determination condition, it is considered that the prediction result p is correctly associated with the observation result q and the correlation pair is recorded in the set of the associated prediction result and the observation result. The above steps are performed cyclically until all or all of the columns in the associated cost matrix S are marked.
上述实施方式,通过判定当前视频帧中目标的预测结果之间未发生遮挡,对观测结果和预测结果之间空间距离特征相似性度量以及外观特征相似性度量进行融合以得到二者的关联代价矩阵,优化求解能够找出正确关联的观测结果和预测结果。In the foregoing implementation manner, by determining that no occlusion occurs between the prediction results of the target in the current video frame, the spatial distance feature similarity measure and the appearance feature similarity measure between the observation result and the prediction result are fused to obtain an associated cost matrix of the two. The optimization solution can find the correct correlation observations and prediction results.
请参阅图6,图6为本发明基于模糊逻辑的视频多目标跟踪方法第四实施例,是在本发明基于模糊逻辑的视频多目标跟踪方法第二实施例中S23的进一步扩 展。Referring to FIG. 6, FIG. 6 is a fourth embodiment of a video multi-target tracking method based on fuzzy logic according to the present invention, which is a further expansion of S23 in a second embodiment of a video multi-target tracking method based on fuzzy logic according to the present invention. exhibition.
在视频帧中不同目标的预测结果之间有遮挡的情况下,由于采用简单的乘性融合策略对两种特征进行融合不能完成预测结果与观测结果之间的关联,这种情况下采用基于模糊逻辑多特征线索加权的融合策略。In the case where there is occlusion between the prediction results of different targets in the video frame, the fusion of the two features cannot be completed due to the simple multiplicative fusion strategy, and the correlation between the prediction result and the observation result cannot be completed. A logical multi-feature clue weighted fusion strategy.
步骤S23进一步包括如下子步骤:Step S23 further includes the following sub-steps:
S23a:计算观测结果和预测结果之间的第三相似性度量。S23a: Calculate a third similarity measure between the observation and the prediction.
在当前视频帧中,由于颜色特征具有较好的抵抗目标形变的能力,但其缺少对预测结果空间结构的描述,且对光照敏感,而边缘特征能够很好地描述人体的边缘,它对光照变化和小量的偏移不敏感,因此颜色与边缘特征具有互补特性,故本发明采用这两种信息融合建立预测结果的外观特征。在本发明中,利用第三相似性度量对观测结果和预测结果之间的距离进行度量,且该第三相似性度量包括外观特征相似性度量、几何形状特征相似性度量、运动特征相似性度量以及空间距离特征相似性度量。In the current video frame, because the color feature has better ability to resist the deformation of the target, but it lacks the description of the spatial structure of the prediction result, and is sensitive to illumination, and the edge feature can well describe the edge of the human body. The variation and the small amount of offset are insensitive, so the color and edge features have complementary characteristics, so the present invention uses these two kinds of information fusion to establish the appearance characteristics of the prediction result. In the present invention, the distance between the observation result and the prediction result is measured by a third similarity measure, and the third similarity measure includes an appearance feature similarity measure, a geometric feature similarity measure, and a motion feature similarity measure. And spatial distance feature similarity measures.
其中,观测结果d与预测结果o之间的外观特征相似性度量fA(·)定义为:Wherein, the appearance feature similarity measure f A (·) between the observation result d and the prediction result o is defined as:
Figure PCTCN2017091575-appb-000007
Figure PCTCN2017091575-appb-000007
其中,ρ(·)为巴氏(Bhattacharyya)系数,Hc(·)为所述当前视频帧图像背景加权的颜色直方图特征,Hg(·)为分块梯度方向直方图特征,
Figure PCTCN2017091575-appb-000008
为方差常量,
Figure PCTCN2017091575-appb-000009
为方差常量。
Where ρ(·) is a Bhattacharyya coefficient, H c (·) is a color histogram feature weighted by the background of the current video frame image, and H g (·) is a block gradient direction histogram feature,
Figure PCTCN2017091575-appb-000008
Is a variance constant,
Figure PCTCN2017091575-appb-000009
Is the variance constant.
观测结果d与预测结果o之间的运动特征相似性度量fM(·)定义为:The motion feature similarity measure f M (·) between the observation d and the prediction result o is defined as:
Figure PCTCN2017091575-appb-000010
Figure PCTCN2017091575-appb-000010
其中,(x′o,y′o)为上一时刻所述预测结果o的中心坐标,(xo,yo)为所述预测结果o的中心坐标,
Figure PCTCN2017091575-appb-000011
为上一时刻所述预测结果o的速度在坐标轴上的投影,
Figure PCTCN2017091575-appb-000012
为方差常量;
Where (x' o , y' o ) is the central coordinate of the prediction result o at the previous moment, and (x o , y o ) is the central coordinate of the prediction result o,
Figure PCTCN2017091575-appb-000011
The projection of the speed of the prediction result o on the coordinate axis for the previous moment,
Figure PCTCN2017091575-appb-000012
Is a variance constant;
观测结果d与预测结果o之间的空间距离特征相似性度量fD(·)定义为:The spatial distance feature similarity measure f D (·) between the observation d and the prediction result o is defined as:
Figure PCTCN2017091575-appb-000013
Figure PCTCN2017091575-appb-000013
其中,||·||2为二范数,(xo,yo)为预测结果o的中心坐标,(xd,yd)为观测结果d的中心坐标,ho为预测结果o的高度,
Figure PCTCN2017091575-appb-000014
为方差常量。
Where ||·|| 2 is a two-norm, (x o , y o ) is the central coordinate of the prediction result o, (x d , y d ) is the central coordinate of the observation d, and h o is the prediction result o height,
Figure PCTCN2017091575-appb-000014
Is the variance constant.
观测结果d与预测结果o之间的外观特征相似性度量fS(·)定义为: The appearance feature similarity measure f S (·) between the observation d and the prediction result o is defined as:
Figure PCTCN2017091575-appb-000015
Figure PCTCN2017091575-appb-000015
其中,hd为观测结果d的高度,
Figure PCTCN2017091575-appb-000016
为方差常量。
Where h d is the height of the observation d
Figure PCTCN2017091575-appb-000016
Is the variance constant.
其中,外观特征相似性度量、几何形状特征相似性度量对应的目标模型与候选模型分别定义为:目标模型:
Figure PCTCN2017091575-appb-000017
候选模型:
Figure PCTCN2017091575-appb-000018
The target model and the candidate model corresponding to the appearance feature similarity measure and the geometric feature similarity measure are respectively defined as: the target model:
Figure PCTCN2017091575-appb-000017
Candidate model:
Figure PCTCN2017091575-appb-000018
为了衡量目标模型和候选模型之间的相似度,本发明采用巴氏(Bhattacharyya)系数来描述,且该巴氏系数定义为:To measure the similarity between the target model and the candidate model, the present invention is described using a Bhattacharyya coefficient, which is defined as:
Figure PCTCN2017091575-appb-000019
Figure PCTCN2017091575-appb-000019
目标的预测结果的运动模型由其质心的坐标和速度来描述。在视频多目标跟踪中,由于相邻两帧视频序列图像间隔时间很短,视频目标运动的机动性不是很大,在大多数情况下,可以假设每一个目标的运动为匀速运动,因此可以为每个目标的预测结果的跟踪矩形框(x,y,w,h)建立基于位置、大小以及速度的运动状态参数模型。则定义卡尔曼滤波器的状态变量Xk为:The motion model of the predicted result of the target is described by the coordinates and velocity of its centroid. In video multi-target tracking, since the interval between adjacent two video sequences is very short, the motion of the video target motion is not very large. In most cases, it can be assumed that the motion of each target is uniform motion, so it can be The tracking rectangle (x, y, w, h) of the prediction result of each target establishes a motion state parameter model based on position, size, and velocity. Then define the state variable X k of the Kalman filter as:
Xk=[x,vx,y,vy]T  (17)X k =[x,v x ,y,v y ] T (17)
其中,x、y分别表示第k帧观测结果的跟踪矩形框质心的横纵坐标,分别表示第k帧观测结果的跟踪矩形框的质心在x轴和y轴方向上的速度。Wherein, x and y respectively represent the horizontal and vertical coordinates of the centroid of the tracking rectangle of the observation result of the kth frame, and respectively represent the velocity of the centroid of the tracking rectangle of the observation result of the kth frame in the x-axis and the y-axis direction.
S23b:采用模糊推理系统模型计算第三相似性度量中每一特征相似性度量的权重值。S23b: The fuzzy inference system model is used to calculate the weight value of each feature similarity measure in the third similarity measure.
本发明中的模糊推理系统主要包含四个基本要素:输入变量的模糊化、建立模糊规则库、模糊推理机、去模糊器(模糊新息精确化输出)。本实施例中,将利用各个特征的相似性度量定义模糊推理系统的输入,通过推理得到每个特征的自适应加权系数。The fuzzy inference system in the present invention mainly includes four basic elements: fuzzification of input variables, establishment of fuzzy rule base, fuzzy inference engine, and defuzzifier (fuzzy innovation precise output). In this embodiment, the input of the fuzzy inference system is defined by the similarity measure of each feature, and the adaptive weighting coefficient of each feature is obtained by inference.
请参阅图7,进一步该步骤S23b还包括如下子步骤:Referring to FIG. 7, the step S23b further includes the following substeps:
S23b1,计算模糊推理系统的输入变量。S23b1, calculating the input variables of the fuzzy inference system.
一并参阅图8,图8为本发明多特征线索融合的结构示意图。以运动特征相似性度量为第一模糊输入变量,其余3个特征的相似性度量均值为第二模糊输入变量,且其余3个特征的相似性度量均值的计算可定义为:Referring to FIG. 8, FIG. 8 is a schematic structural diagram of multi-feature clue fusion according to the present invention. The motion feature similarity measure is used as the first fuzzy input variable, and the similarity measure of the remaining three features is the second fuzzy input variable, and the calculation of the similarity measure mean of the other three features can be defined as:
Figure PCTCN2017091575-appb-000020
Figure PCTCN2017091575-appb-000020
Figure PCTCN2017091575-appb-000021
分别作为模糊逻辑系统的第一及第二模糊输入变量。其中,ei为特征i的相似性度量,
Figure PCTCN2017091575-appb-000022
为t-1时刻的特征k的融合系数,fM(i,j)运动特征相似性度量,
Figure PCTCN2017091575-appb-000023
为余3个特征相似性度量的加权均值。
will
Figure PCTCN2017091575-appb-000021
As the first and second fuzzy input variables of the fuzzy logic system, respectively. Where e i is a measure of similarity of feature i,
Figure PCTCN2017091575-appb-000022
The fusion coefficient of the feature k at time t-1, the f M (i, j) motion feature similarity measure,
Figure PCTCN2017091575-appb-000023
The weighted mean of the remaining three feature similarity measures.
S23b2,确定模糊推理系统的输入变量及输出变量的隶属度函数。 S23b2, determining the input function of the fuzzy inference system and the membership function of the output variable.
一般来说,输出变量的精度受模糊集数量的影响,模糊集越多,输出就越精确,但模糊集越多,算法的计算复杂度就越大,所以通常模糊集数量是由经验选取的。In general, the accuracy of the output variable is affected by the number of fuzzy sets. The more fuzzy sets, the more accurate the output, but the more fuzzy sets, the more computational complexity of the algorithm, so usually the number of fuzzy sets is selected by experience. .
请参阅图9,图9为本发明模糊输入变量fk(i,j)、
Figure PCTCN2017091575-appb-000024
的隶属度函数示意图。
Please refer to FIG. 9. FIG. 9 is a fuzzy input variable f k (i, j) according to the present invention.
Figure PCTCN2017091575-appb-000024
Schematic diagram of the membership function.
输入变量fk(i,j)和
Figure PCTCN2017091575-appb-000025
利用5个语言模糊集{ZE,SP,MP,LP,VP}进行模糊化,其隶属度函数分别用μ0,ZE(i,j)、μ0,SP(i,j)、μ0,MP(i,j)、μ0,LP(i,j)以及μ0,VP(i,j)表示,五个模糊集分别表示零、正小、正中、正大和非常大。。
Input variable f k (i,j) and
Figure PCTCN2017091575-appb-000025
The fuzzification is performed by using five language fuzzy sets {ZE, SP, MP, LP, VP}, and the membership functions are μ 0, ZE (i, j), μ 0, SP (i, j), μ 0, respectively. MP (i, j), μ 0, LP (i, j), and μ 0, VP (i, j) indicate that the five fuzzy sets represent zero, positive, medium, positive, and very large, respectively. .
请参阅图10,图10为本发明输出模糊变量αM的隶属度函数示意图。对于输出模糊变量αM包含五个模糊集:{ZE,SP,MP,LP,VP,EP},EP表示极大模糊集,其隶属度函数分别用μ1,ZE(i,j)、μ1,SP(i,j)、μ1,MP(i,j)、μ1,LP(i,j)、μ1,VP(i,j)以及μ1,EP(i,j)表示。Please refer to FIG. 10. FIG. 10 is a schematic diagram of the membership function of the output fuzzy variable α M according to the present invention. For the output fuzzy variable α M contains five fuzzy sets: {ZE, SP, MP, LP, VP, EP}, EP represents the maximum fuzzy set, and its membership function is μ 1, ZE (i, j), μ 1, SP (i, j), μ 1, MP (i, j), μ 1, LP (i, j), μ 1, VP (i, j), and μ 1, EP (i, j).
S23b3,采用所模糊推理系统的推理规则得到第三相似性度量中每一特征相似性度量的权重值。S23b3 obtains the weight value of each feature similarity measure in the third similarity measure by using the inference rule of the fuzzy inference system.
根据步骤S23b2中的定义的输入变量及输出变量的隶属度函数,模糊推理规则可以如下:,According to the input function of the input variable and the output variable defined in step S23b2, the fuzzy inference rule can be as follows:
规则1:如果fM(i,j)是ZE,并且fM(i,j)是ZE,则αM是EPRule 1: If f M (i,j) is ZE and f M (i,j) is ZE, then α M is EP
规则2:如果fM(i,j)是ZE,并且fM(i,j)是SP,则αM是VPRule 2: If f M (i,j) is ZE and f M (i,j) is SP, then α M is VP
规则3:如果fM(i,j)是ZE,并且fM(i,j)是MP,则αM是LPRule 3: If f M (i,j) is ZE and f M (i,j) is MP, then α M is LP
详细的模糊规则如表1所示:The detailed fuzzy rules are shown in Table 1:
Figure PCTCN2017091575-appb-000026
Figure PCTCN2017091575-appb-000026
在本发明一具体实施例中,以规则1为例,详细给出推理过程:In a specific embodiment of the present invention, rule 1 is taken as an example to give a detailed reasoning process:
a)根据规则1,模糊输入变量fM(i,j)对应的模糊集为ZE,可以根据图9所示模糊隶属函数,利用fM(i,j)的值求出对应的模糊隶属度值
Figure PCTCN2017091575-appb-000027
同样的方法,可以求出模糊输入变量
Figure PCTCN2017091575-appb-000028
对应的模糊隶属度值
Figure PCTCN2017091575-appb-000029
a) According to rule 1, the fuzzy set corresponding to the fuzzy input variable f M (i, j) is ZE, and the corresponding fuzzy membership degree can be obtained by using the value of f M (i, j) according to the fuzzy membership function shown in FIG. 9 . value
Figure PCTCN2017091575-appb-000027
The same method can be used to find fuzzy input variables
Figure PCTCN2017091575-appb-000028
Corresponding fuzzy membership value
Figure PCTCN2017091575-appb-000029
b)利用下式计算出规则1的适用度:b) Calculate the applicability of Rule 1 using the following formula:
Figure PCTCN2017091575-appb-000030
Figure PCTCN2017091575-appb-000030
其中,∧表示取小。Among them, ∧ means take small.
c)根据规则1,对应的模糊输出为EP,则规则1的输出可以用下式计算:c) According to rule 1, the corresponding fuzzy output is EP, then the output of rule 1 can be calculated by:
Figure PCTCN2017091575-appb-000031
Figure PCTCN2017091575-appb-000031
同样的方法,可以计算出所有规则的模糊输出变量。根据表1可知,本申请中M=25。于是,总的模糊输出为:In the same way, you can calculate the fuzzy output variables of all rules. According to Table 1, M=25 in the present application. Thus, the total fuzzy output is:
Figure PCTCN2017091575-appb-000032
Figure PCTCN2017091575-appb-000032
其中,∨表示取大。由于式(20)得到的是一个模糊化的输出,为了得到去模糊化的输出结果可以采用如下方法:Among them, ∨ means take big. Since equation (20) yields a fuzzified output, the following method can be used to obtain the defuzzified output:
Figure PCTCN2017091575-appb-000033
Figure PCTCN2017091575-appb-000033
其中,表示模糊规则m对应输出模糊集合的质心。同样的道理,针对不同特征构建模糊推理系统,可以分别得到几何形状特征、运动方向特征以及空间距离的权重值系数αA、αS以及αDWherein, the fuzzy rule m corresponds to the centroid of the output fuzzy set. For the same reason, the fuzzy inference system is constructed for different features, and the weight value coefficients α A , α S and α D of the geometric shape feature, the motion direction feature and the spatial distance can be obtained respectively.
S23c,对权重值和第三相似性度量进行多特征线索融合,以得到观测结果和预测结果之间的关联代价矩阵。S23c, performing multiple feature clue fusion on the weight value and the third similarity measure to obtain an associative cost matrix between the observation result and the prediction result.
后将所有特征的权重值系数进行归一化,得到当前时刻各特征的融合系数:Then normalize the weight value coefficients of all features to obtain the fusion coefficients of the features at the current time:
Figure PCTCN2017091575-appb-000034
Figure PCTCN2017091575-appb-000034
通过判断各个特征的可信程度,自适应给不同特征分配不同的权重,很好地解决了在复杂背景、互相遮挡情况下的跟踪问题。根据式(21)得到观测结果和预测结果之间的关联代价矩阵,定义为:By judging the credibility of each feature, adaptive assigning different weights to different features, which solves the tracking problem in complex background and mutual occlusion. According to equation (21), the correlation cost matrix between the observation result and the prediction result is obtained, which is defined as:
S=[sij]n×l  (24)S=[s ij ] n×l (24)
其中,{αk}k∈{A,M,S,D}为每一特征相似性度量的融合系数,且满足
Figure PCTCN2017091575-appb-000035
fk(i,j)k∈{A,M,S,D}为观测结果
Figure PCTCN2017091575-appb-000036
和预测结果
Figure PCTCN2017091575-appb-000037
之间的每一特征相似性度量。
Where {α k } k∈{A, M, S, D} is the fusion coefficient of each feature similarity measure and satisfies
Figure PCTCN2017091575-appb-000035
f k (i,j) k∈{A,M,S,D} are observations
Figure PCTCN2017091575-appb-000036
And forecast results
Figure PCTCN2017091575-appb-000037
A measure of similarity between each feature.
S23d,采用贪婪算法对关联代价矩阵进行优化求解,找出关联的观测结果和预测结果。S23d, using the greedy algorithm to optimize the correlation cost matrix to find the associated observations and prediction results.
采用贪婪算法实现预测结果与观测结果之间的正确的关联,从而得到预测结果与观测结果之间的关联对进一步包括:The greedy algorithm is used to achieve the correct correlation between the prediction result and the observation result, so that the correlation between the prediction result and the observation result further includes:
1)找出关联代价矩阵sij中未被标记的所有元素中的最大值。1) Find the maximum of all elements in the associated cost matrix s ij that are not marked.
找出关联代价矩阵sij中未被标记的所有元素中的最大值Spq=max([Sij]n*l),其中,p=1,2,3......n,q=1,2,3......l,并标记该关联代价矩阵S中该最大 值spq所在的第p行及第q列的所有元素。Find the maximum value Spq=max([Sij]n*l) of all elements in the associated cost matrix s ij that are not marked, where p=1, 2, 3...n, q=1 , 2, 3, ..., and mark all elements of the pth row and the qth column where the maximum value s pq is located in the associated cost matrix S.
2)判断最大值是否为所在行列中的最大值,且满足大于第二阈值。2) Determine whether the maximum value is the maximum value in the rank and column, and satisfy the greater than the second threshold.
判断该最大值spq大是否为所在行和所在列中的最大值,即是否满足:spq≥{spj}j=1,2,…l、spq≥{siq}i=1,2,…,r。进一步判断该最大值spq是否大于第二阈值λ2,即预测结果p与观测结果q的关联概率是否大于第二阈值λ2,且该第二阈值的满足λ2∈[0.6,0.9]。It is determined whether the maximum value s pq is the maximum value in the row and the column in which it is located, that is, whether: s pq ≥{s pj } j=1,2,...l , s pq ≥{s iq } i=1, 2,...,r . It is further determined whether the maximum value s pq is greater than the second threshold λ 2 , that is, whether the correlation probability of the prediction result p and the observation result q is greater than the second threshold λ 2 , and the second threshold satisfies λ 2 ∈ [0.6, 0.9].
3)若大于,则观测结果和预测结果正确关联。3) If greater than, the observations are correctly correlated with the predictions.
该最大值spq满足上述判定条件,则认为预测结果p与观测结果q和之间正确关联,则将该关联对记录进已关联预测结果和观测结果的集合中。进一步,若该关联代价矩阵Sij中还存在未被标记的行和列,则继续上述步骤1)。When the maximum value s pq satisfies the above-described determination condition, it is considered that the prediction result p is correctly associated with the observation result q and the correlation pair is recorded in the set of the associated prediction result and the observation result. Further, if there are still unmarked rows and columns in the associated cost matrix Sij, the above step 1) is continued.
上述实施方式,通过判定当前视频帧中目标的预测结果之间发生遮挡,计算预测结果和观测结果之间的第三特征相似性度量,引入模糊推理系统,利用基于模糊逻辑的方法,根据当前跟踪环境自适应给不同特征信息分配不同的权重值,得到多属性特征的加权和融合,构成该帧目标的预测结果与观测结果的关联代价矩阵,再使用贪婪算法优化求解分配,能够有效提高多目标与观测之间的正确关联。In the above implementation manner, by determining that occlusion occurs between the prediction results of the target in the current video frame, calculating a third feature similarity measure between the prediction result and the observation result, introducing a fuzzy inference system, and using the fuzzy logic based method, according to the current tracking The environment adaptively assigns different weight values to different feature information, obtains the weighted sum of the multi-attribute features, constitutes the associated cost matrix of the prediction result of the frame target and the observation result, and then uses the greedy algorithm to optimize the solution allocation, which can effectively improve the multi-objective The correct association with the observations.
请参阅图11,图11为本发明基于模糊逻辑的视频多目标跟踪方法第五实施例的流程示意图,是在本发明视频多目标模糊数据关联方法第一实施例中步骤S3的进一步扩展,本实施例进一步包括:Referring to FIG. 11 , FIG. 11 is a schematic flowchart of a fifth embodiment of a video multi-target tracking method based on the fuzzy logic, which is a further extension of step S3 in the first embodiment of the video multi-target fuzzy data association method of the present invention. Embodiments further include:
S31,通过第一相似性度量,建立终止轨迹片段及新的轨迹片段之间的模糊关联代价矩阵。S31. Establish, by using the first similarity measure, a fuzzy association cost matrix between the terminating trajectory segment and the new trajectory segment.
模糊逻辑数据关联方法能够处理在短时期内出现的高频率遮挡和大量虚假观测条件下的多目标跟踪的数据关联问题,然而在长时间的遮挡以及漏检情况下,一些目标状态长时间得不到更新,其目标运动轨迹很难维持,会出现目标轨迹断批的情况,即同一个目标拥有多条运动轨迹。同时,当新目标进入场景时,需要初始化相应的新的目标轨迹,如果目标离开场景时,也要删除相应的目标轨迹。The fuzzy logic data association method can deal with the data association problem of high-frequency occlusion and multi-target tracking under a large number of false observations in a short period of time. However, in the case of long-term occlusion and missed detection, some target states are not long. To the update, the target motion trajectory is difficult to maintain, and there will be a situation where the target trajectory is broken, that is, the same target has multiple motion trajectories. At the same time, when the new target enters the scene, the corresponding new target trajectory needs to be initialized, and if the target leaves the scene, the corresponding target trajectory is also deleted.
参阅图12,步骤S31进一步包括如下子步骤:Referring to FIG. 12, step S31 further includes the following sub-steps:
S311,建立终止轨迹片段及新的轨迹片段之间的相似性向量。S311, establishing a similarity vector between the ending track segment and the new track segment.
目标的预测结果的终止就是对于场景中离开的目标,或者是静止不动的目标,需要将其从当前的目标跟踪序列中删除。假如目标的估计位置位于视频场景的边缘位置(场景边缘设置为τborder=5),则可以判断为目标离开视频场景,此时将该目标从当前的目标跟踪序列中删除。如果目标的估计位置不在视频场景的边缘,而且目标连续x帧都没有与任何观测进行关联,那么就能推断出目标 静止或者是被遮挡,将该目标从当前的目标跟踪序列中删除。The termination of the target's prediction result is for the target that is left in the scene, or is a stationary target, which needs to be removed from the current target tracking sequence. If the estimated position of the target is at the edge of the video scene (the edge of the scene is set to τ border = 5), then it can be determined that the target leaves the video scene, and the target is deleted from the current target tracking sequence. If the estimated position of the target is not at the edge of the video scene and the target consecutive x frames are not associated with any observations, then the target can be inferred to be stationary or occluded, and the target is removed from the current target tracking sequence.
如果在场景区域内有未关联上的观测,可以通过判断观测结果是否能够关联上目标来确认是否有新的目标出现。在复杂环境下,由于背景干扰、目标自身形变等多种因素的影响,在保持高检测率的条件下,目标检测器不可避免的会产生一些虚假观测,其不会关联上任何已有的目标,这些虚假观测也可能会被错误的初始化为新的目标。一般来说,在连续的几帧内(时间滑动窗内)目标会有重叠的面积和相同的几何尺寸,因此为了能够准确的判断未被关联上的观测是否来源于新的目标,本申请在新目标初始化模块利用连续Tinit帧内的观测结果来判断是否存在矩形框面积重叠以及具有相同的尺寸,定义观测结果之间矩形框的面积重叠率为:If there are unrelated observations in the scene area, it can be confirmed whether a new target appears by judging whether the observation can be associated with the target. In a complex environment, due to various factors such as background interference and target deformation, the target detector will inevitably produce some false observations under the condition of maintaining high detection rate, which will not be associated with any existing targets. These false observations may also be incorrectly initialized to new targets. In general, within a few consecutive frames (within the time sliding window) the target will have overlapping areas and the same geometric dimensions, so in order to be able to accurately determine whether the unrelated observations are derived from new targets, the present application is The new target initialization module uses the observations in the continuous T init frame to determine whether there is an overlap of the rectangular frame areas and the same size, and the area overlap ratio of the rectangular frame between the observation results is defined as:
Figure PCTCN2017091575-appb-000038
Figure PCTCN2017091575-appb-000038
观测结果之间矩形框的尺寸相似度为:The dimensional similarity between the observations is:
Figure PCTCN2017091575-appb-000039
Figure PCTCN2017091575-appb-000039
其中,
Figure PCTCN2017091575-appb-000040
分别为t时刻和t+1时刻的观测值,area(□)表示观测结果的面积,
Figure PCTCN2017091575-appb-000041
表示观测值
Figure PCTCN2017091575-appb-000042
Figure PCTCN2017091575-appb-000043
的重叠面积,h为观测值矩形框的高度值。
among them,
Figure PCTCN2017091575-appb-000040
The observation values at time t and time t+1 are respectively, and area(□) indicates the area of the observation result.
Figure PCTCN2017091575-appb-000041
Representation of observations
Figure PCTCN2017091575-appb-000042
versus
Figure PCTCN2017091575-appb-000043
The overlap area, h is the height value of the observation rectangle.
Figure PCTCN2017091575-appb-000044
Figure PCTCN2017091575-appb-000044
其中,τω和τr分别代表重叠率阈值和尺寸相似度阈值。对于连续init帧内的观测值的面积重叠率和尺寸相似度均大于设定的阈值,即init当大于或等于Tinit时,则将其转化为有效轨迹,即起始一条新的轨迹片段,并将其加入目标跟踪序列中。因此,该方法可以有效剔除目标检测器产生的虚假观测,从而减少了错误的目标轨迹起始。Where τ ω and τ r represent the overlap rate threshold and the size similarity threshold, respectively. The area overlap ratio and the size similarity of the observations in the continuous init frame are greater than the set threshold, that is, when the init is greater than or equal to T init , it is converted into a valid track, that is, a new track segment is started. And add it to the target tracking sequence. Therefore, the method can effectively eliminate false observations generated by the target detector, thereby reducing the erroneous target trajectory start.
其中,由于目标终止轨迹可能是一条轨迹片段或者一条完整的目标轨迹,为了确认目标轨迹的完整性,利用终止轨迹最后的位置来判断轨迹在场景中断开或者离开场景。如果终止轨迹最后的位置在场景内,则其轨迹为终止轨迹片段。同时,当目标的轨迹片段的起始帧为当前时刻时,说明该新的轨迹片段是一个新的观测产生的临时轨迹。Wherein, since the target termination trajectory may be a trajectory segment or a complete target trajectory, in order to confirm the integrity of the target trajectory, the last position of the termination trajectory is used to determine whether the trajectory is disconnected or leaves the scene in the scene. If the last position of the end track is within the scene, its trajectory is the end track segment. Meanwhile, when the start frame of the target track segment is the current time, it indicates that the new track segment is a new observed temporary track.
在本发明一具体实施例中,终止轨迹片段的集合定义为:
Figure PCTCN2017091575-appb-000045
新的轨迹片段的集合定义为:
Figure PCTCN2017091575-appb-000046
其中,na、nb分别表示终止轨迹片段集合及所述新的轨迹片段集合的个数。
In a specific embodiment of the invention, the set of termination trajectory segments is defined as:
Figure PCTCN2017091575-appb-000045
The set of new track fragments is defined as:
Figure PCTCN2017091575-appb-000046
Where n a and n b respectively represent the number of the end track segment set and the new track segment set.
第一相似性度量包括外观相似性度量、形状相似性度量以及运动相似性度量,其中,外观相似性度量定义为: The first similarity measure includes an appearance similarity measure, a shape similarity measure, and a motion similarity measure, wherein the appearance similarity measure is defined as:
Figure PCTCN2017091575-appb-000047
Figure PCTCN2017091575-appb-000047
其中,ρ(·)表示为Bhattacharyya系数,Hc(·)表示背景加权的颜色直方图特征,
Figure PCTCN2017091575-appb-000048
为方差常量,Hg(·)表示方向梯度直方图特征,
Figure PCTCN2017091575-appb-000049
为方差常量;
Where ρ(·) is expressed as a Bhattacharyya coefficient, and H c (·) is a background weighted color histogram feature.
Figure PCTCN2017091575-appb-000048
For the variance constant, H g (·) represents the direction gradient histogram feature,
Figure PCTCN2017091575-appb-000049
Is a variance constant;
形状相似性度量定义为:The shape similarity measure is defined as:
Figure PCTCN2017091575-appb-000050
Figure PCTCN2017091575-appb-000050
其中,hi表示终止轨迹片段Ti在图像中的高度,hj表示新的轨迹片段Tj在图像中的高度,
Figure PCTCN2017091575-appb-000051
为方差常量;
Where h i represents the height of the end track segment T i in the image, and h j represents the height of the new track segment T j in the image,
Figure PCTCN2017091575-appb-000051
Is a variance constant;
运动相似性度量定义为:The motion similarity measure is defined as:
Figure PCTCN2017091575-appb-000052
Figure PCTCN2017091575-appb-000052
其中,G(□)表示高斯分布,∑为高斯分布的方差,Δt是终止轨迹片段Ti最后观测到新的轨迹片段Tj第一个观测的帧间隔、
Figure PCTCN2017091575-appb-000053
vi分别为终止轨迹片段Ti终止位置和速度,
Figure PCTCN2017091575-appb-000054
vj分别为新的轨迹片段起始位置和速度。
Where G(□) represents a Gaussian distribution, ∑ is the variance of the Gaussian distribution, and Δt is the first observed frame interval of the last trajectory segment T i and the new trajectory segment T j is observed.
Figure PCTCN2017091575-appb-000053
v i is the end position and velocity of the terminating track segment T i , respectively.
Figure PCTCN2017091575-appb-000054
v j is the starting position and velocity of the new track segment.
图13为遮挡情况下终止轨迹片段和新的轨迹片段运动相似性度量。假定预测结果的位置和实际观测结果的位置之间的误差满足高斯分布,即当终止轨迹片段的预测位置与新的轨迹片段的实际位置距离越小,则两条轨迹片段之间的运动相似性就越大(例如
Figure PCTCN2017091575-appb-000055
Figure PCTCN2017091575-appb-000056
之间的距离越相近,
Figure PCTCN2017091575-appb-000057
的值就越大)。
Figure 13 is a motion similarity measure for the end track segment and the new track segment in the occlusion case. It is assumed that the error between the position of the predicted result and the position of the actual observation satisfies the Gaussian distribution, that is, the smaller the distance between the predicted position of the terminated trajectory segment and the actual position of the new trajectory segment, the similarity between the two trajectory segments The bigger (for example
Figure PCTCN2017091575-appb-000055
versus
Figure PCTCN2017091575-appb-000056
The closer the distance is,
Figure PCTCN2017091575-appb-000057
The value is larger).
进一步,根据式(1)、式(2)以及式(3)可以计算得到两条轨迹片段间的相似性向量,定义为:Further, according to equations (1), (2), and (3), a similarity vector between two pieces of track segments can be calculated, which is defined as:
Figure PCTCN2017091575-appb-000058
Figure PCTCN2017091575-appb-000058
其中,Λk(Ti,Tj)∈[0,1]3中,τgap为关联的时间间隔阈值,
Figure PCTCN2017091575-appb-000059
表示终止轨迹片段Ti断开的时间帧,
Figure PCTCN2017091575-appb-000060
表示新的轨迹片段Tj起始的时间帧。
Where Λ k (T i , T j ) ∈ [0, 1] 3 , τ gap is the associated time interval threshold,
Figure PCTCN2017091575-appb-000059
a time frame indicating that the trajectory segment T i is broken,
Figure PCTCN2017091575-appb-000060
Indicates the time frame at which the new track segment T j starts.
S312,利用相似性向量计算终止轨迹片段及新的轨迹片段之间的匹配度。S312. Calculate the degree of matching between the end track segment and the new track segment by using the similarity vector.
为了得到任意新的轨迹片段与终止轨迹片段的相似度,本申请采用基于模糊综合函数的模糊性模型来衡量终止轨迹片段和新的轨迹片段之间的匹配度,其定义为:In order to obtain the similarity between any new trajectory segment and the ending trajectory segment, the present application uses a fuzzy comprehensive model based fuzzy model to measure the degree of matching between the ending trajectory segment and the new trajectory segment, which is defined as:
Figure PCTCN2017091575-appb-000061
Figure PCTCN2017091575-appb-000061
其中,∧表示所述匹配度取最小值,∨表示所述匹配度取最大值。Where ∧ indicates that the matching degree takes a minimum value, and ∨ indicates that the matching degree takes a maximum value.
S313,根据匹配度计算终止轨迹片段及新的轨迹片段之间的模糊综合相似 度。S313, calculating a fuzzy comprehensive similarity between the ending trajectory segment and the new trajectory segment according to the matching degree degree.
k时刻终止轨迹片段Ti和新的轨片段Tj之间的模糊综合相似度定义为:The fuzzy comprehensive similarity between the k-terminating trajectory segment T i and the new orbital segment T j is defined as:
Figure PCTCN2017091575-appb-000062
Figure PCTCN2017091575-appb-000062
S314,根据模糊综合相似度建立终止轨迹片段及新的轨迹片段的关联代价矩阵。S314. Establish an associative cost matrix of the terminating trajectory segment and the new trajectory segment according to the fuzzy comprehensive similarity.
终止轨迹片段及新的轨迹片段之间的关联代价矩阵定义为:The associated cost matrix between the terminating trajectory segment and the new trajectory segment is defined as:
Figure PCTCN2017091575-appb-000063
Figure PCTCN2017091575-appb-000063
且两条轨迹片段可以实现关联的前提条件为:The prerequisites for the two track segments to be associated are:
1)时间具有连续性,即相对应的时间帧区间没有发生重叠区域,即
Figure PCTCN2017091575-appb-000064
1) The time has continuity, that is, there is no overlapping area in the corresponding time frame interval, ie
Figure PCTCN2017091575-appb-000064
2)两条轨迹片段之间的时间间隔应该在关联的时间间隔阈值范围之内,即满足
Figure PCTCN2017091575-appb-000065
2) The time interval between the two track segments should be within the associated time interval threshold, ie
Figure PCTCN2017091575-appb-000065
在目标跟踪的过程中,假如目标的预测结果发生遮挡、目标检测误差以及漏检等原因造成了目标的运动轨迹发生断开,那么其断开之后新的轨迹与原始的终止轨迹之间的时间间隔相对来说是比较短的。如果这两条轨迹片段之间的时间间隔相对来说比较长,那么可以认为他们不是来源于同一个目标的。本申请中可以通过设定合理的关联时间间隔阈值τgap能够在一个相对较小的范围之内,将可能会被关联上的轨迹进行关联,这样可以很好的提高了算法的时间效率,也排除了一些不可能成功关联上的轨迹片段。In the process of target tracking, if the target's prediction result is occlusion, target detection error, and missed detection, etc., the target's motion trajectory is disconnected, then the time between the new trajectory and the original termination trajectory after disconnection The interval is relatively short. If the time interval between the two track segments is relatively long, then it can be considered that they are not from the same target. In the present application, the reasonable correlation time interval threshold τ gap can be set to associate the trajectories that may be associated within a relatively small range, which can improve the time efficiency of the algorithm. Excludes some trace segments that cannot be successfully associated.
S32,采用最大模糊综合相似度和阈值判别原则实现终止轨迹片段及新的轨迹片段之间的轨迹关联。S32, using the maximum fuzzy comprehensive similarity and the threshold discriminant principle to implement the trajectory association between the ending trajectory segment and the new trajectory segment.
根据模糊关联代价矩阵U可知,由于目标跟踪环境的复杂性,在轨迹片段关联中为了给出终止轨迹片段Ti与新的轨迹片段Tj之间的相似性判决,需要利用模糊算子去模糊,最大综合相似度表示为:According to the fuzzy correlation cost matrix U, due to the complexity of the target tracking environment, in order to give the similarity judgment between the ending trajectory segment T i and the new trajectory segment T j in the trajectory segment association, it is necessary to use the fuzzy operator to deblur The maximum comprehensive similarity is expressed as:
Figure PCTCN2017091575-appb-000066
Figure PCTCN2017091575-appb-000066
如果in case
μij*≥ε  (29)μ ij* ≥ ε (29)
则终止轨迹片段Ti与新的轨迹片段Tj*关联,并且新的轨迹片段Tj*不再与其他终止轨迹片段Ti关联,否则与为不关联轨迹片段,这里ε为阈值参数,且0≤ε≤1。Then the terminating trajectory segment T i is associated with the new trajectory segment T j* , and the new trajectory segment T j* is no longer associated with other terminating trajectory segments T i , otherwise it is not associated with the trajectory segment, where ε is a threshold parameter, and 0 ≤ ε ≤ 1.
S33,填充关联上的终止轨迹片段及新的轨迹片段之间缺失的轨迹段。S33, filling the trailing track segment on the association and the missing track segment between the new track segments.
由于目标的预测结果之间发生遮挡、目标检测误差以及漏检等原因造成了目标的运动轨迹发生断开,采用上述关联方法可以将两条断开的轨迹关联在一起,但是两条轨迹片段之间往往还缺少若干帧丢失的检测点信息。因此,该目 标不能形成一条完整连续的轨迹,还需要对他们间的空缺处进行预测填充。Due to occlusion, target detection error and missed detection between the target's prediction results, the target's motion trajectory is disconnected. The above correlation method can be used to associate two broken trajectories together, but the two trajectory segments There is often a lack of detection point information for lost frames. Therefore, the item The target does not form a complete continuous trajectory, and it also needs to predict the filling of the gap between them.
参阅图14,步骤S33包括如下子步骤:Referring to Figure 14, step S33 includes the following sub-steps:
S331,对关联上的终止轨迹片段及新的轨迹片段之间的缺失轨迹段进行双向预测,以获取预测点的位置信息。S331: Perform bidirectional prediction on the missing track segment between the associated ending track segment and the new track segment to obtain position information of the predicted point.
图15为获取丢失预测点的位置示意图,Tf是两条断开的轨迹中的前面一条轨迹片段,即终止轨迹片段,Tb是后面的一条轨迹片段,即新的轨迹片段。利用同一目标发生断开的两条轨迹的终止位置、新起始位置以及速度信息,双向连续的预测目标在断开时间间隔内的位置。预测点的位置信息的获取过程如图15所示。pf表示当采用轨迹片段Tf进行正向预测时目标所在的具体位置,pb表示当采用轨迹片段Tb进行反向预测时目标的具体位置,tf表示Tf进行正向预测时当前帧数,tb表示Tb进行反向预测时当前帧数,则获取预测点位置信息的过程如下:15 is a schematic view of the position acquisition loss prediction point, T f in front of the two tracks off a track segment, the track segment is terminated, T b is the back of a track segment, i.e. the new track segment. The position of the two-way continuous predicted target within the disconnection time interval is determined by the end position, the new starting position, and the speed information of the two trajectories where the same target is broken. The process of acquiring the position information of the predicted point is as shown in FIG. p f represents a specific location when using the track segment T f for the target where forward prediction, p b represents a specific location when using the track segment T b of the reverse prediction target, t f T f represents a forward prediction for the current The number of frames, t b represents the current number of frames when T b is used for backward prediction. The process of obtaining the predicted position information is as follows:
1)初始化: 1) Initialization:
2)若tf<tb,则从Pf进行正向预测目标在下一帧中的具体位置:2) If t f <t b , the forward prediction target from P f is in the specific position in the next frame:
pf=pf+vf,tf=tf+1  (30)p f =p f +v f ,t f =t f +1 (30)
从Pb进行反向预测目标在前一帧中的具体位置:The specific position of the reverse prediction target from the previous frame in P b :
pb=pb-vb,tb=tb-1  (31)p b =p b -v b ,t b =t b -1 (31)
重复步骤2),直至tf≥tb,最后得到两条轨迹片段间的缺失点的位置信息。Repeat step 2) until t f ≥ t b , and finally obtain the position information of the missing points between the two track segments.
S332,获取预测点的矩形框信息。S332. Obtain rectangular frame information of the predicted point.
为了能够对跟踪算法的多目标跟踪精度进行评估,还需要获取预测点目标的矩形框的宽高,在本申请中采用平均法来得到预测点的矩形框的宽高,为:In order to be able to evaluate the multi-target tracking accuracy of the tracking algorithm, it is also necessary to obtain the width and height of the rectangular frame of the predicted point target. In this application, the averaging method is used to obtain the width and height of the rectangular frame of the predicted point, which is:
Figure PCTCN2017091575-appb-000068
Figure PCTCN2017091575-appb-000068
Figure PCTCN2017091575-appb-000069
Figure PCTCN2017091575-appb-000069
其中,hk、wk表示第k帧时检测点的矩形框的高度和宽度,
Figure PCTCN2017091575-appb-000070
表示轨迹片段Tf尾部的矩形框的高度和宽度,
Figure PCTCN2017091575-appb-000071
表示轨迹片段Tb头部的矩形框的高度和宽度。
Where h k and w k represent the height and width of the rectangular frame of the detection point in the kth frame.
Figure PCTCN2017091575-appb-000070
The height and width of the rectangle representing the end of the track segment Tf ,
Figure PCTCN2017091575-appb-000071
The height and width of the rectangular box representing the head of the track segment T b .
S333,根据预测点的位置信息及矩形框信息对缺失轨迹段进行填充。S333: Fill the missing track segment according to the position information of the predicted point and the rectangular frame information.
运用上述的预测点填充方法对轨迹片段间的缺失点进行预测填充后,就可以获得目标的一条完整连续的运动轨迹。By using the above predicted point filling method to predict and fill the missing points between the track segments, a complete continuous motion track of the target can be obtained.
在本发明的实际运用中,对已经关联上的目标的预测结果和观测结果采用滤波器进行滤波和预测,以得到目标当前视频帧中的实际轨迹点以及预测结果, 其中,本申请中采用的滤波器可以包括但不限于卡尔曼(Kalman)滤波器。进一步,对没有关联上目标的预测结果进行外推预测,得到其预测结果,实现对多目标的准确跟踪。且目标的预测结果用于下一帧视频帧中的数据关联。In the practical application of the present invention, the prediction result and the observation result of the already associated target are filtered and predicted by a filter to obtain an actual track point and a prediction result in the current video frame of the target, Among them, the filter used in the present application may include, but is not limited to, a Kalman filter. Further, extrapolation prediction is performed on the prediction result without the associated target, and the prediction result is obtained, so as to accurately track the multi-target. And the prediction result of the target is used for data association in the next frame of the video frame.
上述实施方式,对同一目标的断开轨迹间的缺失点进行预测填充,形成完整连续的目标轨迹,能有效解决目标轨迹的平滑与预测、目标轨迹的终止以及新目标轨迹的起始等问题。In the above embodiment, the missing points between the broken trajectories of the same target are predicted and filled, and a complete continuous target trajectory is formed, which can effectively solve the problems of smoothing and predicting the target trajectory, terminating the target trajectory, and starting the new target trajectory.
如图16所示,图16为基于模糊逻辑的视频多目标跟踪装置第一实施例的结构示意图,包括:As shown in FIG. 16, FIG. 16 is a schematic structural diagram of a first embodiment of a video multi-target tracking apparatus based on fuzzy logic, including:
检测模块11,用于对当前视频帧进行在线目标运动检测,检测得到的可能运动对象作为观测结果。The detecting module 11 is configured to perform online target motion detection on the current video frame, and detect the obtained possible moving object as an observation result.
关联模块12,用于对观测结果和目标的预测结果进行数据关联,其中预测结果是至少利用前一视频帧的目标的轨迹进行预测而得到的。The association module 12 is configured to perform data association between the observation result and the prediction result of the target, wherein the prediction result is obtained by predicting at least the trajectory of the target of the previous video frame.
轨迹管理模块13,用于对未被关联上的预测结果及观测结果进行轨迹管理,包括利用未被关联上预测结果获取终止轨迹片段以及利用未被关联上的观测结果获取新的轨迹片段,对终止轨迹片段及新的轨迹片段进行轨迹关联。The trajectory management module 13 is configured to perform trajectory management on the unrelated prediction results and the observation results, including acquiring the ending trajectory segments by using the unrelated prediction results, and acquiring the new trajectory segments by using the unrelated observation results, Terminate the track segment and the new track segment for track association.
如图17所示,图17为本发明基于模糊逻辑的视频多目标跟踪装置第二实施例的结构示意图,包括:处理器110和摄像机120。As shown in FIG. 17, FIG. 17 is a schematic structural diagram of a second embodiment of a video multi-target tracking apparatus based on fuzzy logic according to the present invention, including: a processor 110 and a camera 120.
其中,摄像机120可以为本地摄像机,处理器110通过总线连接摄像机120;摄像机120也可以为远程摄像机,处理器110通过局域网或互联网连接摄像机120。The camera 120 can be a local camera, the processor 110 is connected to the camera 120 through a bus; the camera 120 can also be a remote camera, and the processor 110 is connected to the camera 120 via a local area network or the Internet.
处理器110控制基于模糊逻辑的视频多目标跟踪装置的操作,处理器110还可以称为CPU(Central Processing Unit,中央处理单元)。处理器110可能是一种集成电路芯片,具有信号的处理能力。处理器110还可以是通用处理器、数字信号处理器(DSP)、专用集成电路(ASIC)、现成可编程门阵列(FPGA)或者其他可编程逻辑器件、分立门或者晶体管逻辑器件、分立硬件组件。通用处理器可以是微处理器或者该处理器也可以是任何常规的处理器等。The processor 110 controls the operation of the fuzzy logic based video multi-target tracking device, which may also be referred to as a CPU (Central Processing Unit). Processor 110 may be an integrated circuit chip with signal processing capabilities. The processor 110 can also be a general purpose processor, a digital signal processor (DSP), an application specific integrated circuit (ASIC), an off-the-shelf programmable gate array (FPGA) or other programmable logic device, a discrete gate or transistor logic device, and discrete hardware components. . The general purpose processor may be a microprocessor or the processor or any conventional processor or the like.
基于模糊逻辑的视频多目标跟踪装置可以进一步包括存储器(图中未画出),存储器用于存储处理器110工作所必需的指令及数据,也可以存储传输器120拍摄的视频数据。The video multi-target tracking device based on the fuzzy logic may further include a memory (not shown) for storing instructions and data necessary for the operation of the processor 110, and may also store video data captured by the transmitter 120.
处理器110用于对从摄像机120获取的当前视频帧进行在线目标运动检测,检测得到的可能运动对象作为观测结果;对观测结果和目标的预测结果进行数据关联,其中预测结果是至少利用前一视频帧的目标的轨迹进行预测而得到的;对未被关联上的预测结果及观测结果进行轨迹管理,包括利用未被关联上预测结果获取终止轨迹片段以及利用未被关联上的观测结果获取新的轨迹片段,对终止轨迹片段及新的轨迹片段进行轨迹关联。The processor 110 is configured to perform online target motion detection on the current video frame acquired from the camera 120, and detect the obtained possible moving object as an observation result; perform data association on the observation result and the target prediction result, wherein the prediction result is at least the previous one. The trajectory of the target of the video frame is predicted; the trajectory management is performed on the unconstrained prediction result and the observation result, including the use of the unrelated prediction result to obtain the termination trajectory fragment and the use of the unrelated observation result to acquire the new trajectory The trajectory segment, the trajectory association of the ending trajectory segment and the new trajectory segment.
本发明基于模糊逻辑的视频多目标跟踪装置包括的各部分的功能可参考本 发明基于模糊逻辑的视频多目标跟踪方法各对应实施例中的描述,在此不再赘述。The function of each part included in the video multi-target tracking device based on the fuzzy logic of the present invention can refer to the present The description of the video multi-target tracking method based on the fuzzy logic in the corresponding embodiments is not described herein.
综上所述,本领域技术人员容易理解,本发明提供一种基于模糊逻辑的视频多目标跟踪方法及装置,通过当前视频帧中的观测结果和目标的预测结果进行数据关联,并对未关联上的观测结果和预测结果进行轨迹管理,能有效提高多目标与观测之间的正确关联,对表观相似、频繁交互、遮挡以及背景干扰等情况下的多目标进行准确跟踪,具有较强的鲁棒性和准确性。In summary, those skilled in the art can easily understand that the present invention provides a video multi-target tracking method and apparatus based on fuzzy logic, which performs data association by observing results in a current video frame and prediction results of a target, and is not associated. The observation and prediction results on the trajectory management can effectively improve the correct correlation between multi-objectives and observations, and accurately track multiple targets under the conditions of apparent similarity, frequent interaction, occlusion and background interference. Robustness and accuracy.
以上所述仅为本发明的实施方式,并非因此限制本发明的专利范围,凡是利用本发明说明书及附图内容所作的等效结构或等效流程变换,或直接或间接运用在其他相关的技术领域,均同理包括在本发明的专利保护范围内。 The above is only the embodiment of the present invention, and is not intended to limit the scope of the invention, and the equivalent structure or equivalent process transformations made by the description of the invention and the drawings are directly or indirectly applied to other related technologies. The fields are all included in the scope of patent protection of the present invention.

Claims (11)

  1. 一种基于模糊逻辑的视频多目标跟踪方法,其特征在于,所述方法包括:A video multi-target tracking method based on fuzzy logic, characterized in that the method comprises:
    对当前视频帧进行在线目标运动检测,检测得到的可能运动对象作为观测结果;Performing on-line target motion detection on the current video frame, and detecting the possible moving object as an observation result;
    对所述观测结果和目标的预测结果进行数据关联,其中所述预测结果是至少利用前一视频帧的目标的轨迹进行预测而得到的;Performing data association on the observation result and the prediction result of the target, wherein the prediction result is obtained by predicting at least the trajectory of the target of the previous video frame;
    对未被关联上的所述预测结果及所述观测结果进行轨迹管理,包括利用所述未被关联上所述预测结果获取终止轨迹片段以及利用所述未被关联上的所述观测结果获取新的轨迹片段,对所述终止轨迹片段及所述新的轨迹片段进行轨迹关联。Performing trajectory management on the prediction result and the observation result that are not associated, including acquiring the termination trajectory segment by using the prediction result that is not associated with the prediction result, and acquiring the new trajectory by using the observation result that is not associated a trajectory segment, the trajectory association of the terminating trajectory segment and the new trajectory segment.
  2. 根据权利要求1所述的方法,其特征在于,对所述观测结果和目标的预测结果进行数据关联包括:The method according to claim 1, wherein the data association between the observation result and the prediction result of the target comprises:
    计算所述当前视频帧中不同目标的预测结果之间的遮挡度;Calculating an occlusion degree between prediction results of different targets in the current video frame;
    根据所述遮挡度分别判断每一所述预测结果与其他所述预测结果之间是否发生遮挡;Determining whether occlusion occurs between each of the prediction results and the other prediction results according to the occlusion degree;
    若所述预测结果与任何其他所述预测结果之间均未发生遮挡,则对所述预测结果和所述观测结果进行第一数据关联;若所述预测结果与其他所述预测结果之间发生遮挡,则对所述预测结果和所述观测结果进行第二数据关联,其中,所述第一数据关联和所述第二数据关联不同。If no occlusion occurs between the prediction result and any other of the prediction results, performing a first data association on the prediction result and the observation result; if the prediction result occurs between the prediction result and the other prediction result And occluding, the second data association is performed on the prediction result and the observation result, wherein the first data association and the second data association are different.
  3. 根据权利要求2所述的方法,其特征在于,若所述预测结果与任何其他所述预测结果之间均未发生遮挡,则对所述预测结果和所述观测结果进行第一数据关联包括:The method according to claim 2, wherein if no occlusion occurs between the prediction result and any other of the prediction results, performing a first data association between the prediction result and the observation result comprises:
    计算所述观测结果和所述预测结果之间的第二相似性度量,所述第二相似性度量包括空间距离特征相似性度量以及外观特征相似性度量;Computing a second similarity measure between the observation result and the prediction result, the second similarity measure including a spatial distance feature similarity measure and an appearance feature similarity measure;
    利用所述第二相似性度量计算所述观测结果和所述预测结果之间的关联代价矩阵;Calculating an associated cost matrix between the observation result and the prediction result by using the second similarity measure;
    采用贪婪算法对所述关联代价矩阵进行优化求解,找出关联的所述观测结果和所述预测结果。The correlation cost matrix is optimized by using a greedy algorithm to find the associated observation result and the prediction result.
  4. 根据权利要求2所述的方法,其特征在于,观测结果d与预测结果o之间的所述空间距离特征相似性度量fD(·)定义为:The method according to claim 2, wherein said spatial distance feature similarity measure f D (·) between the observation result d and the prediction result o is defined as:
    Figure PCTCN2017091575-appb-100001
    Figure PCTCN2017091575-appb-100001
    其中,||·||2为二范数,(xo,yo)为所述预测结果o的中心坐标,(xd,yd)为所述观测结果d的中心坐标,ho为所述预测结果o的高度,
    Figure PCTCN2017091575-appb-100002
    为方差常量;
    Where ||·|| 2 is a two-norm, (x o , y o ) is the central coordinate of the prediction result o, and (x d , y d ) is the central coordinate of the observation d, h o is The height of the prediction result o,
    Figure PCTCN2017091575-appb-100002
    Is a variance constant;
    所述观测结果d与所述预测结果o之间的所述外观特征相似性度量fS(·)定义 为:The appearance feature similarity measure f S (·) between the observation d and the prediction result o is defined as:
    Figure PCTCN2017091575-appb-100003
    Figure PCTCN2017091575-appb-100003
    其中,hd为所述观测结果d的高度,
    Figure PCTCN2017091575-appb-100004
    为方差常量;
    Where h d is the height of the observation d
    Figure PCTCN2017091575-appb-100004
    Is a variance constant;
    利用所述第二相似性度量计算所述观测结果和所述预测结果之间的关联代价矩阵包括:Calculating the associated cost matrix between the observation result and the prediction result by using the second similarity measure includes:
    采用乘性融合对所述空间距离特征相似性度量以及外观特征相似性度量进行融合,以得到所述观测结果和所述预测结果之间的关联度,定义为:The spatial distance feature similarity measure and the appearance feature similarity measure are merged by multiplicative fusion to obtain the degree of association between the observation result and the predicted result, which is defined as:
    sij=fD(o,d)×fs(o,d) (3)s ij =f D (o,d)×f s (o,d) (3)
    根据所述关联度得到所述观测结果和所述预测结果之间的关联代价矩阵,定义为:Obtaining an associative cost matrix between the observation result and the prediction result according to the degree of association, which is defined as:
    S=[sij]n×l (4)S=[s ij ] n×l (4)
    其中,i=1,2,…n,j=1,2,…,l;Where i=1,2,...n,j=1,2,...,l;
    所述采用贪婪算法对所述关联代价矩阵进行优化求解,找出关联的观测结果和预测结果包括:The greedy algorithm is used to optimize the correlation cost matrix to find related observation results and prediction results including:
    找出所述关联代价矩阵S中未被标记的所有元素中的最大值;Finding a maximum of all elements in the associated cost matrix S that are not marked;
    判断所述最大值是否为所在行列中的最大值,且大于第一阈值;Determining whether the maximum value is a maximum value in a row and column, and is greater than a first threshold;
    若大于,则所述观测结果和所述预测结果正确关联。If greater, the observation is correctly associated with the prediction.
  5. 根据权利要求2所述的方法,其特征在于,若所述预测结果与其他所述预测结果之间发生遮挡,则对所述预测结果和所述观测结果进行第二数据关联包括:The method according to claim 2, wherein if an occlusion occurs between the prediction result and the other of the prediction results, performing a second data association between the prediction result and the observation result comprises:
    计算所述观测结果和所述预测结果之间的第三相似性度量,所述第三相似性度量包括外观特征相似性度量、几何形状特征相似性度量、运动特征相似性度量以及空间距离特征相似性度量;Calculating a third similarity measure between the observation result and the prediction result, the third similarity measure including an appearance feature similarity measure, a geometric feature similarity measure, a motion feature similarity measure, and a spatial distance feature similarity Sex measure
    采用模糊推理系统模型计算所述第三相似性度量中每一特征相似性度量的权重值;Calculating a weight value of each feature similarity measure in the third similarity measure by using a fuzzy inference system model;
    对所述权重值和所述第三相似性度量进行多特征线索融合,以得到所述观测结果和所述预测结果之间的关联代价矩阵;Performing multi-feature clue fusion on the weight value and the third similarity measure to obtain an associative cost matrix between the observation result and the prediction result;
    采用贪婪算法对所述关联代价矩阵进行优化求解,找出关联的所述观测结果和所述预测结果。The correlation cost matrix is optimized by using a greedy algorithm to find the associated observation result and the prediction result.
  6. 根据权利要求5所述的方法,其特征在于,观测结果d与预测结果o之间的外观特征相似性度量fA(·)定义为:The method according to claim 5, characterized in that the appearance feature similarity measure f A (·) between the observation result d and the prediction result o is defined as:
    Figure PCTCN2017091575-appb-100005
    Figure PCTCN2017091575-appb-100005
    其中,ρ(·)为巴氏系数,Hc(·)为所述当前视频帧图像背景加权的颜色直方图特征,Hg(·)为分块梯度方向直方图特征,
    Figure PCTCN2017091575-appb-100006
    为方差常量,
    Figure PCTCN2017091575-appb-100007
    为方差常量;
    Where ρ(·) is a Barthel's coefficient, H c (·) is a color histogram feature weighted by the background of the current video frame image, and H g (·) is a histogram of the block gradient direction histogram.
    Figure PCTCN2017091575-appb-100006
    Is a variance constant,
    Figure PCTCN2017091575-appb-100007
    Is a variance constant;
    所述观测结果d与所述预测结果o之间的运动特征相似性度量fM(·)定义为:The motion feature similarity measure f M (·) between the observation d and the prediction result o is defined as:
    Figure PCTCN2017091575-appb-100008
    Figure PCTCN2017091575-appb-100008
    其中,(x′o,y′o)为上一时刻所述预测结果o的中心坐标,(xo,yo)为所述预测结果o的中心坐标,
    Figure PCTCN2017091575-appb-100009
    为上一时刻所述预测结果o的速度在坐标轴上的投影,
    Figure PCTCN2017091575-appb-100010
    为方差常量;
    Where (x' o , y' o ) is the central coordinate of the prediction result o at the previous moment, and (x o , y o ) is the central coordinate of the prediction result o,
    Figure PCTCN2017091575-appb-100009
    The projection of the speed of the prediction result o on the coordinate axis for the previous moment,
    Figure PCTCN2017091575-appb-100010
    Is a variance constant;
    所述观测结果d与所述预测结果o之间的所述空间距离特征相似性度量fD(·)定义为:The spatial distance feature similarity measure f D (·) between the observation d and the prediction result o is defined as:
    Figure PCTCN2017091575-appb-100011
    Figure PCTCN2017091575-appb-100011
    其中,||·||2为二范数,(xo,yo)为所述预测结果o的中心坐标,(xd,yd)为所述观测结果d的中心坐标,ho为所述预测结果o的高度,
    Figure PCTCN2017091575-appb-100012
    为方差常量;
    Where ||·|| 2 is a two-norm, (x o , y o ) is the central coordinate of the prediction result o, and (x d , y d ) is the central coordinate of the observation d, h o is The height of the prediction result o,
    Figure PCTCN2017091575-appb-100012
    Is a variance constant;
    所述观测结果d与所述预测结果o之间的所述几何形状特征相似性度量fS(·)定义为:The geometric feature similarity measure f S (·) between the observation d and the predicted o is defined as:
    Figure PCTCN2017091575-appb-100013
    Figure PCTCN2017091575-appb-100013
    其中,hd为所述观测结果d的高度,
    Figure PCTCN2017091575-appb-100014
    为方差常量。
    Where h d is the height of the observation d
    Figure PCTCN2017091575-appb-100014
    Is the variance constant.
  7. 根据权利要求1所述的方法,其特征在于,对所述终止轨迹片段及所述新的轨迹片段进行轨迹关联包括:The method according to claim 1, wherein the trajectory association of the terminating trajectory segment and the new trajectory segment comprises:
    通过第一相似性度量,建立所述终止轨迹片段及所述新的轨迹片段之间的模糊关联代价矩阵;Establishing a fuzzy association cost matrix between the termination trajectory segment and the new trajectory segment by using a first similarity metric;
    采用最大模糊综合相似度和阈值判别原则实现所述终止轨迹片段及所述新的轨迹片段之间的轨迹关联;Implementing a trajectory association between the termination trajectory segment and the new trajectory segment by using a maximum fuzzy comprehensive similarity and a threshold discriminant principle;
    填充关联上的所述终止轨迹片段及所述新的轨迹片段之间缺失的轨迹段。Filling the trailing trajectory segment on the association and the missing trajectory segment between the new trajectory segments.
  8. 根据权利要求7所述的方法,其特征在于,所述通过第一相似性度量,建立所述终止轨迹片段及所述新的轨迹片段之间的模糊关联代价矩阵包括:The method according to claim 7, wherein the establishing a fuzzy association cost matrix between the termination trajectory segment and the new trajectory segment by using the first similarity metric comprises:
    建立所述终止轨迹片段及所述新的轨迹片段之间的相似性向量;Establishing a similarity vector between the termination trajectory segment and the new trajectory segment;
    利用所述相似性向量计算所述终止轨迹片段及所述新的轨迹片段之间的匹配度;Calculating a degree of matching between the terminating trajectory segment and the new trajectory segment by using the similarity vector;
    根据所述匹配度计算所述终止轨迹片段及所述新的轨迹片段之间的模糊综 合相似度;Calculating a fuzzy comprehensive between the end track segment and the new track segment according to the matching degree Similarity
    根据所述模糊综合相似度建立所述终止轨迹片段及所述新的轨迹片段的关联代价矩阵。Establishing an associated cost matrix of the terminating trajectory segment and the new trajectory segment according to the fuzzy comprehensive similarity.
  9. 根据权利要求8所述的方法,其特征在于,所述终止轨迹片段的集合定义为:
    Figure PCTCN2017091575-appb-100015
    所述新的轨迹片段的集合定义为:
    Figure PCTCN2017091575-appb-100016
    其中,na、nb分别表示所述终止轨迹片段集合及所述新的轨迹片段集合的个数;
    The method of claim 8 wherein the set of termination trajectory segments is defined as:
    Figure PCTCN2017091575-appb-100015
    The set of new track segments is defined as:
    Figure PCTCN2017091575-appb-100016
    Where n a and n b respectively represent the set of the end track segment and the number of the new track segment set;
    所述第一相似性度量包括外观相似性度量、形状相似性度量以及运动相似性度量;The first similarity measure includes an appearance similarity measure, a shape similarity measure, and a motion similarity measure;
    所述外观相似性度量定义为:The appearance similarity measure is defined as:
    Figure PCTCN2017091575-appb-100017
    Figure PCTCN2017091575-appb-100017
    其中,ρ(·)表示为Bhattacharyya系数,Hc(·)表示背景加权的颜色直方图特征,
    Figure PCTCN2017091575-appb-100018
    为方差常量,Hg(·)表示方向梯度直方图特征,
    Figure PCTCN2017091575-appb-100019
    为方差常量;
    Where ρ(·) is expressed as a Bhattacharyya coefficient, and H c (·) is a background weighted color histogram feature.
    Figure PCTCN2017091575-appb-100018
    For the variance constant, H g (·) represents the direction gradient histogram feature,
    Figure PCTCN2017091575-appb-100019
    Is a variance constant;
    所述形状相似性度量定义为:The shape similarity measure is defined as:
    Figure PCTCN2017091575-appb-100020
    Figure PCTCN2017091575-appb-100020
    其中,hi表示所述终止轨迹片段Ti在图像中的高度,hj表示所述新的轨迹片段Tj在图像中的高度,
    Figure PCTCN2017091575-appb-100021
    为方差常量;
    Where h i represents the height of the terminating trajectory segment T i in the image, and h j represents the height of the new trajectory segment T j in the image,
    Figure PCTCN2017091575-appb-100021
    Is a variance constant;
    所述运动相似性度量定义为:The motion similarity measure is defined as:
    Figure PCTCN2017091575-appb-100022
    Figure PCTCN2017091575-appb-100022
    其中,
    Figure PCTCN2017091575-appb-100023
    表示高斯分布,∑为所述高斯分布的方差,Δt是所述终止轨迹片段Ti最后观测到所述新的轨迹片段Tj第一个观测的帧间隔、
    Figure PCTCN2017091575-appb-100024
    vi分别为所述终止轨迹片段Ti终止位置和速度,
    Figure PCTCN2017091575-appb-100025
    vj分别为所述新的轨迹片段起始位置和速度;
    among them,
    Figure PCTCN2017091575-appb-100023
    Representing a Gaussian distribution, where ∑ is the variance of the Gaussian distribution, Δt is the first observed frame interval of the new trajectory segment T j that the ending trajectory segment T i finally observed,
    Figure PCTCN2017091575-appb-100024
    v i is the termination position and velocity of the terminating trajectory segment T i , respectively.
    Figure PCTCN2017091575-appb-100025
    v j is the starting position and speed of the new track segment, respectively;
    所述相似性向量定义为:The similarity vector is defined as:
    Figure PCTCN2017091575-appb-100026
    Figure PCTCN2017091575-appb-100026
    其中,Λk(Ti,Tj)∈[0,1]3中,τgap为关联的时间间隔阈值,
    Figure PCTCN2017091575-appb-100027
    表示所述终止轨迹片段Ti断开的时间帧,
    Figure PCTCN2017091575-appb-100028
    表示所述新的轨迹片段Tj起始的时间帧;
    Where Λ k (T i , T j ) ∈ [0, 1] 3 , τ gap is the associated time interval threshold,
    Figure PCTCN2017091575-appb-100027
    a time frame indicating that the terminating trajectory segment T i is disconnected,
    Figure PCTCN2017091575-appb-100028
    a time frame indicating the start of the new track segment T j ;
    所述匹配度定义为:The matching degree is defined as:
    Figure PCTCN2017091575-appb-100029
    Figure PCTCN2017091575-appb-100029
    其中,∧表示所述匹配度取最小值,∨表示所述匹配度取最大值; Where ∧ indicates that the matching degree takes a minimum value, and ∨ indicates that the matching degree takes a maximum value;
    所述模糊综合相似度定义为:The fuzzy comprehensive similarity is defined as:
    Figure PCTCN2017091575-appb-100030
    Figure PCTCN2017091575-appb-100030
    所述关联代价矩阵定义为:The associated cost matrix is defined as:
    Figure PCTCN2017091575-appb-100031
    Figure PCTCN2017091575-appb-100031
  10. 根据权利要求2所述的方法,其特征在于,所述填充关联上的所述终止轨迹片段及所述新的轨迹片段之间的缺失轨迹段包括:The method according to claim 2, wherein the missing track segment between the terminating track segment and the new track segment on the padding association comprises:
    对所述关联上的所述终止轨迹片段及所述新的轨迹片段之间的缺失轨迹段进行双向预测,以获取预测点的位置信息;Performing bidirectional prediction on the missing trajectory segment between the termination trajectory segment and the new trajectory segment on the association to obtain location information of the prediction point;
    获取所述预测点的矩形框信息;Obtaining rectangular frame information of the predicted point;
    根据所述预测点的位置信息及所述矩形框信息对所述缺失轨迹段进行填充。And filling the missing track segment according to the position information of the predicted point and the rectangular frame information.
  11. 一种基于模糊逻辑的视频多目标跟踪的装置,其特征在于,包括:处理器和摄像机,所述处理器连接所述摄像机;An apparatus for video multi-target tracking based on fuzzy logic, comprising: a processor and a camera, wherein the processor is connected to the camera;
    所述处理器用于对对从所述摄像机获取的当前视频帧进行在线目标运动检测,检测得到的可能运动对象作为观测结果;对所述观测结果和目标的预测结果进行数据关联,其中所述预测结果是至少利用前一视频帧的目标的轨迹进行预测而得到的;对未被关联上的所述预测结果及所述观测结果进行轨迹管理,包括利用所述未被关联上所述预测结果获取终止轨迹片段以及利用所述未被关联上的所述观测结果获取新的轨迹片段,对所述终止轨迹片段及所述新的轨迹片段进行轨迹关联。 The processor is configured to perform online target motion detection on the current video frame acquired from the camera, and detect the obtained possible motion object as an observation result; perform data association on the observation result and the target prediction result, where the prediction The result is obtained by predicting at least the trajectory of the target of the previous video frame; performing trajectory management on the unpredicted prediction result and the observation result, including obtaining by using the prediction result that is not associated with the prediction result Terminating the trajectory segment and acquiring a new trajectory segment by using the observation result that is not associated, and performing trajectory association on the terminating trajectory segment and the new trajectory segment.
PCT/CN2017/091575 2017-07-04 2017-07-04 Fuzzy logic based video multi-target tracking method and device WO2019006633A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
PCT/CN2017/091575 WO2019006633A1 (en) 2017-07-04 2017-07-04 Fuzzy logic based video multi-target tracking method and device

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
PCT/CN2017/091575 WO2019006633A1 (en) 2017-07-04 2017-07-04 Fuzzy logic based video multi-target tracking method and device

Publications (1)

Publication Number Publication Date
WO2019006633A1 true WO2019006633A1 (en) 2019-01-10

Family

ID=64949579

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2017/091575 WO2019006633A1 (en) 2017-07-04 2017-07-04 Fuzzy logic based video multi-target tracking method and device

Country Status (1)

Country Link
WO (1) WO2019006633A1 (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112667763A (en) * 2020-12-29 2021-04-16 电子科技大学 Trajectory prediction method based on self-adaptive timestamp and multi-scale feature extraction

Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103281477A (en) * 2013-05-17 2013-09-04 天津大学 Multi-level characteristic data association-based multi-target visual tracking method
CN104091348A (en) * 2014-05-19 2014-10-08 南京工程学院 Multi-target tracking method integrating obvious characteristics and block division templates
WO2016077026A1 (en) * 2014-11-12 2016-05-19 Nec Laboratories America, Inc. Near-online multi-target tracking with aggregated local flow descriptor (alfd)
CN105894542A (en) * 2016-04-26 2016-08-24 深圳大学 Online target tracking method and apparatus
CN106022238A (en) * 2016-05-12 2016-10-12 清华大学 Multi-target tracking method based on sliding window optimization
CN106846373A (en) * 2016-11-16 2017-06-13 浙江工业大学 A kind of mutual occlusion handling method of video object for merging target appearance model and game theory
CN106846361A (en) * 2016-12-16 2017-06-13 深圳大学 Method for tracking target and device based on intuitionistic fuzzy random forest
CN106846355A (en) * 2016-12-16 2017-06-13 深圳大学 Method for tracking target and device based on lifting intuitionistic fuzzy tree

Patent Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103281477A (en) * 2013-05-17 2013-09-04 天津大学 Multi-level characteristic data association-based multi-target visual tracking method
CN104091348A (en) * 2014-05-19 2014-10-08 南京工程学院 Multi-target tracking method integrating obvious characteristics and block division templates
WO2016077026A1 (en) * 2014-11-12 2016-05-19 Nec Laboratories America, Inc. Near-online multi-target tracking with aggregated local flow descriptor (alfd)
CN105894542A (en) * 2016-04-26 2016-08-24 深圳大学 Online target tracking method and apparatus
CN106022238A (en) * 2016-05-12 2016-10-12 清华大学 Multi-target tracking method based on sliding window optimization
CN106846373A (en) * 2016-11-16 2017-06-13 浙江工业大学 A kind of mutual occlusion handling method of video object for merging target appearance model and game theory
CN106846361A (en) * 2016-12-16 2017-06-13 深圳大学 Method for tracking target and device based on intuitionistic fuzzy random forest
CN106846355A (en) * 2016-12-16 2017-06-13 深圳大学 Method for tracking target and device based on lifting intuitionistic fuzzy tree

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
LI JUN;XIE WEI-XIN;LI LIANG-QUN;LIU JUN-BIN;: "Online Multiple Target Tracking Algorithm Based on Fuzzy Spatio-Temporal Cues", ACTA ELECTRONICA SINICA, vol. 45, no. 3, 1 March 2017 (2017-03-01), pages 513 - 519, XP055678763, ISSN: 0372-2112, DOI: 10.3969/j.issn.0372-2112.2017.03.001 *
LI, LIANG-QUN: "A Multiple FCMs Data Association based Algorithm for Mu- lti-target Tracking", PROCEEDINGS 7TH INTERNATIONAL CONFERENCE ON SIGNAL PROCESSING, 2004. PROCEEDINGS. ICSP '04. 2004., 4 September 2004 (2004-09-04), pages 479 - 482, XP010809665, DOI: 10.1109/ICOSP.2004.1452686 *

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112667763A (en) * 2020-12-29 2021-04-16 电子科技大学 Trajectory prediction method based on self-adaptive timestamp and multi-scale feature extraction

Similar Documents

Publication Publication Date Title
CN107545582B (en) Video multi-target tracking method and device based on fuzzy logic
Cannons A review of visual tracking
CN107516321B (en) Video multi-target tracking method and device
WO2020042419A1 (en) Gait-based identity recognition method and apparatus, and electronic device
JP7342919B2 (en) Information processing system, control method, and program
WO2017185688A1 (en) Method and apparatus for tracking on-line target
CN107423686B (en) Video multi-target fuzzy data association method and device
WO2021223367A1 (en) Single lens-based multi-pedestrian online tracking method and apparatus, device, and storage medium
WO2020150896A1 (en) Data association method and apparatus for video target, and storage medium
EP2131328A2 (en) Method for automatic detection and tracking of multiple objects
Di Lascio et al. A real time algorithm for people tracking using contextual reasoning
WO2021007984A1 (en) Target tracking method and apparatus based on tsk fuzzy classifier, and storage medium
WO2018227491A1 (en) Method and device for association of fuzzy data of multiple targets in video
CN101344965A (en) Tracking system based on binocular camera shooting
JP7209115B2 (en) Detection, 3D reconstruction and tracking of multiple rigid objects moving in relatively close proximity
Erdem et al. Visual tracking by fusing multiple cues with context-sensitive reliabilities
Zhou et al. Adaptive fusion of particle filtering and spatio-temporal motion energy for human tracking
Nallasivam et al. Moving human target detection and tracking in video frames
CN113192105A (en) Method and device for tracking multiple persons and estimating postures indoors
Xu et al. A real-time, continuous pedestrian tracking and positioning method with multiple coordinated overhead-view cameras
WO2019006632A1 (en) Video multi-target tracking method and device
Xue et al. Multiple pedestrian tracking under first-person perspective using deep neural network and social force optimization
Wang et al. Effective multiple pedestrian tracking system in video surveillance with monocular stationary camera
Wang et al. Challenge of multi-camera tracking
JP6798609B2 (en) Video analysis device, video analysis method and program

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 17916869

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

32PN Ep: public notification in the ep bulletin as address of the adressee cannot be established

Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205A DATED 12.06.2020)

122 Ep: pct application non-entry in european phase

Ref document number: 17916869

Country of ref document: EP

Kind code of ref document: A1