CN112950674B - Cross-camera track tracking method and device based on cooperation of multiple recognition technologies and storage medium - Google Patents

Cross-camera track tracking method and device based on cooperation of multiple recognition technologies and storage medium Download PDF

Info

Publication number
CN112950674B
CN112950674B CN202110257986.0A CN202110257986A CN112950674B CN 112950674 B CN112950674 B CN 112950674B CN 202110257986 A CN202110257986 A CN 202110257986A CN 112950674 B CN112950674 B CN 112950674B
Authority
CN
China
Prior art keywords
pos
positioning
face
tracked
obj
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202110257986.0A
Other languages
Chinese (zh)
Other versions
CN112950674A (en
Inventor
覃智泉
林淑强
庄毅滨
李立扬
常鹏
李山
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Xiamen Public Security Bureau
Xiamen Meiya Pico Information Co Ltd
Original Assignee
Xiamen Public Security Bureau
Xiamen Meiya Pico Information Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Xiamen Public Security Bureau, Xiamen Meiya Pico Information Co Ltd filed Critical Xiamen Public Security Bureau
Priority to CN202110257986.0A priority Critical patent/CN112950674B/en
Publication of CN112950674A publication Critical patent/CN112950674A/en
Application granted granted Critical
Publication of CN112950674B publication Critical patent/CN112950674B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/20Analysis of motion
    • G06T7/246Analysis of motion using feature-based methods, e.g. the tracking of corners or segments
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/22Matching criteria, e.g. proximity measures
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/25Fusion techniques
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/70Determining position or orientation of objects or cameras
    • G06T7/73Determining position or orientation of objects or cameras using feature-based methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/103Static body considered as a whole, e.g. static pedestrian or occupant recognition
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/161Detection; Localisation; Normalisation
    • G06V40/166Detection; Localisation; Normalisation using acquisition arrangements
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10016Video; Image sequence
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30196Human being; Person
    • G06T2207/30201Face
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30241Trajectory

Abstract

The invention provides a method, a device and a storage medium for tracking a track across cameras based on cooperation of multiple recognition technologies, wherein the method comprises the following steps: acquiring information of an object obj to be tracked and a spatial position Pos 0 The method comprises the steps of carrying out a first treatment on the surface of the Setting a monitoring video range region0 to be analyzed, wherein the region0 is a spatial position Pos 0 Is the circle center and the radius is the range of R; the target to be tracked is co-located based on the information of the target obj to be tracked and the acquired video data in the monitoring video range region0 by using a plurality of recognition technologies to obtain a locating result Pos 1 The method comprises the steps of carrying out a first treatment on the surface of the Obtaining a plurality of positioning results Pos by repeating the setting step and the synergetic positioning step of a plurality of recognition technologies n‑1 ……Pos n Pos is put into 0 ,Pos 1 ...Pos n As a real-time trajectory of the object obj to be tracked. The invention cooperates through a plurality of recognition technologies, has accurate positioning and complete tracking track, uses a small amount of hardware equipment resources in a video relay mode, reduces the hardware cost and provides a specific conflict processing strategy.

Description

Cross-camera track tracking method and device based on cooperation of multiple recognition technologies and storage medium
Technical Field
The invention relates to the technical field of image recognition, in particular to a method and a device for tracking a track across cameras based on cooperation of multiple recognition technologies and a storage medium.
Background
In the field of computer vision, video-based object tracking problems have been an important topic and research hotspot. With the increase of the number and coverage of urban public safety video monitoring cameras, a target track tracking technology based on video images is more valued and applied in the aspects of social management and public safety maintenance.
At present, aiming at the tracking management and control work of key objects, the following methods of the video image target (pedestrian) track tracking technology are adopted: 1. positioning a target object by staring at a video monitoring picture to form track tracking; 2. based on the face snapshot camera, the camera captures face pictures, and whether a target object appears at the position of each face camera is positioned through face recognition comparison, so that track tracking is formed; 3. based on a common video monitoring camera, pedestrian (human body) and appearance attribute features thereof in a video image are extracted and analyzed through a video structuring technology, and then a pedestrian re-recognition technology and a pedestrian appearance attribute feature comparison technology are utilized to locate whether key objects appear at positions of all monitoring cameras.
The three methods of the video image target (pedestrian) track tracking technology have a plurality of defects and bottlenecks, and mainly comprise: the method 1 consumes a great deal of manpower and material resources for staring at the monitoring, has limited human precision and is tired, and the target is easy to lose, so that the method is difficult to popularize and use; the method 2 has mature face recognition technology at present and accurate positioning of the target object, but the method relies on face snapshot cameras, and the actual situations of various places at present are that the deployment density of the face cameras is not high, holes and holes in a plurality of space positions are left, the target object is easy to lose, and finally the track of the target object is incomplete; according to the method 3, a pedestrian (human body) object and attribute characteristics thereof in a video image are extracted through a video structuring technology, and then a pedestrian re-recognition technology and a pedestrian attribute characteristic comparison technology are utilized to locate whether key objects appear at each monitoring camera position.
Disclosure of Invention
The present invention is directed to one or more of the above-mentioned technical drawbacks of the prior art, and designs a specific method to solve the above-mentioned technical problems, and proposes the following technical solutions.
A tracking method of a cross-camera track based on cooperation of multiple recognition technologies comprises the following steps:
an acquisition step of acquiring information of an object obj to be tracked and a spatial position Pos 0
A setting step, namely setting a region0 of a monitoring video range to be analyzed, wherein the region0 is a position Pos in space 0 Is a circle center, and has a radius of R, wherein R is a real number larger than zero;
a step of co-locating the target to be tracked based on the information of the target obj to be tracked and the acquired video data in the monitoring video range region0 by using a plurality of identification technologies to obtain a locating result Pos 1
Obtaining a plurality of positioning results Pos by repeating the setting step and the synergetic positioning step of a plurality of recognition technologies n-1 ……Pos n Pos is put into 0 ,Pos 1 ...Pos n As a real-time trajectory of the object obj to be tracked.
Further, the information of the object obj to be tracked comprises a face image and a whole-body image.
Still further, the plurality of recognition techniques include face recognition positioning and video structured positioning, wherein the face recognition positioning recognizes a face captured by a camera within the monitoring video range based on the face image, and the video structured positioning recognizes video data acquired by a monitoring camera within the monitoring video range based on the whole-body image.
Still further, the video structured locations include pedestrian attribute locations that are located based on appearance local features of pedestrians and pedestrian re-identification locations that are located based on appearance global features of pedestrians.
Further, the operation of co-locating the target to be tracked to obtain a locating result includes: judging whether the positioning results of multiple recognition technologies are consistent, if the positioning results are consistent or only one positioning result is consistent, determining the latest position Pos of obj n And take the previous position Pos n-1 Preserving at the same time in terms of spatial position Pos n The latest position of the target; if the positioning results are inconsistent, combining the positioning results through a preset conflict processing strategy, and determining the latest position Pos of obj n Position Pos n-1 Preserving at the same time in terms of spatial position Pos n The latest position of the target;
the predetermined conflict handling policy is: setting two control thresholds for face recognition positioning: epsilon 1 And epsilon 2 Wherein ε is 1 >ε 2 ,ε 1 The accuracy of the face distribution control early warning result is extremely high, epsilon 2 The accuracy of the face control result is generally high, if the conflict positioning result contains the face result and the face control early warning similarity threshold value is more than or equal to epsilon 1 Then the face positioning result is taken as the latest position Pos of obj n If the conflict positioning result does not contain a human face or contains a human face, and the distribution control early warning similarity epsilon accords with epsilon 2 ≤ε<ε 1 Then confirm the latest position Pos of obj manually n
The invention also provides a device for tracking the track across the cameras based on the cooperation of a plurality of recognition technologies, which comprises:
an acquisition unit for acquiring information of the object obj to be tracked and emptyInter-position Pos 0
The setting unit sets a region0 of the monitoring video to be analyzed, wherein the region0 is a position Pos in space 0 Is a circle center, and has a radius of R, wherein R is a real number larger than zero;
the multiple recognition technology co-locating unit is used for obtaining a locating result Pos by using multiple recognition technologies to co-locate the target to be tracked based on the information of the target obj to be tracked and the acquired video data in the monitoring video range region0 1
Obtaining a plurality of positioning results Pos by repeatedly executing the operations of the setting unit and the co-positioning unit of a plurality of recognition technologies n-1 ……Pos n Pos is put into 0 ,Pos 1 ...Pos n As a real-time trajectory of the object obj to be tracked.
Further, the information of the object obj to be tracked comprises a face image and a whole-body image.
Still further, the plurality of recognition techniques include face recognition positioning and video structured positioning, wherein the face recognition positioning recognizes a face captured by a camera within the monitoring video range based on the face image, and the video structured positioning recognizes video data acquired by a monitoring camera within the monitoring video range based on the whole-body image.
Still further, the video structured locations include pedestrian attribute locations that are located based on appearance local features of pedestrians and pedestrian re-identification locations that are located based on appearance global features of pedestrians.
Further, the operation of co-locating the target to be tracked to obtain a locating result includes: judging whether the positioning results of multiple recognition technologies are consistent, if the positioning results are consistent or only one positioning result is consistent, determining the latest position Pos of obj n And take the previous position Pos n-1 Preserving at the same time in terms of spatial position Pos n The latest position of the target; if the positioning results are inconsistent, combining the positioning results through a preset conflict processing strategyAs a result, the latest position Pos of obj is determined n Position Pos n-1 Preserving at the same time in terms of spatial position Pos n The latest position of the target;
the predetermined conflict handling policy is: setting two control thresholds for face recognition positioning: epsilon 1 And epsilon 2 Wherein ε is 1 >ε 2 ,ε 1 The accuracy of the face distribution control early warning result is extremely high, epsilon 2 The accuracy of the face control result is generally high, if the conflict positioning result contains the face result and the face control early warning similarity threshold value is more than or equal to epsilon 1 Then the face positioning result is taken as the latest position Pos of obj n If the conflict positioning result does not contain a human face or contains a human face, and the distribution control early warning similarity epsilon accords with epsilon 2 ≤ε<ε 1 Then confirm the latest position Pos of obj manually n
The invention also proposes a computer readable storage medium having stored thereon computer program code which, when executed by a computer, performs any of the methods described above.
The invention has the technical effects that: the invention provides a method, a device and a storage medium for tracking a track across cameras based on cooperation of multiple recognition technologies, wherein the method comprises the following steps: an acquisition step of acquiring information of an object obj to be tracked and a spatial position Pos 0 The method comprises the steps of carrying out a first treatment on the surface of the A setting step, namely setting a region0 of a monitoring video range to be analyzed, wherein the region0 is a position Pos in space 0 Is a circle center, and has a radius of R, wherein R is a real number larger than zero; a step of co-locating the target to be tracked based on the information of the target obj to be tracked and the acquired video data in the monitoring video range region0 by using a plurality of identification technologies to obtain a locating result Pos 1 The method comprises the steps of carrying out a first treatment on the surface of the Obtaining a plurality of positioning results Pos by repeating the setting step and the synergetic positioning step of a plurality of recognition technologies n-1 ……Pos n Pos is put into 0 ,Pos 1 ...Pos n As a real-time trajectory of the object obj to be tracked. The invention works cooperatively by a plurality of recognition technologies, fully utilizes the positioning accuracy and universality of the face recognition technologyThe invention achieves the aims of accurate positioning and complete track by adopting a high deployment density of a monitoring camera (video structuring technology) and introducing a conflict processing strategy of a multi-recognition technology means positioning recognition result.
Drawings
Other features, objects and advantages of the present application will become more apparent upon reading of the detailed description of non-limiting embodiments made with reference to the following drawings.
FIG. 1 is a flow chart of a method of tracking a track across cameras based on collaboration of multiple recognition techniques, in accordance with an embodiment of the present invention.
FIG. 2 is a block diagram of a cross-camera trajectory tracking device based on collaboration of multiple recognition techniques in accordance with an embodiment of the invention.
Detailed Description
The present application is described in further detail below with reference to the drawings and examples. It is to be understood that the specific embodiments described herein are merely illustrative of the invention and are not limiting of the invention. It should be noted that, for convenience of description, only the portions related to the present invention are shown in the drawings.
It should be noted that, in the case of no conflict, the embodiments and features in the embodiments may be combined with each other. The present application will be described in detail below with reference to the accompanying drawings in conjunction with embodiments.
Fig. 1 shows a method for tracking a track across cameras based on cooperation of multiple recognition technologies, which comprises the following steps:
an acquisition step S101 of acquiring information of the object obj to be tracked and the spatial position Pos 0 The method comprises the steps of carrying out a first treatment on the surface of the The object to be trackedThe target obj can be a pedestrian, but also can be other objects such as animals, etc., a camera, a snap camera, etc. can be used for acquiring the information of the target obj to be tracked, and the spatial position Pos 0 As the initial position, the information of the object obj to be tracked can be obtained by an image recognition technology or can be manually input, and preferably, the information of the object obj to be tracked comprises a face image and a whole-body image.
A setting step S102, setting a region0 of the monitoring video to be analyzed, wherein the region0 is a spatial position Pos 0 Is a circle center, and has a radius of R, wherein R is a real number larger than zero; the invention aims at the bottleneck that a large amount of hardware server equipment is needed to support in the area range by the current video structuring technology, and provides the method for carrying out the structuring analysis on the monitoring video in the radius R range by taking the target space position as the center, dynamically selecting and changing the monitoring video needing the structuring analysis along with the movement of the positioned target, tracking the target in a video relay mode, reducing the number of the required structuring video, achieving the aim of tracking the target in real time by a small amount of hardware equipment, and reducing the hardware cost of the existing real-time tracking system.
A step S103 of co-locating the target to be tracked based on the information of the target obj to be tracked and the acquired video data in the monitoring video range region0 by using a plurality of identification technologies to obtain a locating result Pos 1 The method comprises the steps of carrying out a first treatment on the surface of the Obtaining a plurality of positioning results Pos by repeating the setting step and the synergetic positioning step of a plurality of recognition technologies n-1 ……Pos n Pos is put into 0 ,Pos 1 ...Pos n As a real-time trajectory of the object obj to be tracked.
In one embodiment, the plurality of recognition techniques includes face recognition localization that recognizes faces captured by cameras within the surveillance video based on the face images and video structured localization that recognizes video data acquired by surveillance cameras within the surveillance video based on the whole-body images.
In one embodiment, the video structured locations include pedestrian attribute locations and pedestrian re-identification locations, the pedestrian attribute locations being located based on appearance local features of the pedestrian, generally appearance local features including, but not limited to, the person's wear (such as coat, pants, hat, etc.), backpack, hairstyle, glasses, etc.; the pedestrian re-identification positioning is based on the global appearance feature of the pedestrian, and in general, the global appearance feature is formed by all local features of the pedestrian and can be expressed in a multi-dimensional vector form.
The invention provides a plurality of identification positioning technologies to cooperatively work, the positioning results are mutually fused, the positions of pedestrians in the cameras are judged, the positioning accuracy of key objects is improved, and a more complete and accurate pedestrian tracking track crossing the cameras is formed. The multi-recognition means cooperative working process of the invention is specifically as follows: the human face picture is extracted according to the provided key object picture and used for carrying out key human dynamic distribution control, the human body picture is extracted for carrying out pedestrian recognition distribution control of the pedestrian re-recognition technology, the human body attribute characteristic is extracted, and the apparent characteristic attribute which is obviously differentiated from the characteristic attribute is selected for characteristic attribute comparison distribution control. And according to the control recognition result, combining with the spatial position information of the camera from which the control early warning picture is derived, the position of the key object (pedestrian) in real time can be known. The multi-recognition technology has a plurality of positioning results, and the positioning results are fused, wherein the fusion method can be as follows: if the intersection is not an empty set, the intersection position is the real-time position of the key object (pedestrian), if the intersection is empty, namely, the control results of a plurality of recognition means have conflicts, a positioning result conflict processing strategy is introduced for processing, and the fusion method of the positioning results is not limited to the method, and the positioning can be performed by adopting a result voting method or a weighting calculation method, which is an important invention point of the invention.
In one embodiment, the operation of co-locating the target to be tracked to obtain a locating result includes: judging multiple kinds of knowledgeWhether the positioning results of other technologies are consistent or not, namely adopting a calculation mode of solving intersection of a plurality of positioning results, if the positioning results are consistent or only one positioning result is adopted, determining the latest position Pos of obj n And take the previous position Pos n-1 Preserving at the same time in terms of spatial position Pos n The latest position of the target; if the positioning results are inconsistent, combining the positioning results through a preset conflict processing strategy, and determining the latest position Pos of obj n Position Pos n-1 Preserving at the same time in terms of spatial position Pos n The latest position of the target;
the predetermined conflict handling policy is: setting two control thresholds for face recognition positioning: epsilon 1 And epsilon 2 Wherein ε is 1 >ε 2 ,ε 1 The accuracy of the face distribution control early warning result is extremely high, epsilon 2 The accuracy of the face control result is generally high, if the conflict positioning result contains the face result and the face control early warning similarity threshold value is more than or equal to epsilon 1 Then the face positioning result is taken as the latest position Pos of obj n If the conflict positioning result does not contain a human face or contains a human face, and the distribution control early warning similarity epsilon accords with epsilon 2 ≤ε<ε 1 Then confirm the latest position Pos of obj manually n . At the same time in spatial position Pos n And setting a monitoring range for the latest position of the target, and updating video real-time data to be analyzed.
The method disclosed by the invention works cooperatively through a plurality of recognition technologies, fully utilizes the positioning accuracy of the face recognition technology, has high deployment density of the common monitoring cameras (video structuring technology), and introduces a conflict processing strategy of the positioning recognition results by a plurality of recognition technology means, so that the aims of positioning accuracy and complete track are achieved. According to the method, the video monitoring range is required to be analyzed through dynamic selection and dynamic change, the function of tracking the target in real time can be achieved through a video relay mode by using a small amount of hardware equipment resources, meanwhile, the face recognition technology, the pedestrian re-recognition technology and the pedestrian appearance characteristic comparison work cooperatively through multiple recognition technology means, multiple positioning results are fused, and the accuracy and the completeness of real-time track tracking of key objects (pedestrians) across cameras are improved.
FIG. 2 illustrates a cross-camera trajectory tracking device based on the cooperation of multiple recognition techniques, the device comprising:
an acquisition unit 201 that acquires information of an object obj to be tracked, and a spatial position Pos 0 The method comprises the steps of carrying out a first treatment on the surface of the The object obj to be tracked can be a pedestrian, but can also be other objects, such as animals, etc., and the information of the object obj to be tracked and the spatial position Pos can be obtained by using a camera, a snapshot camera, etc 0 As the initial position, the information of the object obj to be tracked can be obtained by an image recognition technology or can be manually input, and preferably, the information of the object obj to be tracked comprises a face image and a whole-body image.
The setting unit 202 sets a region0 of the monitoring video to be analyzed, where the region0 is defined by a spatial position Pos 0 Is a circle center, and has a radius of R, wherein R is a real number larger than zero; the invention aims at the bottleneck that a large amount of hardware server equipment is needed to support in the area range by the current video structuring technology, and provides the method for carrying out the structuring analysis on the monitoring video in the radius R range by taking the target space position as the center, dynamically selecting and changing the monitoring video needing the structuring analysis along with the movement of the positioned target, tracking the target in a video relay mode, reducing the number of the required structuring video, achieving the aim of tracking the target in real time by a small amount of hardware equipment, and reducing the hardware cost of the existing real-time tracking system.
The multiple recognition technology co-locating unit 203 performs co-locating on the target to be tracked based on the information of the target obj to be tracked and the acquired video data in the monitoring video range region0 by using multiple recognition technologies to obtain a locating result Pos 1 The method comprises the steps of carrying out a first treatment on the surface of the Obtaining a plurality of positioning results Pos by performing operations of the repeated setting unit and the multiple recognition technology co-positioning unit n-1 ……Pos n Pos is put into 0 ,Pos 1 ...Pos n As a real-time trajectory of the object obj to be tracked.
In one embodiment, the plurality of recognition techniques includes face recognition localization that recognizes faces captured by cameras within the surveillance video based on the face images and video structured localization that recognizes video data acquired by surveillance cameras within the surveillance video based on the whole-body images.
In one embodiment, the video structured locations include pedestrian attribute locations and pedestrian re-identification locations, the pedestrian attribute locations being located based on appearance local features of the pedestrian, generally appearance local features including, but not limited to, the person's wear (such as coat, pants, hat, etc.), backpack, hairstyle, glasses, etc.; the pedestrian re-identification positioning is based on the global appearance feature of the pedestrian, and in general, the global appearance feature is formed by all local features of the pedestrian and can be expressed in a multi-dimensional vector form.
The invention provides a plurality of identification positioning technologies to cooperatively work, the positioning results are mutually fused, the positions of pedestrians in the cameras are judged, the positioning accuracy of key objects is improved, and a more complete and accurate pedestrian tracking track crossing the cameras is formed. The multi-recognition means cooperative working process of the invention is specifically as follows: the human face picture is extracted according to the provided key object picture and used for carrying out key human dynamic distribution control, the human body picture is extracted for carrying out pedestrian recognition distribution control of the pedestrian re-recognition technology, the human body attribute characteristic is extracted, and the apparent characteristic attribute which is obviously differentiated from the characteristic attribute is selected for characteristic attribute comparison distribution control. And according to the control recognition result, combining with the spatial position information of the camera from which the control early warning picture is derived, the position of the key object (pedestrian) in real time can be known. The multi-recognition technology has a plurality of positioning results, and the positioning results are fused, wherein the fusion method can be as follows: if the intersection is not an empty set, the intersection position is the real-time position of the key object (pedestrian), if the intersection is empty, namely, the control results of a plurality of recognition means have conflicts, a positioning result conflict processing strategy is introduced for processing, and the fusion method of the positioning results is not limited to the method, and the positioning can be performed by adopting a result voting method or a weighting calculation method, which is an important invention point of the invention.
In one embodiment, the operation of co-locating the target to be tracked to obtain a locating result includes: judging whether the positioning results of multiple recognition technologies are consistent or not, namely adopting a calculation mode of solving intersection of multiple positioning results, and if the positioning results are consistent or only one positioning result is adopted, determining the latest position Pos of obj n And take the previous position Pos n-1 Preserving at the same time in terms of spatial position Pos n The latest position of the target; if the positioning results are inconsistent, combining the positioning results through a preset conflict processing strategy, and determining the latest position Pos of obj n Position Pos n-1 Preserving at the same time in terms of spatial position Pos n The latest position of the target;
the predetermined conflict handling policy is: setting two control thresholds for face recognition positioning: epsilon 1 And epsilon 2 Wherein ε is 1 >ε 2 ,ε 1 The accuracy of the face distribution control early warning result is extremely high, epsilon 2 The accuracy of the face control result is generally high, if the conflict positioning result contains the face result and the face control early warning similarity threshold value is more than or equal to epsilon 1 Then the face positioning result is taken as the latest position Pos of obj n If the conflict positioning result does not contain a human face or contains a human face, and the distribution control early warning similarity epsilon accords with epsilon 2 ≤ε<ε 1 Then confirm the latest position Pos of obj manually n . At the same time in spatial position Pos n And setting a monitoring range for the latest position of the target, and updating video real-time data to be analyzed.
The device disclosed by the invention works cooperatively through a plurality of recognition technologies, fully utilizes the positioning accuracy of the face recognition technology, has high deployment density of the common monitoring cameras (video structuring technology), and introduces a conflict processing strategy of the positioning recognition results by a plurality of recognition technology means, so that the aims of positioning accuracy and complete track are achieved. The device can achieve the function of tracking the target in real time by using a small amount of hardware equipment resources in a video relay mode through dynamically selecting and dynamically changing the required analysis video monitoring range, meanwhile, the face recognition technology, the pedestrian re-recognition technology and the pedestrian appearance characteristic comparison work cooperatively by a plurality of recognition technology means, various positioning results are fused, and the accuracy and the completeness of real-time track tracking of key objects (pedestrians) across cameras are improved.
For convenience of description, the above method is described as functionally divided into various units. Of course, the functions of each element may be implemented in one or more software and/or hardware elements when implemented in the present application.
From the above description of embodiments, it will be apparent to those skilled in the art that the present application may be implemented in software plus a necessary general purpose hardware platform. Based on such understanding, the technical solutions of the present application may be embodied essentially or in a part contributing to the prior art in the form of a software product, which may be stored in a storage medium, such as a ROM/RAM, a magnetic disk, an optical disk, etc., including several instructions to cause a computer device (which may be a personal computer, a server, or a network device, etc.) to perform the methods described in the embodiments or some parts of the embodiments of the present application.
Finally, what should be said is: the above embodiments are merely for illustrating the technical aspects of the present invention, and it should be understood by those skilled in the art that although the present invention has been described in detail with reference to the above embodiments: modifications and equivalents may be made thereto without departing from the spirit and scope of the invention, which is intended to be encompassed by the claims.

Claims (5)

1. A method for tracking a track across cameras based on cooperation of multiple recognition technologies is characterized by comprising the following steps:
an acquisition step of acquiring information of an object obj to be tracked and a spatial position Pos 0 The information of the object obj to be tracked comprises a face image and a whole-body image;
a setting step of setting a to-be-setThe resolved monitoring video range region0, wherein the region0 is defined by a spatial position Pos 0 Is a circle center, and has a radius of R, wherein R is a real number larger than zero;
a step of co-locating the target to be tracked based on the information of the target obj to be tracked and the acquired video data in the monitoring video range region0 by using a plurality of identification technologies to obtain a locating result Pos 1 The method comprises the steps of carrying out a first treatment on the surface of the The plurality of recognition technologies comprise face recognition positioning and video structuring positioning, wherein the face recognition positioning is used for recognizing a face captured by a camera in the monitoring video range based on the face image, and the video structuring positioning is used for recognizing video data acquired by a monitoring camera in the monitoring video range based on the whole-body image;
obtaining a plurality of positioning results Pos by repeating the setting step and the synergetic positioning step of a plurality of recognition technologies n-1 ……Pos n Pos is put into 0 ,Pos 1 ...Pos n As a real-time track of the object obj to be tracked; the operation of co-locating the target to be tracked to obtain a locating result comprises the following steps: judging whether the positioning results of multiple recognition technologies are consistent, if the positioning results are consistent or only one positioning result is consistent, determining the latest position Pos of obj n And take the previous position Pos n-1 Preserving at the same time in terms of spatial position Pos n The latest position of the target; if the positioning results are inconsistent, combining the positioning results through a preset conflict processing strategy, and determining the latest position Pos of obj n Position Pos n-1 Preserving at the same time in terms of spatial position Pos n The latest position of the target;
the predetermined conflict handling policy is: setting two control thresholds for face recognition positioning: epsilon 1 And epsilon 2 Wherein ε is 1 >ε 2 ,ε 1 The accuracy of the face distribution control early warning result is extremely high, epsilon 2 The accuracy of the face control result is generally high, if the conflict positioning result contains the face result and the face control early warning similarity threshold value is more than or equal to epsilon 1 Then the face positioning result is taken asLatest position Pos of obj n If the conflict positioning result does not contain a human face or contains a human face, and the distribution control early warning similarity epsilon accords with epsilon 2 ≤ε<ε 1 Then confirm the latest position Pos of obj manually n
2. The method of claim 1, wherein the video structured locations include a pedestrian attribute location that locates based on a local feature of appearance of a pedestrian and a pedestrian re-identification location that locates based on a global feature of appearance of a pedestrian.
3. A cross-camera track tracking device based on cooperation of multiple recognition technologies is characterized in that the device comprises:
an acquisition unit for acquiring information of an object obj to be tracked and a spatial position Pos 0 The information of the object obj to be tracked comprises a face image and a whole-body image;
the setting unit sets a region0 of the monitoring video to be analyzed, wherein the region0 is a position Pos in space 0 Is a circle center, and has a radius of R, wherein R is a real number larger than zero;
the multiple recognition technology co-locating unit is used for obtaining a locating result Pos by using multiple recognition technologies to co-locate the target to be tracked based on the information of the target obj to be tracked and the acquired video data in the monitoring video range region0 1 The method comprises the steps of carrying out a first treatment on the surface of the The plurality of recognition technologies comprise face recognition positioning and video structuring positioning, wherein the face recognition positioning is used for recognizing a face captured by a camera in the monitoring video range based on the face image, and the video structuring positioning is used for recognizing video data acquired by a monitoring camera in the monitoring video range based on the whole-body image;
obtaining a plurality of positioning results Pos by repeatedly executing the operations of the setting unit and the co-positioning unit of a plurality of recognition technologies n-1 ……Pos n Pos is put into 0 ,Pos 1 ...Pos n Real-time track as object obj to be trackedA trace; the operation of co-locating the target to be tracked to obtain a locating result comprises the following steps: judging whether the positioning results of multiple recognition technologies are consistent, if the positioning results are consistent or only one positioning result is consistent, determining the latest position Pos of obj n And take the previous position Pos n-1 Preserving at the same time in terms of spatial position Pos n The latest position of the target; if the positioning results are inconsistent, combining the positioning results through a preset conflict processing strategy, and determining the latest position Pos of obj n Position Pos n-1 Preserving at the same time in terms of spatial position Pos n The latest position of the target;
the predetermined conflict handling policy is: setting two control thresholds for face recognition positioning: epsilon 1 And epsilon 2 Wherein ε is 1 >ε 2 ,ε 1 The accuracy of the face distribution control early warning result is extremely high, epsilon 2 The accuracy of the face control result is generally high, if the conflict positioning result contains the face result and the face control early warning similarity threshold value is more than or equal to epsilon 1 Then the face positioning result is taken as the latest position Pos of obj n If the conflict positioning result does not contain a human face or contains a human face, and the distribution control early warning similarity epsilon accords with epsilon 2 ≤ε<ε 1 Then confirm the latest position Pos of obj manually n
4. The apparatus of claim 3, wherein the video structured locations comprise a pedestrian attribute location that locates based on a local feature of appearance of a pedestrian and a pedestrian re-identification location that locates based on a global feature of appearance of a pedestrian.
5. A computer readable storage medium, characterized in that the storage medium has stored thereon a computer program code which, when executed by a computer, performs the method of any of claims 1-2.
CN202110257986.0A 2021-03-09 2021-03-09 Cross-camera track tracking method and device based on cooperation of multiple recognition technologies and storage medium Active CN112950674B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202110257986.0A CN112950674B (en) 2021-03-09 2021-03-09 Cross-camera track tracking method and device based on cooperation of multiple recognition technologies and storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202110257986.0A CN112950674B (en) 2021-03-09 2021-03-09 Cross-camera track tracking method and device based on cooperation of multiple recognition technologies and storage medium

Publications (2)

Publication Number Publication Date
CN112950674A CN112950674A (en) 2021-06-11
CN112950674B true CN112950674B (en) 2024-03-05

Family

ID=76230434

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202110257986.0A Active CN112950674B (en) 2021-03-09 2021-03-09 Cross-camera track tracking method and device based on cooperation of multiple recognition technologies and storage medium

Country Status (1)

Country Link
CN (1) CN112950674B (en)

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113255599B (en) * 2021-06-29 2021-09-24 成都考拉悠然科技有限公司 System and method for user-defined human flow testing face distribution control rate

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2016131300A1 (en) * 2015-07-22 2016-08-25 中兴通讯股份有限公司 Adaptive cross-camera cross-target tracking method and system
CN107679504A (en) * 2017-10-13 2018-02-09 北京奇虎科技有限公司 Face identification method, device, equipment and storage medium based on camera scene
CN110852219A (en) * 2019-10-30 2020-02-28 广州海格星航信息科技有限公司 Multi-pedestrian cross-camera online tracking system
CN110858400A (en) * 2018-08-22 2020-03-03 北京航天长峰科技工业集团有限公司 Method for tracking moving target through monitoring video
CN111079600A (en) * 2019-12-06 2020-04-28 长沙海格北斗信息技术有限公司 Pedestrian identification method and system with multiple cameras
CN111783568A (en) * 2020-06-16 2020-10-16 厦门市美亚柏科信息股份有限公司 Pedestrian re-identification method and device based on belief learning and storage medium
CN111832457A (en) * 2020-07-01 2020-10-27 济南浪潮高新科技投资发展有限公司 Stranger intrusion detection method based on cloud edge cooperation

Family Cites Families (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP3649616A4 (en) * 2017-07-05 2021-04-07 Fastbrick IP Pty Ltd Real time position and orientation tracker
CN109325967B (en) * 2018-09-14 2023-04-07 腾讯科技(深圳)有限公司 Target tracking method, device, medium, and apparatus

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2016131300A1 (en) * 2015-07-22 2016-08-25 中兴通讯股份有限公司 Adaptive cross-camera cross-target tracking method and system
CN107679504A (en) * 2017-10-13 2018-02-09 北京奇虎科技有限公司 Face identification method, device, equipment and storage medium based on camera scene
CN110858400A (en) * 2018-08-22 2020-03-03 北京航天长峰科技工业集团有限公司 Method for tracking moving target through monitoring video
CN110852219A (en) * 2019-10-30 2020-02-28 广州海格星航信息科技有限公司 Multi-pedestrian cross-camera online tracking system
CN111079600A (en) * 2019-12-06 2020-04-28 长沙海格北斗信息技术有限公司 Pedestrian identification method and system with multiple cameras
CN111783568A (en) * 2020-06-16 2020-10-16 厦门市美亚柏科信息股份有限公司 Pedestrian re-identification method and device based on belief learning and storage medium
CN111832457A (en) * 2020-07-01 2020-10-27 济南浪潮高新科技投资发展有限公司 Stranger intrusion detection method based on cloud edge cooperation

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
多目标跨摄像头跟踪技术;焦珊珊;李云波;陈佳林;潘志松;;国防科技(第06期);全文 *

Also Published As

Publication number Publication date
CN112950674A (en) 2021-06-11

Similar Documents

Publication Publication Date Title
CN110717414B (en) Target detection tracking method, device and equipment
CN105404860B (en) Method and apparatus for managing personal information of wandering away
CN106354816B (en) video image processing method and device
Fernandez-Sanjurjo et al. Real-time visual detection and tracking system for traffic monitoring
CN109271554B (en) Intelligent video identification system and application thereof
CN104378582B (en) A kind of intelligent video analysis system and method cruised based on Pan/Tilt/Zoom camera
US20060222244A1 (en) Grouping items in video stream images into events
EP2713308B1 (en) Method and system for using fingerprints to track moving objects in video
CN106709436A (en) Cross-camera suspicious pedestrian target tracking system for rail transit panoramic monitoring
CN110852269B (en) Cross-lens portrait correlation analysis method and device based on feature clustering
WO2020114138A1 (en) Information associated analysis method and apparatus, and storage medium and electronic device
JP2022122858A (en) Method for providing information element that enables identification of unauthorized user in restricted area to device
CN113269091A (en) Personnel trajectory analysis method, equipment and medium for intelligent park
CN112950674B (en) Cross-camera track tracking method and device based on cooperation of multiple recognition technologies and storage medium
CN104765739B (en) Extensive face database search method based on shape space
CN114581990A (en) Intelligent running test method and device
CN110992500A (en) Attendance checking method and device, storage medium and server
CN117392611A (en) Site safety monitoring method, system, equipment and storage medium
CN112052803A (en) Wisdom garden control protection system
Bhattacharya et al. A deep neural network framework for detection and identification of bengal tigers
KR101595334B1 (en) Method and apparatus for movement trajectory tracking of moving object on animal farm
CN109344776A (en) Data processing method
CN114663796A (en) Target person continuous tracking method, device and system
CN116580054A (en) Video data processing method, device, equipment and medium
Xu et al. Vision-IMU based obstacle detection method

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
CB03 Change of inventor or designer information

Inventor after: Tan Zhiquan

Inventor after: Lin Shuqiang

Inventor after: Zhuang Yibin

Inventor after: Li Liyang

Inventor after: Chang Peng

Inventor after: Li Shan

Inventor before: Qin Zhiquan

Inventor before: Lin Shuqiang

Inventor before: Zhuang Yibin

Inventor before: Li Liyang

Inventor before: Chang Peng

Inventor before: Li Shan

CB03 Change of inventor or designer information
GR01 Patent grant
GR01 Patent grant