WO2022227761A1 - Target tracking method and apparatus, electronic device, and storage medium - Google Patents

Target tracking method and apparatus, electronic device, and storage medium Download PDF

Info

Publication number
WO2022227761A1
WO2022227761A1 PCT/CN2022/074956 CN2022074956W WO2022227761A1 WO 2022227761 A1 WO2022227761 A1 WO 2022227761A1 CN 2022074956 W CN2022074956 W CN 2022074956W WO 2022227761 A1 WO2022227761 A1 WO 2022227761A1
Authority
WO
WIPO (PCT)
Prior art keywords
target
position coordinates
target object
current moment
coordinates
Prior art date
Application number
PCT/CN2022/074956
Other languages
French (fr)
Chinese (zh)
Inventor
关英妲
刘文韬
钱晨
Original Assignee
上海商汤智能科技有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 上海商汤智能科技有限公司 filed Critical 上海商汤智能科技有限公司
Publication of WO2022227761A1 publication Critical patent/WO2022227761A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/20Analysis of motion
    • G06T7/246Analysis of motion using feature-based methods, e.g. the tracking of corners or segments
    • G06T7/248Analysis of motion using feature-based methods, e.g. the tracking of corners or segments involving reference images or patches
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/70Determining position or orientation of objects or cameras
    • G06T7/73Determining position or orientation of objects or cameras using feature-based methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10016Video; Image sequence
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30196Human being; Person
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30241Trajectory

Definitions

  • the present disclosure relates to the field of computer vision technology, and in particular, to a target tracking method, apparatus, electronic device, and storage medium.
  • Artificial intelligence technology is playing an increasingly important role in creating intelligent education, entertainment and life.
  • computer vision as one of the key technologies, is widely used.
  • the positioning technology based on computer vision can track the target object in the target place in different scenarios, and determine the trajectory of the target object in the target place.
  • the image of the target site collected by the camera can be used to determine the position of the target object in the target site image at different times, and further track the target object according to the position of the target object at different times.
  • the embodiments of the present disclosure provide at least one target tracking solution.
  • an embodiment of the present disclosure provides a target tracking method, including:
  • the multiple collection devices have different collection perspectives in the target site, and the video images include the sense of the target object in the target site. area of interest;
  • the second position coordinates of the target object at the current moment are determined based on the first position coordinates of the target object and the second position coordinates of the target object at the previous moment.
  • an embodiment of the present disclosure provides a target tracking device, including:
  • an acquisition module configured to acquire the video images at the current moment collected by multiple collection devices set in the target place; the multiple collection devices have different collection perspectives in the target place, and the video images include the target place the region of interest of the target object;
  • a determination module configured to determine the first position coordinates of each of the target objects at the current moment based on the video images at the current moment collected by the multiple collection devices;
  • the tracking module is configured to, for each of the target objects, determine the second position coordinates of the target object at the current moment based on the first position coordinates of the target object and the second position coordinates of the target object at the previous moment.
  • embodiments of the present disclosure provide an electronic device, including: a processor, a memory, and a bus, where the memory stores machine-readable instructions executable by the processor, and when the electronic device runs, the processing The processor and the memory communicate through a bus, and the machine-readable instructions execute the steps of the target tracking method according to the first aspect when the machine-readable instructions are executed by the processor.
  • an embodiment of the present disclosure provides a computer-readable storage medium, where a computer program is stored on the computer-readable storage medium, and the computer program is executed by a processor to execute the target tracking method according to the first aspect. step.
  • an embodiment of the present disclosure provides a computer program product, the computer program product includes a computer program and is stored on a storage medium, and when the computer program is executed by a processor, executes the target tracking method according to the first aspect. step.
  • FIG. 1 shows a flowchart of a target tracking method provided by an embodiment of the present disclosure
  • FIG. 2 shows a flowchart of a method for determining a first position coordinate of a target object provided by an embodiment of the present disclosure
  • FIG. 3 shows a schematic diagram of a detection frame obtained by performing target detection on a video image at the current moment according to an embodiment of the present disclosure
  • FIG. 4 shows a flowchart of a specific method for determining the first position coordinates of the same target object at the current moment provided by an embodiment of the present disclosure
  • FIG. 5 shows a flowchart of a method for determining the second position coordinates of a target object at the current moment provided by an embodiment of the present disclosure
  • FIG. 6 shows a flowchart of a method for determining the coordinates of the observed position of a target object that is missed at the current moment provided by an embodiment of the present disclosure
  • FIG. 7 shows a flowchart of a method for determining trajectory data of each target object provided by an embodiment of the present disclosure
  • FIG. 8 shows a flowchart of a method for revising an identity identifier of a target object that deviates from a target group provided by an embodiment of the present disclosure
  • FIG. 9 shows an early warning method provided by an embodiment of the present disclosure.
  • FIG. 10 shows a schematic diagram of a scene for tracking a target object provided by an embodiment of the present disclosure
  • FIG. 11 shows a schematic structural diagram of a target tracking device provided by an embodiment of the present disclosure
  • FIG. 12 shows a schematic diagram of an electronic device provided by an embodiment of the present disclosure.
  • the target object In many application scenarios, it is usually necessary to track the target object in a place. For example, in a factory, it is necessary to track whether employees have a tendency to enter dangerous areas. In shopping malls, the movements of customers can be tracked. The position of the target object can be determined through the image captured by the camera, and then the tracking of the target object can be completed. However, for some target sites with complex and large areas, in the process of tracking the target object based on the camera, there may be situations where the target object cannot be captured, that is, there is a problem of tracking interruption; Tracking of target objects in these occluded regions is done.
  • the present disclosure provides a target tracking method, in which the acquisition perspectives of the acquisition devices set in the target site are different, and each target object in the target site is simultaneously acquired by at least two acquisition devices.
  • a comprehensive and accurate positioning of the target object in the target place can be completed, and the first position coordinates of the target object at the current moment can be obtained.
  • Further combining the second position coordinates of the target object with higher accuracy at the previous moment and the first position coordinates at the current moment accurately determine the second position coordinates of the target object at the current moment, that is, to complete the target object entering the target place. tracking.
  • the execution body of the positioning method provided by the embodiment of the present disclosure is a computer device with computing capability, and the computer device includes, for example, a server or other processing device.
  • the target tracking method may be implemented by a processor invoking computer-readable instructions stored in a memory.
  • the method includes steps S101-S103.
  • S101 Acquire a video image at the current moment collected by multiple collection devices set in the target site; the multiple collection devices have different collection perspectives in the target site, and the video image includes the region of interest of the target object in the target site.
  • the target location may be a location corresponding to the application scenario.
  • the target place can be the factory; if you need to locate the customers in the mall, the target place can be the shopping mall; if you need to locate the athletes in the gym, the target The venue may be a gymnasium.
  • the region of interest is the region in the video frame where objects that need to be positioned (such as the aforementioned employees, customers and athletes) are located in the target venue.
  • the acquisition device may be a monocular camera or a binocular camera.
  • Multiple acquisition devices can be set up in the target site.
  • the installation positions of multiple collection devices can be determined according to the actual site of the target site.
  • the acquisition angles of the acquisition devices in the target site can be made different, so as to cover the entire area of the target site without leaving a dead angle.
  • too many acquisition devices will lead to too many video images captured at the same time, it will affect the video images. Therefore, when installing the acquisition equipment in the target site, it is necessary to consider the installation angle and quantity of the acquisition equipment.
  • each target object entering the target site can be acquired by two acquisition equipment at the same time, so that Multiple acquisition devices set up in the target site can completely capture the current video image of the entire area of the target site.
  • S102 Determine the first position coordinates of each target object at the current moment based on the video images at the current moment collected by multiple collection devices.
  • the target objects are objects that need to be positioned in the target site, such as the aforementioned employees, customers and athletes.
  • the target site as a factory as an example, each employee in the factory is the target object. Since the multi-view acquisition of the target object is considered when the acquisition device is arranged, the same target object can be acquired by at least two acquisition devices at the same time, and at least two video images can be obtained, each of which involves a sensory image. area of interest.
  • the target site as a factory as an example.
  • the number of employees in the factory is 2, that is, the target audience is 2. Since the same employee needs to be captured by at least 2 capture devices, it can be assumed that there are 2 cameras in the factory. In this case, a possible situation is that at 9:00 am, each of the two cameras captures a video frame, for example, video frame 1 and video frame 2; each video frame involves 2 interested parties. For example, in video screen 1, the area of interest 1 corresponding to employee 1 and the area of interest 2 corresponding to employee 2; in video screen 2, the area of interest 3 corresponding to employee 1 and the area of interest corresponding to employee 2 4.
  • each target object in the target site is collected by at least two collection devices at the same time, for each target object in the target site, the senses including the target object collected by multiple collection devices can be used.
  • the video picture of the region of interest at the current moment is used to determine the first position coordinates of the target object at the current moment.
  • the first position coordinates of the target object may refer to the position coordinates of the target object in a world coordinate system pre-built for the target location.
  • the pixel coordinates of the region of interest in the video image at the current moment in the video image can be used as well as the parameter information of the acquisition device that collects the video image. , and determine the initial position coordinates of the target object corresponding to the region of interest in the video picture in the target place.
  • the initial position coordinates of the same target object in the target objects corresponding to the region of interest determined based on the video images collected by different collection devices at the current moment will have some differences.
  • the initial position coordinates belonging to the same target object may be fused, and then the first position coordinates of the target object at the current moment are determined according to multiple initial position coordinates associated with the same target object at the current moment.
  • the world coordinate system corresponding to the target site may take a fixed position in the target site as the coordinate origin to establish a unique world coordinate system.
  • S103 for each target object, determine the second position coordinates of the target object at the current moment based on the first position coordinates of the target object and the second position coordinates of the target object at the previous moment.
  • the first position coordinates of the target object at the current moment after obtaining the first position coordinates of the target object at the current moment, it can be based on the second position coordinates of the target object in the target place at the previous moment, and the first position of the target object in the target place at the current moment. Coordinates, to perform temporal association on the position of the same target object at different times. Determine the second position coordinates of the target object in the target place at the current moment, and then associate the second position coordinates of each target object in the target place at different times to obtain the movement track of each target object in the target place.
  • the target object determined based on the video image at the current moment collected by the collection device is at the current moment.
  • the first position coordinates also have certain errors. Therefore, in the process of correlating the positions of the same target object at different times, the first position coordinates of the target object are corrected to determine the second position coordinates of the target object with higher accuracy at the current moment.
  • the manner of determining the second position coordinates of the target object at the previous moment is similar to the manner of determining the second position coordinates of the target object at the current moment. Therefore, the present disclosure mainly describes the process of determining the second position coordinates of the target object at the current moment.
  • the collection viewing angles of the collection devices set in the target place are different, so that the target object in the target place can be positioned comprehensively and accurately, and the first position coordinates of the target object at the current moment can be obtained. Further combining the second position coordinates of the target object with higher accuracy at the previous moment and the first position coordinates at the current moment, accurately determine the second position coordinates of the target object at the current moment, that is, to complete the target object entering the target place. tracking.
  • S201 Acquire pixel coordinates of a region of interest in a video image at the current moment collected by multiple collection devices respectively.
  • the region of interest of the target object in the video picture at the current moment may be identified based on a pre-trained neural network for target detection. Further, the pixel coordinates of the set position point in the region of interest in the image coordinate system corresponding to the video screen can be read, and the pixel coordinates corresponding to the set position point can be used as the pixel coordinates of the region of interest.
  • steps S2011 to S2012 may be included when acquiring the pixel coordinates of the region of interest in the video images at the current moment respectively collected by multiple collection devices:
  • S2011 inputting a plurality of video pictures at the current moment into a pre-trained neural network to obtain a target detection frame in each video picture; wherein, the neural network includes a plurality of target detections for detecting regions of interest of target objects of different sizes Sub-network, the region where the target detection frame is located in the video picture is the region of interest.
  • the neural network can detect the region of interest of each target object contained in the video picture at the current moment, and mark each target detection frame.
  • FIG. 3 it is a schematic diagram of the target detection frame included in the video picture at the current moment.
  • the video image at the current moment includes two target detection frames corresponding to the region of interest, including the target detection frame A1B1C1D1 of the region of interest 1 and the target detection frame A2B2C2D2 of the region of interest 2, respectively.
  • a position point can be extracted as the target position point on the target detection frame of each region of interest.
  • the midpoint of the bottom edge of the detection frame is extracted as the target position point, as shown in FIG.
  • the pixel coordinates of the region of interest 2 are represented by the pixel coordinates of the midpoint position point K2 of the bottom edge D2C2 of the target detection frame A2B2C2D2.
  • the neural network used in the embodiments of the present disclosure may include multiple target detection sub-networks for detecting regions of interest of target objects of different sizes.
  • it can be a feature pyramid network.
  • Each target detection sub-network in the feature pyramid network is used to detect and identify a region of interest of a target object of a size corresponding to the target detection sub-network in the video picture at the current moment.
  • the neural network includes a plurality of target detection sub-networks for detecting regions of interest of target objects of different sizes, so that when the neural network is used to perform target detection on the regions of interest of target objects in a video picture , the region of interest of target objects of different sizes in the same video frame can be accurately detected.
  • the parameter information of each acquisition device may include a homography matrix of the acquisition device, wherein the homography matrix may represent the image coordinate system corresponding to the video picture at the current moment collected by the acquisition device and the target where the acquisition device is located.
  • the transformation relationship between the world coordinate systems corresponding to the place In this way, after obtaining the pixel coordinates in the image coordinate system corresponding to the video image of the region of interest at the current moment, the position coordinates of the region of interest in the world coordinate system corresponding to the target site can be determined according to the parameter information of the acquisition device, and the The position coordinates of the region of interest in the world coordinate system are taken as the initial position coordinates of the target object corresponding to the region of interest in the world coordinate system.
  • the initial position coordinates in the target location at the current moment include the following S2021-S2022:
  • the internal parameter matrix of the acquisition device contains (f x , f y ) represents the focal length of the acquisition device, and (c x , c y ) represents the pixel coordinates in the image coordinate system of the center point of the video image captured by the acquisition device at the current moment.
  • the distortion parameters of the acquisition device include radial distortion parameters and tangential distortion coefficients.
  • the internal parameter matrix and distortion parameters of each acquisition device may be predetermined in the manner of Zhang Zhengyou's chessboard calibration. For example, you can take multiple checkerboard images from different angles, detect the feature points in the images, and solve the internal parameter matrix and distortion parameters of the acquisition device according to the pixel coordinates of these feature points in the checkerboard image, and then continuously compare the internal parameter matrix and Distortion parameters are optimized.
  • the same pixel coordinate can be corrected according to the internal parameter matrix and distortion parameters obtained twice adjacently, and whether to end the optimization is determined by the difference between the two corrected pixel coordinates before and after. For example, after the difference is no longer reduced, the optimization can be ended to obtain the internal parameter matrix and distortion parameters of the acquisition device.
  • the homography matrix may represent the conversion relationship between the image coordinate system corresponding to the video frame at the current moment collected by the collecting device and the world coordinate system corresponding to the target location where the collecting device is located.
  • the homography matrix can also be determined when the acquisition device is calibrated in advance. For example, a sample video image with multiple markers can be collected by a collection device, and the intersection of the multiple markers and the ground (the plane where the X and Y axes of the world coordinate system are located) is in the world coordinate system corresponding to the target site. world coordinates. Then, the corrected pixel coordinates corresponding to the intersections of the multiple markers and the ground in the sample video picture are determined according to the above method. Further, the homography matrix of the acquisition device may be determined based on the corrected pixel coordinates and world coordinates corresponding to the plurality of markers respectively.
  • the initial position coordinates of the target object corresponding to the region of interest in the video picture it can be based on the corrected pixel coordinates of the region of interest in the video picture and the homography of the acquisition device that collects the video picture at the current moment. to obtain the position coordinates of the region of interest in the video screen in the world coordinate system corresponding to the target location, and determine the position coordinates of the region of interest in the video screen in the world coordinate system as the target object corresponding to the region of interest The initial position coordinates of .
  • the pixel coordinates are first corrected based on the internal parameter matrix and the distortion coefficient of the capture device that captures the video picture, so that corrected pixels with higher accuracy can be obtained. coordinates, further improving the accuracy of the obtained initial position coordinates of the target object corresponding to the region of interest in the target place.
  • the initial position coordinates of the same target object may be fused to obtain the first position coordinates of the target object.
  • the pixel coordinates of the region of interest in the video picture can be determined first, and then the initial position coordinates of the target object corresponding to the region of interest in the target place can be obtained according to the parameter information of the acquisition device.
  • the initial position coordinates belonging to the same target object among the initial position coordinates in different video pictures are further fused to obtain the first position coordinates with higher accuracy of the target object.
  • each target object is captured by at least two capture devices at the same time, and for each target object, in the case of being captured by different capture devices at the same time, the capture device There is a certain error in the parameter information, and the error between the parameter information of different acquisition devices is different. Therefore, the initial position coordinates of the same target object determined based on video pictures at different current moments may be different. Before fusing the initial position coordinates of the same target object, it is necessary to determine multiple initial position coordinates associated with the same target object.
  • the first two may be fused first.
  • the fused initial position coordinates are obtained, and then the fused initial position coordinates are fused with the third initial position coordinates.
  • the position coordinate obtained by the final fusion is used as the first position coordinate of the target object.
  • the initial position coordinates of the same target object collected by multiple collection devices can be fused, thus, the first position coordinates of the target object with higher accuracy can be obtained.
  • the target object corresponding to each region of interest in the first video picture in any two video pictures is the first target object
  • the second video picture in the arbitrary two video pictures is the first target object.
  • the target object corresponding to each region of interest is a second target object, and for the initial position coordinates of each of the first target objects, determine the initial position coordinates of the first target object and the second current moment in the video picture. a second distance between the initial position coordinates of the second target object;
  • the initial position coordinates of the first target object and the initial position coordinates of the second target object with the minimum second distance from the first target object are taken as multiple initial position coordinates associated with the same target object, the minimum second The distance is less than the second preset fusion distance threshold.
  • a collection devices are set up in the target site, and it is assumed that the video images of the current moment collected by the A collection devices at the same time all include the region of interest of at least one target object, and the initial position of the A group can be obtained at this moment.
  • the set of initial position coordinates of the target object corresponding to the region of interest in the video image captured by the Nth collection device at the current moment where N is an integer greater than or equal to 1 and less than or equal to A.
  • N is an integer greater than or equal to 1 and less than or equal to A.
  • S1 includes initial position coordinates (also referred to as first initial position coordinates) of a first target objects
  • S2 includes b initial position coordinates (also referred to as second initial position coordinates) of second target objects.
  • the Euclidean distance between each first initial position coordinate and each second initial position coordinate can be determined to obtain a distance matrix:
  • d 11 represents the second distance between the first first initial position coordinate in S1 and the first second initial position coordinate in S2;
  • d 1b represents the first first initial position coordinate in S1 and the first second initial position coordinate in S2 The second distance between the b second initial position coordinates;
  • d ij represents the second distance between the i-th first initial position coordinate in S1 and the j-th second initial position coordinate in S2;
  • d a1 represents the second distance in S1 The second distance between the a-th first initial position coordinate and the first second initial position coordinate in S2;
  • d ab represents the a-th first initial position coordinate in S1 and the b-th second initial position coordinate in S2 the second distance between.
  • multiple initial position coordinates associated with the same target object in S1 and S2 can be determined in the following manner, including S30121-S3012:
  • the elements in the current distance matrix include the Euclidean equation between the initial position coordinates of each first target object in S1 and the initial position coordinates of each second target object in S2. distance.
  • S30122 Determine whether the current minimum second distance is less than a second preset fusion distance threshold.
  • the second preset fusion distance may be set empirically. For example, the same target object is photographed by different collection devices in advance, and then multiple initial position coordinates of the same target object in the target site are determined according to the video images collected by different collection devices, according to the distance between the multiple initial position coordinates. to determine the second preset fusion distance threshold.
  • the initial position coordinates of the a-th first target object in S1 and the first second target in S2 can be used.
  • the initial position coordinates of the object serve as the initial position coordinates associated with the same target object.
  • S30124 Set the current minimum second distance in the current distance matrix and all other second distances between any one of the two initial position coordinates associated with the current minimum second distance as the second preset fusion distance After the threshold is set, return to execute S30121 until the current minimum second distance in the current distance matrix is greater than or equal to the second preset fusion distance threshold, obtain all initial position coordinates associated with the same target object in S1 and S2.
  • the current distance matrix is calculated from the initial position coordinates in S1 and S2, and the specific one is a 3 ⁇ 3 matrix:
  • the second preset fusion distance threshold is d th ; assuming that d 11 is the minimum distance in the current matrix and is less than d th , then the first first initial position coordinate in S1 and the first second initial position coordinate in S2, is the associated initial position coordinate of the same target object. Then in the current distance matrix, all other distances calculated from any of the two initial position coordinates are d 12 , d 13 , d 21 , and d 31 . Therefore, according to S30124, in the current matrix, it is necessary to set d 11 , d 12 , d 13 , d 21 , and d 31 to d th ; the set matrix is:
  • the current minimum second distance in the current distance matrix and all other distances between any one of the two initial position coordinates associated with the current minimum second distance are set as the second preset fusion.
  • the element set as the second preset fusion distance threshold can be excluded, thereby improving the search efficiency.
  • the initial position coordinates associated with the same target object after obtaining multiple initial position coordinates associated with the same target object in S1 and S2, it can continue to determine whether there is an association with the same target object based on any two other video images at the current moment.
  • the initial position coordinates of After judging the video images of the current moment collected by the A collection devices at the same time, the different initial position coordinates of each target object involved in the current video images collected by the A collection devices at the same time can be obtained. Then, the initial position coordinates associated with the same target object are fused in different initial position coordinates to obtain the first position of each target object in the target place in the A pieces of video images of the current moment shot by the A collection devices at the same time coordinate.
  • coordinate fusion can be performed on the plurality of initial position coordinates to obtain the updated version of the same target object.
  • Initial position coordinates For the initial position coordinates to be fused in S1 and S2, S2' can be formed with the updated initial position coordinates. Further form a new current distance matrix by the initial position coordinates in S2' and S3, repeat the steps of S30121 to S30124, obtain a plurality of initial position coordinates associated with the same target object in S2' and S3, obtain S3 in the same way '.
  • a new current distance matrix is further formed by the initial position coordinates in S3 ' and S4, and the steps of S30121 to S30124 are repeatedly executed, until after completing the fusion with the initial position coordinates in the last element (that is, SA) in the initial coordinate set, Obtain the first position coordinates of each target object in the target place in the A video images shot by the A collection devices at the same time.
  • any initial position coordinates are detected to be involved in the fusion from the beginning to the end, considering the target location
  • Each target object in is simultaneously collected by at least two collection devices, so any initial position coordinate can be used as the error initial position coordinate for filtering.
  • the initial position coordinates associated with the same target object can be quickly determined , so as to provide a basis for the subsequent determination of the first position coordinates of each target object.
  • the initial position coordinates to be fused refer to the initial position coordinates that do not participate in the fusion.
  • the steps include:
  • the plurality of initial position coordinates associated with the target object A include N pieces. Any initial position coordinate may be used as the first intermediate fusion position coordinate, and the midpoint coordinate of the first intermediate fusion position coordinate and any other initial position coordinate to be fused is determined. Then use the midpoint coordinate as the updated first intermediate fusion position coordinate, and continue to fuse with any other initial position coordinate to be fused. Until there is no initial position coordinate to be fused among the N initial position coordinates, the first position coordinate of the target object A is obtained.
  • multiple initial position coordinates associated with the same target object may be fused in a manner of taking midpoints in sequence, so as to obtain first position coordinates with higher accuracy.
  • a Kalman filter may be introduced here to determine the second position coordinates of the target object at the current moment by means of Kalman filtering.
  • the predicted position coordinates refer to the position coordinates of the target object at the current moment that can be predicted based on the second position coordinates of the previous moment.
  • the coordinates of the observation position may be determined according to the video image at the current moment collected by the collection device, such as the first position coordinates of the target object at the current moment determined above.
  • the embodiment of the present disclosure proposes to jointly determine the observation position coordinates by combining the predicted position coordinates and the first position coordinates determined based on the video images collected by the collection device at the current moment. Finally, the observed position coordinates and the predicted position coordinates can be combined to obtain the second position coordinates of the target object at the current moment.
  • t-1) ATrk(t-1
  • t-1) indicates the predicted position coordinates of the target object at the current moment determined according to the second position coordinates of the target object at the previous moment; Trk(t-1
  • a and B represent the parameters of the Kalman filter matrix, where A represents the state transition matrix in the Kalman filter, and u(t-1) represents the control amount of the Kalman filter at the previous moment, which can be 0.
  • the covariance matrix of the observation position coordinates of the target object at the current moment can be determined according to the following formula (2):
  • t-1) represents the covariance matrix of the coordinates of the observed position of the target object at the current moment, which can represent the uncertainty of the coordinates of the observed position of the target object at the current moment
  • t-1 ) represents the covariance matrix of the second position coordinate of the target object at the last moment, which can represent the uncertainty of the second position coordinate of the target object at the last moment
  • Q represents the system process covariance matrix introduced by the Kalman filter, It is used to represent the error of the state transition matrix compared to the actual process.
  • the predicted position coordinates of the target object at the current moment are obtained, the predicted position coordinates of the target object and the first position coordinates can be combined to determine the observed position coordinates of the target object at the current moment, which will be described in detail later.
  • the second position coordinates of the target object at the current moment can be determined according to the following formula (3) in the Kalman filter formula:
  • t) Trk(t
  • t) represents the second position coordinate of the target object at the current moment
  • Z(t) represents the observation position coordinate of the target object at the current moment
  • K g (t) represents the filter gain matrix in the Kalman filter
  • the filter gain matrix can be determined by the following formula (4):
  • H represents the parameter matrix in the Kalman filter, which represents the observation matrix
  • R represents the known measurement noise covariance in the Kalman filter.
  • the covariance matrix of the observed position coordinates of the target object at the next moment can be determined based on the covariance matrix, so as to determine the second position of the target object at the next moment. Prepare the coordinates.
  • the target object does not have the second position coordinates of the previous moment.
  • the first position coordinate of the target object at the current moment may be directly determined as the second position coordinate of the target object at the current moment.
  • the predicted position coordinates of the target object at the current moment can be determined according to the second position coordinates of the target object at the previous moment, and further combined with the first position coordinates of the target object at the current moment, the current position of the target object can be obtained.
  • the second position coordinate with higher time accuracy.
  • S4022 Determine the predicted position coordinates associated with the same target object and the first midpoint coordinates of the first position coordinates, and use the first midpoint coordinates as the observed position coordinates of the target object at the current moment.
  • the predicted position coordinates of the N target objects at the current moment can be obtained.
  • the first position coordinates of the M target objects in the target place at the current moment can be obtained.
  • the predicted position coordinates and the first position coordinates associated with the same target object may be determined by a distance-based greedy algorithm. Then, the midpoint coordinates of the predicted position coordinates associated with the same target object and the first position coordinates can be further used as the observed position coordinates of the same target object at the current moment.
  • N may be greater than or equal to M.
  • N there may be a target object that is missed in the video image at the current moment collected by the collection device.
  • the video image of a certain target object cannot be captured due to the occlusion of the obstacle in the target area.
  • the first position coordinates of the target object in the target area are determined based on the video image at the current moment, there will be missed detections.
  • the observed position coordinates of the target object at the current moment can be determined by the predicted position coordinates of the target object.
  • determining the observed position coordinates of the target object at the current moment based on the predicted position coordinates and the first position coordinates of the target object at the current moment may include: based on multiple first position coordinates of the target objects at the current moment. a position coordinate and the predicted position coordinate of the target object at the current moment, determine the first position coordinate of the target object; determine the predicted position coordinate of the target object and the first middle of the first position coordinate point coordinates, and the first midpoint coordinates are taken as the coordinates of the observation position of the target object at the current moment.
  • S40212 use the predicted position coordinates and the first position coordinates with the minimum first distance from the predicted position coordinates as the predicted position coordinates and the first position coordinates associated with the same target object, and the minimum first distance is smaller than the first preset fusion distance threshold.
  • the current moment includes N predicted position coordinates and M observed position coordinates
  • the Euclidean distance between each predicted position coordinate and each observed position coordinate is determined according to the N predicted position coordinates and M observed position coordinates
  • l 11 represents the first distance between the first predicted position coordinate in the N predicted position coordinates and the first observed position coordinate in the M observed position coordinates
  • l 1M represents the first predicted position in the N predicted position coordinates The first distance between the position coordinates and the Mth observation position coordinate in the M observation position coordinates
  • l nm represents the nth prediction position coordinate in the N prediction position coordinates and the mth observation position coordinate in the M observation position coordinates
  • l N1 represents the first distance between the Nth predicted position coordinate in the N predicted position coordinates and the first observed position coordinate in the M observed position coordinates;
  • l NM represents the N predicted position coordinates The first distance between the Nth predicted position coordinate in and the Mth observation position coordinate in the M observation position coordinates.
  • predicted position coordinates and the first position coordinates associated with the same target object may be determined according to the above-mentioned method of determining multiple initial position coordinates associated with the same target object, and the specific process will not be repeated here.
  • determining the first position coordinates of the target object based on the plurality of first position coordinates of the target objects at the current moment and the predicted position coordinates of the target objects at the current moment may include: determining a first distance between the predicted position coordinates of the target object and each of the first position coordinates; forming a first position with the smallest first distance from the predicted position coordinates of the target object among the plurality of first position coordinates coordinates, as the first position coordinates with the target object, the minimum first distance is smaller than the first preset fusion distance threshold.
  • the position coordinates of the target object at the current moment predicted according to the position coordinates of the target object at the historical moment, and the first position coordinates of the target object determined according to the video images collected by the acquisition device at the current moment a
  • the position coordinates of the same target object at different times can be quickly obtained, and on the other hand, the observed position coordinates with high accuracy can be obtained.
  • the target tracking method provided by the embodiment of the present disclosure further includes the following S501 to S502:
  • the predicted position coordinates of the missed target object at the current moment are taken as the observed position coordinates of the missed target object at the current moment.
  • the target object A in the video images of the current moment collected by the capture device 1 and the capture device 2 among the multiple capture devices are all blocked.
  • the first position coordinate mark of the target object A is empty, and at this time, the target object A is regarded as the missed target object.
  • the second position coordinates of the target object A at the historical moment will be used. Since the target object A will be collected by the acquisition device in the process of entering the target place, the second position coordinates of the target object A at the historical moment can be determined, so that the target object A can be determined at the current moment according to the method of Kalman filtering.
  • the predicted location coordinates of If the first position coordinates of the target object A at the current moment are empty, the predicted position coordinates of the target object A at the current moment can be directly used as the observed position coordinates at the current moment.
  • the occluded target object when there is an occluded target object in the video image at the current moment collected by the collection device, the occluded target object may be determined based on the second position coordinates of the occluded target object in the historical moment. The coordinates of the observed position at the current moment, so as to determine the second position coordinates of the target object with higher accuracy at the current moment.
  • the target object includes multiple objects.
  • the target tracking method provided by the embodiment of the present disclosure further includes the following S601 to S602:
  • a capturing device for capturing images of employees may be set at the entrance of the factory.
  • Feature extraction is performed based on the collected employee images, for example, facial features and/or human body features in the employee images are extracted.
  • the identity of each employee entering the factory is determined based on the extracted characteristic information and the pre-stored characteristic information of each employee in the employee identity database.
  • the identity identifier associated with the target object may be marked in the map position indicated by the second position coordinates. Then, by connecting the second position coordinates of multiple moments with the same identity identifier, the movement trajectories of different target objects in the map can be obtained.
  • the map may be a pre-built high-resolution map.
  • the pre-built highland map has a corresponding relationship with the target site, and the two can be presented 1:1 in the same coordinate system. Therefore, based on the second position coordinates of the target objects marked with the same identity identifier at multiple times, trajectory data representing the movement trajectory of each target object in the target place can be generated.
  • the movement trajectory of each target object in the target place can be quickly determined according to the identity identifier of the target object and the second position coordinates at different times.
  • clustering these target objects can form a target group. Errors may occur when marking the identity identifier with the second position coordinates of the target objects in the same target group, for example, marking the identity identifier of target object A in the target group to target object B, marking the identity identifier of target object B to Target object A, that is, the serial number problem occurs.
  • marking the identity identifier of target object A in the target group to target object B marking the identity identifier of target object B to Target object A, that is, the serial number problem occurs.
  • the target object A and the target object B belong to the same target group, when a serial number occurs, the distance between the target objects in which the serial number occurs is relatively close, so the impact on the trajectory data is small.
  • the target tracking method further includes the following S701 to S703:
  • S701 based on the second position coordinates of the multiple target objects at the current moment, detect whether there are target objects that deviate from the target group; the target group is obtained by clustering according to the second position coordinates of the multiple target objects at the previous moment.
  • the second position coordinates of the multiple target objects at the previous moment can be clustered to obtain the target group.
  • the distance of the second position coordinates between different target objects in the target group is less than a preset distance threshold for entering the target group.
  • the identity identifier associated with the target object that deviates from the target group when detecting whether the identity identifier associated with the target object that deviates from the target group is accurate, it includes:
  • a video picture of the current moment of the target object deviating from the target group is acquired. Based on the video picture at the current moment, the feature information of the target object deviating from the target group is extracted. Based on the feature information and the pre-stored identity identifier of the target object and the corresponding feature information, it is determined whether the identity identifier associated with the target object deviating from the target group is accurate. For example, the similarity between the feature information of the currently extracted target object that deviates from the target group and the pre-marked feature information associated with the identity identifier of the target object can be determined.
  • the pre-marked identity identifier of the target object is 001.
  • the similarity between the feature information of the deviation target object extracted from the video image at the current moment and the feature information associated with the identity identifier 001 is less than the preset similarity threshold, it is determined that the identity identifier of the target object 001 is inaccurate .
  • the identity identifier of the target object deviating from the target group is inaccurate, it can be based on the extracted feature information of the target object deviating from the target group and the pre-stored feature information of each employee in the employee identity database. , and re-determine the identity identifier of the target object that deviates from the target group.
  • the identity identifier of the target object leaving the target group is re-verified, which can improve the accuracy of the identity identifier of the target object marked at different times. to improve the accuracy of the trajectory data of the target object.
  • the target tracking method proposed in the embodiment of the present disclosure can accurately determine the second position coordinates of each target object in the target place at the current moment, and this method can be applied to various application scenarios. Taking the application in a factory as an example, after obtaining the second position coordinates of the target object in the target place, as shown in FIG. 9 , the positioning method provided by the embodiment of the present disclosure further includes the following S801 to S802:
  • a coordinate range corresponding to a dangerous target area in the factory may be set in advance in the world coordinates corresponding to the target site. Then, it is determined whether there is a target object entering the target area according to the second position coordinates corresponding to each target object in the determined target place at the current moment and the target location in the corresponding coordinate range. Further, when it is determined that there is a target object entering the target area, an early warning prompt is performed.
  • the early warning prompts may include, but are not limited to, sound and light alarm prompts, voice alarm prompts, and the like. Through the early warning prompts, the safety of employees in the target site can be guaranteed and the safety of the target site can be improved.
  • the target object in the target place can be determined based on a preset target area, such as a preset dangerous area Whether to enter the target area, so as to prompt early warning and improve the safety of the target site.
  • a preset target area such as a preset dangerous area Whether to enter the target area
  • the target tracking process provided by the embodiment of the present disclosure is described by taking the target site as a factory and the target object as an employee as an example:
  • the pixel coordinates of the employees involved in the video picture are corrected to obtain the corrected pixel coordinates of the employees involved in the video picture;
  • the employee's associated identity identifier can be marked in the map position indicated by the employee's second position coordinate; The second position coordinates of each moment, and the trajectory data of each employee is generated.
  • the writing order of each step does not mean a strict execution order but constitutes any limitation on the implementation process, and the specific execution order of each step should be based on its function and possible Internal logic is determined.
  • the embodiment of the present disclosure also provides a target tracking device corresponding to the target tracking method. Reference may be made to the implementation of the method, and repeated descriptions will not be repeated.
  • the target tracking apparatus includes:
  • the acquisition module 901 is used to acquire the video images at the current moment collected by multiple collection devices set in the target site; the collection perspectives of the multiple collection devices in the target site are different, and the video images include the region of interest of the target object in the target site ;
  • a determination module 902 configured to determine the first position coordinates of each of the target objects at the current moment based on the video images at the current moment collected by multiple collection devices;
  • the tracking module 903 is configured to, for each of the target objects, determine the second position coordinates of the target object at the current moment based on the first position coordinates of the target object and the second position coordinates of the target object at the previous moment.
  • the tracking module 903 is configured to, for each of the target objects, determine, based on the first position coordinates of the target object and the second position coordinates of the target object at the previous moment, that the target object is The coordinates of the second position at the current moment include:
  • the second position coordinates of the target object at the current moment are determined.
  • the method when the tracking module 903 is used to determine the observed position coordinates of the target object at the current moment based on the predicted position coordinates and the first position coordinates of the target object at the current moment, the method includes:
  • the tracking module 903 when used to determine the predicted position coordinates and the first position coordinates associated with the same target object based on the predicted position coordinates and the first position coordinates of the multiple target objects at the current moment, include:
  • the predicted position coordinates and the first position coordinates with the minimum first distance from the predicted position coordinates are taken as the predicted position coordinates and the first position coordinates associated with the same target object, and the minimum first distance is smaller than the first preset fusion distance threshold. .
  • the tracking module 903 is further configured to:
  • the predicted position coordinates of the missed target object at the current moment are taken as the observed position coordinates of the missed target object at the current moment.
  • the target object includes multiple objects
  • the tracking module 903 is further configured to:
  • the trajectory data of each target object is generated.
  • the tracking module 903 is further configured to:
  • the target group is obtained by clustering according to the second position coordinates of the multiple target objects at the previous moment;
  • the identity identifiers of the target objects deviating from the target group are inaccurate, the identity identifiers associated with the target objects deviating from the target group are corrected.
  • the method includes:
  • the method when the determining module 902 is used to determine the first position coordinates of each target object at the current moment based on the video images at the current moment collected by multiple collection devices, the method includes:
  • the target object corresponding to the region of interest in the video picture is at the current moment. the coordinates of the initial position in the target site;
  • the initial position coordinates belonging to the same target object in the initial position coordinates are fused to obtain the first position coordinates of the target object in the target place at the current moment.
  • the method includes:
  • the determining module 902 is used to determine the target corresponding to the region of interest based on the pixel coordinates of the region of interest in the video image at the current moment collected by each acquisition device and the parameter information of the acquisition device
  • the initial position coordinates of the object in the target location at the current moment include:
  • the determining module 902 when the determining module 902 is used to fuse the initial position coordinates belonging to the same target object in the initial coordinate positions to obtain the first position coordinates of the target object in the target place at the current moment, include:
  • the multiple initial position coordinates associated with the same target object are sequentially fused to obtain the first position coordinates of the target object in the target place at the current moment.
  • determining module 902 when the determining module 902 is used to sequentially fuse multiple initial position coordinates associated with the same target object to obtain the first position coordinates of the target object in the target place at the current moment, include:
  • the first intermediate fusion position coordinates are fused with any other initial position coordinates to be fused to generate the second intermediate fusion position coordinates, and the second intermediate fusion position coordinates are used as the updated first intermediate fusion position coordinates, and return to generate the first intermediate fusion position coordinates.
  • Step 2 The intermediate position coordinates are fused until there is no initial position coordinates to be fused.
  • the determining module 902 when used to fuse the first intermediate fused position coordinates with any other initial position coordinates to be fused to generate the second intermediate fused position coordinates, it includes:
  • the method includes:
  • the target object corresponding to each region of interest in the first video picture in the arbitrary two video pictures is the first target object
  • each of the second video pictures in the arbitrary two video pictures is the first target object.
  • the target object corresponding to each region of interest is a second target object, and the second distance between the initial position coordinates of the first target object and the initial position coordinates of each second target object in the video picture at the second current moment is determined;
  • the initial position coordinates of the first target object and the initial position coordinates of the second target object having the smallest second distance from the first target object are regarded as being associated with the same target object multiple initial position coordinates; the minimum second distance is less than the second preset fusion distance threshold.
  • the determining module 902 is further configured to:
  • an early warning prompt is given.
  • an embodiment of the present disclosure further provides an electronic device 1100 .
  • a schematic structural diagram of the electronic device 1100 provided by an embodiment of the present disclosure includes:
  • Embodiments of the present disclosure further provide a computer-readable storage medium, where a computer program is stored on the computer-readable storage medium, and when the computer program is run by a processor, the steps of the target tracking method described in the above method embodiments are executed.
  • the storage medium may be a volatile or non-volatile computer-readable storage medium.
  • Embodiments of the present disclosure further provide a computer program product, where the computer program product carries program codes, and the instructions included in the program codes can be used to execute the steps of the target tracking method described in the foregoing method embodiments.
  • the computer program product carries program codes
  • the instructions included in the program codes can be used to execute the steps of the target tracking method described in the foregoing method embodiments.
  • the foregoing method please refer to the foregoing method. The embodiments are not repeated here.
  • the above-mentioned computer program product can be specifically implemented by hardware, software or a combination thereof.
  • the computer program product is embodied as a computer storage medium, and in another optional embodiment, the computer program product is embodied as a software product, such as a software development kit (Software Development Kit, SDK), etc. Wait.
  • the units described as separate components may or may not be physically separated, and components displayed as units may or may not be physical units, that is, may be located in one place, or may be distributed to multiple network units. Some or all of the units may be selected according to actual needs to achieve the purpose of the solution in this embodiment.
  • each functional unit in each embodiment of the present disclosure may be integrated into one processing unit, or each unit may exist physically alone, or two or more units may be integrated into one unit.
  • the functions, if implemented in the form of software functional units and sold or used as stand-alone products, may be stored in a processor-executable non-volatile computer-readable storage medium.
  • the computer software products are stored in a storage medium, including Several instructions are used to cause a computer device (which may be a personal computer, a server, or a network device, etc.) to execute all or part of the steps of the methods described in various embodiments of the present disclosure.
  • the aforementioned storage medium includes: U disk, mobile hard disk, read-only memory (Read-Only Memory, ROM), random access memory (Random Access Memory, RAM), magnetic disk or optical disk and other media that can store program codes .

Landscapes

  • Engineering & Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Multimedia (AREA)
  • Image Analysis (AREA)
  • Studio Devices (AREA)

Abstract

The present disclosure provides a target tracking method and apparatus, an electronic device, and a storage medium. The target tracking method comprises: obtaining video pictures at the current moment acquired by a plurality of acquisition devices provided in a target place, the plurality of acquisition devices having different acquisition angles in the target place, and the video pictures comprising regions of interest of target objects in the target place; on the basis of the video pictures at the current moment acquired by the plurality of acquisition devices, determining first position coordinates of each target object at the current moment; and for each target object, on the basis of the first position coordinates of the target object and second position coordinates of the target object at the previous moment, determining second position coordinates of the target object at the current moment.

Description

目标跟踪方法、装置、电子设备及存储介质Target tracking method, device, electronic device and storage medium
相关申请的交叉引用CROSS-REFERENCE TO RELATED APPLICATIONS
本专利申请要求于2021年4月28日提交的、申请号为202110467650.7、发明名称为“一种目标跟踪方法、装置、电子设备及存储介质”的中国专利申请的优先权,该申请以引用的方式并入文本中。This patent application claims the priority of the Chinese patent application filed on April 28, 2021, with the application number of 202110467650.7 and the invention titled "A target tracking method, device, electronic device and storage medium", which is cited by way to be incorporated into the text.
技术领域technical field
本公开涉及计算机视觉技术领域,具体而言,涉及一种目标跟踪方法、装置、电子设备及存储介质。The present disclosure relates to the field of computer vision technology, and in particular, to a target tracking method, apparatus, electronic device, and storage medium.
背景技术Background technique
人工智能技术在打造智能教育,文娱及生活上发挥着越来越重要的作用。其中计算机视觉作为关键的技术之一,应用广泛。比如可以基于计算机视觉的定位技术,对不同场景下的目标场所内的目标对象进行追踪,确定目标场所内目标对象的轨迹。Artificial intelligence technology is playing an increasingly important role in creating intelligent education, entertainment and life. Among them, computer vision, as one of the key technologies, is widely used. For example, the positioning technology based on computer vision can track the target object in the target place in different scenarios, and determine the trajectory of the target object in the target place.
在基于计算机视觉对目标对象进行追踪的过程中,可以通过相机采集的目标场所图像,确定目标场所图像中的目标对象在不同时刻的位置,进一步根据目标对象在不同时刻的位置对目标对象进行追踪。In the process of tracking the target object based on computer vision, the image of the target site collected by the camera can be used to determine the position of the target object in the target site image at different times, and further track the target object according to the position of the target object at different times. .
发明内容SUMMARY OF THE INVENTION
本公开实施例至少提供一种目标追踪方案。The embodiments of the present disclosure provide at least one target tracking solution.
第一方面,本公开实施例提供了一种目标追踪方法,包括:In a first aspect, an embodiment of the present disclosure provides a target tracking method, including:
获取目标场所内设置的多个采集设备采集的当前时刻的视频画面;所述多个采集设备在所述目标场所中的采集视角不同,所述视频画面中包括所述目标场所中目标对象的感兴趣区域;Acquire the video images at the current moment collected by multiple collection devices set in the target site; the multiple collection devices have different collection perspectives in the target site, and the video images include the sense of the target object in the target site. area of interest;
基于所述多个采集设备采集的当前时刻的视频画面,确定各个所述目标对象在当前时刻的第一位置坐标;Determine the first position coordinates of each of the target objects at the current moment based on the video images at the current moment collected by the multiple collection devices;
针对各个所述目标对象,基于该目标对象的所述第一位置坐标和该目标对象在上一时刻的第二位置坐标,确定该目标对象在当前时刻的第二位置坐标。For each of the target objects, the second position coordinates of the target object at the current moment are determined based on the first position coordinates of the target object and the second position coordinates of the target object at the previous moment.
第二方面,本公开实施例提供了一种目标追踪装置,包括:In a second aspect, an embodiment of the present disclosure provides a target tracking device, including:
获取模块,用于获取目标场所内设置的多个采集设备采集的当前时刻的视频画面;所述多个采集设备在所述目标场所中的采集视角不同,所述视频画面中包括所述目标场所中目标对象的感兴趣区域;an acquisition module, configured to acquire the video images at the current moment collected by multiple collection devices set in the target place; the multiple collection devices have different collection perspectives in the target place, and the video images include the target place the region of interest of the target object;
确定模块,用于基于所述多个采集设备采集的当前时刻的视频画面,确定各个所述目标对象在当前时刻的第一位置坐标;A determination module, configured to determine the first position coordinates of each of the target objects at the current moment based on the video images at the current moment collected by the multiple collection devices;
追踪模块,用于针对各个所述目标对象,基于该目标对象的所述第一位置坐标和该目标对象在上一时刻的第二位置坐标,确定该目标对象在当前时刻的第二位置坐标。The tracking module is configured to, for each of the target objects, determine the second position coordinates of the target object at the current moment based on the first position coordinates of the target object and the second position coordinates of the target object at the previous moment.
第三方面,本公开实施例提供了一种电子设备,包括:处理器、存储器和总线,所述存储器存储有所述处理器可执行的机器可读指令,当电子设备运行时,所述处理器与所述存储器之间通过总线通信,所述机器可读指令被所述处理器执行时执行如第一方面所述的目标追踪方法的步骤。In a third aspect, embodiments of the present disclosure provide an electronic device, including: a processor, a memory, and a bus, where the memory stores machine-readable instructions executable by the processor, and when the electronic device runs, the processing The processor and the memory communicate through a bus, and the machine-readable instructions execute the steps of the target tracking method according to the first aspect when the machine-readable instructions are executed by the processor.
第四方面,本公开实施例提供了一种计算机可读存储介质,该计算机可读存储介质上存储有计算机程序,该计算机程序被处理器运行时执行如第一方面所述的目标追踪方法的步骤。In a fourth aspect, an embodiment of the present disclosure provides a computer-readable storage medium, where a computer program is stored on the computer-readable storage medium, and the computer program is executed by a processor to execute the target tracking method according to the first aspect. step.
第五方面,本公开实施例提供了一种计算机程序产品,该计算机程序产品包括计算机程序并存储于存储介质上,该计算机程序被处理器执行时执行如第一方面所述的目标追踪方法的步骤。In a fifth aspect, an embodiment of the present disclosure provides a computer program product, the computer program product includes a computer program and is stored on a storage medium, and when the computer program is executed by a processor, executes the target tracking method according to the first aspect. step.
为使本公开的上述目的、特征和优点能更明显易懂,下文特举一些实施例,并配合所附附图,作详细说明如下。In order to make the above-mentioned objects, features and advantages of the present disclosure more obvious and easy to understand, some embodiments are hereinafter described in detail together with the accompanying drawings.
附图说明Description of drawings
图1示出了本公开实施例所提供的一种目标追踪方法的流程图;FIG. 1 shows a flowchart of a target tracking method provided by an embodiment of the present disclosure;
图2示出了本公开实施例所提供的一种确定目标对象的第一位置坐标的方法流程图;FIG. 2 shows a flowchart of a method for determining a first position coordinate of a target object provided by an embodiment of the present disclosure;
图3示出了本公开实施例所提供的一种针对当前时刻的视频画面进行目标检测得到的检测框示意图;3 shows a schematic diagram of a detection frame obtained by performing target detection on a video image at the current moment according to an embodiment of the present disclosure;
图4示出了本公开实施例所提供的一种确定同一目标对象在当前时刻的第一位置坐标的具体方法流程图;4 shows a flowchart of a specific method for determining the first position coordinates of the same target object at the current moment provided by an embodiment of the present disclosure;
图5示出了本公开实施例所提供的一种确定目标对象在当前时刻的第二位置坐标的方法流程图;5 shows a flowchart of a method for determining the second position coordinates of a target object at the current moment provided by an embodiment of the present disclosure;
图6示出了本公开实施例所提供的一种确定漏检的目标对象在当前时刻的观测位置坐标的方法流程图;FIG. 6 shows a flowchart of a method for determining the coordinates of the observed position of a target object that is missed at the current moment provided by an embodiment of the present disclosure;
图7示出了本公开实施例所提供的一种确定每个目标对象的轨迹数据的方法流程图;FIG. 7 shows a flowchart of a method for determining trajectory data of each target object provided by an embodiment of the present disclosure;
图8示出了本公开实施例所提供的一种针对偏离目标群的目标对象的身份标识符进行修正的方法流程图;FIG. 8 shows a flowchart of a method for revising an identity identifier of a target object that deviates from a target group provided by an embodiment of the present disclosure;
图9示出了本公开实施例所提供的一种预警方法;FIG. 9 shows an early warning method provided by an embodiment of the present disclosure;
图10示出了本公开实施例所提供的一种针对目标对象进行追踪的场景示意图;FIG. 10 shows a schematic diagram of a scene for tracking a target object provided by an embodiment of the present disclosure;
图11示出了本公开实施例所提供的一种目标追踪装置的结构示意图;FIG. 11 shows a schematic structural diagram of a target tracking device provided by an embodiment of the present disclosure;
图12示出了本公开实施例所提供的一种电子设备的示意图。FIG. 12 shows a schematic diagram of an electronic device provided by an embodiment of the present disclosure.
具体实施方式Detailed ways
为使本公开实施例的目的、技术方案和优点更加清楚,下面将结合本公开实施例中附图,对本公开实施例中的技术方案进行清楚、完整地描述,显然,所描述的实施例仅仅是本公开一部分实施例,而不是全部的实施例。通常在此处附图中描述和示出的本公开实施例的组件可以以各种不同的配置来布置和设计。因此,以下对在附图中提供的本公开的实施例的详细描述并非旨在限制要求保护的本公开的范围,而是仅仅表示本公开的选定实施例。基于本公开的实施例,本领域技术人员在没有做出创造性劳动的前提下所获得的所有其他实施例,都属于本公开保护的范围。In order to make the purposes, technical solutions and advantages of the embodiments of the present disclosure more clear, the technical solutions in the embodiments of the present disclosure will be clearly and completely described below with reference to the accompanying drawings in the embodiments of the present disclosure. Obviously, the described embodiments are only These are some, but not all, embodiments of the present disclosure. The components of the disclosed embodiments generally described and illustrated in the drawings herein may be arranged and designed in a variety of different configurations. Therefore, the following detailed description of the embodiments of the disclosure provided in the accompanying drawings is not intended to limit the scope of the disclosure as claimed, but is merely representative of selected embodiments of the disclosure. Based on the embodiments of the present disclosure, all other embodiments obtained by those skilled in the art without creative work fall within the protection scope of the present disclosure.
应注意到:相似的标号和字母在下面的附图中表示类似项,因此,一旦某一项在一个附图中被定义,则在随后的附图中不需要对其进行进一步定义和解释。It should be noted that like numerals and letters refer to like items in the following figures, so once an item is defined in one figure, it does not require further definition and explanation in subsequent figures.
本文中术语“和/或”,仅仅是描述一种关联关系,表示可以存在三种关系,例如,A和/或B,可以表示:单独存在A,同时存在A和B,单独存在B这三种情况。另外,本文中术语“至少一种”表示多种中的任意一种或多种中的至少两种的任意组合,例如,包括A、B、C中的至少一种,可以表示包括从A、B和C构成的集合中选择的任意一个或多个元素。The term "and/or" in this paper only describes an association relationship, which means that there can be three kinds of relationships, for example, A and/or B, which can mean: the existence of A alone, the existence of A and B at the same time, the existence of B alone. a situation. In addition, the term "at least one" herein refers to any combination of any one of the plurality or at least two of the plurality, for example, including at least one of A, B, and C, and may mean including from A, B, and C. Any one or more elements selected from the set of B and C.
在很多应用场景中,通常需要对一个场所内的目标对象进行追踪。比如针对工厂内,需要追踪员工是否有进入危险区域的趋势。在商场中,可以追踪顾客的行动轨迹。可以通过相机采集的图像,确定目标对象的位置,进而完成对目标对象的追踪。然而针对一些场地复杂面积较大的目标场所,在基于相机对目标对象进行追踪的过程中,可能存在无法捕捉到目标对象的情况,即存在追踪中断的问题;还有可能由于存在遮挡区域,无法完成在这些遮挡区域中的目标对象的追踪。In many application scenarios, it is usually necessary to track the target object in a place. For example, in a factory, it is necessary to track whether employees have a tendency to enter dangerous areas. In shopping malls, the movements of customers can be tracked. The position of the target object can be determined through the image captured by the camera, and then the tracking of the target object can be completed. However, for some target sites with complex and large areas, in the process of tracking the target object based on the camera, there may be situations where the target object cannot be captured, that is, there is a problem of tracking interruption; Tracking of target objects in these occluded regions is done.
基于上述研究,本公开提供了一种目标追踪方法,目标场所内设置的采集设备的采集视角不同,且目标场所中的每个目标对象至少被两个采集设备同时采集到。这样可以完成对目标场所中的目标对象进行全面较准确的定位,得到目标对象在当前时刻的第一位置坐标。进一步结合目标对象在上一时刻准确度较高的第二位置坐标和在当前时刻的第一位置坐标,准确确定目标对象在当前时刻的第二位置坐标,即完成对进入目标场所中的目标对象的追踪。Based on the above research, the present disclosure provides a target tracking method, in which the acquisition perspectives of the acquisition devices set in the target site are different, and each target object in the target site is simultaneously acquired by at least two acquisition devices. In this way, a comprehensive and accurate positioning of the target object in the target place can be completed, and the first position coordinates of the target object at the current moment can be obtained. Further combining the second position coordinates of the target object with higher accuracy at the previous moment and the first position coordinates at the current moment, accurately determine the second position coordinates of the target object at the current moment, that is, to complete the target object entering the target place. tracking.
为便于对本实施例进行理解,首先对本公开实施例所公开的一种目标追踪方法进行详细介绍。本公开实施例所提供的定位方法的执行主体为具有计算能力的计算机设备,该计算机设备例如包括:服务器或其它处理设备。在一些可能的实现方式中,该目标追踪方法可以通过处理器调用存储器中存储的计算机可读指令的方式来实现。In order to facilitate the understanding of this embodiment, a target tracking method disclosed in this embodiment of the present disclosure is first introduced in detail. The execution body of the positioning method provided by the embodiment of the present disclosure is a computer device with computing capability, and the computer device includes, for example, a server or other processing device. In some possible implementations, the target tracking method may be implemented by a processor invoking computer-readable instructions stored in a memory.
参见图1所示,为本公开实施例提供的目标追踪方法的流程图,所述方法包括步骤S101~S103。Referring to FIG. 1, which is a flowchart of a target tracking method provided by an embodiment of the present disclosure, the method includes steps S101-S103.
S101,获取目标场所内设置的多个采集设备采集的当前时刻的视频画面;多个采集设备在目标场所中的采集视角不同,视频画面中包括目标场所中目标对象的感兴趣区域。S101: Acquire a video image at the current moment collected by multiple collection devices set in the target site; the multiple collection devices have different collection perspectives in the target site, and the video image includes the region of interest of the target object in the target site.
示例性地,针对不同的应用场景,目标场所可以为与该应用场景对应的场所。比如需要对工厂内的员工进行定位的情况下,目标场所可以为工厂;需要对商场内的顾客进行定位的情况下,目标场所可以为商场;需要对体育馆内的运动员进行定位的情况下,目标场所可以为体育馆。Exemplarily, for different application scenarios, the target location may be a location corresponding to the application scenario. For example, if you need to locate the employees in the factory, the target place can be the factory; if you need to locate the customers in the mall, the target place can be the shopping mall; if you need to locate the athletes in the gym, the target The venue may be a gymnasium.
示例性地,感兴趣区域为目标场所内需要进行定位的对象(比如上述提到的员工、顾客和运动员)在视频画面中所在区域。Exemplarily, the region of interest is the region in the video frame where objects that need to be positioned (such as the aforementioned employees, customers and athletes) are located in the target venue.
示例性地,采集设备可以为单目摄像机或者双目摄像机。目标场所内可以设置多个采集设备。 针对不同的目标场所,可以根据目标场所的实际场地确定多个采集设备的安装位置。比如可以使得采集设备在目标场所中的采集视角不同,以覆盖目标场所的全部区域,不留死角,另外,考虑到采集设备过多会导致同一时刻采集的视频画面过多,因此会影响视频画面的处理速度,故在目标场所中安装采集设备时,需要同时考虑到采集设备的安装角度、以及数量,比如可以使得进入目标场所中的每个目标对象可以同时被两个采集设备采集到,这样目标场所内设置的多个采集设备可以完整的采集到目标场所整个区域的当前视频画面。Exemplarily, the acquisition device may be a monocular camera or a binocular camera. Multiple acquisition devices can be set up in the target site. For different target sites, the installation positions of multiple collection devices can be determined according to the actual site of the target site. For example, the acquisition angles of the acquisition devices in the target site can be made different, so as to cover the entire area of the target site without leaving a dead angle. In addition, considering that too many acquisition devices will lead to too many video images captured at the same time, it will affect the video images. Therefore, when installing the acquisition equipment in the target site, it is necessary to consider the installation angle and quantity of the acquisition equipment. For example, each target object entering the target site can be acquired by two acquisition equipment at the same time, so that Multiple acquisition devices set up in the target site can completely capture the current video image of the entire area of the target site.
S102,基于多个采集设备采集的当前时刻的视频画面,确定各个目标对象在当前时刻的第一位置坐标。S102: Determine the first position coordinates of each target object at the current moment based on the video images at the current moment collected by multiple collection devices.
示例性地,目标对象为目标场所内需要进行定位的对象,比如上述提到的员工、顾客和运动员。以目标场所为工厂为例,工厂内的各个员工即为目标对象。由于采集设备布置时考虑了对目标对象进行多视角采集,所以,同一目标对象在同一个时刻可以至少被两个采集设备采集到,得到至少两张视频画面,每张视频画面中各涉及一个感兴趣区域。Exemplarily, the target objects are objects that need to be positioned in the target site, such as the aforementioned employees, customers and athletes. Taking the target site as a factory as an example, each employee in the factory is the target object. Since the multi-view acquisition of the target object is considered when the acquisition device is arranged, the same target object can be acquired by at least two acquisition devices at the same time, and at least two video images can be obtained, each of which involves a sensory image. area of interest.
示例性地,以目标场所为工厂为例。比如,该工厂内的员工为2个人,即目标对象为2个。由于同一个员工至少需要被2个采集设备采集到,所以可以假设该工厂内设置有2个摄像机。在这种情况下,一种可能的情况是,在早上9点,两个摄像机各采集到一张视频画面,例如,视频画面1和视频画面2;每张视频画面中各涉及2个感兴趣区域,例如,在视频画面1中,员工1对应的感兴趣区域1、员工2对应的感兴趣区域2;在视频画面2中,员工1对应的感兴趣区域3、员工2对应的感兴趣区域4。Illustratively, take the target site as a factory as an example. For example, the number of employees in the factory is 2, that is, the target audience is 2. Since the same employee needs to be captured by at least 2 capture devices, it can be assumed that there are 2 cameras in the factory. In this case, a possible situation is that at 9:00 am, each of the two cameras captures a video frame, for example, video frame 1 and video frame 2; each video frame involves 2 interested parties. For example, in video screen 1, the area of interest 1 corresponding to employee 1 and the area of interest 2 corresponding to employee 2; in video screen 2, the area of interest 3 corresponding to employee 1 and the area of interest corresponding to employee 2 4.
示例性地,考虑到目标场所中的每个目标对象至少被两个采集设备同时采集到,因此针对目标场所中的每个目标对象,均可以按照多个采集设备采集的包含该目标对象的感兴趣区域的当前时刻的视频画面,来确定该目标对象在当前时刻的第一位置坐标。Exemplarily, considering that each target object in the target site is collected by at least two collection devices at the same time, for each target object in the target site, the senses including the target object collected by multiple collection devices can be used. The video picture of the region of interest at the current moment is used to determine the first position coordinates of the target object at the current moment.
示例性地,目标对象的第一位置坐标可以指目标对象在针对目标场所预先构建的世界坐标系中的位置坐标。这样在基于当前时刻的视频画面,确定目标对象的第一位置坐标时,可以基于当前时刻的视频画面中的感兴趣区域在视频画面中的像素坐标,以及采集该视频画面的采集设备的参数信息,确定该视频画面中感兴趣区域对应的目标对象在目标场所中的初始位置坐标。考虑到不同采集设备的参数信息之间存在一些误差,因此基于不同采集设备采集到的当前时刻的视频画面确定的感兴趣区域对应的目标对象中属于同一目标对象的初始位置坐标会有一些差距。可以对属于同一目标对象的初始位置坐标进行融合,然后根据同一目标对象在当前时刻关联的多个初始位置坐标,确定该目标对象在当前时刻的第一位置坐标。Exemplarily, the first position coordinates of the target object may refer to the position coordinates of the target object in a world coordinate system pre-built for the target location. In this way, when determining the first position coordinates of the target object based on the video image at the current moment, the pixel coordinates of the region of interest in the video image at the current moment in the video image can be used as well as the parameter information of the acquisition device that collects the video image. , and determine the initial position coordinates of the target object corresponding to the region of interest in the video picture in the target place. Considering that there are some errors between the parameter information of different collection devices, the initial position coordinates of the same target object in the target objects corresponding to the region of interest determined based on the video images collected by different collection devices at the current moment will have some differences. The initial position coordinates belonging to the same target object may be fused, and then the first position coordinates of the target object at the current moment are determined according to multiple initial position coordinates associated with the same target object at the current moment.
示例性地,目标场所对应的世界坐标系可以以目标场所中一固定位置为坐标原点,建立唯一世界坐标系。比如可以以目标场所地面中心点为坐标系原点,在地面上设定一个方向作为世界坐标系X轴的正方向,在地面上设定与X轴垂直的方向作为世界坐标系Y轴的正方向,将垂直与地面向上的方向作为世界坐标系Z轴的正方向。Exemplarily, the world coordinate system corresponding to the target site may take a fixed position in the target site as the coordinate origin to establish a unique world coordinate system. For example, you can take the center point of the ground of the target site as the origin of the coordinate system, set a direction on the ground as the positive direction of the X-axis of the world coordinate system, and set the direction perpendicular to the X-axis on the ground as the positive direction of the Y-axis of the world coordinate system , take the vertical and ground-up direction as the positive direction of the Z-axis of the world coordinate system.
S103,针对各个目标对象,基于该目标对象的第一位置坐标和该目标对象在上一时刻的第二位置坐标,确定该目标对象在当前时刻的第二位置坐标。S103 , for each target object, determine the second position coordinates of the target object at the current moment based on the first position coordinates of the target object and the second position coordinates of the target object at the previous moment.
示例性地,在得到目标对象在当前时刻的第一位置坐标后,可以基于目标场所中的目标对象在上一时刻的第二位置坐标,以及目标场所中的目标对象在当前时刻的第一位置坐标,对同一目标对象在不同时刻的位置进行时序关联。确定目标场所中的目标对象在当前时刻的第二位置坐标,后续将目标场所中的各目标对象在不同时刻的第二位置坐标进行关联,可以得到各目标对象在目标场所中的移动轨迹。Exemplarily, after obtaining the first position coordinates of the target object at the current moment, it can be based on the second position coordinates of the target object in the target place at the previous moment, and the first position of the target object in the target place at the current moment. Coordinates, to perform temporal association on the position of the same target object at different times. Determine the second position coordinates of the target object in the target place at the current moment, and then associate the second position coordinates of each target object in the target place at different times to obtain the movement track of each target object in the target place.
示例性地,考虑到第一位置坐标是基于采集设备采集到的,考虑到采集设备的外参信息存在一定的误差,因此基于采集设备采集的当前时刻的视频画面确定的目标对象在当前时刻的第一位置坐标也会存在一定的误差。因此在对同一目标对象在不同时刻的位置进行时序关联的过程中,会对目标对象的第一位置坐标进行修正,确定目标对象在当前时刻准确度较高的第二位置坐标。Exemplarily, considering that the first position coordinates are collected based on the collection device, and considering that there is a certain error in the external parameter information of the collection device, the target object determined based on the video image at the current moment collected by the collection device is at the current moment. The first position coordinates also have certain errors. Therefore, in the process of correlating the positions of the same target object at different times, the first position coordinates of the target object are corrected to determine the second position coordinates of the target object with higher accuracy at the current moment.
示例性地,目标对象在上一时刻的第二位置坐标的确定方式与确定目标对象在当前时刻的第二位置坐标的方式相似。因此本公开主要阐述确定目标对象在当前时刻的第二位置坐标的过程进行阐述。Exemplarily, the manner of determining the second position coordinates of the target object at the previous moment is similar to the manner of determining the second position coordinates of the target object at the current moment. Therefore, the present disclosure mainly describes the process of determining the second position coordinates of the target object at the current moment.
本公开实施例中,目标场所内设置的采集设备的采集视角不同,这样可以完成对目标场所中的目标对象进行全面较准确的定位,得到目标对象在当前时刻的第一位置坐标。进一步结合目标对象在上一时刻准确度较高的第二位置坐标和在当前时刻的第一位置坐标,准确确定目标对象在当前时刻的第二位置坐标,即完成对进入目标场所中的目标对象的追踪。In the embodiment of the present disclosure, the collection viewing angles of the collection devices set in the target place are different, so that the target object in the target place can be positioned comprehensively and accurately, and the first position coordinates of the target object at the current moment can be obtained. Further combining the second position coordinates of the target object with higher accuracy at the previous moment and the first position coordinates at the current moment, accurately determine the second position coordinates of the target object at the current moment, that is, to complete the target object entering the target place. tracking.
下面将结合具体实施例对上述S101~S103进行阐述。The foregoing S101 to S103 will be described below with reference to specific embodiments.
针对上述S102,在基于多个采集设备采集的当前时刻的视频画面,确定各个目标对象在当前 时刻的第一位置坐标时,如图2所示,包括以下S201~S203:For the above-mentioned S102, when determining the first position coordinates of each target object at the current moment based on the video pictures at the current moment collected by multiple collection devices, as shown in Figure 2, the following S201 to S203 are included:
S201,获取多个采集设备分别采集的当前时刻的视频画面中的感兴趣区域的像素坐标。S201: Acquire pixel coordinates of a region of interest in a video image at the current moment collected by multiple collection devices respectively.
示例性地,可以基于预先训练的用于进行目标检测的神经网络来识别当前时刻的视频画面中的目标对象的感兴趣区域。进一步可以读取感兴趣区域中设定位置点在视频画面对应的图像坐标系中的像素坐标,将该设定位置点对应的像素坐标作为感兴趣区域的像素坐标。Exemplarily, the region of interest of the target object in the video picture at the current moment may be identified based on a pre-trained neural network for target detection. Further, the pixel coordinates of the set position point in the region of interest in the image coordinate system corresponding to the video screen can be read, and the pixel coordinates corresponding to the set position point can be used as the pixel coordinates of the region of interest.
具体地,在获取多个采集设备分别采集的当前时刻的视频画面中的感兴趣区域的像素坐标时,可以包括以下S2011~S2012:Specifically, the following steps S2011 to S2012 may be included when acquiring the pixel coordinates of the region of interest in the video images at the current moment respectively collected by multiple collection devices:
S2011,将多个当前时刻的视频画面输入预先训练的神经网络,得到每个视频画面中的目标检测框;其中,神经网络包含多个用于检测不同尺寸的目标对象的感兴趣区域的目标检测子网络,在视频画面中目标检测框所在区域即为感兴趣区域。S2011, inputting a plurality of video pictures at the current moment into a pre-trained neural network to obtain a target detection frame in each video picture; wherein, the neural network includes a plurality of target detections for detecting regions of interest of target objects of different sizes Sub-network, the region where the target detection frame is located in the video picture is the region of interest.
S2012,提取每个视频画面中的目标检测框上的目标位置点在该视频画面中的像素坐标,得到该视频画面中的感兴趣区域的像素坐标。S2012, extracting the pixel coordinates of the target position point on the target detection frame in each video picture in the video picture, to obtain the pixel coordinates of the region of interest in the video picture.
示例性地,神经网络可以检测出当前时刻的视频画面中包含的每个目标对象的感兴趣区域,并标记出每个目标检测框。如图3所示,为当前时刻的视频画面中包含的目标检测框的示意图。该当前时刻的视频画面中包含两个感兴趣区域对应的目标检测框,分别包括感兴趣区域1的目标检测框A1B1C1D1和感兴趣区域2的目标检测框A2B2C2D2。可以在每个感兴趣区域的目标检测框上提取一个位置点作为目标位置点。比如提取检测框底边的中点作为目标位置点,如图3中通过目标检测框A1B1C1D1底边D1C1的中点K1的像素坐标表示感兴趣区域1的像素坐标。通过目标检测框A2B2C2D2底边D2C2的中点位置点K2的像素坐标表示感兴趣区域2的像素坐标。Exemplarily, the neural network can detect the region of interest of each target object contained in the video picture at the current moment, and mark each target detection frame. As shown in FIG. 3 , it is a schematic diagram of the target detection frame included in the video picture at the current moment. The video image at the current moment includes two target detection frames corresponding to the region of interest, including the target detection frame A1B1C1D1 of the region of interest 1 and the target detection frame A2B2C2D2 of the region of interest 2, respectively. A position point can be extracted as the target position point on the target detection frame of each region of interest. For example, the midpoint of the bottom edge of the detection frame is extracted as the target position point, as shown in FIG. The pixel coordinates of the region of interest 2 are represented by the pixel coordinates of the midpoint position point K2 of the bottom edge D2C2 of the target detection frame A2B2C2D2.
示例性地,考虑到目标对象在目标场所中的位置是变化的,且目标场所中设置的多个采集设备在目标场所中的采集视角不同,因此,不同采集设备在同一时刻采集的当前时刻的视频画面中包含的目标对象的感兴趣区域的尺寸可能不同。为了能够准确地标记出不同尺寸的目标对象的感兴趣区域的检测框,本公开实施例使用的神经网络可以包含多个用于检测不同尺寸的目标对象的感兴趣区域的目标检测子网络。比如可以是特征金字塔网络。该特征金字塔网络中的每个目标检测子网络用于检测识别出当前时刻的视频画面中与该目标检测子网络对应尺寸的目标对象的感兴趣区域。通过该神经网络,可以准确地检测出同一采集设备采集的当前时刻的视频画面中不同尺寸的目标对象的感兴趣区域。Exemplarily, considering that the position of the target object in the target place changes, and the capture angles of multiple capture devices set in the target place are different in the target place, therefore, the current moment captured by different capture devices at the same time is different. The size of the region of interest of the target object contained in the video frame may vary. In order to accurately mark detection frames of regions of interest of target objects of different sizes, the neural network used in the embodiments of the present disclosure may include multiple target detection sub-networks for detecting regions of interest of target objects of different sizes. For example, it can be a feature pyramid network. Each target detection sub-network in the feature pyramid network is used to detect and identify a region of interest of a target object of a size corresponding to the target detection sub-network in the video picture at the current moment. Through the neural network, the regions of interest of target objects of different sizes in the video picture at the current moment collected by the same collection device can be accurately detected.
本公开实施例中,神经网络中包含多个用于检测不同尺寸的目标对象的感兴趣区域的目标检测子网络,这样在通过神经网络对视频画面中的目标对象的感兴趣区域进行目标检测时,可以准确地检测出同一视频画面中不同尺寸的目标对象的感兴趣区域。In the embodiment of the present disclosure, the neural network includes a plurality of target detection sub-networks for detecting regions of interest of target objects of different sizes, so that when the neural network is used to perform target detection on the regions of interest of target objects in a video picture , the region of interest of target objects of different sizes in the same video frame can be accurately detected.
S202,针对每个采集设备采集的当前时刻的视频画面,基于该视频画面中的感兴趣区域的像素坐标和该采集设备的参数信息,确定该视频画面中感兴趣区域对应的目标对象在当前时刻下在目标场所中的初始位置坐标。S202, for the video picture at the current moment collected by each acquisition device, based on the pixel coordinates of the region of interest in the video picture and the parameter information of the acquisition device, determine the target object corresponding to the region of interest in the video picture at the current moment The coordinates of the initial position in the target location.
示例性地,每个采集设备的参数信息可以包含采集设备的单应性矩阵,其中,单应性矩阵可以表示采集设备采集的当前时刻的视频画面对应的图像坐标系和采集设备所处的目标场所对应的世界坐标系之间的转换关系。这样,在得到感兴趣区域在当前时刻的视频画面对应的图像坐标系中的像素坐标后,可以根据采集设备的参数信息,确定感兴趣区域在目标场所对应的世界坐标系中的位置坐标,将感兴趣区域在世界坐标系中的位置坐标作为该感兴趣区域对应的目标对象在世界坐标系中的初始位置坐标。Exemplarily, the parameter information of each acquisition device may include a homography matrix of the acquisition device, wherein the homography matrix may represent the image coordinate system corresponding to the video picture at the current moment collected by the acquisition device and the target where the acquisition device is located. The transformation relationship between the world coordinate systems corresponding to the place. In this way, after obtaining the pixel coordinates in the image coordinate system corresponding to the video image of the region of interest at the current moment, the position coordinates of the region of interest in the world coordinate system corresponding to the target site can be determined according to the parameter information of the acquisition device, and the The position coordinates of the region of interest in the world coordinate system are taken as the initial position coordinates of the target object corresponding to the region of interest in the world coordinate system.
具体地,在针对每个采集设备采集的当前时刻的视频画面,基于该视频画面中的感兴趣区域的像素坐标和该采集设备的参数信息,确定该视频画面中感兴趣区域对应的目标对象在当前时刻下在目标场所中的初始位置坐标时,包括以下S2021~S2022:Specifically, in the video picture at the current moment collected by each acquisition device, based on the pixel coordinates of the region of interest in the video picture and the parameter information of the acquisition device, it is determined that the target object corresponding to the region of interest in the video picture is in The initial position coordinates in the target location at the current moment include the following S2021-S2022:
S2021,针对每个采集设备,基于该采集设备的内参矩阵和畸变参数,对该采集设备采集的视频画面中的感兴趣区域的像素坐标进行修正,得到该视频画面中的感兴趣区域的修正像素坐标。S2021, for each acquisition device, based on the internal parameter matrix and the distortion parameter of the acquisition device, correct the pixel coordinates of the region of interest in the video picture collected by the acquisition device, and obtain the corrected pixel of the region of interest in the video picture coordinate.
示例性地,采集设备的内参矩阵包含
Figure PCTCN2022074956-appb-000001
(f x,f y)表示采集设备的焦距,(c x,c y)表示采集设备采集的当前时刻的视频画面的中心点在图像坐标系中的像素坐标。采集设备的畸变参数包含径向畸变参数和切向畸变系数。在预先得到每个采集设备的内参矩阵和畸变系数后,可以根 据该采集设备的内参矩阵和畸变系数对该采集设备采集的当前时刻的视频画面中的感兴趣区域的像素坐标进行去畸变处理。比如可以通过Opencv软件中的去畸变函数,得到该采集设备采集的当前时刻的视频画面中的感兴趣区域的修正像素坐标。
Exemplarily, the internal parameter matrix of the acquisition device contains
Figure PCTCN2022074956-appb-000001
(f x , f y ) represents the focal length of the acquisition device, and (c x , c y ) represents the pixel coordinates in the image coordinate system of the center point of the video image captured by the acquisition device at the current moment. The distortion parameters of the acquisition device include radial distortion parameters and tangential distortion coefficients. After obtaining the internal parameter matrix and distortion coefficient of each acquisition device in advance, the pixel coordinates of the region of interest in the video image at the current moment collected by the acquisition device can be dedistorted according to the internal parameter matrix and distortion coefficient of the acquisition device. For example, the corrected pixel coordinates of the region of interest in the video image captured by the acquisition device at the current moment can be obtained through the de-distortion function in the Opencv software.
示例性地,可以按照张正友棋盘标定的方式预先确定每个采集设备的内参矩阵和畸变参数。比如可以从不同角度拍摄多张棋盘格图像,检测出图像中的特征点,根据这些特征点在棋盘格图像中的像素坐标,求解出采集设备的内参矩阵和畸变参数,然后不断对内参矩阵和畸变参数进行优化。在优化过程中,可以根据相邻两次得到的内参矩阵和畸变参数对同一像素坐标进行修正处理,通过前后两次的修正像素坐标的差异确定是否结束优化。比如可以在该差异不再降低后,结束优化得到采集设备的内参矩阵和畸变参数。Exemplarily, the internal parameter matrix and distortion parameters of each acquisition device may be predetermined in the manner of Zhang Zhengyou's chessboard calibration. For example, you can take multiple checkerboard images from different angles, detect the feature points in the images, and solve the internal parameter matrix and distortion parameters of the acquisition device according to the pixel coordinates of these feature points in the checkerboard image, and then continuously compare the internal parameter matrix and Distortion parameters are optimized. In the optimization process, the same pixel coordinate can be corrected according to the internal parameter matrix and distortion parameters obtained twice adjacently, and whether to end the optimization is determined by the difference between the two corrected pixel coordinates before and after. For example, after the difference is no longer reduced, the optimization can be ended to obtain the internal parameter matrix and distortion parameters of the acquisition device.
S2022,基于预先确定的该采集设备的单应性矩阵和该视频画面中的感兴趣区域的修正像素坐标,确定该视频画面中的感兴趣区域对应的目标对象的初始位置坐标。S2022 , based on the predetermined homography matrix of the acquisition device and the corrected pixel coordinates of the region of interest in the video picture, determine the initial position coordinates of the target object corresponding to the region of interest in the video picture.
示例性地,单应性矩阵可以表示采集设备采集的当前时刻的视频画面对应的图像坐标系和采集设备位于的目标场所对应的世界坐标系之间的转换关系。该单应性矩阵同样可以预先对采集设备进行标定时确定。比如可以通过采集设备采集带有多个标志物的样本视频画面,预先确定多个标志物与地面(世界坐标系X轴和Y轴所在的平面)的交点在目标场所对应的世界坐标系中的世界坐标。然后根据上述方式确定多个标志物与地面的交点在样本视频画面中对应的修正像素坐标。进一步可以基于多个标志物分别对应的修正像素坐标和世界坐标,确定该采集设备的单应性矩阵。Exemplarily, the homography matrix may represent the conversion relationship between the image coordinate system corresponding to the video frame at the current moment collected by the collecting device and the world coordinate system corresponding to the target location where the collecting device is located. The homography matrix can also be determined when the acquisition device is calibrated in advance. For example, a sample video image with multiple markers can be collected by a collection device, and the intersection of the multiple markers and the ground (the plane where the X and Y axes of the world coordinate system are located) is in the world coordinate system corresponding to the target site. world coordinates. Then, the corrected pixel coordinates corresponding to the intersections of the multiple markers and the ground in the sample video picture are determined according to the above method. Further, the homography matrix of the acquisition device may be determined based on the corrected pixel coordinates and world coordinates corresponding to the plurality of markers respectively.
示例性地,在确定视频画面中的感兴趣区域对应的目标对象的初始位置坐标时,可以根据视频画面中的感兴趣区域的修正像素坐标和采集该当前时刻的视频画面的采集设备的单应性矩阵,得到视频画面中的感兴趣区域在目标场所对应的世界坐标系中的位置坐标,将视频画面中的感兴趣区域在世界坐标系中的位置坐标确定为该感兴趣区域对应的目标对象的初始位置坐标。Exemplarily, when determining the initial position coordinates of the target object corresponding to the region of interest in the video picture, it can be based on the corrected pixel coordinates of the region of interest in the video picture and the homography of the acquisition device that collects the video picture at the current moment. to obtain the position coordinates of the region of interest in the video screen in the world coordinate system corresponding to the target location, and determine the position coordinates of the region of interest in the video screen in the world coordinate system as the target object corresponding to the region of interest The initial position coordinates of .
本公开实施例中,在得到视频画面中感兴趣区域的像素坐标后,先基于采集该视频画面的采集设备的内参矩阵和畸变系数对像素坐标进行修正,从而可以得到准确度较高的修正像素坐标,进一步提高得到的感兴趣区域对应的目标对象在目标场所中的初始位置坐标的准确度。In the embodiment of the present disclosure, after obtaining the pixel coordinates of the region of interest in the video picture, the pixel coordinates are first corrected based on the internal parameter matrix and the distortion coefficient of the capture device that captures the video picture, so that corrected pixels with higher accuracy can be obtained. coordinates, further improving the accuracy of the obtained initial position coordinates of the target object corresponding to the region of interest in the target place.
S203,对初始坐标位置中属于同一目标对象的初始位置坐标进行融合,得到该目标对象在当前时刻下在目标场所中的第一位置坐标。S203 , fuse the initial position coordinates of the same target object in the initial coordinate positions to obtain the first position coordinates of the target object in the target place at the current moment.
示例性地,考虑到不同采集设备的参数信息之间存在一些误差,因此基于不同采集设备采集到的当前时刻的视频画面确定的属于同一目标对象的初始位置坐标会有一些差距。可以对属于同一目标对象的初始位置坐标进行融合,得到该目标对象的第一位置坐标。Exemplarily, considering that there are some errors between the parameter information of different collection devices, there will be some differences in the initial position coordinates of the same target object determined based on the video images collected by different collection devices at the current moment. The initial position coordinates belonging to the same target object may be fused to obtain the first position coordinates of the target object.
本公开实施例中,提出可以先确定感兴趣区域在视频画面中的像素坐标,然后再根据采集设备的参数信息,得到感兴趣区域对应的目标对象在目标场所中的初始位置坐标。进一步对不同视频画面中的初始位置坐标中属于同一目标对象的初始位置坐标进行融合,得到该目标对象准确度较高的第一位置坐标。In the embodiment of the present disclosure, it is proposed that the pixel coordinates of the region of interest in the video picture can be determined first, and then the initial position coordinates of the target object corresponding to the region of interest in the target place can be obtained according to the parameter information of the acquisition device. The initial position coordinates belonging to the same target object among the initial position coordinates in different video pictures are further fused to obtain the first position coordinates with higher accuracy of the target object.
具体地,针对上述S203,在对初始坐标位置中属于同一目标对象的初始位置坐标进行融合,得到该目标对象在当前时刻下在目标场所中的第一位置坐标时,如图4所示,可以包括以下S301~S302:Specifically, for the above S203, when the initial position coordinates belonging to the same target object in the initial coordinate positions are fused to obtain the first position coordinates of the target object in the target place at the current moment, as shown in FIG. Including the following S301~S302:
S301,基于多个当前时刻的视频画面确定的感兴趣区域对应的目标对象的初始位置坐标,确定初始位置坐标中与同一目标对象关联的多个初始位置坐标。S301 , based on the initial position coordinates of the target object corresponding to the region of interest determined from multiple video images at the current moment, determine a plurality of initial position coordinates associated with the same target object in the initial position coordinates.
示例性地,根据上文提到的目标场所中的每个目标对象至少被两个采集设备同时采集到,针对每个目标对象,在同一时刻被不同的采集设备拍摄到的情况下,采集设备的参数信息存在一定的误差,且不同的采集设备的参数信息之间的误差不同。因此,基于不同的当前时刻的视频画面确定的同一目标对象的初始位置坐标可能不同。在对同一目标对象的初始位置坐标进行融合之前,需要先确定与同一目标对象关联的多个初始位置坐标。Exemplarily, according to the above-mentioned target location, each target object is captured by at least two capture devices at the same time, and for each target object, in the case of being captured by different capture devices at the same time, the capture device There is a certain error in the parameter information, and the error between the parameter information of different acquisition devices is different. Therefore, the initial position coordinates of the same target object determined based on video pictures at different current moments may be different. Before fusing the initial position coordinates of the same target object, it is necessary to determine multiple initial position coordinates associated with the same target object.
S302,将与同一目标对象关联的多个初始位置坐标进行依次融合,得到同一目标对象在当前时刻下在目标场所中的第一位置坐标。S302 , successively fuse multiple initial position coordinates associated with the same target object to obtain the first position coordinates of the same target object in the target place at the current moment.
示例性地,假设与同一目标对象关联的多个初始位置坐标包含N个,可以先将前两个进行融合。得到融合后的初始位置坐标,然后将融合后的初始位置坐标与第三个初始位置坐标进行融合。直至与最后一个初始位置坐标进行融合后,将最终融合得到的位置坐标作为该目标对象的第一位置坐标。Exemplarily, assuming that the multiple initial position coordinates associated with the same target object include N, the first two may be fused first. The fused initial position coordinates are obtained, and then the fused initial position coordinates are fused with the third initial position coordinates. After being fused with the last initial position coordinate, the position coordinate obtained by the final fusion is used as the first position coordinate of the target object.
本公开实施例中,考虑到基于不同采集设备采集的视频画面确定的同一目标对象的初始位置坐标会存在一些误差,因此可以通过对多个采集设备采集的同一目标对象的初始位置坐标进行融合,从而可以得到该目标对象准确度较高的第一位置坐标。In the embodiment of the present disclosure, considering that there may be some errors in the initial position coordinates of the same target object determined based on video images collected by different collection devices, the initial position coordinates of the same target object collected by multiple collection devices can be fused, Thus, the first position coordinates of the target object with higher accuracy can be obtained.
在一种实施方式中,针对上述S301,在基于多个当前时刻的视频画面确定的感兴趣区域对应 的目标对象的初始位置坐标,确定初始坐标中与同一目标对象关联的多个初始位置坐标时,包括以下S3011~S3012:In an embodiment, for the above S301, when the initial position coordinates of the target object corresponding to the region of interest determined based on multiple video images at the current moment are determined, when multiple initial position coordinates associated with the same target object in the initial coordinates are determined , including the following S3011~S3012:
S3011,针对当前时刻的任意两个视频画面,确定任意两个视频画面中第一视频画面中每个感兴趣区域对应的目标对象为第一目标对象,该任意两个视频画面中第二视频画面中每个感兴趣区域对应的目标对象为第二目标对象,针对每个所述第一目标对象的初始位置坐标,确定该第一目标对象的初始位置坐标与第二当前时刻的视频画面中各个第二目标对象的初始位置坐标之间的第二距离;S3011, with respect to any two video pictures at the current moment, determine that the target object corresponding to each region of interest in the first video picture in any two video pictures is the first target object, and the second video picture in the arbitrary two video pictures is the first target object. The target object corresponding to each region of interest is a second target object, and for the initial position coordinates of each of the first target objects, determine the initial position coordinates of the first target object and the second current moment in the video picture. a second distance between the initial position coordinates of the second target object;
S3012,将该第一目标对象的初始位置坐标和与该第一目标对象具有最小第二距离的第二目标对象的初始位置坐标,作为与同一目标对象关联的多个初始位置坐标,最小第二距离小于第二预设融合距离阈值。S3012, the initial position coordinates of the first target object and the initial position coordinates of the second target object with the minimum second distance from the first target object are taken as multiple initial position coordinates associated with the same target object, the minimum second The distance is less than the second preset fusion distance threshold.
示例性地,比如目标场所中设置A个采集设备,假设在同一时刻A个采集设备采集的当前时刻的视频画面中均包含至少一个目标对象的感兴趣区域,在该时刻可以得到A组初始位置坐标,A组初始位置坐标构成初始坐标集合s={S1,S2,S3,......SA},其中,S1、S2、S3...SA依次表示为A个采集设备中的第一个采集设备、第二个采集设备、第三个采集设备至第A个采集设备拍摄的当前时刻的视频画面中的目标对象的初始位置坐标集合,简言之,SN表示A个采集设备中第N个采集设备拍摄的当前时刻的视频画面中感兴趣区域对应的目标对象的初始位置坐标集合,N为大于等于1且小于等于A的整数。下面将以下当前时刻的任意两个视频画面为第一个采集设备和第二个采集设备采集的当前时刻的视频画面为例,说明如何确定与同一目标对象关联的多个初始位置坐标:Exemplarily, for example, A collection devices are set up in the target site, and it is assumed that the video images of the current moment collected by the A collection devices at the same time all include the region of interest of at least one target object, and the initial position of the A group can be obtained at this moment. Coordinates, the initial position coordinates of the A group constitute the initial coordinate set s={S1, S2, S3, ...... SA}, where S1, S2, S3 ...... SA are sequentially expressed as the No. 1 in the A collection device The set of initial position coordinates of the target object in the video picture at the current moment captured by one acquisition device, the second acquisition device, the third acquisition device to the A-th acquisition device. The set of initial position coordinates of the target object corresponding to the region of interest in the video image captured by the Nth collection device at the current moment, where N is an integer greater than or equal to 1 and less than or equal to A. The following is an example of how to determine multiple initial position coordinates associated with the same target object by taking any two video images at the current moment as the video images at the current moment captured by the first capture device and the second capture device:
示例性地,S1中包含a个第一目标对象的初始位置坐标(也称作第一初始位置坐标),S2中包含b个第二目标对象的初始位置坐标(也称作第二初始位置坐标),可以确定每个第一初始位置坐标和各个第二初始位置坐标之间的欧式距离,得到距离矩阵:Exemplarily, S1 includes initial position coordinates (also referred to as first initial position coordinates) of a first target objects, and S2 includes b initial position coordinates (also referred to as second initial position coordinates) of second target objects. ), the Euclidean distance between each first initial position coordinate and each second initial position coordinate can be determined to obtain a distance matrix:
Figure PCTCN2022074956-appb-000002
Figure PCTCN2022074956-appb-000002
其中,d 11表示S1中第1个第一初始位置坐标和S2中第1个第二初始位置坐标之间的第二距离;d 1b表示S1中第1个第一初始位置坐标和S2中第b个第二初始位置坐标之间的第二距离;d ij表示S1中第i个第一初始位置坐标和S2中第j个第二初始位置坐标之间的第二距离;d a1表示S1中第a个第一初始位置坐标和S2中第1个第二初始位置坐标之间的第二距离;d ab表示S1中第a个第一初始位置坐标和S2中第b个第二初始位置坐标之间的第二距离。 Wherein, d 11 represents the second distance between the first first initial position coordinate in S1 and the first second initial position coordinate in S2; d 1b represents the first first initial position coordinate in S1 and the first second initial position coordinate in S2 The second distance between the b second initial position coordinates; d ij represents the second distance between the i-th first initial position coordinate in S1 and the j-th second initial position coordinate in S2; d a1 represents the second distance in S1 The second distance between the a-th first initial position coordinate and the first second initial position coordinate in S2; d ab represents the a-th first initial position coordinate in S1 and the b-th second initial position coordinate in S2 the second distance between.
示例性地,具体在操作时,可以按照以下方式确定S1和S2中与同一目标对象关联的多个初始位置坐标,包括S30121~S3012:Exemplarily, during operation, multiple initial position coordinates associated with the same target object in S1 and S2 can be determined in the following manner, including S30121-S3012:
S30121,在当前距离矩阵中的元素中查找当前最小第二距离;S30121, find the current minimum second distance in the elements in the current distance matrix;
示例性地,在首次查找最小第二距离的情况下,当前距离矩阵中的元素包含S1中每个第一目标对象的初始位置坐标和S2中各个第二目标对象的初始位置坐标之间的欧式距离。Exemplarily, in the case of finding the minimum second distance for the first time, the elements in the current distance matrix include the Euclidean equation between the initial position coordinates of each first target object in S1 and the initial position coordinates of each second target object in S2. distance.
S30122,判断当前最小第二距离是否小于第二预设融合距离阈值。S30122: Determine whether the current minimum second distance is less than a second preset fusion distance threshold.
示例性地,第二预设融合距离可以根据经验设定。比如预先通过不同采集设备针对同一目标对象进行拍摄,然后根据不同采集设备采集的视频画面分别确定出该同一目标对象在目标场所中的多个初始位置坐标,根据多个初始位置坐标之间的距离来确定该第二预设融合距离阈值。Exemplarily, the second preset fusion distance may be set empirically. For example, the same target object is photographed by different collection devices in advance, and then multiple initial position coordinates of the same target object in the target site are determined according to the video images collected by different collection devices, according to the distance between the multiple initial position coordinates. to determine the second preset fusion distance threshold.
S30123,在确定该当前最小第二距离小于第二预设融合距离阈值的情况下,确定该当前最小第二距离关联的两个初始位置坐标为同一目标对象关联的初始位置坐标。S30123: In the case where it is determined that the current minimum second distance is less than the second preset fusion distance threshold, determine that the two initial position coordinates associated with the current minimum second distance are initial position coordinates associated with the same target object.
示例性地,比如确定出d a1为当前最小距离,且d a1小于第二预设融合距离阈值,可以将S1中第a个第一目标对象的初始位置坐标和S2中第1个第二目标对象的初始位置坐标作为与同一目标对象关联的初始位置坐标。 Exemplarily, for example, it is determined that d a1 is the current minimum distance, and d a1 is less than the second preset fusion distance threshold, the initial position coordinates of the a-th first target object in S1 and the first second target in S2 can be used. The initial position coordinates of the object serve as the initial position coordinates associated with the same target object.
S30124,将当前距离矩阵中的当前最小第二距离,以及与当前最小第二距离关联的两个初始位置坐标中任一初始位置坐标之间的所有其它第二距离设置为第二预设融合距离阈值后,返回执行S30121,直至当前距离矩阵中的当前最小第二距离大于等于第二预设融合距离阈值的情况下,得到S1和S2中所有与同一目标对象关联的初始位置坐标。S30124: Set the current minimum second distance in the current distance matrix and all other second distances between any one of the two initial position coordinates associated with the current minimum second distance as the second preset fusion distance After the threshold is set, return to execute S30121 until the current minimum second distance in the current distance matrix is greater than or equal to the second preset fusion distance threshold, obtain all initial position coordinates associated with the same target object in S1 and S2.
示例性地,假设当前距离矩阵由S1和S2中初始位置坐标计算得到,具体一个为3×3矩阵:Exemplarily, it is assumed that the current distance matrix is calculated from the initial position coordinates in S1 and S2, and the specific one is a 3×3 matrix:
Figure PCTCN2022074956-appb-000003
Figure PCTCN2022074956-appb-000003
第二预设融合距离阈值为d th;假设d 11为当前矩阵中最小距离且小于d th,那么S1中的第1个第一初始位置坐标以及S2中的第1个第二初始位置坐标,为同一目标对象的关联初始位置坐标。则在当前距离矩阵中,与这两个初始位置坐标中任一初始位置坐标计算出的所有其它距离为d 12、d 13、d 21、d 31。所以,根据S30124,在当前矩阵中,需要把d 11、d 12、d 13、d 21、d 31均设置为d th;设置后的矩阵为: The second preset fusion distance threshold is d th ; assuming that d 11 is the minimum distance in the current matrix and is less than d th , then the first first initial position coordinate in S1 and the first second initial position coordinate in S2, is the associated initial position coordinate of the same target object. Then in the current distance matrix, all other distances calculated from any of the two initial position coordinates are d 12 , d 13 , d 21 , and d 31 . Therefore, according to S30124, in the current matrix, it is necessary to set d 11 , d 12 , d 13 , d 21 , and d 31 to d th ; the set matrix is:
Figure PCTCN2022074956-appb-000004
Figure PCTCN2022074956-appb-000004
之后返回执行S30121。Then, it returns to execute S30121.
示例性地,在将当前距离矩阵中的当前最小第二距离,以及与当前最小第二距离关联的两个初始位置坐标中任一初始位置坐标之间的所有其它距离设置为第二预设融合距离阈值后,在继续查找当前最小第二距离的过程中,可以排除被设置为第二预设融合距离阈值的元素,从而提高搜索效率。Exemplarily, the current minimum second distance in the current distance matrix and all other distances between any one of the two initial position coordinates associated with the current minimum second distance are set as the second preset fusion. After the distance threshold is determined, in the process of continuing to search for the current minimum second distance, the element set as the second preset fusion distance threshold can be excluded, thereby improving the search efficiency.
示例性地,在一种实施方式中,在得到S1和S2中与同一目标对象关联的多个初始位置坐标后,可以继续基于其它任意两张当前时刻的视频画面确定是否存在与同一目标对象关联的初始位置坐标。直至判断完A个采集设备在在同一时刻采集的当前时刻的视频画面后,可以得到A个采集设备在同一时刻采集的当前时刻视频画面中涉及的各个目标对象的不同初始位置坐标。然后在不同初始位置坐标中对与同一目标对象关联的初始位置坐标进行融合,得到A个采集设备在同一时刻拍摄的A张当前时刻的视频画面中的各目标对象在目标场所中的第一位置坐标。Exemplarily, in an implementation manner, after obtaining multiple initial position coordinates associated with the same target object in S1 and S2, it can continue to determine whether there is an association with the same target object based on any two other video images at the current moment. The initial position coordinates of . After judging the video images of the current moment collected by the A collection devices at the same time, the different initial position coordinates of each target object involved in the current video images collected by the A collection devices at the same time can be obtained. Then, the initial position coordinates associated with the same target object are fused in different initial position coordinates to obtain the first position of each target object in the target place in the A pieces of video images of the current moment shot by the A collection devices at the same time coordinate.
示例性地,在另一种实施方式中,在得到S1和S2中与同一目标对象关联的多个初始位置坐标后,可以对多个初始位置坐标进行坐标融合,得到该同一目标对象更新后的初始位置坐标。针对S1和S2中待融合的初始位置坐标,可以与更新后的初始位置坐标构成S2’。进一步通过S2’和S3中的初始位置坐标构成新的当前距离矩阵,重复执行S30121至S30124的步骤,得到S2’和S3中与同一目标对象关联的多个初始位置坐标,按照相同的方式得到S3’。进一步通过S3’和S4中的初始位置坐标构成新的当前距离矩阵,重复执行S30121至S30124的步骤,直至完成与初始坐标集合中的最后一个元素(即SA)中的初始位置坐标的融合后,得到A个采集设备在同一时刻拍摄的A张视频画面中的各目标对象在目标场所中的第一位置坐标。Exemplarily, in another embodiment, after obtaining multiple initial position coordinates associated with the same target object in S1 and S2, coordinate fusion can be performed on the plurality of initial position coordinates to obtain the updated version of the same target object. Initial position coordinates. For the initial position coordinates to be fused in S1 and S2, S2' can be formed with the updated initial position coordinates. Further form a new current distance matrix by the initial position coordinates in S2' and S3, repeat the steps of S30121 to S30124, obtain a plurality of initial position coordinates associated with the same target object in S2' and S3, obtain S3 in the same way '. A new current distance matrix is further formed by the initial position coordinates in S3 ' and S4, and the steps of S30121 to S30124 are repeatedly executed, until after completing the fusion with the initial position coordinates in the last element (that is, SA) in the initial coordinate set, Obtain the first position coordinates of each target object in the target place in the A video images shot by the A collection devices at the same time.
特别地,在直至完成与初始坐标集合中的最后一个元素(即SA)中的初始位置坐标的融合后,若检测到存在任一初始位置坐标从开始到结束均为参与融合,考虑到目标场所中的每个目标对象至少被两个采集设备同时采集到,因此可以将该任一初始位置坐标作为误差初始位置坐标进行过滤。In particular, until the fusion with the initial position coordinates in the last element in the initial coordinate set (ie SA) is completed, if any initial position coordinates are detected to be involved in the fusion from the beginning to the end, considering the target location Each target object in is simultaneously collected by at least two collection devices, so any initial position coordinate can be used as the error initial position coordinate for filtering.
本公开实施例中,根据当前时刻的任意两张视频画面中感兴趣区域对应的目标对象的初始位置坐标,和第二预设融合距离阈值,可以快速确定出与同一目标对象关联的初始位置坐标,从而为后续确定各目标对象的第一位置坐标提供依据。In the embodiment of the present disclosure, according to the initial position coordinates of the target object corresponding to the region of interest in any two video images at the current moment, and the second preset fusion distance threshold, the initial position coordinates associated with the same target object can be quickly determined , so as to provide a basis for the subsequent determination of the first position coordinates of each target object.
针对上述S302,在对同一目标对象关联的多个初始位置坐标进行依次融合,得到同一目标对象在当前时刻下在目标场所中的第一位置坐标时,可以包括以下S3021~S3022:For the above S302, when the multiple initial position coordinates associated with the same target object are sequentially fused to obtain the first position coordinates of the same target object in the target place at the current moment, the following steps S3021 to S3022 may be included:
S3021,从同一目标对象关联的多个初始位置坐标中选取任一初始位置坐标,将选取的任一初始位置坐标作为第一中间融合位置坐标;S3021, select any initial position coordinate from a plurality of initial position coordinates associated with the same target object, and use any selected initial position coordinate as the first intermediate fusion position coordinate;
S3022,将第一中间融合位置坐标与其它任一待融合的初始位置坐标进行融合,生成第二中间融合位置坐标;将第二中间融合位置坐标作为更新后的第一中间融合位置坐标,并返回生成第二中间融合位置坐标的步骤,直到不存在待融合的初始位置坐标。S3022, fuse the first intermediate fusion position coordinates with any other initial position coordinates to be fused to generate second intermediate fusion position coordinates; use the second intermediate fusion position coordinates as the updated first intermediate fusion position coordinates, and return The step of generating the second intermediate fused position coordinates until there are no initial position coordinates to be fused.
其中,待融合的初始位置坐标是指未参与融合的初始位置坐标。The initial position coordinates to be fused refer to the initial position coordinates that do not participate in the fusion.
示例性地,在将第一中间融合位置坐标与其它任一待融合的初始位置坐标进行融合,生成第二中间融合位置坐标时,包括:Exemplarily, when the first intermediate fusion position coordinates are fused with any other initial position coordinates to be fused to generate the second intermediate fusion position coordinates, the steps include:
确定第一中间融合位置坐标与其它任一待融合的初始位置坐标的中点坐标,将该中点坐标作为生成的第二中间融合位置坐标。Determine the midpoint coordinate of the first intermediate fusion position coordinate and any other initial position coordinate to be fused, and use the midpoint coordinate as the generated second intermediate fusion position coordinate.
示例性地,结合上述实施例若确定出与目标对象A关联的多个初始位置坐标包含N个。可以将任一初始位置坐标作为第一中间融合位置坐标,确定该第一中间融合位置坐标与其它任一待融合的初始位置坐标的中点坐标。然后将该中点坐标作为更新后的第一中间融合位置坐标,继续与其它 任一待融合的初始位置坐标进行融合。直到N个初始位置坐标中不存在待融合的初始位置坐标后,得到目标对象A的第一位置坐标。Exemplarily, in combination with the above embodiments, if it is determined that the plurality of initial position coordinates associated with the target object A include N pieces. Any initial position coordinate may be used as the first intermediate fusion position coordinate, and the midpoint coordinate of the first intermediate fusion position coordinate and any other initial position coordinate to be fused is determined. Then use the midpoint coordinate as the updated first intermediate fusion position coordinate, and continue to fuse with any other initial position coordinate to be fused. Until there is no initial position coordinate to be fused among the N initial position coordinates, the first position coordinate of the target object A is obtained.
本公开实施例中,针对与同一目标对象关联的多个初始位置坐标,可以按照依次取中点的方式融合,从而得到准确度较高的第一位置坐标。In this embodiment of the present disclosure, multiple initial position coordinates associated with the same target object may be fused in a manner of taking midpoints in sequence, so as to obtain first position coordinates with higher accuracy.
在一种可能的实施方式中,在基于获取到的目标对象的第一位置坐标和目标对象在上一时刻的第二位置坐标,确定目标对象在当前时刻的第二位置坐标时,如图5所示,可以包括以下S401~S403:In a possible implementation manner, when determining the second position coordinates of the target object at the current moment based on the obtained first position coordinates of the target object and the second position coordinates of the target object at the previous moment, as shown in FIG. 5 As shown, the following S401 to S403 may be included:
S401,基于目标对象在上一时刻的第二位置坐标,确定目标对象在当前时刻的预测位置坐标;S401, based on the second position coordinates of the target object at the previous moment, determine the predicted position coordinates of the target object at the current moment;
S402,基于目标对象在当前时刻的预测位置坐标和第一位置坐标,确定目标对象在当前时刻的观测位置坐标;S402, based on the predicted position coordinates and the first position coordinates of the target object at the current moment, determine the observed position coordinates of the target object at the current moment;
S403,基于目标对象在当前时刻的预测位置坐标和观测位置坐标,确定目标对象在当前时刻的第二位置坐标。S403 , based on the predicted position coordinates and the observed position coordinates of the target object at the current moment, determine the second position coordinates of the target object at the current moment.
示例性地,这里可以引入卡尔曼滤波器通过卡尔曼滤波的方式来确定目标对象在当前时刻的第二位置坐标。在基于卡尔曼滤波方式确定目标对象在当前时刻准确度较高的第二位置坐标的过程中,需要确定观测位置坐标和预测位置坐标。其中预测位置坐标是指可以基于上一时刻的第二位置坐标预测得到的目标对象在当前时刻的位置坐标。观测位置坐标可以根据采集设备采集的当前时刻的视频画面来确定,比如上述确定的目标对象在当前时刻的第一位置坐标。但是考虑到第一位置坐标可能存在误差,本公开实施例提出结合预测位置坐标和基于采集设备采集的当前时刻的视频画面确定的第一位置坐标来共同确定观测位置坐标。最后可以结合观测位置坐标和预测位置坐标,得到目标对象在当前时刻的第二位置坐标。Exemplarily, a Kalman filter may be introduced here to determine the second position coordinates of the target object at the current moment by means of Kalman filtering. In the process of determining the second position coordinates of the target object with higher accuracy at the current moment based on the Kalman filtering method, it is necessary to determine the observed position coordinates and the predicted position coordinates. The predicted position coordinates refer to the position coordinates of the target object at the current moment that can be predicted based on the second position coordinates of the previous moment. The coordinates of the observation position may be determined according to the video image at the current moment collected by the collection device, such as the first position coordinates of the target object at the current moment determined above. However, considering that there may be errors in the first position coordinates, the embodiment of the present disclosure proposes to jointly determine the observation position coordinates by combining the predicted position coordinates and the first position coordinates determined based on the video images collected by the collection device at the current moment. Finally, the observed position coordinates and the predicted position coordinates can be combined to obtain the second position coordinates of the target object at the current moment.
具体地,在基于目标对象在上一时刻的第二位置坐标,确定目标对象在当前时刻的预测位置坐标时,可以根据卡尔曼滤波公式中的以下公式(1)来确定:Specifically, when determining the predicted position coordinates of the target object at the current moment based on the second position coordinates of the target object at the previous moment, it can be determined according to the following formula (1) in the Kalman filter formula:
Trk(t|t-1)=ATrk(t-1|t-1)+Bu(t-1)+W(t-1)            (1);Trk(t|t-1)=ATrk(t-1|t-1)+Bu(t-1)+W(t-1)    (1);
其中,Trk(t|t-1)表示根据目标对象在上一时刻的第二位置坐标,确定的目标对象在当前时刻的预测位置坐标;Trk(t-1|t-1)表示目标对象在上一时刻的第二位置坐标;W(t-1)表示预测目标对象在当前时刻的预测位置坐标过程中的白噪声,表示预测位置坐标的误差量;A和B表示卡尔曼滤波器的参数矩阵,其中,A表示卡尔曼滤波器中的状态转移矩阵,u(t-1)表示在上一时刻对卡尔曼滤波的控制量,可以为0。Among them, Trk(t|t-1) indicates the predicted position coordinates of the target object at the current moment determined according to the second position coordinates of the target object at the previous moment; Trk(t-1|t-1) indicates that the target object is in The second position coordinate of the previous moment; W(t-1) represents the white noise in the process of predicting the target object's predicted position coordinates at the current moment, and represents the error amount of the predicted position coordinates; A and B represent the parameters of the Kalman filter matrix, where A represents the state transition matrix in the Kalman filter, and u(t-1) represents the control amount of the Kalman filter at the previous moment, which can be 0.
另外,进一步地,在得到目标对象在当前时刻的观测位置坐标后,可以根据以下公式(2)确定目标对象在当前时刻的观测位置坐标的协方差矩阵:In addition, further, after obtaining the observation position coordinates of the target object at the current moment, the covariance matrix of the observation position coordinates of the target object at the current moment can be determined according to the following formula (2):
P(t|t-1)=AP(t-1|t-1)A T+Q             (2); P(t|t-1)=AP(t-1|t-1)A T +Q (2);
其中,P(t|t-1)表示目标对象在当前时刻的观测位置坐标的协方差矩阵,可以表示目标对象在当前时刻的观测位置坐标的不确定度;P(t-1|t-1)表示目标对象在上一时刻的第二位置坐标的协方差矩阵,可以表示目标对象在上一时刻的第二位置坐标的不确定度;Q表示卡尔曼滤波器引入的系统过程协方差矩阵,用来表示状态转移矩阵相比实际过程的误差。Among them, P(t|t-1) represents the covariance matrix of the coordinates of the observed position of the target object at the current moment, which can represent the uncertainty of the coordinates of the observed position of the target object at the current moment; P(t-1|t-1 ) represents the covariance matrix of the second position coordinate of the target object at the last moment, which can represent the uncertainty of the second position coordinate of the target object at the last moment; Q represents the system process covariance matrix introduced by the Kalman filter, It is used to represent the error of the state transition matrix compared to the actual process.
示例性地,在得到目标对象在当前时刻的预测位置坐标后,可以结合目标对象的预测位置坐标和第一位置坐标,确定目标对象在当前时刻的观测位置坐标,具体将在后续进行阐述。Exemplarily, after the predicted position coordinates of the target object at the current moment are obtained, the predicted position coordinates of the target object and the first position coordinates can be combined to determine the observed position coordinates of the target object at the current moment, which will be described in detail later.
在得到目标对象在当前时刻的预测位置坐标和观测位置坐标后,可以根据卡尔曼滤波公式中的以下公式(3)来确定目标对象在当前时刻的第二位置坐标:After obtaining the predicted position coordinates and the observed position coordinates of the target object at the current moment, the second position coordinates of the target object at the current moment can be determined according to the following formula (3) in the Kalman filter formula:
Trk(t|t)=Trk(t|t-1)+K g(t)(z(t)-HTrk(t|t-1))         (3); Trk(t|t)=Trk(t|t-1)+K g (t)(z(t)-HTrk(t|t-1)) (3);
其中,Trk(t|t)表示目标对象在当前时刻的第二位置坐标;Z(t)表示目标对象在当前时刻的观测位置坐标;K g(t)表示卡尔曼滤波器中的滤波增益矩阵,该滤波增益矩阵可以通过以下公式(4)确定: Among them, Trk(t|t) represents the second position coordinate of the target object at the current moment; Z(t) represents the observation position coordinate of the target object at the current moment; K g (t) represents the filter gain matrix in the Kalman filter , the filter gain matrix can be determined by the following formula (4):
Figure PCTCN2022074956-appb-000005
Figure PCTCN2022074956-appb-000005
其中,H表示卡尔曼滤波器中的参数矩阵,表示观测矩阵;R表示卡尔曼滤波器中已知的测量噪声协方差。Among them, H represents the parameter matrix in the Kalman filter, which represents the observation matrix; R represents the known measurement noise covariance in the Kalman filter.
进一步地,可以基于滤波增益矩阵后续还需要确定目标对象在下一时刻的第二位置坐标,因此需要确定目标对象在当前时刻的第二位置坐标的协方差矩阵P(t|t),具体可以通过以下公式(5)进行确定:Further, based on the filtering gain matrix, the second position coordinate of the target object at the next moment needs to be determined, so it is necessary to determine the covariance matrix P(t|t) of the second position coordinate of the target object at the current moment. The following formula (5) is determined:
P(t|t)=(I-K g(t)H)P(t|t-1)            (5); P(t|t)=(IK g (t)H)P(t|t-1) (5);
在得到目标对象在当前时刻的第二位置坐标的协方差矩阵后,可以基于该协方差矩阵确定目标对象在下一时刻的观测位置坐标的协方差矩阵,为确定目标对象在下一时刻的第二位置坐标做准备。After obtaining the covariance matrix of the second position coordinates of the target object at the current moment, the covariance matrix of the observed position coordinates of the target object at the next moment can be determined based on the covariance matrix, so as to determine the second position of the target object at the next moment. Prepare the coordinates.
示例性地,如果当前时刻为采集的初始时刻,则目标对象没有上一时刻的第二位置坐标。在这种情况下,可以将目标对象在当前时刻的第一位置坐标直接确定为目标对象在当前时刻的第二位置坐标。Exemplarily, if the current moment is the initial moment of acquisition, the target object does not have the second position coordinates of the previous moment. In this case, the first position coordinate of the target object at the current moment may be directly determined as the second position coordinate of the target object at the current moment.
本公开实施例中,可以根据目标对象在上一时刻的第二位置坐标,确定目标对象在当前时刻的预测位置坐标,进一步结合目标对象在当前时刻的第一位置坐标,可以得到目标对象在当前时刻准确度较高的第二位置坐标。In the embodiment of the present disclosure, the predicted position coordinates of the target object at the current moment can be determined according to the second position coordinates of the target object at the previous moment, and further combined with the first position coordinates of the target object at the current moment, the current position of the target object can be obtained. The second position coordinate with higher time accuracy.
具体地,目标对象包含多个,针对上述S402,在基于目标对象在当前时刻的预测位置坐标和第一位置坐标,确定目标对象在当前时刻的观测位置坐标时,可以包括以下S4021~S4022:Specifically, there are multiple target objects. For the above S402, when determining the observed position coordinates of the target object at the current moment based on the predicted position coordinates and the first position coordinates of the target object at the current moment, the following steps S4021 to S4022 may be included:
S4021,基于多个目标对象在当前时刻的预测位置坐标和第一位置坐标,确定与同一目标对象关联的预测位置坐标和第一位置坐标。S4021 , based on the predicted position coordinates and the first position coordinates of the multiple target objects at the current moment, determine the predicted position coordinates and the first position coordinates associated with the same target object.
S4022,确定与同一目标对象关联的预测位置坐标和第一位置坐标的第一中点坐标,将该第一中点坐标作为该目标对象在当前时刻的观测位置坐标。S4022: Determine the predicted position coordinates associated with the same target object and the first midpoint coordinates of the first position coordinates, and use the first midpoint coordinates as the observed position coordinates of the target object at the current moment.
示例性地,根据上一时刻目标场所中包含的N个目标对象的第二位置坐标,可以得到N个目标对象在当前时刻的预测位置坐标。另外基于多个采集设备采集的当前时刻的视频画面,可以得到目标场所中的M个目标对象在当前时刻的第一位置坐标。在N个预测位置坐标和M个第一位置坐标中,可以基于距离的贪心算法确定与同一目标对象关联的预测位置坐标和第一位置坐标。然后进一步可以将与同一目标对象关联的预测位置坐标和第一位置坐标的中点坐标,作为该同一目标对象在当前时刻的观测位置坐标。Exemplarily, according to the second position coordinates of the N target objects included in the target place at the previous moment, the predicted position coordinates of the N target objects at the current moment can be obtained. In addition, based on the video images at the current moment collected by multiple collection devices, the first position coordinates of the M target objects in the target place at the current moment can be obtained. Among the N predicted position coordinates and the M first position coordinates, the predicted position coordinates and the first position coordinates associated with the same target object may be determined by a distance-based greedy algorithm. Then, the midpoint coordinates of the predicted position coordinates associated with the same target object and the first position coordinates can be further used as the observed position coordinates of the same target object at the current moment.
示例性地,N可以大于或等于M。在N大于M的情况下,可能在采集设备采集的当前时刻的视频画面中,存在漏检的目标对象。比如由于目标场所中的障碍物遮挡导致无法采集到某个目标对象的视频画面,这样在基于当前时刻的视频画面确定目标场所中的目标对象的第一位置坐标的情况下,会存在漏检的情况。此时可以通过该目标对象的预测位置坐标来确定该目标对象在当前时刻的观测位置坐标。Illustratively, N may be greater than or equal to M. In the case where N is greater than M, there may be a target object that is missed in the video image at the current moment collected by the collection device. For example, the video image of a certain target object cannot be captured due to the occlusion of the obstacle in the target area. In this case, if the first position coordinates of the target object in the target area are determined based on the video image at the current moment, there will be missed detections. Happening. In this case, the observed position coordinates of the target object at the current moment can be determined by the predicted position coordinates of the target object.
本公开实施例中,基于目标对象在当前时刻的预测位置坐标和第一位置坐标,确定目标对象在当前时刻的观测位置坐标,可以包括:基于多个目标对象在所述当前时刻的多个第一位置坐标和该目标对象在所述当前时刻的所述预测位置坐标,确定该目标对象的第一位置坐标;确定该目标对象的所述预测位置坐标和所述第一位置坐标的第一中点坐标,将该第一中点坐标作为该目标对象在所述当前时刻的观测位置坐标。In this embodiment of the present disclosure, determining the observed position coordinates of the target object at the current moment based on the predicted position coordinates and the first position coordinates of the target object at the current moment may include: based on multiple first position coordinates of the target objects at the current moment. a position coordinate and the predicted position coordinate of the target object at the current moment, determine the first position coordinate of the target object; determine the predicted position coordinate of the target object and the first middle of the first position coordinate point coordinates, and the first midpoint coordinates are taken as the coordinates of the observation position of the target object at the current moment.
具体地,针对上述S4021,在基于多个目标对象在当前时刻的预测位置坐标和第一位置坐标,确定与同一目标对象关联的预测位置坐标和第一位置坐标时,可以包括以下S40211~S40212:Specifically, for the above S4021, when determining the predicted position coordinates and the first position coordinates associated with the same target object based on the predicted position coordinates and the first position coordinates of the multiple target objects at the current moment, the following steps S40211 to S40212 may be included:
S40211,针对每个预测位置坐标,确定该预测位置坐标和各个第一位置坐标之间的第一距离;S40211, for each predicted position coordinate, determine the first distance between the predicted position coordinate and each first position coordinate;
S40212,将该预测位置坐标和与该预测位置坐标具有最小第一距离的第一位置坐标,作为与同一目标对象关联的预测位置坐标和第一位置坐标,最小第一距离小于第一预设融合距离阈值。S40212, use the predicted position coordinates and the first position coordinates with the minimum first distance from the predicted position coordinates as the predicted position coordinates and the first position coordinates associated with the same target object, and the minimum first distance is smaller than the first preset fusion distance threshold.
示例性地,比如当前时刻包含N个预测位置坐标以及M个观测位置坐标,根据N个预测位置坐标以及M个观测位置坐标,确定每个预测位置坐标和各个观测位置坐标之间的欧式距离,得到距离矩阵:Exemplarily, for example, the current moment includes N predicted position coordinates and M observed position coordinates, and the Euclidean distance between each predicted position coordinate and each observed position coordinate is determined according to the N predicted position coordinates and M observed position coordinates, Get the distance matrix:
Figure PCTCN2022074956-appb-000006
Figure PCTCN2022074956-appb-000006
其中,l 11表示N个预测位置坐标中第1个预测位置坐标和M个观测位置坐标中第1个观测位置坐标之间的第一距离;l 1M表示N个预测位置坐标中第1个预测位置坐标和M个观测位置坐标 中第M个观测位置坐标之间的第一距离;l nm表示N个预测位置坐标中第n个预测位置坐标和M个观测位置坐标中第m个观测位置坐标之间的第一距离;l N1表示N个预测位置坐标中第N个预测位置坐标和M个观测位置坐标中第1个观测位置坐标之间的第一距离;l NM表示N个预测位置坐标中第N个预测位置坐标和M个观测位置坐标中第M个观测位置坐标之间的第一距离。 Wherein, l 11 represents the first distance between the first predicted position coordinate in the N predicted position coordinates and the first observed position coordinate in the M observed position coordinates; l 1M represents the first predicted position in the N predicted position coordinates The first distance between the position coordinates and the Mth observation position coordinate in the M observation position coordinates; l nm represents the nth prediction position coordinate in the N prediction position coordinates and the mth observation position coordinate in the M observation position coordinates The first distance between; l N1 represents the first distance between the Nth predicted position coordinate in the N predicted position coordinates and the first observed position coordinate in the M observed position coordinates; l NM represents the N predicted position coordinates The first distance between the Nth predicted position coordinate in and the Mth observation position coordinate in the M observation position coordinates.
进一步地,可以按照上述确定与同一目标对象关联的多个初始位置坐标的方式,确定与同一目标对象关联的预测位置坐标和第一位置坐标,具体过程在此不再赘述。Further, the predicted position coordinates and the first position coordinates associated with the same target object may be determined according to the above-mentioned method of determining multiple initial position coordinates associated with the same target object, and the specific process will not be repeated here.
本公开实施例中,基于多个目标对象在所述当前时刻的多个第一位置坐标和该目标对象在所述当前时刻的预测位置坐标,确定该目标对象的第一位置坐标,可以包括:确定该目标对象的预测位置坐标和各个第一位置坐标之间的第一距离;将所述多个第一位置坐标中与该目标对象的所述预测位置坐标构成最小第一距离的第一位置坐标,作为与该目标对象的所述第一位置坐标,所述最小第一距离小于第一预设融合距离阈值。In this embodiment of the present disclosure, determining the first position coordinates of the target object based on the plurality of first position coordinates of the target objects at the current moment and the predicted position coordinates of the target objects at the current moment may include: determining a first distance between the predicted position coordinates of the target object and each of the first position coordinates; forming a first position with the smallest first distance from the predicted position coordinates of the target object among the plurality of first position coordinates coordinates, as the first position coordinates with the target object, the minimum first distance is smaller than the first preset fusion distance threshold.
本公开实施例中,结合根据目标对象在历史时刻的位置坐标预测的目标对象在当前时刻的预测位置坐标,以及根据采集设备采集的当前时刻的视频画面确定的目标对象的第一位置坐标,一方面可以快速得到同一目标对象在不同时刻的位置坐标,另一方面可以得到准确度较高的观测位置坐标。In the embodiment of the present disclosure, combining the predicted position coordinates of the target object at the current moment predicted according to the position coordinates of the target object at the historical moment, and the first position coordinates of the target object determined according to the video images collected by the acquisition device at the current moment, a On the one hand, the position coordinates of the same target object at different times can be quickly obtained, and on the other hand, the observed position coordinates with high accuracy can be obtained.
在一种实施方式中,如图6所示,本公开实施例提供的目标追踪方法还包括以下S501~S502:In an implementation manner, as shown in FIG. 6 , the target tracking method provided by the embodiment of the present disclosure further includes the following S501 to S502:
S501,确定当前时刻的视频画面中是否存在漏检的目标对象,其中,漏检的目标对象在当前时刻具有预测位置坐标,且在当前时刻的第一位置坐标为空;S501, determine whether there is an undetected target object in the video picture at the current moment, wherein the missed target object has a predicted position coordinate at the current moment, and the first position coordinate at the current moment is empty;
S502,在确定存在漏检的目标对象的情况下,将漏检的目标对象在当前时刻的预测位置坐标作为漏检的目标对象在当前时刻的观测位置坐标。S502 , when it is determined that there is an undetected target object, the predicted position coordinates of the missed target object at the current moment are taken as the observed position coordinates of the missed target object at the current moment.
示例性地,考虑到在目标场所中的目标对象较多的情况下,不同目标对象之间容易发生拥堵的情况,这样可能存在某个时刻不同的目标对象之间发生遮挡导致采集设备采集的视频画面中存在漏检的情况。比如多个采集设备中采集设备1和采集设备2采集的当前时刻的视频画面中的目标对象A均被遮挡,该情况可以将基于采集设备1和采集设备2采集的当前时刻的视频画面确定出的目标对象A的第一位置坐标记为空,此时将目标对象A作为漏检的目标对象。Exemplarily, considering that when there are many target objects in the target site, congestion is likely to occur between different target objects, so there may be occlusion between different target objects at a certain moment, resulting in the video captured by the capture device. There is a missed detection in the screen. For example, the target object A in the video images of the current moment collected by the capture device 1 and the capture device 2 among the multiple capture devices are all blocked. The first position coordinate mark of the target object A is empty, and at this time, the target object A is regarded as the missed target object.
示例性地,在使用卡尔曼滤波方式确定目标对象A在当前时刻的预测位置坐标的情况下,会使用目标对象A在历史时刻的第二位置坐标。由于有目标对象A在进入目标场所的过程中,会被采集设备采集到,因此可以确定目标对象A在历史时刻的第二位置坐标,这样按照卡尔曼滤波的方式可以确定目标对象A在当前时刻的预测位置坐标。如果目标对象A在当前时刻的第一位置坐标为空,可以直接将目标对象A在当前时刻的预测位置坐标作为其在当前时刻的观测位置坐标。Exemplarily, in the case of determining the predicted position coordinates of the target object A at the current moment using the Kalman filtering method, the second position coordinates of the target object A at the historical moment will be used. Since the target object A will be collected by the acquisition device in the process of entering the target place, the second position coordinates of the target object A at the historical moment can be determined, so that the target object A can be determined at the current moment according to the method of Kalman filtering. The predicted location coordinates of . If the first position coordinates of the target object A at the current moment are empty, the predicted position coordinates of the target object A at the current moment can be directly used as the observed position coordinates at the current moment.
本公开实施例中,在采集设备采集的当前时刻的视频画面中存在被遮挡的目标对象的情况下,可以基于被遮挡的目标对象在历史时刻中的第二位置坐标确定被遮挡的目标对象在当前时刻的观测位置坐标,以便确定目标对象在当前时刻准确度较高的第二位置坐标。In the embodiment of the present disclosure, when there is an occluded target object in the video image at the current moment collected by the collection device, the occluded target object may be determined based on the second position coordinates of the occluded target object in the historical moment. The coordinates of the observed position at the current moment, so as to determine the second position coordinates of the target object with higher accuracy at the current moment.
在一种实施方式中,目标对象包含多个,如图7所示,本公开实施例提供的目标追踪方法还包括以下S601~S602:In one embodiment, the target object includes multiple objects. As shown in FIG. 7 , the target tracking method provided by the embodiment of the present disclosure further includes the following S601 to S602:
S601,在确定目标对象在当前时刻的第二位置坐标后,在第二位置坐标指示的地图位置中标记与目标对象关联的身份标识符;S601, after determining the second position coordinates of the target object at the current moment, mark the identity identifier associated with the target object in the map position indicated by the second position coordinates;
S602,基于标记同一身份标识符的目标对象在多个时刻的第二位置坐标,生成每个目标对象的轨迹数据。S602, based on the second position coordinates of the target objects marked with the same identity identifier at multiple times, generate trajectory data of each target object.
示例性地,以目标场所为工厂,目标对象为进入工厂中的员工为例,可以在工厂进口处设置用于采集员工图像的采集设备。并基于采集的员工图像进行特征提取,比如提取员工图像中的人脸特征和/或人体特征。基于提取的特征信息和预先存储的员工身份库中每个员工的特征信息,确定进入工厂中的每个员工的身份。在对目标对象进行追踪过程中,确定出目标对象在当前时刻的第二位置坐标后,可以在第二位置坐标指示的地图位置中标记与目标对象关联的身份标识符。然后连接具有同一身份标识符的多个时刻的第二位置坐标,可以得到不同目标对象在地图中的移动轨迹。Exemplarily, taking the target place as a factory and the target objects as employees entering the factory as an example, a capturing device for capturing images of employees may be set at the entrance of the factory. Feature extraction is performed based on the collected employee images, for example, facial features and/or human body features in the employee images are extracted. The identity of each employee entering the factory is determined based on the extracted characteristic information and the pre-stored characteristic information of each employee in the employee identity database. In the process of tracking the target object, after determining the second position coordinates of the target object at the current moment, the identity identifier associated with the target object may be marked in the map position indicated by the second position coordinates. Then, by connecting the second position coordinates of multiple moments with the same identity identifier, the movement trajectories of different target objects in the map can be obtained.
示例性地,地图可以为预先构建高精度地图。预先构建的高境地地图与目标场所具有对应关系,两者在相同的坐标系中可以按照1:1进行呈现。因此这里可以基于标记同一身份标识符的目标对象在多个时刻的第二位置坐标,生成用于表示各目标对象在目标场所中的移动轨迹的轨迹数据。Illustratively, the map may be a pre-built high-resolution map. The pre-built highland map has a corresponding relationship with the target site, and the two can be presented 1:1 in the same coordinate system. Therefore, based on the second position coordinates of the target objects marked with the same identity identifier at multiple times, trajectory data representing the movement trajectory of each target object in the target place can be generated.
本公开实施例中,可以根据目标对象的身份标识符以及不同时刻的第二位置坐标,快速确定每个目标对象在目标场所中的移动轨迹。In the embodiment of the present disclosure, the movement trajectory of each target object in the target place can be quickly determined according to the identity identifier of the target object and the second position coordinates at different times.
示例性地,在目标场所中存在部分目标对象之间的距离较近的情况下,对这部分目标对象进 行聚类可以构成目标群。同一目标群中的目标对象的第二位置坐标标记身份标识符时可能发生错误,比如将目标群中的目标对象A的身份标识符标记给目标对象B,将目标对象B的身份标识符标记给目标对象A,即发生串号问题。在目标对象A和目标对象B属于同一目标群的情况下,发生串号时,因为发生串号的目标对象之间的距离较近,因此对轨迹数据的影响较小。但是在发生串号的目标对象远离目标群的情况下,若该目标对象的身份标识符发生错误,最终确定的该目标对象的轨迹数据也会发生错误。因此在一种实施方式中,在确定目标对象在当前时刻的第二位置坐标之后,如图8所示,本公开实施例提供的目标追踪方法还包括以下S701~S703:Exemplarily, in the case that there are some target objects in the target place that are close in distance, clustering these target objects can form a target group. Errors may occur when marking the identity identifier with the second position coordinates of the target objects in the same target group, for example, marking the identity identifier of target object A in the target group to target object B, marking the identity identifier of target object B to Target object A, that is, the serial number problem occurs. When the target object A and the target object B belong to the same target group, when a serial number occurs, the distance between the target objects in which the serial number occurs is relatively close, so the impact on the trajectory data is small. However, when the target object with serial number is far away from the target group, if the identity identifier of the target object is wrong, the final determined trajectory data of the target object will also be wrong. Therefore, in an implementation manner, after determining the second position coordinates of the target object at the current moment, as shown in FIG. 8 , the target tracking method provided by the embodiment of the present disclosure further includes the following S701 to S703:
S701,基于多个目标对象在当前时刻的第二位置坐标,检测是否存在偏离目标群的目标对象;目标群为根据多个目标对象在上一时刻的第二位置坐标进行聚类得到的。S701 , based on the second position coordinates of the multiple target objects at the current moment, detect whether there are target objects that deviate from the target group; the target group is obtained by clustering according to the second position coordinates of the multiple target objects at the previous moment.
示例性地,可以根据聚类算法(Density-Based Spatial Clustering of Applications with Noise,DBSCAN),对多个目标对象在上一时刻的第二位置坐标进行聚类,得到目标群。目标群中不同目标对象之间的第二位置坐标的距离小于预先设定的进入目标群的距离阈值。Exemplarily, according to a clustering algorithm (Density-Based Spatial Clustering of Applications with Noise, DBSCAN), the second position coordinates of the multiple target objects at the previous moment can be clustered to obtain the target group. The distance of the second position coordinates between different target objects in the target group is less than a preset distance threshold for entering the target group.
示例性地,基于多个目标对象在当前时刻的第二位置坐标和预先设定的离开目标群的距离阈值,可以判断是否存在偏离所属目标群的目标对象。Exemplarily, based on the second position coordinates of the multiple target objects at the current moment and a preset distance threshold from the target group, it may be determined whether there are target objects that deviate from the target group to which they belong.
S702,在确定存在偏离目标群的目标对象情况下,检测偏离目标群的目标对象关联的身份标识符是否准确。S702, in the case where it is determined that there is a target object deviating from the target group, detect whether the identity identifier associated with the target object deviating from the target group is accurate.
具体地,在检测偏离目标群的目标对象关联的身份标识符是否准确时,包括:Specifically, when detecting whether the identity identifier associated with the target object that deviates from the target group is accurate, it includes:
S7021,提取偏离目标群的目标对象的特征信息;S7021, extract the feature information of the target object deviating from the target group;
S7022,基于偏离目标群的目标对象的特征信息,以及预先保存的进入目标场所中的各目标对象的特征信息和身份标识符之间的映射关系,检测偏离目标群的目标对象关联的身份标识符是否准确。S7022, based on the feature information of the target objects that deviate from the target group and the pre-stored mapping relationship between the feature information of each target object entering the target place and the identity identifiers, detect the identity identifiers associated with the target objects that deviate from the target group Is it accurate.
示例性地,在确定存在偏离目标群的目标对象的情况下,获取该偏离目标群的目标对象的当前时刻的视频画面。基于该当前时刻的视频画面,提取该偏离目标群的目标对象的特征信息。基于该特征信息和预先存储的该目标对象的身份标识符和对应的特征信息,确定该偏离目标群的目标对象关联的身份标识符是否准确。比如可以确定当前提取的偏离目标群的目标对象的特征信息和预先标记的该目标对象的身份标识符关联的特征信息之间的相似度。在确定该相似度达到预设相似度阈值的情况下,确定该偏离目标群的目标对象关联的身份标识符准确。相反,在确定该相似度未达到预设相似度阈值的情况下,确定该偏离目标群的目标对象关联的身份标识符不准确。Exemplarily, when it is determined that there is a target object deviating from the target group, a video picture of the current moment of the target object deviating from the target group is acquired. Based on the video picture at the current moment, the feature information of the target object deviating from the target group is extracted. Based on the feature information and the pre-stored identity identifier of the target object and the corresponding feature information, it is determined whether the identity identifier associated with the target object deviating from the target group is accurate. For example, the similarity between the feature information of the currently extracted target object that deviates from the target group and the pre-marked feature information associated with the identity identifier of the target object can be determined. When it is determined that the similarity reaches a preset similarity threshold, it is determined that the identity identifier associated with the target object deviating from the target group is accurate. On the contrary, if it is determined that the similarity does not reach the preset similarity threshold, it is determined that the identity identifier associated with the target object deviating from the target group is inaccurate.
比如,检测到存在偏离目标群的目标对象,预先对该目标对象标记的身份标识符为001。在确定当前时刻的视频画面提取的该偏离目标对象的特征信息和身份标识符001关联的特征信息之间的相似度小于预设相似度阈值的情况下,确定目标对象001的身份标识符不准确。For example, if it is detected that there is a target object that deviates from the target group, the pre-marked identity identifier of the target object is 001. In the case where it is determined that the similarity between the feature information of the deviation target object extracted from the video image at the current moment and the feature information associated with the identity identifier 001 is less than the preset similarity threshold, it is determined that the identity identifier of the target object 001 is inaccurate .
S703,在确定偏离目标群的目标对象的身份标识符不准确的情况下,对偏离目标群的目标对象关联的身份标识符进行修正。S703 , in the case that the identity identifiers of the target objects deviating from the target group are determined to be inaccurate, correct the identity identifiers associated with the target objects deviating from the target group.
示例性地,在确定偏离目标群的目标对象的身份标识符不准确的情况下,可以基于提取的该偏离目标群的目标对象的特征信息和预先存储的员工身份库中每个员工的特征信息,重新确定该偏离目标群的目标对象的身份标识符。Exemplarily, in the case of determining that the identity identifier of the target object deviating from the target group is inaccurate, it can be based on the extracted feature information of the target object deviating from the target group and the pre-stored feature information of each employee in the employee identity database. , and re-determine the identity identifier of the target object that deviates from the target group.
本公开实施例中,在检测到存在离开目标群的目标对象的情况下,对离开目标群的目标对象的身份标识符进行重新验证,可以提高在不同时刻标记的目标对象的身份标识符的准确度,从而提高目标对象的轨迹数据的准确度。In the embodiment of the present disclosure, when it is detected that there is a target object leaving the target group, the identity identifier of the target object leaving the target group is re-verified, which can improve the accuracy of the identity identifier of the target object marked at different times. to improve the accuracy of the trajectory data of the target object.
本公开实施例提出的目标追踪方法可以准确地确定目标场所中各目标对象在当前时刻的第二位置坐标,该方式可以应用于多种应用场景。以应用于工厂为例,在得到目标对象在目标场所中的第二位置坐标之后,如图9所示,本公开实施例提供的定位方法还包括以下S801~S802:The target tracking method proposed in the embodiment of the present disclosure can accurately determine the second position coordinates of each target object in the target place at the current moment, and this method can be applied to various application scenarios. Taking the application in a factory as an example, after obtaining the second position coordinates of the target object in the target place, as shown in FIG. 9 , the positioning method provided by the embodiment of the present disclosure further includes the following S801 to S802:
S801,基于目标场所中的各目标对象分别对应的第二位置坐标,以及预先设定的目标区域,确定是否存在进入目标区域的目标对象;S801, based on the second position coordinates corresponding to each target object in the target place and the preset target area, determine whether there is a target object entering the target area;
S802,在确定存在进入目标区域的目标对象的情况下,进行预警提示。S802, if it is determined that there is a target object entering the target area, perform an early warning prompt.
示例性地,目标场所为工厂的情况下,可以预先在目标场所对应的世界坐标下中设定工厂内存在危险的目标区域对应的坐标范围。然后根据确定的目标场所中的各目标对象在当前时刻分别对应的第二位置坐标以及目标位于对应的坐标范围,确定是否存在进入目标区域的目标对象。进一步在确定存在进入目标区域的目标对象的情况下,进行预警提示。Exemplarily, in the case where the target site is a factory, a coordinate range corresponding to a dangerous target area in the factory may be set in advance in the world coordinates corresponding to the target site. Then, it is determined whether there is a target object entering the target area according to the second position coordinates corresponding to each target object in the determined target place at the current moment and the target location in the corresponding coordinate range. Further, when it is determined that there is a target object entering the target area, an early warning prompt is performed.
示例性地,预警提示可以包括但不限于声光报警提示、语音报警提示等。通过预警提示,可以保障目标场所中员工的安全,提高目标场所的安全性。Exemplarily, the early warning prompts may include, but are not limited to, sound and light alarm prompts, voice alarm prompts, and the like. Through the early warning prompts, the safety of employees in the target site can be guaranteed and the safety of the target site can be improved.
本公开实施例中,在得到目标场所中的各目标对象准确度较高的第二位置坐标后,可以基于预先设定的目标区域,比如预先设定的危险区域,判断目标场所中的目标对象是否进入目标区域, 以便及时预警提示,提高目标场所的安全性。In the embodiment of the present disclosure, after obtaining the second position coordinates of each target object in the target place with high accuracy, the target object in the target place can be determined based on a preset target area, such as a preset dangerous area Whether to enter the target area, so as to prompt early warning and improve the safety of the target site.
下面结合图10,以目标场所为工厂,目标对象为员工为例,对本公开实施例提供的目标追踪过程进行阐述:Below in conjunction with FIG. 10 , the target tracking process provided by the embodiment of the present disclosure is described by taking the target site as a factory and the target object as an employee as an example:
1)针对工厂进行采集设备安装,比如在工厂安装多个相机;为了实现对场景内目标的精准定位,并保证算法通用性及鲁棒性,使得不同的采集设备在工厂中的采集视角不同,确保进入工厂中的每个员工至少被两个采集设备同时采集到;1) Install the acquisition equipment for the factory, such as installing multiple cameras in the factory; in order to achieve accurate positioning of the target in the scene, and ensure the versatility and robustness of the algorithm, different acquisition equipment has different acquisition perspectives in the factory. Ensure that each employee entering the factory is simultaneously captured by at least two capture devices;
2)使用张正友标定方式确定每个相机的内参矩阵和畸变系数;2) Use Zhang Zhengyou's calibration method to determine the internal parameter matrix and distortion coefficient of each camera;
3)在工厂内设置多个标志物,确定标志物与地面的交点在工厂对应的世界坐标系中的位置坐标;并根据相机的内参矩阵和畸变系数确定标志物与地面的交点在样本视频画面中的修正像素坐标;并根据交点在世界坐标系中的位置坐标和在样本视频画面中的修正像素坐标,确定每个相机的单应性矩阵;3) Set up multiple markers in the factory, and determine the position coordinates of the intersection of the marker and the ground in the world coordinate system corresponding to the factory; and determine the intersection of the marker and the ground in the sample video screen according to the camera's internal parameter matrix and distortion coefficient. The corrected pixel coordinates in ; and the homography matrix of each camera is determined according to the position coordinates of the intersection in the world coordinate system and the corrected pixel coordinates in the sample video screen;
4)针对进入工厂中的每个员工进行特征检测,比如可以包含图10中的人体检测和人脸识别,得到每个员工的特征信息;并基于提取的特征信息和预先构建预先存储的员工身份库中每个员工的特征信息,确定进入工厂的每个员工的身份标识符;4) Perform feature detection for each employee entering the factory, such as the human body detection and face recognition in Figure 10, to obtain the feature information of each employee; and based on the extracted feature information and pre-built pre-stored employee identities The characteristic information of each employee in the library to determine the identity identifier of each employee entering the factory;
5)针对采集设备采集的当前时刻的视频画面,使用加入特征金字塔的神经网络进行目标检测,得到每张当前时刻的视频画面中涉及的员工的像素坐标;5) For the video picture of the current moment collected by the acquisition device, use the neural network added to the feature pyramid to perform target detection, and obtain the pixel coordinates of the employees involved in the video picture of each current moment;
6)根据采集该张视频画面的相机的内参矩阵和畸变系数,对该张视频画面中涉及的员工的像素坐标进行修正,得到该张视频画面中涉及的员工的修正像素坐标;6) According to the internal parameter matrix and the distortion coefficient of the camera that collects the video picture, the pixel coordinates of the employees involved in the video picture are corrected to obtain the corrected pixel coordinates of the employees involved in the video picture;
7)根据采集该张视频画面的相机的单应性矩阵和该张视频画面中涉及的员工的修正像素坐标,确定该张视频画面中涉及的员工在工厂中的初始位置坐标;7) According to the homography matrix of the camera that collects the video picture and the corrected pixel coordinates of the employee involved in the video picture, determine the initial position coordinates of the employee involved in the video picture in the factory;
8)对同一时刻采集到的视频画面中涉及同一员工的初始位置坐标进行融合,得到工厂中的员工在该时刻的第一位置坐标;8) Fusion of the initial position coordinates of the same employee involved in the video images collected at the same moment, to obtain the first position coordinates of the employees in the factory at this moment;
9)根据确定的员工在上一时刻的第二位置坐标和员工在当前时刻的第一位置坐标,确定员工在当前时刻的第二位置坐标,具体过程详见上文;9) According to the determined second position coordinates of the employee at the previous moment and the first position coordinates of the employee at the current moment, determine the second position coordinates of the employee at the current moment. For the specific process, please refer to the above;
10)每次在确定员工在当前时刻的第二位置坐标的同时,可以在员工的第二位置坐标指示的地图位置中标记员工关联的身份标识符;进一步基于标记同一身份标识符的员工在多个时刻的第二位置坐标,生成每个员工的轨迹数据。10) When determining the second position coordinates of the employee at the current moment, the employee's associated identity identifier can be marked in the map position indicated by the employee's second position coordinate; The second position coordinates of each moment, and the trajectory data of each employee is generated.
本领域技术人员可以理解,在具体实施方式的上述方法中,各步骤的撰写顺序并不意味着严格的执行顺序而对实施过程构成任何限定,各步骤的具体执行顺序应当以其功能和可能的内在逻辑确定。Those skilled in the art can understand that in the above method of the specific implementation, the writing order of each step does not mean a strict execution order but constitutes any limitation on the implementation process, and the specific execution order of each step should be based on its function and possible Internal logic is determined.
基于同一技术构思,本公开实施例中还提供了与目标追踪方法对应的目标追踪装置,由于本公开实施例中的装置解决问题的原理与本公开实施例上述目标追踪方法相似,因此装置的实施可以参见方法的实施,重复之处不再赘述。Based on the same technical concept, the embodiment of the present disclosure also provides a target tracking device corresponding to the target tracking method. Reference may be made to the implementation of the method, and repeated descriptions will not be repeated.
参照图11所示,为本公开实施例提供的一种目标追踪装置900的示意图,该目标追踪装置包括:Referring to FIG. 11 , which is a schematic diagram of a target tracking apparatus 900 according to an embodiment of the present disclosure, the target tracking apparatus includes:
获取模块901,用于获取目标场所内设置的多个采集设备采集的当前时刻的视频画面;多个采集设备在目标场所中的采集视角不同,视频画面中包括目标场所中目标对象的感兴趣区域;The acquisition module 901 is used to acquire the video images at the current moment collected by multiple collection devices set in the target site; the collection perspectives of the multiple collection devices in the target site are different, and the video images include the region of interest of the target object in the target site ;
确定模块902,用于基于多个采集设备采集的当前时刻的视频画面,确定各个所述目标对象在当前时刻的第一位置坐标;A determination module 902, configured to determine the first position coordinates of each of the target objects at the current moment based on the video images at the current moment collected by multiple collection devices;
追踪模块903,用于针对各个所述目标对象,基于该目标对象的第一位置坐标和该目标对象在上一时刻的第二位置坐标,确定该目标对象在当前时刻的第二位置坐标。The tracking module 903 is configured to, for each of the target objects, determine the second position coordinates of the target object at the current moment based on the first position coordinates of the target object and the second position coordinates of the target object at the previous moment.
在一种可能的实施方式中,追踪模块903在用于针对各个所述目标对象,基于该目标对象的第一位置坐标和该目标对象在上一时刻的第二位置坐标,确定该目标对象在当前时刻的第二位置坐标时,包括:In a possible implementation manner, the tracking module 903 is configured to, for each of the target objects, determine, based on the first position coordinates of the target object and the second position coordinates of the target object at the previous moment, that the target object is The coordinates of the second position at the current moment include:
基于目标对象在上一时刻的第二位置坐标,确定目标对象在当前时刻的预测位置坐标;Determine the predicted position coordinates of the target object at the current moment based on the second position coordinates of the target object at the previous moment;
基于该目标对象在当前时刻的预测位置坐标和第一位置坐标,确定该目标对象在当前时刻的观测位置坐标;Based on the predicted position coordinates and the first position coordinates of the target object at the current moment, determine the observed position coordinates of the target object at the current moment;
基于该目标对象在当前时刻的预测位置坐标和观测位置坐标,确定该目标对象在当前时刻的第二位置坐标。Based on the predicted position coordinates and the observed position coordinates of the target object at the current moment, the second position coordinates of the target object at the current moment are determined.
在一种可能的实施方式中,追踪模块903在用于基于该目标对象在当前时刻的预测位置坐标和第一位置坐标,确定该目标对象在当前时刻的观测位置坐标时,包括:In a possible implementation manner, when the tracking module 903 is used to determine the observed position coordinates of the target object at the current moment based on the predicted position coordinates and the first position coordinates of the target object at the current moment, the method includes:
基于多个目标对象在当前时刻的预测位置坐标和第一位置坐标,确定与同一目标对象关联的预测位置坐标和第一位置坐标;Determine the predicted position coordinates and the first position coordinates associated with the same target object based on the predicted position coordinates and the first position coordinates of the multiple target objects at the current moment;
确定与同一目标对象关联的预测位置坐标和第一位置坐标的第一中点坐标,将该第一中点坐标作为该目标对象在当前时刻的观测位置坐标。Determine the predicted position coordinates associated with the same target object and the first midpoint coordinates of the first position coordinates, and use the first midpoint coordinates as the observed position coordinates of the target object at the current moment.
在一种可能的实施方式中,追踪模块903在用于基于多个目标对象在当前时刻的预测位置坐标和第一位置坐标,确定与同一目标对象关联的预测位置坐标和第一位置坐标时,包括:In a possible implementation manner, when the tracking module 903 is used to determine the predicted position coordinates and the first position coordinates associated with the same target object based on the predicted position coordinates and the first position coordinates of the multiple target objects at the current moment, include:
针对每个预测位置坐标,确定该预测位置坐标和各个第一位置坐标之间的第一距离;For each predicted position coordinate, determine a first distance between the predicted position coordinate and each first position coordinate;
将该预测位置坐标和与该预测位置坐标具有最小第一距离的第一位置坐标,作为与同一目标对象关联的预测位置坐标和第一位置坐标,最小第一距离小于第一预设融合距离阈值。The predicted position coordinates and the first position coordinates with the minimum first distance from the predicted position coordinates are taken as the predicted position coordinates and the first position coordinates associated with the same target object, and the minimum first distance is smaller than the first preset fusion distance threshold. .
在一种可能的实施方式中,追踪模块903还用于:In a possible implementation manner, the tracking module 903 is further configured to:
确定当前时刻的视频画面中是否存在漏检的目标对象,其中,漏检的目标对象在当前时刻具有预测位置坐标,且在当前时刻的第一位置坐标为空;Determine whether there is an undetected target object in the video picture at the current moment, wherein the missed target object has a predicted position coordinate at the current moment, and the first position coordinate at the current moment is empty;
在确定存在漏检的目标对象的情况下,将漏检的目标对象在当前时刻的预测位置坐标作为漏检的目标对象在当前时刻的观测位置坐标。When it is determined that there is an undetected target object, the predicted position coordinates of the missed target object at the current moment are taken as the observed position coordinates of the missed target object at the current moment.
在一种可能的实施方式中,目标对象包含多个,追踪模块903还用于:In a possible implementation manner, the target object includes multiple objects, and the tracking module 903 is further configured to:
在确定目标对象在当前时刻的第二位置坐标后,在第二位置坐标指示的地图位置中标记与目标对象关联的身份标识符;After determining the second position coordinates of the target object at the current moment, marking the identity identifier associated with the target object in the map position indicated by the second position coordinates;
基于标记同一身份标识符的目标对象在多个时刻的第二位置坐标,生成每个目标对象的轨迹数据。Based on the second position coordinates of the target objects marked with the same identity identifier at multiple times, the trajectory data of each target object is generated.
在一种可能的实施方式中,在确定目标对象在当前时刻的第二位置坐标之后,追踪模块903还用于:In a possible implementation manner, after determining the second position coordinates of the target object at the current moment, the tracking module 903 is further configured to:
基于多个目标对象在当前时刻的第二位置坐标,检测是否存在偏离目标群的目标对象;目标群为根据多个目标对象在上一时刻的第二位置坐标进行聚类得到的;Based on the second position coordinates of the multiple target objects at the current moment, detect whether there are target objects that deviate from the target group; the target group is obtained by clustering according to the second position coordinates of the multiple target objects at the previous moment;
在确定存在偏离目标群的目标对象情况下,检测偏离目标群的目标对象关联的身份标识符是否准确;In the case of determining that there is a target object deviating from the target group, detecting whether the identity identifier associated with the target object deviating from the target group is accurate;
在确定偏离目标群的目标对象的身份标识符不准确的情况下,对偏离目标群的目标对象关联的身份标识符进行修正。If it is determined that the identity identifiers of the target objects deviating from the target group are inaccurate, the identity identifiers associated with the target objects deviating from the target group are corrected.
在一种可能的实施方式中,追踪模块903在用于检测偏离目标群的目标对象关联的身份标识符是否准确时,包括:In a possible implementation manner, when the tracking module 903 is used to detect whether the identity identifier associated with the target object deviating from the target group is accurate, the method includes:
提取偏离目标群的目标对象的特征信息;Extract feature information of target objects that deviate from the target group;
基于偏离目标群的目标对象的特征信息,以及预先保存的进入目标场所中的各目标对象的特征信息和身份标识符之间的映射关系,检测偏离目标群的目标对象关联的身份标识符是否准确。Based on the feature information of the target objects that deviate from the target group, and the pre-saved mapping relationship between the feature information of each target object entering the target place and the identity identifiers, it is detected whether the identity identifiers associated with the target objects that deviate from the target group are accurate. .
在一种可能的实施方式中,确定模块902在用于基于多个采集设备采集的当前时刻的视频画面,确定各个所述目标对象在当前时刻的第一位置坐标时,包括:In a possible implementation manner, when the determining module 902 is used to determine the first position coordinates of each target object at the current moment based on the video images at the current moment collected by multiple collection devices, the method includes:
获取多个采集设备分别采集的当前时刻的视频画面中的感兴趣区域的像素坐标;Acquiring the pixel coordinates of the region of interest in the video image at the current moment collected by multiple collection devices respectively;
针对每个采集设备,基于该采集设备采集的当前时刻的视频画面中的感兴趣区域的像素坐标和该采集设备的参数信息,确定该视频画面中感兴趣区域对应的目标对象在当前时刻下在目标场所中的初始位置坐标;For each acquisition device, based on the pixel coordinates of the region of interest in the video picture collected by the acquisition device at the current moment and the parameter information of the acquisition device, it is determined that the target object corresponding to the region of interest in the video picture is at the current moment. the coordinates of the initial position in the target site;
对初始位置坐标中属于同一目标对象的初始位置坐标进行融合,得到该目标对象在当前时刻下在目标场所中的第一位置坐标。The initial position coordinates belonging to the same target object in the initial position coordinates are fused to obtain the first position coordinates of the target object in the target place at the current moment.
在一种可能的实施方式中,确定模块902在用于获取多个采集设备分别采集的当前时刻的视频画面中的感兴趣区域的像素坐标时,包括:In a possible implementation manner, when the determining module 902 is used to acquire the pixel coordinates of the region of interest in the video images at the current moment respectively collected by multiple collection devices, the method includes:
将多个当前时刻的视频画面输入预先训练的神经网络,得到每个当前时刻的视频画面中的目标检测框;其中,神经网络包含多个用于检测不同尺寸的目标对象的感兴趣区域的目标检测子网络;Inputting multiple video images at the current moment into a pre-trained neural network to obtain target detection frames in the video images at each current moment; wherein the neural network includes multiple targets for detecting regions of interest of target objects of different sizes detect subnets;
提取每个当前时刻的视频画面中的目标检测框上的目标位置点在该当前时刻的视频画面中的像素坐标,得到该当前时刻的视频画面中的感兴趣区域的像素坐标。Extract the pixel coordinates of the target position point on the target detection frame in the video picture at the current moment in the video picture at the current moment, and obtain the pixel coordinates of the region of interest in the video picture at the current moment.
在一种可能的实施方式中,确定模块902在用于基于每个采集设备采集的当前时刻的视频画面中的感兴趣区域的像素坐标和该采集设备的参数信息,确定感兴趣区域对应的目标对象在当前时刻下在目标场所中的初始位置坐标时,包括:In a possible implementation manner, the determining module 902 is used to determine the target corresponding to the region of interest based on the pixel coordinates of the region of interest in the video image at the current moment collected by each acquisition device and the parameter information of the acquisition device The initial position coordinates of the object in the target location at the current moment include:
基于每个采集设备的内参矩阵和畸变参数,对该采集设备采集的视频画面中感兴趣区域的像素坐标进行修正,得到该视频画面中的感兴趣区域的修正像素坐标;Based on the internal parameter matrix and distortion parameters of each acquisition device, correct the pixel coordinates of the region of interest in the video picture collected by the acquisition device, and obtain the corrected pixel coordinates of the region of interest in the video picture;
基于预先确定的该采集设备的单应性矩阵和该采集设备采集的当前时刻的视频画面中的感兴趣区域的修正像素坐标,确定该视频画面中的感兴趣区域对应的目标对象的初始位置坐标。Based on the predetermined homography matrix of the acquisition device and the corrected pixel coordinates of the region of interest in the video image at the current moment collected by the acquisition device, determine the initial position coordinates of the target object corresponding to the region of interest in the video image .
在一种可能的实施方式中,确定模块902在用于对初始坐标位置中属于同一目标对象的初始位置坐标进行融合,得到该目标对象在当前时刻下在目标场所中的第一位置坐标时,包括:In a possible implementation manner, when the determining module 902 is used to fuse the initial position coordinates belonging to the same target object in the initial coordinate positions to obtain the first position coordinates of the target object in the target place at the current moment, include:
基于多个当前时刻的视频画面确定的初始位置坐标,确定与同一目标对象关联的多个初始位置坐标;Determine a plurality of initial position coordinates associated with the same target object based on the initial position coordinates determined by a plurality of video images at the current moment;
将与同一目标对象关联的多个初始位置坐标进行依次融合,得到该目标对象在当前时刻下在目标场所中的第一位置坐标。The multiple initial position coordinates associated with the same target object are sequentially fused to obtain the first position coordinates of the target object in the target place at the current moment.
在一种可能的实施方式中,确定模块902在用于将与同一目标对象关联的多个初始位置坐标进行依次融合,得到该目标对象在当前时刻下在目标场所中的第一位置坐标时,包括:In a possible implementation manner, when the determining module 902 is used to sequentially fuse multiple initial position coordinates associated with the same target object to obtain the first position coordinates of the target object in the target place at the current moment, include:
从该目标对象关联的多个初始位置坐标中选取任一初始位置坐标,将选取的任一初始位置坐标作为第一中间融合位置坐标;Select any initial position coordinate from a plurality of initial position coordinates associated with the target object, and use any selected initial position coordinate as the first intermediate fusion position coordinate;
将第一中间融合位置坐标与其它任一待融合的初始位置坐标进行融合,生成第二中间融合位置坐标,将第二中间融合位置坐标作为更新后的第一中间融合位置坐标,并返回生成第二中间融合位置坐标的步骤,直到不存在待融合的初始位置坐标。The first intermediate fusion position coordinates are fused with any other initial position coordinates to be fused to generate the second intermediate fusion position coordinates, and the second intermediate fusion position coordinates are used as the updated first intermediate fusion position coordinates, and return to generate the first intermediate fusion position coordinates. Step 2: The intermediate position coordinates are fused until there is no initial position coordinates to be fused.
在一种可能的实施方式中,确定模块902在用于将第一中间融合位置坐标与其它任一待融合的初始位置坐标进行融合,生成第二中间融合位置坐标时,包括:In a possible implementation manner, when the determining module 902 is used to fuse the first intermediate fused position coordinates with any other initial position coordinates to be fused to generate the second intermediate fused position coordinates, it includes:
确定第一中间融合位置坐标与其它任一待融合的初始位置坐标的中点坐标,将该中点坐标作为生成的第二中间融合位置坐标。Determine the midpoint coordinate of the first intermediate fusion position coordinate and any other initial position coordinate to be fused, and use the midpoint coordinate as the generated second intermediate fusion position coordinate.
在一种可能的实施方式中,确定模块902在用于基于多个当前时刻的视频画面确定的初始位置坐标,确定与同一目标对象关联的多个初始位置坐标时,包括:In a possible implementation manner, when the determining module 902 is used to determine a plurality of initial position coordinates associated with the same target object based on the initial position coordinates determined based on a plurality of video images at the current moment, the method includes:
针对当前时刻的任意两个视频画面,确定任意两个视频画面中第一视频画面中每个感兴趣区域对应的目标对象为第一目标对象,该任意两个视频画面中第二视频画面中每个感兴趣区域对应的目标对象为第二目标对象,确定该第一目标对象的初始位置坐标与第二当前时刻的视频画面中各个第二目标对象的初始位置坐标之间的第二距离;For any two video pictures at the current moment, determine the target object corresponding to each region of interest in the first video picture in the arbitrary two video pictures as the first target object, and each of the second video pictures in the arbitrary two video pictures is the first target object. The target object corresponding to each region of interest is a second target object, and the second distance between the initial position coordinates of the first target object and the initial position coordinates of each second target object in the video picture at the second current moment is determined;
针对每个第一目标对象的初始位置坐标,将该第一目标对象的初始位置坐标和与该第一目标对象具有最小第二距离的第二目标对象的初始位置坐标,作为与同一目标对象关联的多个初始位置坐标;最小第二距离小于第二预设融合距离阈值。For the initial position coordinates of each first target object, the initial position coordinates of the first target object and the initial position coordinates of the second target object having the smallest second distance from the first target object are regarded as being associated with the same target object multiple initial position coordinates; the minimum second distance is less than the second preset fusion distance threshold.
在一种可能的实施方式中,在追踪模块903确定目标对象在当前时刻的第二位置坐标后,确定模块902还用于:In a possible implementation manner, after the tracking module 903 determines the second position coordinates of the target object at the current moment, the determining module 902 is further configured to:
基于目标场所中的各目标对象分别对应的第二位置坐标,以及预先设定的目标区域,确定是否存在进入目标区域的目标对象;Determine whether there is a target object entering the target area based on the second position coordinates corresponding to each target object in the target place and the preset target area;
在确定存在进入目标区域的目标对象的情况下,进行预警提示。When it is determined that there is a target object entering the target area, an early warning prompt is given.
关于装置中的各模块的处理流程、以及各模块之间的交互流程的描述可以参照上述方法实施例中的相关说明,这里不再详述。For the description of the processing flow of each module in the apparatus and the interaction flow between the modules, reference may be made to the relevant descriptions in the foregoing method embodiments, which will not be described in detail here.
对应于图1中的目标追踪方法,本公开实施例还提供了一种电子设备1100,如图12所示,为本公开实施例提供的电子设备1100结构示意图,包括:Corresponding to the target tracking method in FIG. 1 , an embodiment of the present disclosure further provides an electronic device 1100 . As shown in FIG. 12 , a schematic structural diagram of the electronic device 1100 provided by an embodiment of the present disclosure includes:
处理器111、存储器112、和总线113;存储器112用于存储执行指令,包括内存1121和外部存储器1122;这里的内存1121也称内存储器,用于暂时存放处理器111中的运算数据,以及与硬盘等外部存储器1122交换的数据;处理器111通过内存1121与外部存储器1122进行数据交换;当电子设备1100运行时,处理器111与存储器112之间通过总线113通信,使得处理器111执行以下指令:获取目标场所内设置的多个采集设备采集的当前时刻的视频画面;多个采集设备在目标场所中的采集视角不同,视频画面中包括目标场所中目标对象的感兴趣区域;基于多个采集设备采集的当前时刻的视频画面,确定目标场所中的目标对象在当前时刻的第一位置坐标;针对各个目标对象,基于该目标对象的第一位置坐标和该目标对象在上一时刻的第二位置坐标,确定该目标对象在当前时刻的第二位置坐标。The processor 111, the memory 112, and the bus 113; the memory 112 is used to store the execution instructions, including the memory 1121 and the external memory 1122; the memory 1121 here is also called the internal memory, which is used to temporarily store the operation data in the processor 111, and The data exchanged by the external memory 1122 such as the hard disk; the processor 111 exchanges data with the external memory 1122 through the memory 1121; when the electronic device 1100 is running, the processor 111 and the memory 112 communicate through the bus 113, so that the processor 111 executes the following instructions : Obtain the video images at the current moment collected by multiple collection devices set up in the target site; the collection perspectives of multiple collection devices in the target site are different, and the video images include the area of interest of the target object in the target site; The video screen at the current moment collected by the device determines the first position coordinates of the target object in the target place at the current moment; for each target object, based on the first position coordinates of the target object and the second position of the target object at the previous moment. Position coordinates, which determine the second position coordinates of the target object at the current moment.
本公开实施例还提供一种计算机可读存储介质,该计算机可读存储介质上存储有计算机程序,该计算机程序被处理器运行时执行上述方法实施例中所述的目标追踪方法的步骤。其中,该存储介质可以是易失性或非易失的计算机可读取存储介质。Embodiments of the present disclosure further provide a computer-readable storage medium, where a computer program is stored on the computer-readable storage medium, and when the computer program is run by a processor, the steps of the target tracking method described in the above method embodiments are executed. Wherein, the storage medium may be a volatile or non-volatile computer-readable storage medium.
本公开实施例还提供一种计算机程序产品,该计算机程序产品承载有程序代码,所述程序代码包括的指令可用于执行上述方法实施例中所述的目标追踪方法的步骤,具体可参见上述方法实施例,在此不再赘述。Embodiments of the present disclosure further provide a computer program product, where the computer program product carries program codes, and the instructions included in the program codes can be used to execute the steps of the target tracking method described in the foregoing method embodiments. For details, please refer to the foregoing method. The embodiments are not repeated here.
其中,上述计算机程序产品可以具体通过硬件、软件或其结合的方式实现。在一个可选实施例中,所述计算机程序产品具体体现为计算机存储介质,在另一个可选实施例中,计算机程序产品具体体现为软件产品,例如软件开发包(Software Development Kit,SDK)等等。Wherein, the above-mentioned computer program product can be specifically implemented by hardware, software or a combination thereof. In an optional embodiment, the computer program product is embodied as a computer storage medium, and in another optional embodiment, the computer program product is embodied as a software product, such as a software development kit (Software Development Kit, SDK), etc. Wait.
所属领域的技术人员可以清楚地了解到,为描述的方便和简洁,上述描述的系统和装置的具体工作过程,可以参考前述方法实施例中的对应过程,在此不再赘述。在本公开所提供的几个实施 例中,应该理解到,所揭露的系统、装置和方法,可以通过其它的方式实现。以上所描述的装置实施例仅仅是示意性的,例如,所述单元的划分,仅仅为一种逻辑功能划分,实际实现时可以有另外的划分方式,又例如,多个单元或组件可以结合或者可以集成到另一个系统,或一些特征可以忽略,或不执行。另一点,所显示或讨论的相互之间的耦合或直接耦合或通信连接可以是通过一些通信接口,装置或单元的间接耦合或通信连接,可以是电性,机械或其它的形式。Those skilled in the art can clearly understand that, for the convenience and brevity of description, for the specific working process of the system and device described above, reference may be made to the corresponding process in the foregoing method embodiments, which will not be repeated here. In the several embodiments provided by the present disclosure, it should be understood that the disclosed system, apparatus and method may be implemented in other manners. The apparatus embodiments described above are only illustrative. For example, the division of the units is only a logical function division. In actual implementation, there may be other division methods. For example, multiple units or components may be combined or Can be integrated into another system, or some features can be ignored, or not implemented. On the other hand, the shown or discussed mutual coupling or direct coupling or communication connection may be through some communication interfaces, indirect coupling or communication connection of devices or units, which may be in electrical, mechanical or other forms.
所述作为分离部件说明的单元可以是或者也可以不是物理上分开的,作为单元显示的部件可以是或者也可以不是物理单元,即可以位于一个地方,或者也可以分布到多个网络单元上。可以根据实际的需要选择其中的部分或者全部单元来实现本实施例方案的目的。The units described as separate components may or may not be physically separated, and components displayed as units may or may not be physical units, that is, may be located in one place, or may be distributed to multiple network units. Some or all of the units may be selected according to actual needs to achieve the purpose of the solution in this embodiment.
另外,在本公开各个实施例中的各功能单元可以集成在一个处理单元中,也可以是各个单元单独物理存在,也可以两个或两个以上单元集成在一个单元中。In addition, each functional unit in each embodiment of the present disclosure may be integrated into one processing unit, or each unit may exist physically alone, or two or more units may be integrated into one unit.
所述功能如果以软件功能单元的形式实现并作为独立的产品销售或使用时,可以存储在一个处理器可执行的非易失的计算机可读取存储介质中。基于这样的理解,本公开的技术方案本质上或者说对现有技术做出贡献的部分或者该技术方案的部分可以以软件产品的形式体现出来,该计算机软件产品存储在一个存储介质中,包括若干指令用以使得一台计算机设备(可以是个人计算机,服务器,或者网络设备等)执行本公开各个实施例所述方法的全部或部分步骤。而前述的存储介质包括:U盘、移动硬盘、只读存储器(Read-Only Memory,ROM)、随机存取存储器(Random Access Memory,RAM)、磁碟或者光盘等各种可以存储程序代码的介质。The functions, if implemented in the form of software functional units and sold or used as stand-alone products, may be stored in a processor-executable non-volatile computer-readable storage medium. Based on such understanding, the technical solutions of the present disclosure can be embodied in the form of software products in essence, or the parts that contribute to the prior art or the parts of the technical solutions. The computer software products are stored in a storage medium, including Several instructions are used to cause a computer device (which may be a personal computer, a server, or a network device, etc.) to execute all or part of the steps of the methods described in various embodiments of the present disclosure. The aforementioned storage medium includes: U disk, mobile hard disk, read-only memory (Read-Only Memory, ROM), random access memory (Random Access Memory, RAM), magnetic disk or optical disk and other media that can store program codes .
最后应说明的是:以上所述实施例,仅为本公开的具体实施方式,用以说明本公开的技术方案,而非对其限制,本公开的保护范围并不局限于此,尽管参照前述实施例对本公开进行了详细的说明,本领域的普通技术人员应当理解:任何熟悉本技术领域的技术人员在本公开揭露的技术范围内,其依然可以对前述实施例所记载的技术方案进行修改或可轻易想到变化,或者对其中部分技术特征进行等同替换;而这些修改、变化或者替换,并不使相应技术方案的本质脱离本公开实施例技术方案的精神和范围,都应涵盖在本公开的保护范围之内。因此,本公开的保护范围应所述以权利要求的保护范围为准。Finally, it should be noted that the above-mentioned embodiments are only specific implementations of the present disclosure, and are used to illustrate the technical solutions of the present disclosure, but not to limit them. The protection scope of the present disclosure is not limited to this, although the aforementioned The embodiments describe the present disclosure in detail, and those skilled in the art should understand that: any person skilled in the art can still modify the technical solutions described in the foregoing embodiments within the technical scope disclosed by the present disclosure. Or can easily think of changes, or equivalently replace some of the technical features; and these modifications, changes or replacements do not make the essence of the corresponding technical solutions deviate from the spirit and scope of the technical solutions of the embodiments of the present disclosure, and should be covered in the present disclosure. within the scope of protection. Therefore, the protection scope of the present disclosure should be based on the protection scope of the claims.

Claims (20)

  1. 一种目标追踪方法,包括:A target tracking method comprising:
    获取目标场所内设置的多个采集设备采集的当前时刻的视频画面;所述多个采集设备在所述目标场所中的采集视角不同,所述视频画面中包括所述目标场所中的目标对象的感兴趣区域;Obtain the video images at the current moment collected by multiple collection devices set in the target place; the multiple collection devices have different capture angles in the target place, and the video images include the image of the target object in the target place. area of interest;
    基于所述多个采集设备采集的所述当前时刻的视频画面,确定各个所述目标对象在所述当前时刻的第一位置坐标;determining the first position coordinates of each of the target objects at the current moment based on the video images at the current moment collected by the multiple collection devices;
    针对各个所述目标对象,基于该目标对象的所述第一位置坐标和该目标对象在上一时刻的第二位置坐标,确定该目标对象在当前时刻的第二位置坐标。For each of the target objects, the second position coordinates of the target object at the current moment are determined based on the first position coordinates of the target object and the second position coordinates of the target object at the previous moment.
  2. 根据权利要求1所述的目标追踪方法,其特征在于,基于该目标对象的所述第一位置坐标和该目标对象在上一时刻的第二位置坐标,确定该目标对象在所述当前时刻的第二位置坐标,包括:The target tracking method according to claim 1, characterized in that, based on the first position coordinates of the target object and the second position coordinates of the target object at the previous moment, determining the position of the target object at the current moment Second location coordinates, including:
    基于该目标对象在上一时刻的第二位置坐标,确定该目标对象在所述当前时刻的预测位置坐标;Based on the second position coordinates of the target object at the previous moment, determine the predicted position coordinates of the target object at the current moment;
    基于该目标对象在所述当前时刻的所述预测位置坐标和所述第一位置坐标,确定该目标对象在所述当前时刻的观测位置坐标;Determine the observed position coordinates of the target object at the current moment based on the predicted position coordinates and the first position coordinates of the target object at the current moment;
    基于该目标对象在所述当前时刻的所述预测位置坐标和所述观测位置坐标,确定该目标对象在所述当前时刻的第二位置坐标。Based on the predicted position coordinates and the observed position coordinates of the target object at the current moment, the second position coordinates of the target object at the current moment are determined.
  3. 根据权利要求2所述的目标追踪方法,其特征在于,基于该目标对象在所述当前时刻的所述预测位置坐标和所述第一位置坐标,确定该目标对象在所述当前时刻的观测位置坐标,包括:The target tracking method according to claim 2, wherein the observed position of the target object at the current moment is determined based on the predicted position coordinates and the first position coordinates of the target object at the current moment Coordinates, including:
    基于所述目标对象在所述当前时刻的多个第一位置坐标和该目标对象在所述当前时刻的所述预测位置坐标,确定该目标对象的第一位置坐标;determining the first position coordinates of the target object based on a plurality of first position coordinates of the target object at the current moment and the predicted position coordinates of the target object at the current moment;
    确定该目标对象的所述预测位置坐标和所述第一位置坐标的第一中点坐标,将该第一中点坐标作为该目标对象在所述当前时刻的观测位置坐标。The predicted position coordinates of the target object and the first midpoint coordinates of the first position coordinates are determined, and the first midpoint coordinates are used as the observed position coordinates of the target object at the current moment.
  4. 根据权利要求3所述的目标追踪方法,其特征在于,基于所述目标对象在所述当前时刻的多个第一位置坐标和该目标对象在所述当前时刻的预测位置坐标,确定该目标对象的第一位置坐标,包括:The target tracking method according to claim 3, wherein the target object is determined based on a plurality of first position coordinates of the target object at the current moment and the predicted position coordinates of the target object at the current moment The first position coordinates of , including:
    确定该目标对象的所述预测位置坐标和各个所述第一位置坐标之间的第一距离;determining a first distance between the predicted position coordinates of the target object and each of the first position coordinates;
    将所述多个第一位置坐标中与该目标对象的所述预测位置坐标构成最小第一距离的第一位置坐标,作为与该目标对象的所述第一位置坐标,所述最小第一距离小于第一预设融合距离阈值。Taking the first position coordinate that forms the minimum first distance with the predicted position coordinate of the target object among the plurality of first position coordinates as the first position coordinate with the target object, the minimum first distance is less than the first preset fusion distance threshold.
  5. 根据权利要求2至4任一所述的目标追踪方法,其特征在于,所述目标追踪方法还包括:The target tracking method according to any one of claims 2 to 4, wherein the target tracking method further comprises:
    确定所述当前时刻的视频画面中是否存在漏检的目标对象,其中,所述漏检的目标对象在所述当前时刻具有预测位置坐标,且在所述当前时刻的第一位置坐标为空;determining whether there is an undetected target object in the video picture at the current moment, wherein the missed target object has a predicted position coordinate at the current moment, and the first position coordinate at the current moment is empty;
    在确定存在漏检的目标对象的情况下,将所述漏检的目标对象在所述当前时刻的所述预测位置坐标作为所述漏检的目标对象在所述当前时刻的观测位置坐标。In the case where it is determined that there is a target object that is missed, the predicted position coordinates of the target object that is missed at the current moment are used as the coordinates of the observed position of the target object that is missed at the current moment.
  6. 根据权利要求1至5任一所述的目标追踪方法,其特征在于,所述目标追踪方法还包括:The target tracking method according to any one of claims 1 to 5, wherein the target tracking method further comprises:
    在确定各所述目标对象在所述当前时刻的第二位置坐标后,在所述第二位置坐标指示的地图位置中标记与各所述目标对象关联的身份标识符;After determining the second position coordinates of each of the target objects at the current moment, marking the identity identifier associated with each of the target objects in the map position indicated by the second position coordinates;
    基于标记同一身份标识符的目标对象在多个时刻的第二位置坐标,生成每个所述目标对象的轨迹数据。Based on the second position coordinates of the target objects marked with the same identity identifier at multiple times, the trajectory data of each target object is generated.
  7. 根据权利要求6所述的目标追踪方法,其特征在于,在确定各所述目标对象在所述当前时刻的第二位置坐标之后,所述目标追踪方法还包括:The target tracking method according to claim 6, wherein after determining the second position coordinates of each of the target objects at the current moment, the target tracking method further comprises:
    基于各所述目标对象在所述当前时刻的第二位置坐标,检测是否存在偏离目标群的目标对象;所述目标群为根据各所述目标对象在上一时刻的第二位置坐标进行聚类得到的;Based on the second position coordinates of each target object at the current moment, it is detected whether there are target objects that deviate from the target group; the target group is clustered according to the second position coordinates of each target object at the previous moment owned;
    在确定存在偏离目标群的目标对象情况下,检测所述偏离目标群的目标对象关联的身份标识符是否准确;In the case of determining that there is a target object deviating from the target group, detecting whether the identity identifier associated with the target object deviating from the target group is accurate;
    在确定所述偏离目标群的目标对象的身份标识符不准确的情况下,对所述偏离目标群的目标对象关联的身份标识符进行修正。When it is determined that the identity identifiers of the target objects deviating from the target group are inaccurate, the identity identifiers associated with the target objects deviating from the target group are corrected.
  8. 根据权利要求7所述的目标追踪方法,其特征在于,检测所述偏离目标群的目标对象关联的身份标识符是否准确,包括:The target tracking method according to claim 7, wherein detecting whether the identity identifier associated with the target object deviating from the target group is accurate, comprising:
    提取所述偏离目标群的目标对象的特征信息;extracting feature information of the target object deviating from the target group;
    基于所述偏离目标群的目标对象的特征信息,以及预先保存的进入所述目标场所中的各对象的特征信息和身份标识符之间的映射关系,检测所述偏离目标群的目标对象关联的身份标识符是否准确。Based on the feature information of the target objects that deviate from the target group, and the pre-stored mapping relationship between the feature information of each object entering the target place and the identity identifiers, detecting the relationship between the target objects that deviate from the target group Whether the identity identifier is accurate.
  9. 根据权利要求1所述的目标追踪方法,其特征在于,基于所述多个采集设备采集的所述当前时刻的视频画面,确定各个所述目标对象在所述当前时刻的第一位置坐标,包括:The target tracking method according to claim 1, wherein the first position coordinates of each target object at the current moment are determined based on the video images at the current moment collected by the multiple collection devices, comprising: :
    获取所述多个采集设备分别采集的所述当前时刻的视频画面中所述感兴趣区域的像素坐标;Acquiring the pixel coordinates of the region of interest in the video picture at the current moment respectively collected by the multiple collection devices;
    针对所述多个采集设备里的每一个,基于该采集设备采集的视频画面中感兴趣区域的像素坐标和该采集设备的参数信息,确定该视频画面中感兴趣区域对应的目标对象在所述当前时刻下在所述目标场所中的初始位置坐标;For each of the plurality of acquisition devices, based on the pixel coordinates of the region of interest in the video picture collected by the acquisition device and the parameter information of the acquisition device, it is determined that the target object corresponding to the region of interest in the video picture is in the video picture. The initial position coordinates in the target place at the current moment;
    对所述初始位置坐标中属于同一目标对象的初始位置坐标进行融合,得到该目标对象在当前时刻下在所述目标场所中的第一位置坐标。The initial position coordinates belonging to the same target object in the initial position coordinates are fused to obtain the first position coordinates of the target object in the target place at the current moment.
  10. 根据权利要求9所述的目标追踪方法,其特征在于,获取所述多个采集设备分别采集的所述当前时刻的视频画面中所述感兴趣区域的像素坐标,包括:The target tracking method according to claim 9, wherein acquiring the pixel coordinates of the region of interest in the video images at the current moment respectively collected by the multiple collection devices comprises:
    将所述当前时刻的视频画面分别输入预先训练的神经网络,Inputting the video images at the current moment into the pre-trained neural network respectively,
    针对所述当前时刻的视频画面中的每一个,For each of the video pictures at the current moment,
    得到该视频画面中的目标检测框;Obtain the target detection frame in the video image;
    提取该视频画面中的目标检测框上的目标位置点在该视频画面中的像素坐标,得到该视频画面中感兴趣区域的像素坐标。The pixel coordinates of the target position point on the target detection frame in the video picture are extracted in the video picture, and the pixel coordinates of the region of interest in the video picture are obtained.
  11. 根据权利要求9或10所述的目标追踪方法,其特征在于,基于该采集设备采集的视频画面中感兴趣区域的像素坐标和该采集设备的参数信息,确定该视频画面中感兴趣区域对应的目标对象在所述当前时刻下在所述目标场所中的初始位置坐标,包括:The target tracking method according to claim 9 or 10, wherein, based on the pixel coordinates of the region of interest in the video picture collected by the acquisition device and the parameter information of the acquisition device, determine the corresponding region of interest in the video picture. The initial position coordinates of the target object in the target location at the current moment, including:
    基于该采集设备的内参矩阵和畸变参数,对该视频画面中感兴趣区域的像素坐标进行修正,得到该视频画面中感兴趣区域的修正像素坐标;Based on the internal parameter matrix and the distortion parameter of the acquisition device, the pixel coordinates of the region of interest in the video picture are corrected to obtain the corrected pixel coordinates of the region of interest in the video picture;
    基于预先确定的该采集设备的单应性矩阵和该视频画面中感兴趣区域的的修正像素坐标,确定该视频画面中感兴趣区域对应的目标对象的初始位置坐标。Based on the predetermined homography matrix of the acquisition device and the corrected pixel coordinates of the region of interest in the video picture, the initial position coordinates of the target object corresponding to the region of interest in the video picture are determined.
  12. 根据权利要求9至10任一所述的目标追踪方法,其特征在于,对所述初始位置坐标中属于同一目标对象的初始位置坐标进行融合,得到该目标对象在所述当前时刻下在所述目标场所中的第一位置坐标,包括:The target tracking method according to any one of claims 9 to 10, wherein the initial position coordinates belonging to the same target object in the initial position coordinates are fused to obtain the target object at the current moment in the said target object. The coordinates of the first location in the target site, including:
    基于所述初始位置坐标,确定与同一目标对象关联的多个初始位置坐标;Based on the initial position coordinates, determine a plurality of initial position coordinates associated with the same target object;
    将与该目标对象关联的所述多个初始位置坐标进行依次融合,得到该目标对象在所述当前时刻下在所述目标场所中的第一位置坐标。The plurality of initial position coordinates associated with the target object are sequentially fused to obtain the first position coordinates of the target object in the target place at the current moment.
  13. 根据权利要求12所述的目标追踪方法,其特征在于,将与该目标对象关联的所述多个初始位置坐标进行依次融合,得到该目标对象在所述当前时刻下在所述目标场所中的第一位置坐标,包括:The target tracking method according to claim 12, wherein the plurality of initial position coordinates associated with the target object are sequentially fused to obtain the target object's position in the target place at the current moment. The first position coordinates, including:
    从该目标对象关联的所述多个初始位置坐标中选取任一初始位置坐标,将选取的任一初始位置坐标作为第一中间融合位置坐标;Select any initial position coordinate from the plurality of initial position coordinates associated with the target object, and use the selected initial position coordinate as the first intermediate fusion position coordinate;
    将所述第一中间融合位置坐标与所述多个初始位置坐标中其它任一待融合的初始位置坐标进行融合,生成第二中间融合位置坐标,将所述第二中间融合位置坐标作为更新后的所述第一中间融合位置坐标,并返回生成所述第二中间融合位置坐标的步骤,直到所述多个初始位置坐标中不存在待融合的初始位置坐标。The first intermediate fusion position coordinates are fused with any other initial position coordinates to be fused in the plurality of initial position coordinates to generate the second intermediate fusion position coordinates, and the second intermediate fusion position coordinates are used as the updated and returning to the step of generating the second intermediate fusion position coordinates, until there is no initial position coordinate to be fused in the plurality of initial position coordinates.
  14. 根据权利要求13所述的目标追踪方法,其特征在于,将所述第一中间融合位置坐标与所述多个初始位置坐标中其它任一待融合的初始位置坐标进行融合,生成第二中间融合位置坐标,包括:The target tracking method according to claim 13, wherein the first intermediate fusion position coordinate is fused with any other initial position coordinate to be fused among the plurality of initial position coordinates to generate a second intermediate fusion Location coordinates, including:
    确定所述第一中间融合位置坐标与所述多个初始位置坐标中其它任一待融合的初始位置坐标的中点坐标,将该中点坐标作为所述第二中间融合位置坐标。Determine a midpoint coordinate of the first intermediate fusion position coordinate and any other initial position coordinate to be fused among the plurality of initial position coordinates, and use the midpoint coordinate as the second intermediate fusion position coordinate.
  15. 根据权利要求12至14任一所述的目标追踪方法,其特征在于,基于所述初始位置坐标,确定与同一目标对象关联的多个初始位置坐标,包括:The target tracking method according to any one of claims 12 to 14, wherein, based on the initial position coordinates, determining a plurality of initial position coordinates associated with the same target object, comprising:
    针对所述当前时刻的视频画面中的任意两个视频画面,将所述任意两个视频画面中第一视频画面中的感兴趣区域对应的目标对象确定为第一目标对象,所述任意两张视频画面中第二视频画面中的感兴趣区域对应的目标对象确定为第二目标对象;For any two video pictures in the video pictures at the current moment, the target object corresponding to the region of interest in the first video picture in the any two video pictures is determined as the first target object, and the arbitrary two video pictures are determined as the first target object. The target object corresponding to the region of interest in the second video picture in the video picture is determined as the second target object;
    在所述第一视频画面中确定每个所述第一目标对象的初始位置坐标;determining the initial position coordinates of each of the first target objects in the first video frame;
    在所述第二视频画面中确定每个所述第二目标对象的初始位置坐标;determining the initial position coordinates of each of the second target objects in the second video frame;
    针对每个所述第一目标对象的初始位置坐标,For the initial position coordinates of each of the first target objects,
    确定该第一目标对象的初始位置坐标与每个所述第二目标对象的初始位置坐标之间的第二距离;determining a second distance between the initial position coordinates of the first target object and the initial position coordinates of each of the second target objects;
    确定与该第一目标对象具有最小第二距离的第二目标对象与该第一目标对象为同一目标对象;其中,所述最小第二距离小于第二预设融合距离阈值;It is determined that a second target object with a minimum second distance from the first target object is the same target object as the first target object; wherein the minimum second distance is less than a second preset fusion distance threshold;
    将该第一目标对象的初始位置坐标和与该第一目标对象具有最小第二距离的第二目标对象的初始位置坐标,作为与同一目标对象关联的多个初始位置坐标。The initial position coordinates of the first target object and the initial position coordinates of the second target object having the smallest second distance from the first target object are used as a plurality of initial position coordinates associated with the same target object.
  16. 根据权利要求1至15任一所述的目标追踪方法,其特征在于,在确定各个所述目标对象在所述当前时刻的第二位置坐标后,所述目标追踪方法还包括:The target tracking method according to any one of claims 1 to 15, wherein after determining the second position coordinates of each of the target objects at the current moment, the target tracking method further comprises:
    基于所述目标场所中的各所述目标对象分别对应的第二位置坐标,以及预先设定的目标区域,确定是否存在进入所述目标区域的目标对象;determining whether there is a target object entering the target area based on the second position coordinates corresponding to each of the target objects in the target place and a preset target area;
    在确定存在进入所述目标区域的目标对象的情况下,进行预警提示。When it is determined that there is a target object entering the target area, an early warning prompt is performed.
  17. 一种目标追踪装置,包括:A target tracking device, comprising:
    获取模块,用于获取目标场所内设置的多个采集设备采集的当前时刻的视频画面;所述多个采集设备在所述目标场所中的采集视角不同,所述视频画面中包括所述目标场所中目标对象的感兴趣区域;an acquisition module, configured to acquire the video images at the current moment collected by multiple collection devices set in the target place; the multiple collection devices have different collection perspectives in the target place, and the video images include the target place the region of interest of the target object;
    确定模块,用于基于所述多个采集设备采集的所述当前时刻的视频画面,确定各个所述目标对象在当前时刻的第一位置坐标;a determination module, configured to determine the first position coordinates of each of the target objects at the current moment based on the video images at the current moment collected by the multiple collection devices;
    追踪模块,用于针对各个所述目标对象,基于该目标对象的所述第一位置坐标和该目标对象在上一时刻的第二位置坐标,确定该目标对象在当前时刻的第二位置坐标。The tracking module is configured to, for each of the target objects, determine the second position coordinates of the target object at the current moment based on the first position coordinates of the target object and the second position coordinates of the target object at the previous moment.
  18. 一种电子设备,包括:处理器、存储器和总线,所述存储器存储有所述处理器可执行的机器可读指令,当电子设备运行时,所述处理器与所述存储器之间通过总线通信,所述机器可读指令被所述处理器执行时执行如权利要求1至16任一所述的目标追踪方法的步骤。An electronic device, comprising: a processor, a memory and a bus, the memory stores machine-readable instructions executable by the processor, and when the electronic device is running, the processor and the memory communicate through the bus , the machine-readable instructions execute the steps of the target tracking method according to any one of claims 1 to 16 when the machine-readable instructions are executed by the processor.
  19. 一种计算机可读存储介质,该计算机可读存储介质上存储有计算机程序,该计算机程序被处理器运行时执行如权利要求1至16任一项所述的目标追踪方法的步骤。A computer-readable storage medium storing a computer program on the computer-readable storage medium, the computer program executing the steps of the target tracking method according to any one of claims 1 to 16 when the computer program is run by a processor.
  20. 一种计算机程序产品,该计算机程序产品包括存储于存储介质中的计算机程序,该计算机程序被处理器执行时执行如权利要求1至16任一项所述的目标追踪方法。A computer program product, the computer program product comprising a computer program stored in a storage medium, the computer program executing the target tracking method according to any one of claims 1 to 16 when the computer program is executed by a processor.
PCT/CN2022/074956 2021-04-28 2022-01-29 Target tracking method and apparatus, electronic device, and storage medium WO2022227761A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN202110467650.7 2021-04-28
CN202110467650.7A CN113129339B (en) 2021-04-28 2021-04-28 Target tracking method and device, electronic equipment and storage medium

Publications (1)

Publication Number Publication Date
WO2022227761A1 true WO2022227761A1 (en) 2022-11-03

Family

ID=76781059

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2022/074956 WO2022227761A1 (en) 2021-04-28 2022-01-29 Target tracking method and apparatus, electronic device, and storage medium

Country Status (3)

Country Link
CN (1) CN113129339B (en)
TW (1) TW202244847A (en)
WO (1) WO2022227761A1 (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN116453029A (en) * 2023-06-16 2023-07-18 济南东庆软件技术有限公司 Building fire environment detection method based on image data

Families Citing this family (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113129339B (en) * 2021-04-28 2023-03-10 北京市商汤科技开发有限公司 Target tracking method and device, electronic equipment and storage medium
CN113759314A (en) * 2021-09-01 2021-12-07 浙江讯飞智能科技有限公司 Sound source visualization method, device and system and computer readable storage medium
CN113823029A (en) * 2021-10-29 2021-12-21 北京市商汤科技开发有限公司 Video processing method and device, electronic equipment and storage medium

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20170344832A1 (en) * 2012-11-28 2017-11-30 Innovative Alert Systems Inc. System and method for event monitoring and detection
CN111277745A (en) * 2018-12-04 2020-06-12 北京奇虎科技有限公司 Target person tracking method and device, electronic equipment and readable storage medium
CN111563919A (en) * 2020-04-03 2020-08-21 深圳市优必选科技股份有限公司 Target tracking method and device, computer readable storage medium and robot
CN112653848A (en) * 2020-12-23 2021-04-13 北京市商汤科技开发有限公司 Display method and device in augmented reality scene, electronic equipment and storage medium
CN113129339A (en) * 2021-04-28 2021-07-16 北京市商汤科技开发有限公司 Target tracking method and device, electronic equipment and storage medium

Family Cites Families (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107481327B (en) * 2017-09-08 2019-03-15 腾讯科技(深圳)有限公司 About the processing method of augmented reality scene, device, terminal device and system
CN110544273B (en) * 2018-05-29 2022-08-05 杭州海康机器人技术有限公司 Motion capture method, device and system
CN110544278B (en) * 2018-05-29 2022-09-16 杭州海康机器人技术有限公司 Rigid body motion capture method and device and AGV pose capture system
CN110517292A (en) * 2019-08-29 2019-11-29 京东方科技集团股份有限公司 Method for tracking target, device, system and computer readable storage medium
CN110738143B (en) * 2019-09-27 2023-06-02 Oppo广东移动通信有限公司 Positioning method and device, equipment and storage medium
CN110728717B (en) * 2019-09-27 2022-07-15 Oppo广东移动通信有限公司 Positioning method and device, equipment and storage medium
CN111539991B (en) * 2020-04-28 2023-10-20 北京市商汤科技开发有限公司 Target tracking method and device and storage medium
CN112330717B (en) * 2020-11-11 2023-03-10 北京市商汤科技开发有限公司 Target tracking method and device, electronic equipment and storage medium
CN112330721B (en) * 2020-11-11 2023-02-17 北京市商汤科技开发有限公司 Three-dimensional coordinate recovery method and device, electronic equipment and storage medium

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20170344832A1 (en) * 2012-11-28 2017-11-30 Innovative Alert Systems Inc. System and method for event monitoring and detection
CN111277745A (en) * 2018-12-04 2020-06-12 北京奇虎科技有限公司 Target person tracking method and device, electronic equipment and readable storage medium
CN111563919A (en) * 2020-04-03 2020-08-21 深圳市优必选科技股份有限公司 Target tracking method and device, computer readable storage medium and robot
CN112653848A (en) * 2020-12-23 2021-04-13 北京市商汤科技开发有限公司 Display method and device in augmented reality scene, electronic equipment and storage medium
CN113129339A (en) * 2021-04-28 2021-07-16 北京市商汤科技开发有限公司 Target tracking method and device, electronic equipment and storage medium

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
GUAN YINGDA; YE ZUOCHANG; WANG YAN: "Object Detection with Extended Attention and Spatial Information", 2020 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), IEEE, 19 July 2020 (2020-07-19), pages 1 - 8, XP033833937, DOI: 10.1109/IJCNN48605.2020.9206729 *
YANG GUANG: "Geometric Features Recognition of UAV Swarm Targets", CHINESE MASTER'S THESES FULL-TEXT DATABASE - UNIVERSITY OF ELECTRONIC SCIENCE AND TECHNOLOGY OF CHINA, 15 August 2020 (2020-08-15), pages 1 - 85, XP055981058 *

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN116453029A (en) * 2023-06-16 2023-07-18 济南东庆软件技术有限公司 Building fire environment detection method based on image data
CN116453029B (en) * 2023-06-16 2023-08-29 济南东庆软件技术有限公司 Building fire environment detection method based on image data

Also Published As

Publication number Publication date
CN113129339B (en) 2023-03-10
TW202244847A (en) 2022-11-16
CN113129339A (en) 2021-07-16

Similar Documents

Publication Publication Date Title
WO2022227761A1 (en) Target tracking method and apparatus, electronic device, and storage medium
Walch et al. Image-based localization using lstms for structured feature correlation
LU102028B1 (en) Multiple view multiple target tracking method and system based on distributed camera network
Torabi et al. An iterative integrated framework for thermal–visible image registration, sensor fusion, and people tracking for video surveillance applications
JP4852765B2 (en) Estimating connection relationship between distributed cameras and connection relationship estimation program
US9165190B2 (en) 3D human pose and shape modeling
Nam et al. Intelligent video surveillance system: 3-tier context-aware surveillance system with metadata
Calderara et al. Bayesian-competitive consistent labeling for people surveillance
CN110969118B (en) Track monitoring system and method
TW201118803A (en) Person-tracing apparatus and person-tracing program
CN102612704A (en) Method of providing a descriptor for at least one feature of an image and method of matching features
CN102243765A (en) Multi-camera-based multi-objective positioning tracking method and system
García et al. Tracking people motion based on extended condensation algorithm
CN111860352A (en) Multi-lens vehicle track full-tracking system and method
WO2022227462A1 (en) Positioning method and apparatus, electronic device, and storage medium
JP2013242728A (en) Image monitoring device
CN112950717A (en) Space calibration method and system
WO2023015938A1 (en) Three-dimensional point detection method and apparatus, electronic device, and storage medium
Choe et al. Traffic analysis with low frame rate camera networks
Wang et al. Effective multiple pedestrian tracking system in video surveillance with monocular stationary camera
Bazzani et al. A comparison of multi hypothesis kalman filter and particle filter for multi-target tracking
CN113793362A (en) Pedestrian track extraction method and device based on multi-lens video
JP2019121019A (en) Information processing device, three-dimensional position estimation method, computer program, and storage medium
Prokaj Exploitation of wide area motion imagery
WO2022126668A1 (en) Method for pedestrian identification in public places and human flow statistics system

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 22794242

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 22794242

Country of ref document: EP

Kind code of ref document: A1