CN112037259A - System and method for tracking dynamic target - Google Patents

System and method for tracking dynamic target Download PDF

Info

Publication number
CN112037259A
CN112037259A CN202010879875.9A CN202010879875A CN112037259A CN 112037259 A CN112037259 A CN 112037259A CN 202010879875 A CN202010879875 A CN 202010879875A CN 112037259 A CN112037259 A CN 112037259A
Authority
CN
China
Prior art keywords
target
tracking
tracking result
dynamic
module
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202010879875.9A
Other languages
Chinese (zh)
Inventor
王丹
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing Geekplus Technology Co Ltd
Original Assignee
Beijing Geekplus Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing Geekplus Technology Co Ltd filed Critical Beijing Geekplus Technology Co Ltd
Priority to CN202010879875.9A priority Critical patent/CN112037259A/en
Publication of CN112037259A publication Critical patent/CN112037259A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/20Analysis of motion
    • G06T7/246Analysis of motion using feature-based methods, e.g. the tracking of corners or segments
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/70Determining position or orientation of objects or cameras
    • G06T7/73Determining position or orientation of objects or cameras using feature-based methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10016Video; Image sequence
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10024Color image
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30196Human being; Person
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V2201/00Indexing scheme relating to image or video recognition or understanding
    • G06V2201/07Target detection

Abstract

The embodiment of the application discloses a system and a method for tracking a dynamic target. The system comprises: an image acquisition module configured to acquire an RGBD image of the dynamic target; the dynamic target detection module is connected with the image acquisition module and is configured to perform target detection on the RGBD image to obtain an initial position of a target component; and the component position tracking module is connected with the dynamic target detection module and is configured to determine the estimated position of the target in the RGBD image of the target frame according to the target motion information between at least two RGBD images so as to determine the component tracking result of the dynamic target. By executing the technical scheme, the accuracy and robustness of dynamic target tracking can be improved.

Description

System and method for tracking dynamic target
Technical Field
The embodiment of the application relates to the technical field of image processing, in particular to a system and a method for tracking a dynamic target.
Background
With the rapid development of the technology level, transportation equipment is widely applied in the field of transportation of goods. In the order-to-person scenario, there is a transportation scenario where a pedestrian is tracked by a transportation device and a corresponding item is transported for the pedestrian. In this case, the transportation device needs to continuously acquire images of pedestrians and determine its own tracking target from the images. However, the pedestrian is often in motion and may be accompanied by bending over from the side, raising the arm, and the like, and may also be blocked by other objects, or the pedestrian moves out of the image range, thereby causing a situation where the pedestrian cannot track normally. This not only makes the unable normal motion of haulage equipment, still influences pedestrian's use experience simultaneously, causes pedestrian's economic loss even easily.
Disclosure of Invention
The embodiment of the application provides a system and a method for tracking a dynamic target, which can improve the accuracy and robustness of tracking the dynamic target.
In a first aspect, an embodiment of the present application provides a tracking system for a dynamic target, where the system includes:
an image acquisition module configured to acquire an RGBD image of the dynamic target;
the dynamic target detection module is connected with the image acquisition module and is configured to perform target detection on the RGBD image to obtain an initial position of a target component;
and the component position tracking module is connected with the dynamic target detection module and is configured to determine the estimated position of the target in the RGBD image of the target frame according to the target motion information between at least two RGBD images so as to determine the component tracking result of the dynamic target.
Further, the dynamic target detection module is further configured to: carrying out target detection on the RGBD image to obtain the overall position of a target;
the system also includes an overall position tracking module, coupled to the dynamic target detection module, configured to:
and determining at least one overall tracking result matched with the overall position in the RGBD image of the target frame.
Further, the system further includes a tracking fusion module, connected to the component position tracking module and to the global position tracking module, configured to:
and determining a final tracking result of the dynamic target according to the component tracking result and the at least one integral tracking result.
Further, the tracking fusion module is configured to:
if at least two integral tracking results exist, selecting the integral tracking result with the maximum superposition coefficient with the component tracking result as a final tracking result;
and if an integral tracking result exists, taking the integral tracking result as a final tracking result when the superposition coefficient of the integral tracking result and the part tracking result accords with a set threshold value.
Further, the tracking fusion module is further configured to:
and if the integral tracking result does not exist, taking the part tracking result as a final tracking result.
Further, the dynamic target detection module is configured to:
carrying out target detection by adopting a preset detection network model to obtain the initial position of a target component; wherein the components include one or more of a head, a torso, arms, and legs.
Further, the component location tracking module is configured to:
determining target motion information between at least two frames of RGBD images by adopting an optical flow method; and determining the estimated position of the target in the RGBD image of the target frame according to the target motion information so as to determine a component tracking result of the dynamic target.
Further, an overall location tracking module configured to:
determining a target detection result in a target frame RGBD image, and determining a matching confidence of a detected target and a tracking target template of the whole position;
and determining at least one overall tracking result according to the matching confidence.
In a second aspect, an embodiment of the present application provides a method for tracking a dynamic target, where the method includes:
obtaining an RGBD image of a dynamic target;
carrying out target detection on the RGBD image to obtain an initial position of a part of a target;
and determining the estimated position of the target in the RGBD image of the target frame according to the target motion information between at least two frames of RGBD images so as to determine the part tracking result of the dynamic target.
Further, the method further comprises:
carrying out target detection on the RGBD image to obtain the overall position of a target;
and determining at least one overall tracking result matched with the overall position in the RGBD image of the target frame.
Further, the method further comprises:
and determining a final tracking result of the dynamic target according to the component tracking result and the at least one integral tracking result.
Further, determining a final tracking result of the dynamic target according to the component tracking result and the at least one overall tracking result, including:
if at least two integral tracking results exist, selecting the integral tracking result with the maximum superposition coefficient with the component tracking result as a final tracking result;
and if an integral tracking result exists, taking the integral tracking result as a final tracking result when the superposition coefficient of the integral tracking result and the part tracking result accords with a set threshold value.
Further, determining a final tracking result of the dynamic target according to the component tracking result and the at least one overall tracking result, further comprising:
and if the integral tracking result does not exist, taking the part tracking result as a final tracking result.
Further, performing target detection on the RGBD image to obtain an initial position of a component of a target, including:
carrying out target detection by adopting a preset detection network model to obtain the initial position of a target component; wherein the components include one or more of a head, a torso, arms, and legs.
Further, determining an estimated position of the target in the RGBD image of the target frame according to the target motion information between the at least two RGBD images to determine a component tracking result of the dynamic target, includes:
determining target motion information between at least two frames of RGBD images by adopting an optical flow method; and determining the estimated position of the target in the RGBD image of the target frame according to the target motion information so as to determine a component tracking result of the dynamic target.
Further, determining at least one overall tracking result matched with the overall position in the RGBD image of the target frame comprises the following steps:
determining a target detection result in a target frame RGBD image, and determining a matching confidence of a detected target and a tracking target template of the whole position;
and determining at least one overall tracking result according to the matching confidence.
The technical scheme provided by the embodiment of the application comprises the following steps: an image acquisition module configured to acquire an RGBD image of the dynamic target; the dynamic target detection module is connected with the image acquisition module and is configured to perform target detection on the RGBD image to obtain an initial position of a target component; and the component position tracking module is connected with the dynamic target detection module and is configured to determine the estimated position of the target in the RGBD image of the target frame according to the target motion information between at least two RGBD images so as to determine the component tracking result of the dynamic target. By adopting the technical scheme provided by the application, the accuracy and robustness of dynamic target tracking can be improved.
Drawings
FIG. 1 is a schematic illustration of an order-to-person operation process provided in an embodiment of the present invention;
FIG. 2 is a schematic structural view of a shelf with one-way openings according to an embodiment of the present invention;
FIG. 3 is a schematic illustration of an order-to-person control process provided in an embodiment of the present invention;
FIG. 4 is a schematic diagram of a dynamic target tracking system provided by an embodiment of the present application;
FIG. 5 is a schematic diagram of a dynamic target tracking method provided in an embodiment of the present application;
fig. 6 is a schematic diagram of a tracking method for a dynamic target provided in an embodiment of the present application.
Detailed Description
The present application will be described in further detail with reference to the following drawings and examples. It is to be understood that the specific embodiments described herein are merely illustrative of the application and are not limiting of the application. It should be further noted that, for the convenience of description, only some of the structures related to the present application are shown in the drawings, not all of the structures.
Before discussing exemplary embodiments in more detail, it should be noted that some exemplary embodiments are described as processes or methods depicted as flowcharts. Although a flowchart may describe the steps as a sequential process, many of the steps can be performed in parallel, concurrently or simultaneously. In addition, the order of the steps may be rearranged. The process may be terminated when its operations are completed, but may have additional steps not included in the figure. The processes may correspond to methods, functions, procedures, subroutines, and the like.
FIG. 1 is a schematic diagram of an order-to-person process provided in an embodiment of the invention. Referring to fig. 1, the memory area 110 includes: the shelf 111 is provided with a plurality of shelves 111, various articles are placed on the shelves 111, and the shelves 111 are arranged in an array form, for example, like shelves in supermarkets where various commodities are placed. The shelf 111 may be a container having a compartment through which items can be accessed, such as a storage box, wherein the shelf 111 includes a plurality of compartments and four floor-standing support columns, at least one compartment is disposed on the compartments of the shelf 111, and one or more items may be placed in the compartment. In addition, the shelf 111 may be one-way open, for example, fig. 2 is a schematic structural diagram of a one-way open shelf provided in an embodiment of the present invention, such as the one-way open shelf 111 shown in fig. 2, and may also be a two-way open shelf. The shelf 111 may be a movable shelf or a fixed shelf.
In an order-to-person scenario, the control system may send a corresponding order to the transport 120. The staff determines the order that needs to be completed by the transportation device 120 through the display of the transportation device 120 or an intelligent terminal carried by the staff. It can be understood that, if the staff uses the intelligent terminal carried in the hand to view the order or views the order through the display of the transportation device 120, before picking the goods contained in the order, the staff needs to be connected to the transportation device 120, that is, the staff is bound to the corresponding transportation device 120, specifically, the staff may be connected to the transportation device 120 by scanning a code, inputting a serial number, or connecting to a card.
After the shipping equipment 120 receives the order, the shipping equipment 120 needs to track the staff's completion of picking the items in the order. The transportation device 120 needs to determine the staff as a tracking target to track the staff in real time during walking. In tracking the staff, the staff may pick the goods and place the goods on the transportation device 120. In addition, the staff member can interact with the transportation device 120 according to the display of the transportation device 120, for example, to know the specific content of the order and so on.
FIG. 3 is a schematic diagram of an order to person control process provided in an embodiment of the present invention. As shown in fig. 3, the control system 130 is in wireless communication with the transportation device 120, the customer generates an order through the console, the order is transmitted to the control system 130, the control system 130 responds to the order and issues the order to the transportation device 120, and the transportation device 120 performs the task of the order under the control of the control system 130 and with the assistance of the staff. For example, the transportation device 120 receives the order task, binds with a worker in the warehouse, that is, the bound worker is responsible for picking the order task received by the transportation device, and the transportation device 120 may travel along the aisle space in the middle of the rack array, move to the position of the rack 111, receive the item picked by the worker, and track the worker for real-time transportation, and finally transport the picked item to the designated position.
It can be seen that after a worker removes an item from the shelf 111, the worker may be placed directly on the transport 120 and without control of the transport 120, the transport 120 may track the worker until the order assigned by the transport 120 is completed. It can be understood that, similar to the above process, when the goods are restocked, the goods restocking here means that after the number of a certain commodity in the warehouse is lower than a set certain threshold, the commodity needs to be restocked, a certain amount of the commodity is loaded by the transportation device 120, and the commodity is carried to the target storage shelf, and the worker puts the commodity on the transportation device 120 onto the target storage shelf to complete the restocking of the commodity. Compared with the traditional manual goods taking or replenishing mode, the scheme from the order to the person is more convenient to realize, and the goods shelf is not required to be improved while the manpower is saved, so that the goods taking and replenishing method can be suitable for the goods storage and retrieval of various storage warehouses.
The control system 130 may be a software system with data storage and information processing capabilities running on a control server, and may be connected to the transport equipment 120, a hardware input system, or other software systems via wireless or wired connections. The control system 130 may include one or more control servers, which may be a centralized control architecture or a distributed computing architecture. The control server has a processor 1301 and a memory 1302, and may have an order pool 1303 in the memory 1302.
The transport apparatus 120 may have only a tracking function, or both a tracking function and a navigation function. Additionally, the transport device 120 may have other desired and/or necessary functions, such as obstacle avoidance functionality, and the like.
For the tracking function, the logistics robot may include: a main controller; a movable base; an information acquisition device adapted to acquire image information and depth information of an object present within its field of view; the information acquisition device may comprise an RGBD camera adapted to acquire image information and depth information of objects present in its field of view, resulting in an RGBD image.
Alternatively or additionally, the information acquisition device may comprise a monocular camera adapted to acquire image information of objects present within its field of view and a lidar adapted to acquire depth information of objects present within its field of view.
The main controller may acquire image information and depth information of the tracking target from the RGBD camera. For example, in the case that the recognition target is beyond the view range of the RGBD camera, the master controller cannot acquire valid information from the RGBD camera for a period of time, for example, 1 second, to recognize the tracking target, and it is determined that the tracking target is lost. A plurality of RGBD cameras may be provided as necessary. In addition, the number of the monocular cameras and the number of the laser radars may be one or more, as necessary.
It is contemplated that the information acquisition device may acquire visual information, including image information and depth information, of objects present within its field of view in real time; the main controller can acquire the information acquired by the information acquisition device in real time, identify the tracked target in real time, determine the position information of the tracked target in real time and plan the tracking route in real time.
It is contemplated that the master controller may obtain the identity information of the tracking target in a variety of ways, wired or wirelessly, for example, the master controller may obtain the identity information of the tracking target from any device to which the master controller may be coupled or in communication with, such as a server, a mobile device, such as a mobile device of an operator, and the like. The master controller can be connected and communicate with such devices via cables, or via Bluetooth, as desiredTM、WiFi、ZigBeeTMOr other wireless communication protocol to communicate with such devices.
Typically, the tracking targets are personnel authorized to use the transport apparatus 120, such as the operator of the transport apparatus 120. The master controller may be configured to authorize the tracking target based on the acquired identity information of the tracking target. The identity information may be any information that can uniquely identify the identity of the tracking target, such as but not limited to: a username and password of the tracked target, a fingerprint of the tracked target, a facial image of the tracked target, etc.
The identifying characteristics of the tracking target may be associated with identity information of the tracking target. The identifying feature may be any feature worn or carried by the tracked object itself or on its body suitable for identifying the tracked object. In the case where the tracking target is an operator, the identification features may be, for example, but not limited to: the body type, the work number carried on the back of the garment, the two-dimensional code carried on the back of the garment, the color of the garment, and the like.
Fig. 4 is a schematic diagram of a system for tracking a dynamic target provided in an embodiment of the present application, where the embodiment is applicable to a situation where a transportation device tracks the dynamic target, the system may perform the method for tracking a dynamic target provided in an embodiment of the present application, and the system may be implemented by software and/or hardware.
As shown in fig. 4, the tracking system of the dynamic target includes:
an image acquisition module 410 configured to acquire an RGBD image of the dynamic target;
a dynamic target detection module 420, connected to the image acquisition module 410, configured to perform target detection on the RGBD image, so as to obtain an initial position of a target component;
and the component position tracking module 430 is connected to the dynamic target detection module 420, and configured to determine an estimated position of the target in the RGBD image of the target frame according to the target motion information between at least two RGBD images, so as to determine a component tracking result of the dynamic target.
The present solution may be executed by a robot having a transportation function, and specifically, may be a mobile device having a full-automatic driving function, a semi-automatic driving function, or an auxiliary driving function, such as an automobile having an automatic driving mode, an AGV (Automated Guided Vehicle), a robot, and the like. In this case, an AGV is taken as an example for explanation.
In scenarios where AGVs are utilized to perform tasks, multiple AGVs within the same scenario are typically scheduled by a unified management platform to enable the entire scenario to proceed in order. For example, when the AGVs are used to pick goods, all AGVs performing tasks or being ready to perform tasks in the storage area need to be scheduled and controlled. The task to be executed can be automatically imported by the management platform, and can also be input by a worker through an input device. When the management platform receives one or more tasks to be executed, the management platform determines corresponding AGVs to complete each task according to a scheduling algorithm or rule of the management platform, and of course, the management platform can also control the corresponding AGVs to execute new tasks according to the AGVs designated by a worker. This requires the routing of the AGVs that received the task (i.e., the AGVs that are to perform the task). The path planning for the AGV to execute the task may be performed by a path planning method commonly used in the art. Specifically, the selectable path is determined according to the starting point and the destination of the AGV, and then an optimal path is selected as the planned path after weighing is carried out based on time, path factors and the busyness degree of the path.
The image obtaining module 410 may be a depth camera disposed on the AGV, and the camera may obtain not only color information of the image but also depth information of each pixel point in the image. Therefore, an RGBD image of the dynamic target can be acquired by the camera. Wherein it is understood that the dynamic object may be a pedestrian. The dynamic target may be an adult or a child.
The dynamic target detection module 420 may be configured to perform target detection on the acquired RGBD image to obtain an initial position of a component of the target. It is understood that a certain algorithm may be adopted to implement the target detection on the RGBD image, for example, a human form in the image is recognized through a human form recognition technology, and if a plurality of human forms are included in the RGBD image, each human form may be marked for subsequent confirmation whether a dynamic target is included therein.
In this embodiment, optionally, the dynamic target detection module 420 is configured to:
carrying out target detection by adopting a preset detection network model to obtain the initial position of a target component; wherein the components include one or more of a head, a torso, arms, and legs.
According to the scheme, the pedestrian detection is carried out by utilizing the improved CenterNet detection network, and the positions of pedestrians and parts of the pedestrians are obtained. The method is mainly used for obtaining the positions of the pedestrians and the parts thereof in the initial frame and the positions of the pedestrians and the parts thereof in the subsequent new image frame.
The component position tracking module 430 may determine an estimated position of the target in the RGBD image of the target frame according to the target motion information between at least two RGBD images to determine a component tracking result of the dynamic target.
The motion information may be the amount of motion of each part. Specifically, the position variation amount of the pixel point at each position from the previous RGBD image array to the next RGBD image frame may be used. Because each part is a part of a human body, the motion conditions between two adjacent frames of RGBD images are often relatively close, and therefore, the estimated positions of each part in the RGBD image of the target frame can be determined according to the motion information seen by at least two frames of RGBD images.
It is understood that at least two frames of RGBD images may be two consecutive frames or may be non-consecutive, for example, the 2 nd frame RGBD image and the 4 th frame RGBD image in the video stream, and the determined target frame RGBD image may be the 6 th frame RGBD image. Specifically, the calculation of the action amount may be performed by taking two non-consecutive frames if the movement of the staff is relatively slow in the scene, which may be determined according to the need or the actual scene. Thus, the calculation amount can be reduced, and the calculation load can be reduced.
In this embodiment, optionally, the component position tracking module 430 is configured to:
determining target motion information between at least two frames of RGBD images by adopting an optical flow method; and determining the estimated position of the target in the RGBD image of the target frame according to the target motion information so as to determine a component tracking result of the dynamic target.
Optical flow is the motion of an object, scene, or camera as it moves between two consecutive frames of images. The method is a two-dimensional vector field of an image in the process of translation, is a speed field for representing three-dimensional motion of an object point through a two-dimensional image, and reflects image change formed by motion in a tiny time interval so as to determine the motion direction and the motion rate of the image point.
By adopting an optical flow method, the estimated position of the target in the RGBD image of the target frame can be determined more accurately.
It will be appreciated that where the components include a head, torso, arms, and legs, the central position of each component may be determined as the position of the dynamic target.
According to the technical scheme, by means of component detection, the problem that the dynamic target cannot be tracked due to the fact that the body state of the dynamic target changes and the image range moves under the condition that the whole dynamic target range is tracked can be solved. The robustness of tracking the dynamic target is improved.
In a possible embodiment, optionally, the dynamic object detection module is further configured to: carrying out target detection on the RGBD image to obtain the overall position of a target;
the system also includes an overall position tracking module, coupled to the dynamic target detection module, configured to:
and determining at least one overall tracking result matched with the overall position in the RGBD image of the target frame.
Wherein the global position may be determined based on a global tracking target template of the dynamic target. The overall position may be a supplement to the position of the component, or may be used alone for performing dynamic target tracking, and in some cases, if it is difficult to perform dynamic target tracking based on the position of the component, the overall position tracking method provided in the present solution may be adopted.
In this embodiment, specifically, the overall position tracking module is configured to:
determining a target detection result in a target frame RGBD image, and determining a matching confidence of a detected target and a tracking target template of the whole position;
and determining at least one overall tracking result according to the matching confidence.
It can be understood that, in the RGBD image of the target frame, there may be an overall position of the dynamic target, and there may also be an overall position of another dynamic target, so that, in the overall position detection process, the pedestrian detection module in the previous step may be used to obtain a pedestrian detection result, each detection result is compared with the tracking target template through a matching network, a matching confidence is obtained, the first three largest matching results are selected, and the position and the confidence are stored.
On the basis of the above technical solution, optionally, the system further includes a tracking fusion module, connected to the component position tracking module, and connected to the overall position tracking module, and configured to:
and determining a final tracking result of the dynamic target according to the component tracking result and the at least one integral tracking result.
Under the condition of simultaneously adopting the whole position and the component position for detection, the final tracking result of the dynamic target can be determined according to the two detection results. For example, when one of the detections is successful and the other detection is failed, the position obtained by the successful detection may be used as the final tracking result.
Specifically, the tracking fusion module is configured to:
if at least two integral tracking results exist, selecting the integral tracking result with the maximum superposition coefficient with the component tracking result as a final tracking result;
and if an integral tracking result exists, taking the integral tracking result as a final tracking result when the superposition coefficient of the integral tracking result and the part tracking result accords with a set threshold value.
When two or more overall tracking results exist, which one of the overall tracking results is the real tracking target can be determined according to the component tracking results. When only one overall tracking result exists, the overall tracking result can be used as a final tracking result when the superposition coefficient of the overall tracking result and the part tracking result meets a set threshold value. For example, if the coincidence coefficient of the two is 20%, it is highly likely that the overall tracking result and the part tracking result are not the same target. Only when the coincidence coefficient of the two reaches 50% or higher, the final tracking result of the dynamic target can be determined. It is to be understood that if the overall tracking result and the part tracking result are represented by point locations, the coincidence coefficient of the two may be determined by determining the distance between the two point locations.
In another possible embodiment, the tracking fusion module is further configured to:
and if the integral tracking result does not exist, taking the part tracking result as a final tracking result.
If the overall tracking result is not detected, which indicates that the dynamic target is partially or completely out of the image range at the moment, the final tracking result of the dynamic target can be determined only through the component tracking result. If the dynamic tracking result can not be detected, the current frame image can be skipped, the images of the subsequent preset number of frames can be detected, and if the overall position or the part position can not be detected, the lost dynamic tracking target can be determined.
Fig. 5 is a schematic diagram of a tracking method for a dynamic target provided in an embodiment of the present application. As shown in fig. 5, the tracking method of the dynamic target is executed by a management platform, and includes:
and S510, acquiring an RGBD image of the dynamic target.
S520, carrying out target detection on the RGBD image to obtain the initial position of a target component.
S530, according to the target motion information between at least two frames of RGBD images, determining the estimated position of the target in the RGBD image of the target frame so as to determine the part tracking result of the dynamic target.
On the basis of the above technical solution, optionally, the method further includes:
carrying out target detection on the RGBD image to obtain the overall position of a target;
and determining at least one overall tracking result matched with the overall position in the RGBD image of the target frame.
On the basis of the above technical solution, optionally, the method further includes:
and determining a final tracking result of the dynamic target according to the component tracking result and the at least one integral tracking result.
On the basis of the above technical solution, optionally, determining a final tracking result of the dynamic target according to the component tracking result and the at least one overall tracking result, includes:
if at least two integral tracking results exist, selecting the integral tracking result with the maximum superposition coefficient with the component tracking result as a final tracking result;
and if an integral tracking result exists, taking the integral tracking result as a final tracking result when the superposition coefficient of the integral tracking result and the part tracking result accords with a set threshold value.
On the basis of the above technical solution, optionally, determining a final tracking result of the dynamic target according to the component tracking result and the at least one overall tracking result, further comprising:
and if the integral tracking result does not exist, taking the part tracking result as a final tracking result.
On the basis of the foregoing technical solution, optionally, performing target detection on the RGBD image to obtain an initial position of a component of a target, including:
carrying out target detection by adopting a preset detection network model to obtain the initial position of a target component; wherein the components include one or more of a head, a torso, arms, and legs.
On the basis of the above technical solution, optionally, determining an estimated position of a target in the RGBD image of the target frame according to target motion information between at least two RGBD images to determine a component tracking result of the dynamic target, includes:
determining target motion information between at least two frames of RGBD images by adopting an optical flow method; and determining the estimated position of the target in the RGBD image of the target frame according to the target motion information so as to determine a component tracking result of the dynamic target.
On the basis of the foregoing technical solution, optionally, determining at least one overall tracking result matched with the overall position in the RGBD image of the target frame includes:
determining a target detection result in a target frame RGBD image, and determining a matching confidence of a detected target and a tracking target template of the whole position;
and determining at least one overall tracking result according to the matching confidence.
The method can be executed by the system provided by any embodiment of the application, and has the corresponding execution steps and beneficial effects.
Fig. 6 is a schematic diagram of a tracking method for a dynamic target provided in an embodiment of the present application. As shown in fig. 6, the method for tracking a dynamic target includes:
the method utilizes appearance information, spatial information and motion consistency information. The person as a whole divides the body of the person into different parts and combines the whole and part tracking results for tracking. The target and the components thereof are taken as tracking objects, the target and the components thereof are described through depth features, and the consistency of motion among the components is judged by utilizing optical flow. The tracking accuracy and the tracking robustness can be guaranteed.
The pedestrian is divided into different parts, such as: head, torso, arms, legs.
The pedestrian detection system is composed of a pedestrian detection module, a pedestrian integral tracking module, a tracking module based on component motion and a tracking fusion module.
1. A pedestrian detection module to: and carrying out pedestrian detection based on the RGBD image to obtain the positions of pedestrians and parts thereof in the image. The method utilizes the improved CenterNet detection network to detect the pedestrian and obtain the positions of the pedestrian and the parts thereof. The method is mainly used for obtaining the positions of the pedestrians and the parts thereof (tracking target templates) in the initial frame and the positions of the pedestrians and the parts thereof in the subsequent new image frame.
2. The integral pedestrian tracking module is used for: and in a new frame of image, obtaining a pedestrian detection result by using the pedestrian detection module in the previous step, comparing each detection result with the tracking target template through a matching network to obtain a matching confidence coefficient, selecting the first three largest matching results and storing the position and the confidence coefficient. This step may yield a pedestrian location result that is most similar to the target template. The step can process the situation that the pedestrian disappears due to the appearance of the picture or the occlusion.
3. A component motion based tracking module to: and obtaining the position of the target frame part according to the relative motion condition between two continuous frame images. The method is realized by an optical flow method. The optical flow is a movement amount of a pixel point representing the same object (object) in one frame of a video image to a target frame, and is expressed by a two-dimensional vector. From a priori knowledge, it is known that the optical flow behavior of the same part of the human body in the previous and subsequent frames should be the same. Therefore, the position of the previous frame component in the target frame can be obtained through the position of the previous frame component and the optical flow operation information between two frames, and the component tracking result can be obtained. The center of all the parts is the overall tracking result of the pedestrian.
4. A tracking fusion module to: and combining the tracking results of the step 2 and the step 3 to obtain a final pedestrian tracking result:
(1) when the tracking result exists in the step 2, the result closest to the tracking result in the step 3 is selected as the final tracking result. The step can filter the false detection in the step 2, and improve the stability and accuracy of tracking.
(2) And when the matching of the target in the step 2 fails due to the deformation or the shielding of the human body, the result in the step 3 is used as a tracking result, and the robustness of the tracking result is improved.
The invention comprehensively utilizes the appearance information, the space information and the motion consistency information of the target, can be suitable for tracking the pedestrian appearing after disappearance, the pedestrian deforming, shielding and the like under various conditions, and simultaneously considers both the accuracy and the robustness.
It is to be noted that the foregoing is only illustrative of the preferred embodiments of the present application and the technical principles employed. It will be understood by those skilled in the art that the present application is not limited to the particular embodiments described herein, but is capable of various obvious changes, rearrangements and substitutions as will now become apparent to those skilled in the art without departing from the scope of the application. Therefore, although the present application has been described in more detail with reference to the above embodiments, the present application is not limited to the above embodiments, and may include other equivalent embodiments without departing from the spirit of the present application, and the scope of the present application is determined by the scope of the appended claims.

Claims (10)

1. A system for tracking a dynamic target, comprising:
an image acquisition module configured to acquire an RGBD image of the dynamic target;
the dynamic target detection module is connected with the image acquisition module and is configured to perform target detection on the RGBD image to obtain an initial position of a target component;
and the component position tracking module is connected with the dynamic target detection module and is configured to determine the estimated position of the target in the RGBD image of the target frame according to the target motion information between at least two RGBD images so as to determine the component tracking result of the dynamic target.
2. The system of claim 1, wherein the dynamic object detection module is further configured to: carrying out target detection on the RGBD image to obtain the overall position of a target;
the system also includes an overall position tracking module, coupled to the dynamic target detection module, configured to:
and determining at least one overall tracking result matched with the overall position in the RGBD image of the target frame.
3. The system of claim 2, further comprising a tracking fusion module coupled to the component position tracking module and coupled to the global position tracking module, configured to:
and determining a final tracking result of the dynamic target according to the component tracking result and the at least one integral tracking result.
4. The system of claim 3, wherein the trace fusion module is configured to:
if at least two integral tracking results exist, selecting the integral tracking result with the maximum superposition coefficient with the component tracking result as a final tracking result;
and if an integral tracking result exists, taking the integral tracking result as a final tracking result when the superposition coefficient of the integral tracking result and the part tracking result accords with a set threshold value.
5. The system of claim 4, wherein the trace fusion module is further configured to:
and if the integral tracking result does not exist, taking the part tracking result as a final tracking result.
6. The system of claim 1, wherein the dynamic object detection module is configured to:
carrying out target detection by adopting a preset detection network model to obtain the initial position of a target component; wherein the components include one or more of a head, a torso, arms, and legs.
7. The system of claim 1, wherein the component location tracking module is configured to:
determining target motion information between at least two frames of RGBD images by adopting an optical flow method; and determining the estimated position of the target in the RGBD image of the target frame according to the target motion information so as to determine a component tracking result of the dynamic target.
8. The system of claim 2, wherein the global position tracking module is configured to:
determining a target detection result in a target frame RGBD image, and determining a matching confidence of a detected target and a tracking target template of the whole position;
and determining at least one overall tracking result according to the matching confidence.
9. A method for tracking a dynamic target, the method comprising:
obtaining an RGBD image of a dynamic target;
carrying out target detection on the RGBD image to obtain an initial position of a part of a target;
and determining the estimated position of the target in the RGBD image of the target frame according to the target motion information between at least two frames of RGBD images so as to determine the part tracking result of the dynamic target.
10. The method of claim 9, further comprising:
carrying out target detection on the RGBD image to obtain the overall position of a target;
and determining at least one overall tracking result matched with the overall position in the RGBD image of the target frame.
CN202010879875.9A 2020-08-27 2020-08-27 System and method for tracking dynamic target Pending CN112037259A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202010879875.9A CN112037259A (en) 2020-08-27 2020-08-27 System and method for tracking dynamic target

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202010879875.9A CN112037259A (en) 2020-08-27 2020-08-27 System and method for tracking dynamic target

Publications (1)

Publication Number Publication Date
CN112037259A true CN112037259A (en) 2020-12-04

Family

ID=73586692

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202010879875.9A Pending CN112037259A (en) 2020-08-27 2020-08-27 System and method for tracking dynamic target

Country Status (1)

Country Link
CN (1) CN112037259A (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN116883915A (en) * 2023-09-06 2023-10-13 常州星宇车灯股份有限公司 Target detection method and system based on front and rear frame image association

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104835182A (en) * 2015-06-03 2015-08-12 上海建炜信息技术有限公司 Method for realizing dynamic object real-time tracking by using camera
WO2015161697A1 (en) * 2014-04-25 2015-10-29 深港产学研基地 Method and system for tracking moving object applied to man-machine interaction
CN108269269A (en) * 2016-12-30 2018-07-10 纳恩博(北京)科技有限公司 Method for tracking target and device
CN110287855A (en) * 2019-06-21 2019-09-27 上海闪马智能科技有限公司 A kind of pedestrian track acquisition methods and system
CN110516620A (en) * 2019-08-29 2019-11-29 腾讯科技(深圳)有限公司 Method for tracking target, device, storage medium and electronic equipment
WO2020107524A1 (en) * 2018-11-27 2020-06-04 上海芯仑光电科技有限公司 Target tracking method and computing device

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2015161697A1 (en) * 2014-04-25 2015-10-29 深港产学研基地 Method and system for tracking moving object applied to man-machine interaction
CN104835182A (en) * 2015-06-03 2015-08-12 上海建炜信息技术有限公司 Method for realizing dynamic object real-time tracking by using camera
CN108269269A (en) * 2016-12-30 2018-07-10 纳恩博(北京)科技有限公司 Method for tracking target and device
WO2020107524A1 (en) * 2018-11-27 2020-06-04 上海芯仑光电科技有限公司 Target tracking method and computing device
CN110287855A (en) * 2019-06-21 2019-09-27 上海闪马智能科技有限公司 A kind of pedestrian track acquisition methods and system
CN110516620A (en) * 2019-08-29 2019-11-29 腾讯科技(深圳)有限公司 Method for tracking target, device, storage medium and electronic equipment

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
方明等: "《视频图像光流场估计技术及应用》", 31 May 2019, 国防工业出版社, pages: 86 *

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN116883915A (en) * 2023-09-06 2023-10-13 常州星宇车灯股份有限公司 Target detection method and system based on front and rear frame image association
CN116883915B (en) * 2023-09-06 2023-11-21 常州星宇车灯股份有限公司 Target detection method and system based on front and rear frame image association

Similar Documents

Publication Publication Date Title
JP7341652B2 (en) Information processing device, information processing method, program, and system
US10810544B2 (en) Distributed autonomous robot systems and methods
US20170270579A1 (en) Robotic equipment for the location of items in a shop and operating process thereof
US11915192B2 (en) Systems, devices, and methods for scanning a shopping space
CN106347919A (en) Automatic warehousing system
US20170097232A1 (en) Using sensor-based observations of agents in an environment to estimate the pose of an object in the environment and to estimate an uncertainty measure for the pose
US11961303B1 (en) Agent re-verification and resolution using imaging
JP6722348B2 (en) Integrated obstacle detection and payload centering sensor system
CN110945449B (en) Real-time supervision type machine learning system and method for field environment
Novoselov et al. Development of the method local navigation of mobile robot a based on the tags with QR code and wireless sensor network
CN110816522B (en) Vehicle attitude control method, apparatus, and computer-readable storage medium
CN112037259A (en) System and method for tracking dynamic target
KR20190078828A (en) Apparatus for controlling flight of unmanned aerial vehicle and method thereof
CN111975776A (en) Robot movement tracking system and method based on deep learning and Kalman filtering
US20230211987A1 (en) Pathfinding using centerline heuristics for an autonomous mobile robot
US20230161356A1 (en) Method of updating map in fusion slam and robot implementing same
EP4348544A1 (en) Using slam 3d information to optimize training and use of deep neural networks for recognition and tracking of 3d object
US20230039143A1 (en) Own-position estimating device, moving body, own-position estimating method, and own-position estimating program
US20180299901A1 (en) Hybrid Remote Retrieval System
CN113632029A (en) Information processing device, program, and information processing method
RU2658092C2 (en) Method and navigation system of the mobile object using three-dimensional sensors
Aquino et al. Image Processing Based Human Following Cart Using 360° Camera
US20230266770A1 (en) Movable platform for taking inventory and/or performing other actions on objects
WO2022193762A1 (en) Method and device for controlling transfer robot
Mronga et al. MARLIN: A cloud integrated robotic solution to support intralogistics in retail

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
CB02 Change of applicant information
CB02 Change of applicant information

Address after: 100020 Room 101, 1st floor, building 1, No.36 courtyard, Chuangyuan Road, Chaoyang District, Beijing

Applicant after: Beijing jizhijia Technology Co.,Ltd.

Address before: 100020 Room 101, 1st floor, building 1, No.36 courtyard, Chuangyuan Road, Chaoyang District, Beijing

Applicant before: Beijing Geekplus Technology Co.,Ltd.