WO2020258889A1 - Tracking method of video tracking device, and video tracking device - Google Patents

Tracking method of video tracking device, and video tracking device Download PDF

Info

Publication number
WO2020258889A1
WO2020258889A1 PCT/CN2020/075481 CN2020075481W WO2020258889A1 WO 2020258889 A1 WO2020258889 A1 WO 2020258889A1 CN 2020075481 W CN2020075481 W CN 2020075481W WO 2020258889 A1 WO2020258889 A1 WO 2020258889A1
Authority
WO
WIPO (PCT)
Prior art keywords
tracking
video image
target
video
tracking algorithm
Prior art date
Application number
PCT/CN2020/075481
Other languages
French (fr)
Chinese (zh)
Inventor
高宗伟
Original Assignee
杭州海康微影传感科技有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 杭州海康微影传感科技有限公司 filed Critical 杭州海康微影传感科技有限公司
Publication of WO2020258889A1 publication Critical patent/WO2020258889A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/20Analysis of motion
    • G06T7/246Analysis of motion using feature-based methods, e.g. the tracking of corners or segments
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/70Determining position or orientation of objects or cameras
    • G06T7/73Determining position or orientation of objects or cameras using feature-based methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10016Video; Image sequence

Definitions

  • This application relates to the field of security technology, and in particular to a tracking method of a video tracking device and a video tracking device.
  • Video tracking devices such as existing video trackers usually use a single tracking algorithm for target tracking.
  • the main tracking algorithms are contrast tracking algorithm, correlation tracking algorithm and binary tracking algorithm.
  • the related tracking algorithm can track multiple types of targets. When the tracked target has no boundaries, the motion is not very strong, and the scene is more complicated, the tracking effect of the related tracking algorithm is good.
  • the contrast tracking algorithm can track fast-moving targets and is highly adaptable to changes in target posture.
  • the binary tracking algorithm can automatically detect the target, the tracking gate is adaptive to the target size, the closed loop speed is fast, the tracking is stable, and it is suitable for tracking the air target.
  • each tracking algorithm has its focus on application scenarios, when a video tracking device uses a single tracking algorithm for target tracking, the tracking effect of the video tracking device will be different when used in different scenarios.
  • the purpose of this application is to provide a tracking method for a video tracking device and a video tracking device, which can adaptively adjust the tracking algorithm according to scene changes, thereby ensuring the tracking effect.
  • an embodiment of the present application provides a tracking method for a video tracking device, the video tracking device supports multiple tracking algorithms; the method includes:
  • the shooting angle of the imaging device is adjusted according to the distance, so that the tracking target is at the center of the next video image captured by the imaging device.
  • an embodiment of the present application provides a video tracking device that supports multiple tracking algorithms;
  • the video tracking device includes a non-transitory computer-readable storage medium and a processor, wherein:
  • the non-transitory computer-readable storage medium is used to store instructions that can be executed by the processor, and when the instructions are executed by the processor, the processor is caused to:
  • the shooting angle of the imaging device is adjusted according to the distance, so that the tracking target is at the center of the next video image captured by the imaging device.
  • the tracking algorithm suitable for the frame of video image is adaptively selected based on the parameters that characterize the scene complexity of the frame of video image , Based on the selected tracking algorithm to track the tracking target in the frame of the video image, and adjust the shooting angle of the imaging device according to the distance between the tracking target in the video image and the center of the video image, so that the tracking target is in the next video shot by the imaging device The center position of the image. Since the tracking algorithm can be switched at any time according to the parameters that characterize the scene complexity of each frame of video image, it can well adapt to the changes in the shooting scene of the imaging device, thereby ensuring the video tracking effect.
  • FIG. 1 is a schematic diagram of the architecture of a video tracking system provided by an embodiment of the present application
  • FIG. 2 is a flowchart of a tracking method of a video tracking device provided by an embodiment of the present application
  • Fig. 3 is a schematic structural diagram of a video tracking device provided by an embodiment of the present application.
  • the tracking algorithm is adaptively adjusted according to the different complexity of each frame of video image captured by the imaging device, so as to avoid the difference in tracking effect caused by the change of the imaging device shooting scene.
  • Fig. 1 is a schematic diagram of the architecture of a video tracking system provided by an embodiment of the present application.
  • the video tracking system includes: imaging equipment, video tracking equipment, display terminals, and follow-up control equipment. The following are introduced separately.
  • the imaging device is integrated with an imaging movement, which is used to shoot videos, and the imaging device sends each frame of video image taken by the imaging movement to the video tracking device.
  • the imaging device will not only send each frame of video image captured by the imaging core to the video tracking device, but also characterize each frame of video image.
  • the parameters of the scene complexity are sent to the video tracking device.
  • the parameters that characterize the scene complexity of each frame of video image may be represented by the definition evaluation parameter of each frame of video image.
  • the imaging core integrated with the imaging device may include a thermal imaging core and a visible light imaging core. Both the thermal imaging core and the visible light imaging core are used to capture video images, the former captures the thermal imaging video image, and the latter captures the visible light video image.
  • the imaging device can send both the thermal imaging video image captured by the thermal imaging core and the visible light video image captured by the visible light imaging core to the video tracking device.
  • the imaging device needs to send the thermal imaging video image to the video tracking device and also need to send the parameters that characterize the scene complexity of the thermal imaging video image to the video. Tracking equipment.
  • the video tracking device performs target tracking on the visible light video image
  • the imaging device needs to send the visible light video image to the video tracking device and also need to send the parameters representing the scene complexity of the visible light video image to the video tracking device.
  • Video tracking equipment can be implemented using a Digital Signal Processing (DSP) chip and supports multiple tracking algorithms.
  • DSP Digital Signal Processing
  • the video tracking device For each frame of video image sent by the imaging device, the video tracking device selects an appropriate tracking algorithm for target tracking based on the parameters that characterize the scene complexity of each frame of video image, and obtains the position information of the tracking target in each frame of video image, according to the above The position information obtains the tracking result, and then adjusts the shooting angle of the imaging device according to the tracking result, so that the tracking target is at the center position of the video image subsequently captured by the imaging device.
  • the video tracking device can send the tracking result to the follow-up control device, and the follow-up control device adjusts the shooting angle of the imaging device.
  • the tracking result may be: the distance between the tracking target and the center of the video image in the video image calculated based on the position information.
  • the video tracking device performs target tracking on the thermal imaging video image sent by the imaging device, and the visible light video image sent by the imaging device is directly sent to the display device for display.
  • the imaging device and the video tracking device need to maintain timing synchronization.
  • the video tracking device can use the synchronization signal of the thermal imaging core or the visible light imaging core in the imaging device as the synchronization source, and use the aforementioned synchronization source as the basic working sequence of the video tracking device, so as to ensure that it is connected to the thermal imaging core or The visible light imaging movement is strictly synchronized.
  • the video tracking equipment generates the working timing of each unit circuit on this basis, forming a unified timing synchronization system.
  • the follow-up control device executes the follow-up control operation according to the tracking result sent by the video tracking device. For example, the shooting angle of the imaging device is directly adjusted, or the shooting angle of the imaging device is adjusted by moving and/or rotating the follow-up control device itself, so as to ensure that the tracking target is located in the center of the video image captured by the imaging device.
  • Figure 2 is an example of target tracking of thermal imaging video images captured by the thermal imaging core in the imaging device.
  • the video tracking device can also capture the visible light captured by the visible light imaging core in the imaging device. Video image for target tracking.
  • the solution provided by the embodiment of the present application is applied to a video tracking device for tracking a tracking target in a video image.
  • FIG. 2 is a flowchart of a tracking method of a video tracking device provided in an embodiment of the present application. As shown in FIG. 2, the method includes the following steps 201 to 205.
  • Step 201 Obtain a video image and a parameter that characterizes the scene complexity of the video image from the imaging device.
  • the video tracking device can also obtain video images from the imaging device frame by frame.
  • the target tracking method is the same, and all of them can be implemented through step 201 to step 205 provided in the embodiment of the present application.
  • the definition evaluation parameter of the video image may be used as a parameter that characterizes the scene complexity of the video image.
  • the definition evaluation parameter of the video image can be expressed by using an active image format descriptor (Active Format Description, AFD).
  • AFD Active Format Description
  • the size of the AFD value can represent the complexity of the scene corresponding to the video image to a certain extent. Specifically, the larger the AFD value, the higher the scene complexity of the video image. Conversely, the smaller the AFD value, the lower the scene complexity of the video image. According to the AFD value of the video image, the complexity of the scene can be determined.
  • the value range of the AFD value of the first frame of video image is [5000, 10000] and the value range of the AFD value of the second frame of video image is [500, 1000], then
  • the scene complexity of one frame of video image is higher than the scene complexity of the second frame of video image.
  • the AFD value of a frame of video image can be obtained through the thermal imaging core.
  • the processor of the thermal imaging core performs high-frequency filtering processing on the parameters that characterize the clarity of a frame of video image to obtain a set of data. Therefore, each data in the group of data can be used to characterize the clarity of the video image. The larger the data in the group of data, the higher the clarity of the video image. On the contrary, the smaller the data in the group of data, the higher the clarity of the video image. The lower the level of clarity.
  • Each data in this set of data can be referred to as the aforementioned AFD value.
  • the processor of the thermal imaging core may be a Field Programmable Gate Array (FPGA).
  • FPGA Field Programmable Gate Array
  • the imaging device may integrate a thermal imaging core, a visible light core, or other imaging cores capable of performing video shooting, and use the integrated imaging core to capture video images and send each frame of the captured video images to Video tracking equipment.
  • the imaging device also analyzes each frame of video image captured by the imaging core to determine its definition evaluation parameters, and sends the definition evaluation parameters of each frame of video image as a parameter that characterizes the scene complexity of each frame of video image to the video. Tracking equipment.
  • only one imaging core is integrated in the imaging device. After the imaging device sends each frame of video image captured by the imaging core to the video tracking device, the video tracking device needs to track each frame of video image on the one hand. On the other hand, it can also send each frame of video image to the display terminal for display.
  • multiple imaging cores can be integrated in the imaging device, the imaging device sends each frame of video image captured by one of the imaging cores to the video tracking device, and the video tracking device performs target tracking on each frame of video image;
  • each frame of video images captured by other imaging cores can be sent to a video tracking device, and the video tracking device performs other video processing on each frame of video images captured by other imaging cores, for example, sending them to a display terminal for display.
  • a thermal imaging core and a visible light imaging core are integrated in the imaging device.
  • the imaging device integrating the visible light imaging core and the thermal imaging core performs video shooting of the current scene, where the thermal imaging core in the imaging device captures the thermal imaging video image of the current scene, and the visible light imaging core in the imaging device captures Obtain the visible light video image of the current scene.
  • the imaging device also analyzes the thermal imaging video image of the current scene to determine its sharpness evaluation parameters.
  • the imaging device sends each frame of thermal imaging video image captured by the thermal imaging core and each frame of visible light video image captured by the visible light imaging core to the video tracking device, and also uses the definition evaluation parameters of each frame of thermal imaging video image as a characterization The parameters of the scene complexity of each thermal imaging video image are sent to the video tracking device.
  • Step 202 Select a tracking algorithm suitable for the video image from among multiple tracking algorithms based on the parameter that characterizes the scene complexity of the video image.
  • the video tracking device can support multiple tracking algorithms such as contrast tracking algorithms, related tracking algorithms, and binary tracking algorithms.
  • the parameter value range corresponding to each tracking algorithm can be preset.
  • the scene complexity of the applicable shooting scene is consistent with the scene complexity represented by the parameter value range corresponding to each tracking algorithm.
  • the definition evaluation parameter of the video image can be used as a parameter that characterizes the scene complexity of the video image.
  • the preset value range of the parameter corresponding to each tracking algorithm is the definition corresponding to each tracking algorithm Evaluation parameter value range.
  • the AFD value output by the imaging device can be divided into multiple levels according to the order of magnitude:...n-1th level, nth level, n+1th level..., where n represents the serial number of the level.
  • n represents the serial number of the level.
  • the AFD value is near the n-1th level, that is, when the AFD value is near 40,000, for example, 40,000 ⁇ 15,000, it means that the current scene is a simple scene, such as a sky scene, a sea scene, and the video tracking device selects binary tracking
  • the algorithm is more suitable for target tracking.
  • the video tracking device chooses the contrast tracking algorithm for target tracking; when the AFD value is in the nth
  • the +1 level that is, when the AFD value is near 100000, for example, 100000 ⁇ 15000, it indicates that the current scene is very complicated, and it is more appropriate for the video tracking device to select the relative tracking algorithm for target tracking.
  • the following parameter value ranges can be set:..., (25000, 55000), (55000, 85000), (85000, 115000)..., among them, the parameter value range corresponding to the binary tracking algorithm Yes (25000, 55000), the parameter value range corresponding to the contrast tracking algorithm is (55000, 85000), and the parameter value range corresponding to the correlation tracking algorithm is (85000, 115000).
  • the video tracking device determines the binary tracking algorithm as a tracking algorithm suitable for the frame of thermal imaging video image, and uses the binary tracking algorithm to track the tracking target in the frame of thermal imaging video image.
  • the video tracking device can evaluate the definition of the frame of video image after obtaining a frame of video image and the definition evaluation parameters of the frame of video image from the imaging device The parameter is compared with the parameter value range corresponding to each tracking algorithm, and the parameter value range corresponding to the definition evaluation parameter of the frame of video image is found, and the tracking algorithm corresponding to the parameter value range is determined as suitable for the frame Video image tracking algorithm.
  • the parameter value range corresponding to the definition evaluation parameter of the video image may be the parameter value range to which the definition evaluation parameter of the video image belongs.
  • this step 202 a specific method for selecting a tracking algorithm suitable for a video image from among multiple tracking algorithms based on a parameter that characterizes the scene complexity of the video image is shown in the following S11-S12.
  • the parameter value ranges corresponding to various tracking algorithms are: the parameter value range corresponding to algorithm 1 (25000, 55000), the parameter value range corresponding to algorithm 2 ( 55000, 85000), the parameter value range corresponding to Algorithm 3 (85000, 115000).
  • the value range of the parameter to which 44000 belongs is: (25000, 55000).
  • the parameter value range of 44000 is: (25000, 55000)
  • the tracking algorithm corresponding to the parameter value range of (25000, 55000) is Algorithm 1, so it is suitable for video images
  • the tracking algorithm is Algorithm 1.
  • Step 203 Determine the tracking target in the video image according to the tracking algorithm suitable for the video image.
  • a frame of video image includes more than one type of object.
  • the contrast tracking algorithm when using the contrast tracking algorithm to track a frame of video image, it is necessary to extract the tracking information for the contrast tracking algorithm corresponding to each detected target from the frame of the video image, for example, edge information, contour The length, area, center of gravity, and/or centroid, etc., and then compare the tracking information of all detected targets with the tracking information of the detected target in the previous frame of video image to find out the tracking information and the detected target in the previous frame of video image The detection target with the greatest matching degree of tracking information is determined as the tracking target in the frame of video image.
  • the frame of video image before extracting the tracking information for the contrast tracking algorithm corresponding to each detection target from a frame of video image, the frame of video image can also be preprocessed, and the specific method is shown as X1-X3.
  • Gaussian filtering can be used to achieve noise removal processing, and Gaussian filtering of the video image can remove scattered noise in the video image.
  • Edge detection is mainly to identify pixels with obvious brightness changes in video images.
  • the algorithm for edge detection can include two types of edge detection algorithms based on search and zero-crossing.
  • X3. Determine the segmentation threshold range of the edge-detected video image, and perform binarization processing on the edge-detected video image according to the segmentation threshold range.
  • the determination of the segmentation threshold range is for the subsequent binarization of the edge-detected video image.
  • the segmentation threshold range can be determined according to the gray limit value of the video image and the maximum video signal amplitude.
  • the gray limit value can be the maximum gray value or the minimum gray value.
  • P is The gray limit value of the video image;
  • V P is the maximum video signal amplitude of the video image, for example, 700 mV;
  • is the preset contrast parameter value, and the value range can be 5% to 15% of the value of V P.
  • the gray values of pixels with gray values lower than T V_min can be uniformly set to 0, and the gray values of pixels with gray values greater than T V_max can be uniform. Set to 255, and the gray value of the pixel with gray value between T V_min and T V_max can remain unchanged.
  • the tracking information for the contrast tracking algorithm corresponding to each detection target can be extracted from the preprocessed video image, so that according to the extracted tracking information of all detection targets, from the video image
  • the tracking target is filtered out of all detection targets.
  • the tracking target in the video image is determined according to the tracking algorithm suitable for the video image, which specifically includes:
  • the tracking target is selected from all the detected targets.
  • the contrast tracking algorithm when used to track the target, it is necessary to use the tracking information of the detected target in the previous frame of video image, but for the first frame of video image to start tracking the target, there is no previous frame.
  • the tracking target can be determined in the first frame of video image by manual designation.
  • the binary tracking algorithm when using the binary tracking algorithm to track a frame of video image, it is necessary to extract the tracking information for the binary tracking algorithm corresponding to each detection target from the video image, for example, the aforementioned tracking information It can include edge information, contour length, area, center of gravity, and/or centroid, and then compare the tracking information of all detected targets with the tracking information of the detected target in the previous frame of video image to find out the tracking information and the previous frame of video
  • the detection target in the image with the greatest matching degree of the tracking information of the detection target is determined as the tracking target in the frame of the video image.
  • the frame of video image before extracting the tracking information for the contrast tracking algorithm corresponding to each detection target from a frame of video image, the frame of video image can also be preprocessed, and the preprocessing method is the same as that in the contrast tracking algorithm . After the preprocessing, it is also necessary to perform area filling for each detection target in the frame of the video image after the preprocessing. The area filling can more highlight each detection target in the video image.
  • the tracking information for the binary tracking algorithm corresponding to each detection target can be extracted from the preprocessed and area filled video image, so as to correspond to all the extracted detection targets
  • the tracking information is selected from all the detected targets in the frame of video image.
  • the tracking target in the video image is determined according to the tracking algorithm suitable for the video image, which specifically includes the following S31 and S32.
  • the tracking target is selected from all the detection targets.
  • the previous frame of video image may be manually designated to determine the tracking target in the first frame of video image.
  • the video image can also be preprocessed, and the preprocessing method is the same as the preprocessing method in the contrast tracking algorithm.
  • the pre-selected template image containing the tracking target can be preset or extracted from previous video images.
  • the image containing the tracking target is extracted from the video image where the tracking target appears for the first time as the template image , Or extract the image containing the tracking target from the previous frame of video image as a template image.
  • the tracking target in the video image is determined according to the tracking algorithm suitable for the video image, which specifically includes the following S41.
  • a detection target with the greatest degree of matching with the template image is selected from all detection targets in the video image, and the screened detection target is determined as the tracking target of the video image.
  • Step 204 Extract the position information of the tracking target in the video image, and calculate the distance between the tracking target and the center of the video image based on the position information.
  • the location information of the tracking target mainly includes information such as the height, width, coordinates of the tracking target.
  • the height and width of the tracking target can be determined according to the projection of the video image on the x-axis and y-axis in the three-dimensional coordinate system. For example, the projection on the x-axis falls into the interval [x1, x2], on the y-axis If the projection falls within the interval [y1, y2], it can be determined that the width of the detection target is x2-x1, the height is y2-y1, and the center point coordinates of the tracking target are: ((x1+x2)/2, (y1+y2 )/2).
  • the distance between the tracking target and the center of the video image can be obtained by calculating the distance between the coordinates.
  • the center of gravity and/or centroid of the tracking target can also be used as the location information of the tracking target.
  • the location of a specific point on the tracking target can also be used as the location information of the tracking target. Such as tracking a corner point or protruding end point on the edge of the target.
  • Step 205 Adjust the shooting angle of the imaging device according to the above distance, so that the tracking target is at the center of the next video image shot by the imaging device.
  • the specific implementation of adjusting the shooting angle of the imaging device according to the above distance is: sending the distance between the tracking target and the center of the video image to the follow-up control device equipped with the video tracking device, so that the follow-up control device is based on the foregoing The distance performs the follow-up control operation to adjust the shooting angle of the imaging device.
  • the position between the imaging device and the video tracking device is very close or directly integrated, and both are installed on the follow-up control device and move with the movement of the follow-up control device.
  • the video tracking device After the video tracking device determines the distance between the tracking target in the video image of the current scene and the center of the video image, it can send this distance to the follow-up control device, and the follow-up control device can drive the imaging device to move by controlling its own movement, or directly The imaging device is controlled to rotate or move so that the tracking target is located at the center of the next frame of video image captured by the imaging device.
  • the tracking algorithm suitable for the frame of video image is adaptively selected based on the parameter that characterizes the scene complexity of the frame of video image.
  • the tracking algorithm tracks the tracking target in this frame of video image, and adjusts the shooting angle of the imaging device according to the distance between the tracking target in the video image and the center of the video image, so that the tracking target is at the center of the next video image captured by the imaging device position. Since the tracking algorithm can be switched at any time according to the parameters that characterize the scene complexity of each frame of video image, it can well adapt to the changes in the shooting scene of the imaging device, thereby ensuring the video tracking effect.
  • the tracking method of the video tracking device provided by the embodiment of the application is described in detail above, and the embodiment of the application also provides a video tracking device, which is described in detail below with reference to FIG. 3.
  • FIG. 3 is a schematic structural diagram of a video tracking device provided by an embodiment of the present application.
  • the video tracking device 300 includes a processor 301 and a non-transitory computer-readable storage medium 302, wherein,
  • the non-transitory computer-readable storage medium 302 is configured to store instructions that can be executed by the processor 301, and when the instructions are executed by the processor 301, the processor 301 is caused to:
  • the shooting angle of the imaging device is adjusted according to the distance, so that the tracking target is at the center of the next video image captured by the imaging device.
  • an imaging core is integrated in the imaging device
  • the processor 301 when acquiring a video image and a parameter characterizing the scene complexity of the video image from an imaging device, includes:
  • the method when the processor 301 selects a tracking algorithm suitable for the video image from among the multiple tracking algorithms based on the parameter, the method includes:
  • the tracking algorithm corresponding to the parameter value range to which the parameter belongs is determined as the tracking algorithm suitable for the video image.
  • the multiple tracking algorithms include: a contrast tracking algorithm, a binary tracking algorithm, and a related tracking algorithm;
  • the processor 301 determines the tracking target in the video image according to the tracking algorithm suitable for the video image, it includes:
  • the method includes:
  • the detection target with the greatest degree of matching with the template image is selected from all the detection targets of the video image, and the screened detection target is determined as the video image. Track the target.
  • the processor 301 when the processor 301 adjusts the shooting angle of the imaging device according to the distance, it includes: sending the distance to a follow-up control equipped with the video tracking device Device, so that the follow-up control device performs a follow-up control operation according to the distance, thereby adjusting the shooting angle of the imaging device.
  • the video tracking device when the video tracking device provided by the foregoing embodiments performs target tracking, for each frame of video image captured by the imaging device, adaptive selection is applied to the frame of video image based on the parameter that characterizes the scene complexity of the frame of video image.
  • Tracking algorithm based on the selected tracking algorithm to track the tracking target in the frame of video image, and adjust the shooting angle of the imaging device according to the distance between the tracking target in the video image and the center of the video image, so that the tracking target is under the shooting of the imaging device The center position of a video image. Since the video tracking device can switch the tracking algorithm at any time according to the parameters characterizing the scene complexity of each frame of video image, it can well adapt to changes in the shooting scene of the imaging device, thereby ensuring the video tracking effect.
  • an embodiment of the present application also provides a tracking device of the video tracking device.
  • the following describes the tracking device of the video tracking device provided in the embodiment of the present application.
  • the aforementioned video tracking device supports multiple tracking algorithms; the tracking device of the aforementioned video tracking device includes:
  • An information acquisition module for acquiring a video image from an imaging device and parameters that characterize the scene complexity of the video image
  • An algorithm selection module configured to select a tracking algorithm suitable for the video image among the multiple tracking algorithms based on the parameter
  • a target determination module configured to determine a tracking target in the video image according to a tracking algorithm applicable to the video image
  • a distance calculation module configured to extract position information of the tracking target in the video image, and calculate the distance between the tracking target and the center of the video image based on the position information
  • the angle adjustment module is configured to adjust the shooting angle of the imaging device according to the distance, so that the tracking target is at the center of the next video image shot by the imaging device.
  • an imaging core is integrated in the imaging device
  • the information acquisition module is specifically used for:
  • the algorithm selection module is specifically used for:
  • the tracking algorithm corresponding to the parameter value range to which the parameter belongs is determined as the tracking algorithm suitable for the video image.
  • the multiple tracking algorithms include: a contrast tracking algorithm, a binary tracking algorithm, and a related tracking algorithm;
  • the target determination module is specifically used for:
  • the target determination module is specifically used for:
  • the detection target with the greatest degree of matching with the template image is selected from all detection targets in the video image, and the selected detection target is determined as the video image Tracking target.
  • the angle adjustment module is specifically used for:
  • the distance is sent to a follow-up control device equipped with the video tracking device, so that the follow-up control device performs a follow-up control operation according to the distance, thereby adjusting the shooting angle of the imaging device.
  • the tracking algorithm suitable for the frame of video image is adaptively selected based on the parameter that characterizes the scene complexity of the frame of video image.
  • the tracking algorithm tracks the tracking target in this frame of video image, and adjusts the shooting angle of the imaging device according to the distance between the tracking target in the video image and the center of the video image, so that the tracking target is at the center of the next video image captured by the imaging device position. Since the tracking algorithm can be switched at any time according to the parameters that characterize the scene complexity of each frame of video image, it can well adapt to the changes in the shooting scene of the imaging device, thereby ensuring the video tracking effect.
  • an embodiment of the present application also provides a computer-readable storage medium in which a computer program is stored, and the computer program is executed by a processor to realize this The steps of the tracking method of the video tracking device described in the application embodiment.
  • an embodiment of the present application also provides a computer program product containing instructions that, when it runs on a computer, causes the computer to perform the tracking of the video tracking device described in the embodiment of the application. method.
  • the computer program product includes one or more computer instructions.
  • the computer may be a general-purpose computer, a special-purpose computer, a computer network, or other programmable devices.
  • the computer instructions may be stored in a computer-readable storage medium or transmitted from one computer-readable storage medium to another computer-readable storage medium. For example, the computer instructions may be transmitted from a website, computer, server, or data center.
  • the computer-readable storage medium may be any available medium that can be accessed by a computer or a data storage device such as a server or data center integrated with one or more available media.
  • the usable medium may be a magnetic medium (for example, a floppy disk, a hard disk, and a magnetic tape), an optical medium (for example, a DVD), or a semiconductor medium (for example, a solid state disk (SSD)).

Landscapes

  • Engineering & Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Multimedia (AREA)
  • Studio Devices (AREA)
  • Image Analysis (AREA)

Abstract

Embodiments of the present application provide a tracking method of a video tracking device, and a video tracking device. The video tracking device supports multiple tracking algorithms. The method comprises: obtaining a video image and a parameter representing the scene complexity of the video image from an imaging device; selecting a tracking algorithm suitable for the video image from the multiple tracking algorithms on the basis of the parameter; determining a tracking target in the video image according to the tracking algorithm suitable for the video image; extracting position information of the tracking target in the video image, and calculating the distance from the tracking target to the center of the video image on the basis of the position information; and adjusting a capturing angle of the imaging device according to the distance, so that the tracking target is located at the center of the next video image captured by the imaging device. By applying the solution provided in the embodiments of the present application for target tracking, adaptive tracking algorithm adjustment can be achieved depending on scene change, thereby ensuring a tracking effect.

Description

一种视频跟踪设备的跟踪方法和视频跟踪设备Tracking method of video tracking equipment and video tracking equipment
本申请要求于2019年6月25日提交中国专利局、申请号为201910555419.6发明名称为“一种视频跟踪设备的跟踪方法和视频跟踪设备”的中国专利申请的优先权,其全部内容通过引用结合在本申请中。This application claims the priority of a Chinese patent application filed with the Chinese Patent Office on June 25, 2019 with the application number 201910555419.6 and the invention titled "A tracking method and video tracking device for a video tracking device", the entire content of which is incorporated by reference In this application.
技术领域Technical field
本申请涉及安防技术领域,特别涉及一种视频跟踪设备的跟踪方法和视频跟踪设备。This application relates to the field of security technology, and in particular to a tracking method of a video tracking device and a video tracking device.
背景技术Background technique
现有视频跟踪器等视频跟踪设备通常采用单一跟踪算法进行目标跟踪。主要的跟踪算法有对比度跟踪算法、相关跟踪算法和二值跟踪算法。其中,相关跟踪算法,可以跟踪多种类型的目标,当被跟踪目标无边界且运动不是很强烈、场景较为复杂的情况下,相关跟踪算法的跟踪效果不错。对比度跟踪算法,可以跟踪快速运动的目标,对目标姿态变化的适应性强。二值跟踪算法,可以自动检测目标,跟踪波门自适应目标大小,闭环速度快、跟踪稳定,适应于对空中目标的跟踪。Video tracking devices such as existing video trackers usually use a single tracking algorithm for target tracking. The main tracking algorithms are contrast tracking algorithm, correlation tracking algorithm and binary tracking algorithm. Among them, the related tracking algorithm can track multiple types of targets. When the tracked target has no boundaries, the motion is not very strong, and the scene is more complicated, the tracking effect of the related tracking algorithm is good. The contrast tracking algorithm can track fast-moving targets and is highly adaptable to changes in target posture. The binary tracking algorithm can automatically detect the target, the tracking gate is adaptive to the target size, the closed loop speed is fast, the tracking is stable, and it is suitable for tracking the air target.
由于每种跟踪算法都有其侧重的应用场景,当视频跟踪设备采用单一跟踪算法进行目标跟踪时,会造成视频跟踪设备在不同场景使用时跟踪效果出现差异。Since each tracking algorithm has its focus on application scenarios, when a video tracking device uses a single tracking algorithm for target tracking, the tracking effect of the video tracking device will be different when used in different scenarios.
发明内容Summary of the invention
有鉴于此,本申请的目的在于提供一种视频跟踪设备的跟踪方法和视频跟踪设备,以能够根据场景变化自适应调整跟踪算法,从而保证跟踪效果。In view of this, the purpose of this application is to provide a tracking method for a video tracking device and a video tracking device, which can adaptively adjust the tracking algorithm according to scene changes, thereby ensuring the tracking effect.
为了达到上述目的,本申请实施例提供了如下技术方案:In order to achieve the foregoing objectives, the embodiments of the present application provide the following technical solutions:
第一方面,本申请实施例提供了一种视频跟踪设备的跟踪方法,所述视频跟踪设备支持多种跟踪算法;所述方法包括:In the first aspect, an embodiment of the present application provides a tracking method for a video tracking device, the video tracking device supports multiple tracking algorithms; the method includes:
从成像设备获取视频图像及表征所述视频图像的场景复杂度的参数;Acquiring a video image from an imaging device and a parameter that characterizes the scene complexity of the video image;
基于所述参数在所述多种跟踪算法中选择适用于所述视频图像的跟踪算法;Selecting a tracking algorithm suitable for the video image from the multiple tracking algorithms based on the parameter;
根据适用于所述视频图像的跟踪算法确定所述视频图像中的跟踪目标;Determining a tracking target in the video image according to a tracking algorithm applicable to the video image;
提取所述跟踪目标在所述视频图像中的位置信息,基于所述位置信息计算所述跟踪目标与所述视频图像的中心的距离;Extracting position information of the tracking target in the video image, and calculating the distance between the tracking target and the center of the video image based on the position information;
根据所述距离调整所述成像设备的拍摄角度,以使所述跟踪目标处于所述成像设备拍摄的下一视频图像的中心。The shooting angle of the imaging device is adjusted according to the distance, so that the tracking target is at the center of the next video image captured by the imaging device.
第二方面,本申请实施例提供了一种视频跟踪设备,所述视频跟踪设备支持多种跟踪算法;所述视频跟踪设备包括非瞬时性计算机可读存储介质和处理器,其中,In a second aspect, an embodiment of the present application provides a video tracking device that supports multiple tracking algorithms; the video tracking device includes a non-transitory computer-readable storage medium and a processor, wherein:
所述非瞬时性计算机可读存储介质,用于存储可以被所述处理器执行的指令,在所述指令由所述处理器执行时,使得所述处理器:The non-transitory computer-readable storage medium is used to store instructions that can be executed by the processor, and when the instructions are executed by the processor, the processor is caused to:
从成像设备获取视频图像及表征所述视频图像的场景复杂度的参数;Acquiring a video image from an imaging device and a parameter that characterizes the scene complexity of the video image;
基于所述参数在所述多种跟踪算法中选择适用于所述视频图像的跟踪算法;Selecting a tracking algorithm suitable for the video image from the multiple tracking algorithms based on the parameter;
根据适用于所述视频图像的跟踪算法确定所述视频图像中的跟踪目标;Determining a tracking target in the video image according to a tracking algorithm applicable to the video image;
提取所述跟踪目标在所述视频图像中的位置信息,基于所述位置信息计算所述跟踪目标与所述视频图像的中心的距离;Extracting position information of the tracking target in the video image, and calculating the distance between the tracking target and the center of the video image based on the position information;
根据所述距离调整所述成像设备的拍摄角度,以使所述跟踪目标处于所述成像设备拍摄的下一视频图像的中心。The shooting angle of the imaging device is adjusted according to the distance, so that the tracking target is at the center of the next video image captured by the imaging device.
由上面的技术方案可知,本申请实施例提供的方案中,对于成像设备拍摄的每帧视频图像,基于表征该帧视频图像的场景复杂度的参数自适应选择适用于该帧视频图像的跟踪算法,基于选择的跟踪算法对该帧视频图像中的跟踪目标进行跟踪,并根据视频图像中的跟踪目标与视频图像中心的距离调整成像设备的拍摄角度,使得跟踪目标处于成像设备拍摄的下一视频图像的中心位置。由于可以随时根据表征每帧视频图像的场景复杂度的参数切换跟踪算法,因此可以很好的适应成像设备拍摄场景的变化,从而保证了视频跟踪效果。It can be seen from the above technical solutions that, in the solution provided by the embodiments of the present application, for each frame of video image taken by the imaging device, the tracking algorithm suitable for the frame of video image is adaptively selected based on the parameters that characterize the scene complexity of the frame of video image , Based on the selected tracking algorithm to track the tracking target in the frame of the video image, and adjust the shooting angle of the imaging device according to the distance between the tracking target in the video image and the center of the video image, so that the tracking target is in the next video shot by the imaging device The center position of the image. Since the tracking algorithm can be switched at any time according to the parameters that characterize the scene complexity of each frame of video image, it can well adapt to the changes in the shooting scene of the imaging device, thereby ensuring the video tracking effect.
附图说明Description of the drawings
为了更清楚地说明本申请实施例和现有技术的技术方案,下面对实施例和现有技术中所需要使用的附图作简单地介绍,显而易见地,下面描述中的附图仅仅是本申请的一些实施例,对于本领域普通技术人员来讲,在不付出 创造性劳动的前提下,还可以根据这些附图获得其他的附图。In order to explain the embodiments of the present application and the technical solutions of the prior art more clearly, the following briefly introduces the drawings needed in the embodiments and the prior art. Obviously, the drawings in the following description are only the present For some of the embodiments of the application, for those of ordinary skill in the art, other drawings can be obtained from these drawings without creative work.
图1是本申请实施例提供的视频跟踪系统的架构示意图;FIG. 1 is a schematic diagram of the architecture of a video tracking system provided by an embodiment of the present application;
图2是本申请实施例提供的视频跟踪设备的跟踪方法流程图;FIG. 2 is a flowchart of a tracking method of a video tracking device provided by an embodiment of the present application;
图3是本申请实施例提供的视频跟踪设备的结构示意图。Fig. 3 is a schematic structural diagram of a video tracking device provided by an embodiment of the present application.
具体实施方式Detailed ways
为使本申请的目的、技术方案、及优点更加清楚明白,以下参照附图并举实施例,对本申请进一步详细说明。显然,所描述的实施例仅仅是本申请一部分实施例,而不是全部的实施例。基于本申请中的实施例,本领域普通技术人员在没有作出创造性劳动前提下所获得的所有其他实施例,都属于本申请保护的范围。In order to make the purpose, technical solutions, and advantages of the present application clearer, the following further describes the present application in detail with reference to the drawings and embodiments. Obviously, the described embodiments are only a part of the embodiments of the present application, rather than all the embodiments. Based on the embodiments in this application, all other embodiments obtained by those of ordinary skill in the art without creative work fall within the protection scope of this application.
现有技术中,不同跟踪算法适用的场景不同,不同场景大多是由场景的复杂度区分的。本申请实施例中,根据成像设备拍摄的每帧视频图像的不同复杂度自适应调整跟踪算法,避免因成像设备拍摄场景变化而出现的跟踪效果差异。In the prior art, different tracking algorithms are applicable to different scenarios, and different scenarios are mostly distinguished by the complexity of the scenarios. In the embodiment of the present application, the tracking algorithm is adaptively adjusted according to the different complexity of each frame of video image captured by the imaging device, so as to avoid the difference in tracking effect caused by the change of the imaging device shooting scene.
参见图1,图1是本申请实施例提供的视频跟踪系统的架构示意图,视频跟踪系统包括:成像设备、视频跟踪设备、显示终端、随动控制设备。以下分别进行介绍。Refer to Fig. 1, which is a schematic diagram of the architecture of a video tracking system provided by an embodiment of the present application. The video tracking system includes: imaging equipment, video tracking equipment, display terminals, and follow-up control equipment. The following are introduced separately.
1)成像设备1) Imaging equipment
成像设备集成有成像机芯,成像机芯用于拍摄视频,成像设备将成像机芯拍摄的每帧视频图像发送到视频跟踪设备。另外,为了使视频跟踪设备能够根据拍摄场景的变化自适应的进行跟踪算法切换,成像设备除将成像机芯拍摄的每帧视频图像发送到视频跟踪设备之外,还将表征各帧视频图像的场景复杂度的参数发送到视频跟踪设备。一个实施例中,表征各帧视频图像的场景复杂度的参数可以使用各帧视频图像的清晰度评价参数表示。The imaging device is integrated with an imaging movement, which is used to shoot videos, and the imaging device sends each frame of video image taken by the imaging movement to the video tracking device. In addition, in order to enable the video tracking device to adaptively switch the tracking algorithm according to the changes in the shooting scene, the imaging device will not only send each frame of video image captured by the imaging core to the video tracking device, but also characterize each frame of video image. The parameters of the scene complexity are sent to the video tracking device. In an embodiment, the parameters that characterize the scene complexity of each frame of video image may be represented by the definition evaluation parameter of each frame of video image.
一个实施例中,成像设备集成的成像机芯可以包括热成像机芯和可见光成像机芯。热成像机芯和可见光成像机芯均用于拍摄视频图像,前者拍摄得到热成像视频图像,后者拍摄得到可见光视频图像。成像设备可以将热成像机芯拍摄的热成像视频图像和可见光成像机芯拍摄的可见光视频图像均发送到视频跟踪设备。另外,如果后续视频跟踪设备是对热成像视频图像进行目 标跟踪,则成像设备除发送热成像视频图像到视频跟踪设备之外,还需要将表征热成像视频图像的场景复杂度的参数发送到视频跟踪设备。同理,如果视频跟踪设备是对可见光视频图像进行目标跟踪,则成像设备除发送可见光视频图像到视频跟踪设备之外,还需要将表征可见光视频图像的场景复杂度的参数发送到视频跟踪设备。In one embodiment, the imaging core integrated with the imaging device may include a thermal imaging core and a visible light imaging core. Both the thermal imaging core and the visible light imaging core are used to capture video images, the former captures the thermal imaging video image, and the latter captures the visible light video image. The imaging device can send both the thermal imaging video image captured by the thermal imaging core and the visible light video image captured by the visible light imaging core to the video tracking device. In addition, if the subsequent video tracking device performs target tracking on the thermal imaging video image, the imaging device needs to send the thermal imaging video image to the video tracking device and also need to send the parameters that characterize the scene complexity of the thermal imaging video image to the video. Tracking equipment. In the same way, if the video tracking device performs target tracking on the visible light video image, the imaging device needs to send the visible light video image to the video tracking device and also need to send the parameters representing the scene complexity of the visible light video image to the video tracking device.
2)视频跟踪设备2) Video tracking equipment
视频跟踪设备可使用数字信号处理器(Digital Signal Processing,DSP)芯片实现,支持多种跟踪算法。Video tracking equipment can be implemented using a Digital Signal Processing (DSP) chip and supports multiple tracking algorithms.
视频跟踪设备对于成像设备发送的每帧视频图像,基于表征各帧视频图像的场景复杂度的参数选择合适的跟踪算法进行目标跟踪,得到跟踪目标在各帧视频图像中的位置信息,从而根据上述位置信息得到跟踪结果,然后根据跟踪结果调整成像设备的拍摄角度,使得跟踪目标处于成像设备后续拍摄的视频图像的中心位置。具体实现中,视频跟踪设备可以将跟踪结果发送到随动控制设备,由随动控制设备对成像设备的拍摄角度进行调整。For each frame of video image sent by the imaging device, the video tracking device selects an appropriate tracking algorithm for target tracking based on the parameters that characterize the scene complexity of each frame of video image, and obtains the position information of the tracking target in each frame of video image, according to the above The position information obtains the tracking result, and then adjusts the shooting angle of the imaging device according to the tracking result, so that the tracking target is at the center position of the video image subsequently captured by the imaging device. In specific implementation, the video tracking device can send the tracking result to the follow-up control device, and the follow-up control device adjusts the shooting angle of the imaging device.
其中,上述跟踪结果可以是:基于上述位置信息计算得到的视频图像中跟踪目标与视频图像的中心之间的距离。Wherein, the tracking result may be: the distance between the tracking target and the center of the video image in the video image calculated based on the position information.
一个实施例中,视频跟踪设备对成像设备发送的热成像视频图像进行目标跟踪,而对于成像设备发送的可见光视频图像,则是直接发送到显示设备进行显示。一个实施例中,为了保证视频图像被同步处理,成像设备和视频跟踪设备需保持时序同步。具体实现中,视频跟踪设备可以以成像设备中的热成像机芯或可见光成像机芯的同步信号为同步源,将上述同步源作为视频跟踪设备的基本工作时序,这样保证与热成像机芯或可见光成像机芯严格同步。视频跟踪设备在此基础上产生各单元电路的工作时序,形成一个统一的时序同步系统。In one embodiment, the video tracking device performs target tracking on the thermal imaging video image sent by the imaging device, and the visible light video image sent by the imaging device is directly sent to the display device for display. In one embodiment, in order to ensure that the video images are processed synchronously, the imaging device and the video tracking device need to maintain timing synchronization. In specific implementation, the video tracking device can use the synchronization signal of the thermal imaging core or the visible light imaging core in the imaging device as the synchronization source, and use the aforementioned synchronization source as the basic working sequence of the video tracking device, so as to ensure that it is connected to the thermal imaging core or The visible light imaging movement is strictly synchronized. The video tracking equipment generates the working timing of each unit circuit on this basis, forming a unified timing synchronization system.
3)随动控制设备3) Follow-up control equipment
随动控制设备根据视频跟踪设备发送的跟踪结果执行随动控制操作。例如直接对成像设备的拍摄角度进行调整,或者通过移动和/或旋转随动控制设备自身来实现对成像设备的拍摄角度的调整,以此保证跟踪目标位于成像设备拍摄的视频图像的中心。The follow-up control device executes the follow-up control operation according to the tracking result sent by the video tracking device. For example, the shooting angle of the imaging device is directly adjusted, or the shooting angle of the imaging device is adjusted by moving and/or rotating the follow-up control device itself, so as to ensure that the tracking target is located in the center of the video image captured by the imaging device.
以下结合图2,对本申请实施例提供的视频跟踪设备的跟踪方法进行详细说明。图2是以对成像设备中的热成像机芯拍摄的热成像视频图像进行目标跟踪为例展开说明的,在实际实现中,视频跟踪设备也可以对成像设备中的可见光成像机芯拍摄的可见光视频图像进行目标跟踪。The following describes in detail the tracking method of the video tracking device provided by the embodiment of the present application with reference to FIG. 2. Figure 2 is an example of target tracking of thermal imaging video images captured by the thermal imaging core in the imaging device. In actual implementation, the video tracking device can also capture the visible light captured by the visible light imaging core in the imaging device. Video image for target tracking.
另外,本申请实施例提供的方案应用于视频跟踪设备,用于对视频图像中的跟踪目标进行跟踪。In addition, the solution provided by the embodiment of the present application is applied to a video tracking device for tracking a tracking target in a video image.
参见图2,图2是本申请实施例提供的视频跟踪设备的跟踪方法流程图,如图2所示,该方法包括如下步骤201-步骤205。Refer to FIG. 2, which is a flowchart of a tracking method of a video tracking device provided in an embodiment of the present application. As shown in FIG. 2, the method includes the following steps 201 to 205.
步骤201、从成像设备获取视频图像及表征视频图像的场景复杂度的参数。Step 201: Obtain a video image and a parameter that characterizes the scene complexity of the video image from the imaging device.
由于成像设备是按照时间顺序依次拍摄视频图像的,所以视频跟踪设备也可以逐帧从成像设备获取视频图像。对于视频跟踪设备从成像设备获取的每一视频图像而言,进行目标跟踪的方法均相同,均可以通过本申请实施例提供的步骤201-步骤205实现。Since the imaging device sequentially shoots video images in chronological order, the video tracking device can also obtain video images from the imaging device frame by frame. For each video image acquired by the video tracking device from the imaging device, the target tracking method is the same, and all of them can be implemented through step 201 to step 205 provided in the embodiment of the present application.
一个实施例中,可以将视频图像的清晰度评价参数作为表征视频图像的场景复杂度的参数。视频图像的清晰度评价参数可以使用活动图像格式描述符(Active Format Description,AFD)表示。AFD值的大小可以在一定程度上代表视频图像所对应场景的复杂度。具体地,AFD值越大则视频图像的场景复杂度越高,反之,AFD值越小则视频图像的场景复杂度越低。根据视频图像的AFD值的大小,可以确定其场景复杂度的高低。例如,有两帧视频图像,第一帧视频图像的AFD值所属的取值范围是[5000,10000],第二帧视频图像的AFD值所属的取值范围是[500,1000],则第一帧视频图像的场景复杂度要高于第二帧视频图像的场景复杂度。In an embodiment, the definition evaluation parameter of the video image may be used as a parameter that characterizes the scene complexity of the video image. The definition evaluation parameter of the video image can be expressed by using an active image format descriptor (Active Format Description, AFD). The size of the AFD value can represent the complexity of the scene corresponding to the video image to a certain extent. Specifically, the larger the AFD value, the higher the scene complexity of the video image. Conversely, the smaller the AFD value, the lower the scene complexity of the video image. According to the AFD value of the video image, the complexity of the scene can be determined. For example, there are two frames of video images, the value range of the AFD value of the first frame of video image is [5000, 10000], and the value range of the AFD value of the second frame of video image is [500, 1000], then The scene complexity of one frame of video image is higher than the scene complexity of the second frame of video image.
其中,一帧视频图像的AFD值可以通过热成像机芯得到。具体的,热成像机芯的处理器对表征一帧视频图像清晰度的参数进行高频滤波处理,得到一组数据。因此,该组数据中的各个数据可以用来表征视频图像的清晰程度,该组数据中数据越大,表明视频图像的清晰程度越高,反之,该组数据中数据越小,表明视频图像的清晰程度越低。该组数据中的各个数据可以被称为上述AFD值。Among them, the AFD value of a frame of video image can be obtained through the thermal imaging core. Specifically, the processor of the thermal imaging core performs high-frequency filtering processing on the parameters that characterize the clarity of a frame of video image to obtain a set of data. Therefore, each data in the group of data can be used to characterize the clarity of the video image. The larger the data in the group of data, the higher the clarity of the video image. On the contrary, the smaller the data in the group of data, the higher the clarity of the video image. The lower the level of clarity. Each data in this set of data can be referred to as the aforementioned AFD value.
一个实施例中,上述热成像机芯的处理器可以是现场可编程逻辑门阵列 (Field Programmable Gate Array,FPGA)。In an embodiment, the processor of the thermal imaging core may be a Field Programmable Gate Array (FPGA).
一个实施例中,成像设备中可以集成热成像机芯、可见光机芯、或其它能够执行视频拍摄的成像机芯,并利用集成的成像机芯拍摄视频图像并将拍摄的每帧视频图像发送到视频跟踪设备。另外,成像设备还对成像机芯拍摄得到的每帧视频图像进行分析确定其清晰度评价参数,将每帧视频图像的清晰度评价参数作为表征每帧视频图像的场景复杂度的参数发送到视频跟踪设备。In one embodiment, the imaging device may integrate a thermal imaging core, a visible light core, or other imaging cores capable of performing video shooting, and use the integrated imaging core to capture video images and send each frame of the captured video images to Video tracking equipment. In addition, the imaging device also analyzes each frame of video image captured by the imaging core to determine its definition evaluation parameters, and sends the definition evaluation parameters of each frame of video image as a parameter that characterizes the scene complexity of each frame of video image to the video. Tracking equipment.
一个实施例中,成像设备中仅集成一种成像机芯,成像设备将由该成像机芯拍摄的每帧视频图像发送到视频跟踪设备后,视频跟踪设备一方面要对每帧视频图像进行跟踪处理,另一方面还可以将每帧视频图像发送到显示终端进行显示。In one embodiment, only one imaging core is integrated in the imaging device. After the imaging device sends each frame of video image captured by the imaging core to the video tracking device, the video tracking device needs to track each frame of video image on the one hand. On the other hand, it can also send each frame of video image to the display terminal for display.
另一个实施例中,成像设备中可以集成多种成像机芯,成像设备将其中一种成像机芯拍摄的每帧视频图像发送到视频跟踪设备,视频跟踪设备对每帧视频图像进行目标跟踪;另外,还可将其它成像机芯拍摄的每帧视频图像发送到视频跟踪设备,视频跟踪设备对由其它成像机芯拍摄的每帧视频图像进行其它视频处理,例如发送到显示终端进行显示。In another embodiment, multiple imaging cores can be integrated in the imaging device, the imaging device sends each frame of video image captured by one of the imaging cores to the video tracking device, and the video tracking device performs target tracking on each frame of video image; In addition, each frame of video images captured by other imaging cores can be sent to a video tracking device, and the video tracking device performs other video processing on each frame of video images captured by other imaging cores, for example, sending them to a display terminal for display.
在成像设备中集成多种成像机芯时,例如,成像设备中同时集成有热成像机芯和可见光成像机芯。集成了可见光成像机芯和热成像机芯的成像设备对当前场景进行视频拍摄,其中,成像设备中的热成像机芯拍摄得到当前场景的热成像视频图像,成像设备中的可见光成像机芯拍摄得到当前场景的可见光视频图像。另外,成像设备还对当前场景的热成像视频图像进行分析确定其清晰度评价参数。成像设备会将热成像机芯拍摄的每帧热成像视频图像和可见光成像机芯拍摄的每帧可见光视频图像发送到视频跟踪设备,另外还将每帧热成像视频图像的清晰度评价参数作为表征每帧热成像视频图像的场景复杂度的参数发送到视频跟踪设备。When multiple imaging cores are integrated in an imaging device, for example, a thermal imaging core and a visible light imaging core are integrated in the imaging device. The imaging device integrating the visible light imaging core and the thermal imaging core performs video shooting of the current scene, where the thermal imaging core in the imaging device captures the thermal imaging video image of the current scene, and the visible light imaging core in the imaging device captures Obtain the visible light video image of the current scene. In addition, the imaging device also analyzes the thermal imaging video image of the current scene to determine its sharpness evaluation parameters. The imaging device sends each frame of thermal imaging video image captured by the thermal imaging core and each frame of visible light video image captured by the visible light imaging core to the video tracking device, and also uses the definition evaluation parameters of each frame of thermal imaging video image as a characterization The parameters of the scene complexity of each thermal imaging video image are sent to the video tracking device.
步骤202、基于表征视频图像的场景复杂度的参数在多种跟踪算法中选择适用于视频图像的跟踪算法。Step 202: Select a tracking algorithm suitable for the video image from among multiple tracking algorithms based on the parameter that characterizes the scene complexity of the video image.
为了使视频跟踪设备能够根据拍摄场景的变化切换跟踪算法,视频跟踪设备可以支持对比度跟踪算法、相关跟踪算法、二值跟踪算法等多种跟踪算 法。In order to enable the video tracking device to switch tracking algorithms according to changes in the shooting scene, the video tracking device can support multiple tracking algorithms such as contrast tracking algorithms, related tracking algorithms, and binary tracking algorithms.
由于不同的跟踪算法适用于不同的拍摄场景,不同拍摄场景是根据场景复杂度区分的,鉴于此,一个实施例中,可以预先设置每种跟踪算法对应的参数取值范围,每种跟踪算法所适用拍摄场景的场景复杂度和每种跟踪算法对应的参数取值范围所代表的场景复杂度保持一致。Since different tracking algorithms are suitable for different shooting scenes, different shooting scenes are distinguished according to the complexity of the scene. In view of this, in one embodiment, the parameter value range corresponding to each tracking algorithm can be preset. The scene complexity of the applicable shooting scene is consistent with the scene complexity represented by the parameter value range corresponding to each tracking algorithm.
例如,可以将视频图像的清晰度评价参数作为表征视频图像的场景复杂度的参数,这种情况下,预先设置的每种跟踪算法对应的参数取值范围,为每种跟踪算法对应的清晰度评价参数取值范围。For example, the definition evaluation parameter of the video image can be used as a parameter that characterizes the scene complexity of the video image. In this case, the preset value range of the parameter corresponding to each tracking algorithm is the definition corresponding to each tracking algorithm Evaluation parameter value range.
举例来说,可以根据数量级将成像设备输出的AFD值划分为多个等级:……第n-1等级、第n等级、第n+1等级……,其中,n表示等级的序号。假设第n-1等级是AFD=40000,第n等级是AFD=70000,第n+1等级是AFD=100000。当AFD值在第n-1等级附近时,也就是,AFD值在40000附近时,例如40000±15000,说明当前场景为简单场景,如天空场景、海天场景,此时视频跟踪设备选择二值跟踪算法进行目标跟踪较为合适。当AFD值在第n等级附近时,也就是,AFD值在70000附近时,例如70000±15000,说明当前场景较为复杂,视频跟踪设备选择对比度跟踪算法进行目标跟踪较为合适;当AFD值在第n+1等级附近时,也就是,AFD值在100000附近时,例如100000±15000,说明当前场景非常复杂,视频跟踪设备选择相对跟踪算法进行目标跟踪比较合适。For example, the AFD value output by the imaging device can be divided into multiple levels according to the order of magnitude:...n-1th level, nth level, n+1th level..., where n represents the serial number of the level. Assume that the n-1th level is AFD=40,000, the nth level is AFD=70000, and the n+1th level is AFD=100,000. When the AFD value is near the n-1th level, that is, when the AFD value is near 40,000, for example, 40,000±15,000, it means that the current scene is a simple scene, such as a sky scene, a sea scene, and the video tracking device selects binary tracking The algorithm is more suitable for target tracking. When the AFD value is near the nth level, that is, when the AFD value is near 70,000, for example, 70,000±15000, it indicates that the current scene is more complicated, and the video tracking device chooses the contrast tracking algorithm for target tracking; when the AFD value is in the nth When the +1 level is near, that is, when the AFD value is near 100000, for example, 100000±15000, it indicates that the current scene is very complicated, and it is more appropriate for the video tracking device to select the relative tracking algorithm for target tracking.
根据上述AFD值的等级,可设置如下参数取值范围:……、(25000,55000)、(55000、85000)、(85000,115000)……,其中,二值跟踪算法对应的参数取值范围是(25000,55000),对比度跟踪算法对应的参数取值范围是(55000、85000),相关跟踪算法对应的参数取值范围是(85000,115000)。因此,如果视频跟踪设备从成像设备接收的一帧热成像视频图像的AFD值是50000,则由于50000属于参数取值范围(25000,55000),此参数取值范围对应于二值跟踪算法,因此视频跟踪设备将二值跟踪算法确定为适用于该帧热成像视频图像的跟踪算法,使用二值跟踪算法对该帧热成像视频图像中的跟踪目标进行跟踪。According to the level of the above AFD value, the following parameter value ranges can be set:..., (25000, 55000), (55000, 85000), (85000, 115000)..., among them, the parameter value range corresponding to the binary tracking algorithm Yes (25000, 55000), the parameter value range corresponding to the contrast tracking algorithm is (55000, 85000), and the parameter value range corresponding to the correlation tracking algorithm is (85000, 115000). Therefore, if the AFD value of a frame of thermal imaging video image received by the video tracking device from the imaging device is 50000, since 50000 belongs to the parameter value range (25000, 55000), the parameter value range corresponds to the binary tracking algorithm, so The video tracking device determines the binary tracking algorithm as a tracking algorithm suitable for the frame of thermal imaging video image, and uses the binary tracking algorithm to track the tracking target in the frame of thermal imaging video image.
由于预先设置了每种跟踪算法对应的参数取值范围,则视频跟踪设备从 成像设备获取了一帧视频图像及该帧视频图像的清晰度评价参数之后,可以将该帧视频图像的清晰度评价参数与每种跟踪算法对应的参数取值范围进行比较,找出该帧视频图像的清晰度评价参数对应的参数取值范围,将对应于该参数取值范围的跟踪算法确定为适用于该帧视频图像的跟踪算法。Since the parameter value range corresponding to each tracking algorithm is preset, the video tracking device can evaluate the definition of the frame of video image after obtaining a frame of video image and the definition evaluation parameters of the frame of video image from the imaging device The parameter is compared with the parameter value range corresponding to each tracking algorithm, and the parameter value range corresponding to the definition evaluation parameter of the frame of video image is found, and the tracking algorithm corresponding to the parameter value range is determined as suitable for the frame Video image tracking algorithm.
具体的,上述视频图像的清晰度评价参数对应的参数取值范围可以是视频图像的清晰度评价参数所属的参数取值范围。Specifically, the parameter value range corresponding to the definition evaluation parameter of the video image may be the parameter value range to which the definition evaluation parameter of the video image belongs.
鉴于上述情况,一个实施例中,本步骤202中,基于表征视频图像的场景复杂度的参数在多种跟踪算法中选择适用于视频图像的跟踪算法的具体方式如下S11-S12所示。In view of the foregoing, in one embodiment, in this step 202, a specific method for selecting a tracking algorithm suitable for a video image from among multiple tracking algorithms based on a parameter that characterizes the scene complexity of the video image is shown in the following S11-S12.
S11、比较上述参数与预先设置的每种跟踪算法对应的参数取值范围。S11. Compare the above-mentioned parameters with the parameter value ranges corresponding to each preset tracking algorithm.
通过对上述参数和各种跟踪算法对应的参数取值范围进行比较,可以确定出上述参数属于哪一个参数取值范围。By comparing the above parameters with the parameter value ranges corresponding to various tracking algorithms, it can be determined which parameter value range the above parameters belong to.
例如,假设上述参数为AFD值,且AFD值为44000,各种跟踪算法对应的参数取值范围为:算法1对应的参数取值范围(25000,55000)、算法2对应的参数取值范围(55000、85000)、算法3对应的参数取值范围(85000,115000)。For example, assuming that the above parameters are AFD values, and the AFD value is 44000, the parameter value ranges corresponding to various tracking algorithms are: the parameter value range corresponding to algorithm 1 (25000, 55000), the parameter value range corresponding to algorithm 2 ( 55000, 85000), the parameter value range corresponding to Algorithm 3 (85000, 115000).
则44000所属的参数取值范围为:(25000,55000)。The value range of the parameter to which 44000 belongs is: (25000, 55000).
S12、将对应于上述参数所属参数取值范围的跟踪算法确定为适用于视频图像的跟踪算法。S12. Determine the tracking algorithm corresponding to the parameter value range to which the above-mentioned parameter belongs as the tracking algorithm suitable for the video image.
在上述S11所示的举例中,44000所属的参数取值范围为:(25000,55000),对应于(25000,55000)这一参数取值范围的跟踪算法为算法1,所以,适用于视频图像的跟踪算法为算法1。In the example shown in S11 above, the parameter value range of 44000 is: (25000, 55000), and the tracking algorithm corresponding to the parameter value range of (25000, 55000) is Algorithm 1, so it is suitable for video images The tracking algorithm is Algorithm 1.
步骤203、根据适用于视频图像的跟踪算法确定视频图像中的跟踪目标。Step 203: Determine the tracking target in the video image according to the tracking algorithm suitable for the video image.
通常情况下,一帧视频图像中包括不止一种物体,为了甄别出其中的跟踪目标,需要将所有物体作为检测目标,逐个辨别是否是跟踪目标。Generally, a frame of video image includes more than one type of object. In order to identify the tracking target in it, it is necessary to use all the objects as the detection target, and identify whether they are tracking targets one by one.
以下结合具体跟踪算法进行介绍:The following describes the specific tracking algorithm:
I)对比度跟踪算法I) Contrast tracking algorithm
对于对比度跟踪算法,在利用对比度跟踪算法对一帧视频图像进行目标跟踪时,需要从该帧视频图像中提取出每个检测目标对应的用于对比度跟踪 算法的跟踪信息,例如,边缘信息、轮廓长度、面积、重心、和/或形心等,然后将所有检测目标的跟踪信息与前一帧视频图像中检测目标的跟踪信息进行比较,找出跟踪信息与前一帧视频图像中检测目标的跟踪信息的匹配度最大的检测目标,将找出的检测目标确定为该帧视频图像中的跟踪目标。For the contrast tracking algorithm, when using the contrast tracking algorithm to track a frame of video image, it is necessary to extract the tracking information for the contrast tracking algorithm corresponding to each detected target from the frame of the video image, for example, edge information, contour The length, area, center of gravity, and/or centroid, etc., and then compare the tracking information of all detected targets with the tracking information of the detected target in the previous frame of video image to find out the tracking information and the detected target in the previous frame of video image The detection target with the greatest matching degree of tracking information is determined as the tracking target in the frame of video image.
其中,在从一帧视频图像中提取每个检测目标对应的用于对比度跟踪算法的跟踪信息之前,还可以对该帧视频图像进行预处理,具体方法如下X1-X3所示。Among them, before extracting the tracking information for the contrast tracking algorithm corresponding to each detection target from a frame of video image, the frame of video image can also be preprocessed, and the specific method is shown as X1-X3.
X1、对视频图像进行去噪声处理。X1. De-noise processing on the video image.
例如,本步骤可通过高斯滤波实现去噪声处理,对视频图像进行高斯滤波能够去掉视频图像中的散点噪声。For example, in this step, Gaussian filtering can be used to achieve noise removal processing, and Gaussian filtering of the video image can remove scattered noise in the video image.
X2、对经过去噪声处理的视频图像进行边缘检测。X2. Perform edge detection on the video image that has been denoised.
边缘检测主要是标识视频图像中亮度变化明显的像素点。例如,实现边缘检测的算法可以包括基于搜索和基于零交叉两类边缘检测算法。Edge detection is mainly to identify pixels with obvious brightness changes in video images. For example, the algorithm for edge detection can include two types of edge detection algorithms based on search and zero-crossing.
X3、确定经边缘检测后的视频图像的分割阈值范围,根据该分割阈值范围对经边缘检测的视频图像进行二值化处理。X3. Determine the segmentation threshold range of the edge-detected video image, and perform binarization processing on the edge-detected video image according to the segmentation threshold range.
确定分割阈值范围是为了后续对经边缘检测的视频图像进行二值化处理。例如,可以根据视频图像的灰度极限值、最大视频信号幅度确定分割阈值范围。其中,上述灰度极限值可以是最大灰度值或最小灰度值。具体地,分割阈值范围的下限T V_min可采用T V_min=P-α·V P计算得到;分割阈值范围的上限T V_max可采用公式T V_max=P+α·V P计算得到;其中,P为视频图像的灰度极限值;V P为视频图像的最大视频信号幅度,例如,700mV;α为预设的对比度参数值,取值范围可以为V P取值的5%到15%。 The determination of the segmentation threshold range is for the subsequent binarization of the edge-detected video image. For example, the segmentation threshold range can be determined according to the gray limit value of the video image and the maximum video signal amplitude. Wherein, the gray limit value can be the maximum gray value or the minimum gray value. Specifically, the lower limit T V_min of the segmentation threshold range can be calculated using T V_min =P-α·V P ; the upper limit T V_max of the segmentation threshold range can be calculated using the formula T V_max =P+α·V P ; where P is The gray limit value of the video image; V P is the maximum video signal amplitude of the video image, for example, 700 mV; α is the preset contrast parameter value, and the value range can be 5% to 15% of the value of V P.
在对经边缘检测的视频图像进行二值化处理时,可以将灰度值低于T V_min的像素点的灰度值统一设置为0,灰度值大于T V_max的像素点的灰度值统一设置为255,而灰度值介于T V_min和T V_max之间的像素点的灰度值则可以保持不变。 When the edge-detected video image is binarized, the gray values of pixels with gray values lower than T V_min can be uniformly set to 0, and the gray values of pixels with gray values greater than T V_max can be uniform. Set to 255, and the gray value of the pixel with gray value between T V_min and T V_max can remain unchanged.
在对视频图像进行预处理之后,可从经预处理的视频图像中提取出每个检测目标对应的用于对比度跟踪算法的跟踪信息,从而根据提取的所有检测目标的跟踪信息,从视频图像的所有检测目标中筛选出跟踪目标。After the video image is preprocessed, the tracking information for the contrast tracking algorithm corresponding to each detection target can be extracted from the preprocessed video image, so that according to the extracted tracking information of all detection targets, from the video image The tracking target is filtered out of all detection targets.
因此,当适用于视频图像的跟踪算法为对比度跟踪算法时,根据适用于 视频图像的跟踪算法确定视频图像中的跟踪目标,具体包括:Therefore, when the tracking algorithm suitable for the video image is a contrast tracking algorithm, the tracking target in the video image is determined according to the tracking algorithm suitable for the video image, which specifically includes:
S21、针对视频图像中的每一检测目标,提取该检测目标的用于对比度跟踪算法的跟踪信息;S21: For each detection target in the video image, extract tracking information of the detection target used for the contrast tracking algorithm;
S22、根据所有检测目标的用于对比度跟踪算法的跟踪信息,从所有检测目标中筛选出跟踪目标。S22: According to the tracking information used for the contrast tracking algorithm of all the detected targets, the tracking target is selected from all the detected targets.
由以上描述可以看出,采用对比度跟踪算法对目标进行跟踪时,需要使用前一帧视频图像中检测目标的跟踪信息,但是对于开始进行目标跟踪的第一帧视频图像而言,并不存在前一帧视频图像,鉴于此,一个实施例中,可以采用人工手动指定的方式在上述第一帧视频图像中确定跟踪目标。It can be seen from the above description that when the contrast tracking algorithm is used to track the target, it is necessary to use the tracking information of the detected target in the previous frame of video image, but for the first frame of video image to start tracking the target, there is no previous frame. One frame of video image. In view of this, in one embodiment, the tracking target can be determined in the first frame of video image by manual designation.
II)二值跟踪算法II) Binary tracking algorithm
对于二值跟踪算法,在利用二值跟踪算法对一帧视频图像进行跟踪处理时,需要从视频图像中提取出每个检测目标对应的用于二值跟踪算法的跟踪信息,例如,上述跟踪信息可以包括边缘信息、轮廓长度、面积、重心、和/或形心,然后将所有检测目标的跟踪信息与前一帧视频图像中检测目标的跟踪信息进行比较,找出跟踪信息与前一帧视频图像中检测目标的跟踪信息的匹配度最大的检测目标,将找出的检测目标确定为该帧视频图像中的跟踪目标。For the binary tracking algorithm, when using the binary tracking algorithm to track a frame of video image, it is necessary to extract the tracking information for the binary tracking algorithm corresponding to each detection target from the video image, for example, the aforementioned tracking information It can include edge information, contour length, area, center of gravity, and/or centroid, and then compare the tracking information of all detected targets with the tracking information of the detected target in the previous frame of video image to find out the tracking information and the previous frame of video The detection target in the image with the greatest matching degree of the tracking information of the detection target is determined as the tracking target in the frame of the video image.
其中,在从一帧视频图像中提取每个检测目标对应的用于对比度跟踪算法的跟踪信息之前,还可以对该帧视频图像进行预处理,预处理方法与对比度跟踪算法中的预处理方法相同。在预处理之后,还需要对经预处理后的该帧视频图像中的各个检测目标进行区域填充,区域填充可以更凸显出视频图像中的各检测目标。Among them, before extracting the tracking information for the contrast tracking algorithm corresponding to each detection target from a frame of video image, the frame of video image can also be preprocessed, and the preprocessing method is the same as that in the contrast tracking algorithm . After the preprocessing, it is also necessary to perform area filling for each detection target in the frame of the video image after the preprocessing. The area filling can more highlight each detection target in the video image.
在对视频图像进行预处理和区域填充之后,可从经预处理和区域填充的视频图像中提取出每个检测目标对应的用于二值跟踪算法的跟踪信息,从而根据提取的所有检测目标对应的跟踪信息,从该帧视频图像的所有检测目标中筛选出跟踪目标。After the video image is preprocessed and area filled, the tracking information for the binary tracking algorithm corresponding to each detection target can be extracted from the preprocessed and area filled video image, so as to correspond to all the extracted detection targets The tracking information is selected from all the detected targets in the frame of video image.
因此,当适用于视频图像的跟踪算法为二值跟踪算法时,根据适用于视频图像的跟踪算法确定视频图像中的跟踪目标,具体包括如下S31和S32。Therefore, when the tracking algorithm suitable for the video image is a binary tracking algorithm, the tracking target in the video image is determined according to the tracking algorithm suitable for the video image, which specifically includes the following S31 and S32.
S31、针对视频图像中的每一检测目标,提取该检测目标对应的用于二值 跟踪算法的跟踪信息;S31. For each detection target in the video image, extract tracking information corresponding to the detection target for the binary tracking algorithm;
S32、根据所有检测目标的用于二值跟踪算法的跟踪信息,从所有检测目标中筛选出跟踪目标。S32. According to the tracking information used for the binary tracking algorithm of all the detection targets, the tracking target is selected from all the detection targets.
由以上描述可以看出,采用二值跟踪算法对目标进行跟踪时,需要使用前一帧视频图像中检测目标的跟踪信息,但是对于开始进行目标跟踪的第一帧视频图像而言,并不存在前一帧视频图像,鉴于此,一个实施例中,可以采用人工手动指定的方式在上述第一帧视频图像中确定跟踪目标。It can be seen from the above description that when a binary tracking algorithm is used to track a target, it is necessary to use the tracking information of the detected target in the previous frame of video image, but for the first frame of video image to start tracking the target, there is no In view of this, in one embodiment, the previous frame of video image may be manually designated to determine the tracking target in the first frame of video image.
III)相关跟踪算法III) Related tracking algorithms
对于相关跟踪算法,不需要从视频图像中提取出每个检测目标的跟踪信息,只需将预先选定的包含有跟踪目标的模板图像与视频图像中的各个检测目标进行匹配,将匹配度最大的检测目标作为视频图像中的跟踪目标。其中,将预先选定的包含有跟踪目标的模板图像与视频图像中的各个检测目标进行匹配之前,还可以对视频图像进行预处理,预处理方法与对比度跟踪算法中的预处理方法相同。For related tracking algorithms, there is no need to extract the tracking information of each detection target from the video image, just match the preselected template image containing the tracking target with each detection target in the video image to maximize the degree of matching The detection target is used as the tracking target in the video image. Among them, before matching the pre-selected template image containing the tracking target with each detection target in the video image, the video image can also be preprocessed, and the preprocessing method is the same as the preprocessing method in the contrast tracking algorithm.
这里,预先选定的包含有跟踪目标的模板图像,可以预先设定,也可以从之前的视频图像中提取得到,例如从首次出现跟踪目标的视频图像中提取出包含跟踪目标的图像作为模板图像,或者从前一帧视频图像中提取出包含跟踪目标的图像作为模板图像。Here, the pre-selected template image containing the tracking target can be preset or extracted from previous video images. For example, the image containing the tracking target is extracted from the video image where the tracking target appears for the first time as the template image , Or extract the image containing the tracking target from the previous frame of video image as a template image.
因此,当适用于视频图像的跟踪算法为相关跟踪算法时,根据适用于视频图像的跟踪算法确定视频图像中的跟踪目标,具体包括如下S41。Therefore, when the tracking algorithm suitable for the video image is the related tracking algorithm, the tracking target in the video image is determined according to the tracking algorithm suitable for the video image, which specifically includes the following S41.
S41、依据预先选定的包含有跟踪目标的模板图像,从视频图像的所有检测目标中筛选出与该模板图像匹配度最大的检测目标,将筛选出的检测目标确定为视频图像的跟踪目标。S41. According to a pre-selected template image containing a tracking target, a detection target with the greatest degree of matching with the template image is selected from all detection targets in the video image, and the screened detection target is determined as the tracking target of the video image.
步骤204、提取跟踪目标在视频图像中的位置信息,基于该位置信息计算该跟踪目标与视频图像的中心的距离。Step 204: Extract the position information of the tracking target in the video image, and calculate the distance between the tracking target and the center of the video image based on the position information.
一个实施例中,跟踪目标的位置信息主要包括跟踪目标的高度、宽度、坐标等信息。其中,跟踪目标的高度和宽度,可以根据视频图像在三维坐标系中的x轴和y轴上的投影确定,例如,在x轴上的投影落入区间[x1、x2],在y轴上的投影落入区间[y1、y2],则可以确定检测目标的宽度是x2-x1,高 度是y2-y1,跟踪目标的中心点坐标为:((x1+x2)/2,(y1+y2)/2)。In an embodiment, the location information of the tracking target mainly includes information such as the height, width, coordinates of the tracking target. Among them, the height and width of the tracking target can be determined according to the projection of the video image on the x-axis and y-axis in the three-dimensional coordinate system. For example, the projection on the x-axis falls into the interval [x1, x2], on the y-axis If the projection falls within the interval [y1, y2], it can be determined that the width of the detection target is x2-x1, the height is y2-y1, and the center point coordinates of the tracking target are: ((x1+x2)/2, (y1+y2 )/2).
确定了视频图像中跟踪目标的位置信息后,由于视频图像的中心的坐标是已知的,因此,通过坐标间的距离计算即可得出跟踪目标与视频图像的中心的距离。After the position information of the tracking target in the video image is determined, since the coordinates of the center of the video image are known, the distance between the tracking target and the center of the video image can be obtained by calculating the distance between the coordinates.
另外,一个实施例中,也可以将跟踪目标的重心和/或形心作为跟踪目标的位置信息使用,再者,也可以将跟踪目标上特定点的位置作为跟踪目标的位置信息使用,特定点如跟踪目标的边缘上某个拐角点或突出的端点等。In addition, in one embodiment, the center of gravity and/or centroid of the tracking target can also be used as the location information of the tracking target. Furthermore, the location of a specific point on the tracking target can also be used as the location information of the tracking target. Such as tracking a corner point or protruding end point on the edge of the target.
步骤205、根据上述距离调整成像设备的拍摄角度,以使跟踪目标处于成像设备拍摄的下一视频图像的中心。Step 205: Adjust the shooting angle of the imaging device according to the above distance, so that the tracking target is at the center of the next video image shot by the imaging device.
一个实施例中,根据上述距离调整成像设备的拍摄角度的具体实现方式为:将跟踪目标与视频图像的中心的距离发送到搭载视频跟踪设备的随动控制设备,以使随动控制设备根据上述距离执行随动控制操作、进而调整成像设备的拍摄角度。In one embodiment, the specific implementation of adjusting the shooting angle of the imaging device according to the above distance is: sending the distance between the tracking target and the center of the video image to the follow-up control device equipped with the video tracking device, so that the follow-up control device is based on the foregoing The distance performs the follow-up control operation to adjust the shooting angle of the imaging device.
一种情况下,成像设备和视频跟踪设备之间的位置是非常接近或直接集成在一起的,并且均安装在随动控制设备上,随着随动控制设备的移动而移动。In one case, the position between the imaging device and the video tracking device is very close or directly integrated, and both are installed on the follow-up control device and move with the movement of the follow-up control device.
视频跟踪设备确定当前场景的视频图像中的跟踪目标与视频图像的中心的距离后,可以将此距离发送到随动控制设备,而随动控制设备可以通过控制自身移动带动成像设备移动,或直接控制成像设备旋转或移动,使得跟踪目标位于成像设备拍摄的下一帧视频图像的中心。After the video tracking device determines the distance between the tracking target in the video image of the current scene and the center of the video image, it can send this distance to the follow-up control device, and the follow-up control device can drive the imaging device to move by controlling its own movement, or directly The imaging device is controlled to rotate or move so that the tracking target is located at the center of the next frame of video image captured by the imaging device.
由以上可见,上述各个实施例提供的方案中,对于成像设备拍摄的每帧视频图像,基于表征该帧视频图像的场景复杂度的参数自适应选择适用于该帧视频图像的跟踪算法,基于选择的跟踪算法对该帧视频图像中的跟踪目标进行跟踪,并根据视频图像中的跟踪目标与视频图像中心的距离调整成像设备的拍摄角度,使得跟踪目标处于成像设备拍摄的下一视频图像的中心位置。由于可以随时根据表征每帧视频图像的场景复杂度的参数切换跟踪算法,因此可以很好的适应成像设备拍摄场景的变化,从而保证了视频跟踪效果。It can be seen from the above that, in the solutions provided by the foregoing embodiments, for each frame of video image captured by the imaging device, the tracking algorithm suitable for the frame of video image is adaptively selected based on the parameter that characterizes the scene complexity of the frame of video image. The tracking algorithm tracks the tracking target in this frame of video image, and adjusts the shooting angle of the imaging device according to the distance between the tracking target in the video image and the center of the video image, so that the tracking target is at the center of the next video image captured by the imaging device position. Since the tracking algorithm can be switched at any time according to the parameters that characterize the scene complexity of each frame of video image, it can well adapt to the changes in the shooting scene of the imaging device, thereby ensuring the video tracking effect.
以上对本申请实施例提供的视频跟踪设备的跟踪方法进行了详细说明,本申请实施例还提供了一种视频跟踪设备,以下结合图3进行详细说明。The tracking method of the video tracking device provided by the embodiment of the application is described in detail above, and the embodiment of the application also provides a video tracking device, which is described in detail below with reference to FIG. 3.
参见图3,图3是本申请实施例提供的视频跟踪设备的结构示意图,如图3所示,该视频跟踪设备300包括处理器301和非瞬时性计算机可读存储介质302,其中,Referring to FIG. 3, FIG. 3 is a schematic structural diagram of a video tracking device provided by an embodiment of the present application. As shown in FIG. 3, the video tracking device 300 includes a processor 301 and a non-transitory computer-readable storage medium 302, wherein,
所述非瞬时性计算机可读存储介质302,用于存储可以被所述处理器301执行的指令,在所述指令被所述处理器301执行时,使得所述处理器301:The non-transitory computer-readable storage medium 302 is configured to store instructions that can be executed by the processor 301, and when the instructions are executed by the processor 301, the processor 301 is caused to:
从成像设备获取视频图像及表征所述帧视频图像的场景复杂度的参数;Acquiring a video image from an imaging device and a parameter characterizing the scene complexity of the frame video image;
基于所述参数在所述多种跟踪算法中选择适用于所述视频图像的跟踪算法;Selecting a tracking algorithm suitable for the video image from the multiple tracking algorithms based on the parameter;
根据适用于所述视频图像的跟踪算法确定所述视频图像中的跟踪目标;Determining a tracking target in the video image according to a tracking algorithm applicable to the video image;
提取所述跟踪目标在所述视频图像中的位置信息,基于所述位置信息计算所述跟踪目标与所述视频图像的中心的距离;Extracting position information of the tracking target in the video image, and calculating the distance between the tracking target and the center of the video image based on the position information;
根据所述距离调整所述成像设备的拍摄角度,以使所述跟踪目标处于所述成像设备拍摄的下一视频图像的中心。The shooting angle of the imaging device is adjusted according to the distance, so that the tracking target is at the center of the next video image captured by the imaging device.
一个实施例中,图3所示设备中,所述成像设备中集成了成像机芯;In one embodiment, in the device shown in FIG. 3, an imaging core is integrated in the imaging device;
所述处理器301,从成像设备获取视频图像及表征所述视频图像的场景复杂度的参数时,包括:The processor 301, when acquiring a video image and a parameter characterizing the scene complexity of the video image from an imaging device, includes:
接收成像设备发送的由所述成像机芯拍摄的该帧视频图像,并接收所述成像设备发送的所述视频图像的清晰度评价参数,其中,所述清晰度评价参数为表征所述视频图像的场景复杂度的参数。Receive the frame of the video image taken by the imaging core sent by the imaging device, and receive the definition evaluation parameter of the video image sent by the imaging device, wherein the definition evaluation parameter represents the video image The parameter of the scene complexity.
一个实施例中,图3所示设备中,所述处理器301,基于所述参数在所述多种跟踪算法中选择适用于所述视频图像的跟踪算法时,包括:In one embodiment, in the device shown in FIG. 3, when the processor 301 selects a tracking algorithm suitable for the video image from among the multiple tracking algorithms based on the parameter, the method includes:
比较所述参数与预先设置的每种跟踪算法对应的参数取值范围;Compare the parameter value range corresponding to each preset tracking algorithm with the parameter;
将对应于所述参数所属参数取值范围的跟踪算法确定为适用于所述视频图像的跟踪算法。The tracking algorithm corresponding to the parameter value range to which the parameter belongs is determined as the tracking algorithm suitable for the video image.
一个实施例中,图3所示设备中,所述多种跟踪算法包括:对比度跟踪算法、二值跟踪算法、和相关跟踪算法;In one embodiment, in the device shown in FIG. 3, the multiple tracking algorithms include: a contrast tracking algorithm, a binary tracking algorithm, and a related tracking algorithm;
适用于所述视频图像的目标跟踪算法为对比度跟踪算法或二值跟踪算法时,所述处理器301根据适用于所述视频图像的跟踪算法确定所述视频图像中的跟踪目标时,包括:When the target tracking algorithm suitable for the video image is a contrast tracking algorithm or a binary tracking algorithm, when the processor 301 determines the tracking target in the video image according to the tracking algorithm suitable for the video image, it includes:
针对所述视频图像中的每一检测目标,提取该检测目标对应的用于所述目标跟踪算法的跟踪信息;For each detection target in the video image, extract tracking information corresponding to the detection target for the target tracking algorithm;
根据所有检测目标的用于所述目标跟踪算法的跟踪信息,从所有检测目标中筛选出跟踪目标;Screening out the tracking target from all the detection targets according to the tracking information of all the detection targets used in the target tracking algorithm;
所述目标跟踪算法为相关跟踪算法时,所述处理器301根据适用于所述视频图像的跟踪算法确定所述视频图像中的跟踪目标时,包括:When the target tracking algorithm is a related tracking algorithm, when the processor 301 determines the tracking target in the video image according to the tracking algorithm applicable to the video image, the method includes:
依据预先选定的包含有跟踪目标的模板图像,从所述视频图像的所有检测目标中筛选出与所述模板图像匹配度最大的检测目标,将筛选出的检测目标确定为所述视频图像的跟踪目标。According to the pre-selected template image containing the tracking target, the detection target with the greatest degree of matching with the template image is selected from all the detection targets of the video image, and the screened detection target is determined as the video image. Track the target.
一个实施例中,图3所示设备中,所述处理器301,根据所述距离调整所述成像设备的拍摄角度时,包括:将所述距离发送到搭载所述视频跟踪设备的随动控制设备,以使所述随动控制设备根据所述距离执行随动控制操作、进而调整所述成像设备的拍摄角度。In one embodiment, in the device shown in FIG. 3, when the processor 301 adjusts the shooting angle of the imaging device according to the distance, it includes: sending the distance to a follow-up control equipped with the video tracking device Device, so that the follow-up control device performs a follow-up control operation according to the distance, thereby adjusting the shooting angle of the imaging device.
由以上可见,上述各个实施例提供的视频跟踪设备进行目标跟踪时,对于成像设备拍摄的每帧视频图像,基于表征该帧视频图像的场景复杂度的参数自适应选择适用于该帧视频图像的跟踪算法,基于选择的跟踪算法对该帧视频图像中的跟踪目标进行跟踪,并根据视频图像中的跟踪目标与视频图像中心的距离调整成像设备的拍摄角度,使得跟踪目标处于成像设备拍摄的下一视频图像的中心位置。由于视频跟踪设备可以随时根据表征每帧视频图像的场景复杂度的参数切换跟踪算法,因此可以很好的适应成像设备拍摄场景的变化,从而保证了视频跟踪效果。It can be seen from the above that, when the video tracking device provided by the foregoing embodiments performs target tracking, for each frame of video image captured by the imaging device, adaptive selection is applied to the frame of video image based on the parameter that characterizes the scene complexity of the frame of video image. Tracking algorithm, based on the selected tracking algorithm to track the tracking target in the frame of video image, and adjust the shooting angle of the imaging device according to the distance between the tracking target in the video image and the center of the video image, so that the tracking target is under the shooting of the imaging device The center position of a video image. Since the video tracking device can switch the tracking algorithm at any time according to the parameters characterizing the scene complexity of each frame of video image, it can well adapt to changes in the shooting scene of the imaging device, thereby ensuring the video tracking effect.
与上述视频跟踪设备的跟踪方法相对应,本申请实施例还提供了一种视频跟踪设备的跟踪装置。下面对本申请实施例提供的视频跟踪设备的跟踪装置进行说明。Corresponding to the tracking method of the video tracking device described above, an embodiment of the present application also provides a tracking device of the video tracking device. The following describes the tracking device of the video tracking device provided in the embodiment of the present application.
一个实施例中,上述视频跟踪设备支持多种跟踪算法;上述视频跟踪设备的跟踪装置包括:In one embodiment, the aforementioned video tracking device supports multiple tracking algorithms; the tracking device of the aforementioned video tracking device includes:
信息获取模块,用于从成像设备获取视频图像及表征所述视频图像的场景复杂度的参数;An information acquisition module for acquiring a video image from an imaging device and parameters that characterize the scene complexity of the video image;
算法选择模块,用于基于所述参数在所述多种跟踪算法中选择适用于所 述视频图像的跟踪算法;An algorithm selection module, configured to select a tracking algorithm suitable for the video image among the multiple tracking algorithms based on the parameter;
目标确定模块,用于根据适用于所述视频图像的跟踪算法确定所述视频图像中的跟踪目标;A target determination module, configured to determine a tracking target in the video image according to a tracking algorithm applicable to the video image;
距离计算模块,用于提取所述跟踪目标在所述视频图像中的位置信息,基于所述位置信息计算所述跟踪目标与所述视频图像的中心的距离;A distance calculation module, configured to extract position information of the tracking target in the video image, and calculate the distance between the tracking target and the center of the video image based on the position information;
角度调整模块,用于根据所述距离调整所述成像设备的拍摄角度,以使所述跟踪目标处于所述成像设备拍摄的下一视频图像的中心。The angle adjustment module is configured to adjust the shooting angle of the imaging device according to the distance, so that the tracking target is at the center of the next video image shot by the imaging device.
一个实施例中,所述成像设备中集成了成像机芯;In one embodiment, an imaging core is integrated in the imaging device;
所述信息获取模块,具体用于:The information acquisition module is specifically used for:
接收成像设备发送的由所述成像机芯拍摄的视频图像,并接收所述成像设备发送的所述视频图像的清晰度评价参数,其中,所述清晰度评价参数为表征所述视频图像的场景复杂度的参数。Receive a video image taken by the imaging core sent by an imaging device, and receive a sharpness evaluation parameter of the video image sent by the imaging device, wherein the sharpness evaluation parameter is a scene characterizing the video image Complexity parameter.
一个实施例中,所述算法选择模块,具体用于:In an embodiment, the algorithm selection module is specifically used for:
比较所述参数与预先设置的每种跟踪算法对应的参数取值范围;Compare the parameter value range corresponding to each preset tracking algorithm with the parameter;
将对应于所述参数所属参数取值范围的跟踪算法确定为适用于所述视频图像的跟踪算法。The tracking algorithm corresponding to the parameter value range to which the parameter belongs is determined as the tracking algorithm suitable for the video image.
一个实施例中,所述多种跟踪算法包括:对比度跟踪算法、二值跟踪算法和相关跟踪算法;In an embodiment, the multiple tracking algorithms include: a contrast tracking algorithm, a binary tracking algorithm, and a related tracking algorithm;
适用于所述视频图像的目标跟踪算法为对比度跟踪算法或二值跟踪算法时,所述目标确定模块,具体用于:When the target tracking algorithm suitable for the video image is a contrast tracking algorithm or a binary tracking algorithm, the target determination module is specifically used for:
针对所述视频图像中的每一检测目标,提取该检测目标对应的用于所述目标跟踪算法的跟踪信息;For each detection target in the video image, extract tracking information corresponding to the detection target for the target tracking algorithm;
根据所有检测目标的用于所述目标跟踪算法的跟踪信息,从所有检测目标中筛选出跟踪目标;Screening out the tracking target from all the detection targets according to the tracking information of all the detection targets used in the target tracking algorithm;
所述目标跟踪算法为相关跟踪算法时,所述目标确定模块,具体用于:When the target tracking algorithm is a related tracking algorithm, the target determination module is specifically used for:
依据预先选定的包含有跟踪目标的模板图像,从所述视频图像的所有检测目标中筛选出与所述模板图像匹配度最大的检测目标,将筛选出的检测目标确定为所述视频图像中的跟踪目标。According to the pre-selected template image containing the tracking target, the detection target with the greatest degree of matching with the template image is selected from all detection targets in the video image, and the selected detection target is determined as the video image Tracking target.
一个实施例中,所述角度调整模块,具体用于:In an embodiment, the angle adjustment module is specifically used for:
将所述距离发送到搭载所述视频跟踪设备的随动控制设备,以使所述随动控制设备根据所述距离执行随动控制操作、进而调整所述成像设备的拍摄角度。The distance is sent to a follow-up control device equipped with the video tracking device, so that the follow-up control device performs a follow-up control operation according to the distance, thereby adjusting the shooting angle of the imaging device.
由以上可见,上述各个实施例提供的方案中,对于成像设备拍摄的每帧视频图像,基于表征该帧视频图像的场景复杂度的参数自适应选择适用于该帧视频图像的跟踪算法,基于选择的跟踪算法对该帧视频图像中的跟踪目标进行跟踪,并根据视频图像中的跟踪目标与视频图像中心的距离调整成像设备的拍摄角度,使得跟踪目标处于成像设备拍摄的下一视频图像的中心位置。由于可以随时根据表征每帧视频图像的场景复杂度的参数切换跟踪算法,因此可以很好的适应成像设备拍摄场景的变化,从而保证了视频跟踪效果。It can be seen from the above that, in the solutions provided by the foregoing embodiments, for each frame of video image captured by the imaging device, the tracking algorithm suitable for the frame of video image is adaptively selected based on the parameter that characterizes the scene complexity of the frame of video image. The tracking algorithm tracks the tracking target in this frame of video image, and adjusts the shooting angle of the imaging device according to the distance between the tracking target in the video image and the center of the video image, so that the tracking target is at the center of the next video image captured by the imaging device position. Since the tracking algorithm can be switched at any time according to the parameters that characterize the scene complexity of each frame of video image, it can well adapt to the changes in the shooting scene of the imaging device, thereby ensuring the video tracking effect.
与上述视频跟踪设备的跟踪方法相对应,本申请实施例还提供了一种计算机可读存储介质,所述计算机可读存储介质内存储有计算机程序,所述计算机程序被处理器执行时实现本申请实施例所述的视频跟踪设备的跟踪方法步骤。Corresponding to the tracking method of the video tracking device described above, an embodiment of the present application also provides a computer-readable storage medium in which a computer program is stored, and the computer program is executed by a processor to realize this The steps of the tracking method of the video tracking device described in the application embodiment.
与上述视频跟踪设备的跟踪方法相对应,本申请实施例还提供了一种包含指令的计算机程序产品,当其在计算机上运行时,使得计算机执行本申请实施例所述的视频跟踪设备的跟踪方法。Corresponding to the tracking method of the video tracking device described above, an embodiment of the present application also provides a computer program product containing instructions that, when it runs on a computer, causes the computer to perform the tracking of the video tracking device described in the embodiment of the application. method.
在上述实施例中,可以全部或部分地通过软件、硬件、固件或者其任意组合来实现。当使用软件实现时,可以全部或部分地以计算机程序产品的形式实现。所述计算机程序产品包括一个或多个计算机指令。在计算机上加载和执行所述计算机程序指令时,全部或部分地产生按照本申请实施例所述的流程或功能。所述计算机可以是通用计算机、专用计算机、计算机网络、或者其他可编程装置。所述计算机指令可以存储在计算机可读存储介质中,或者从一个计算机可读存储介质向另一个计算机可读存储介质传输,例如,所述计算机指令可以从一个网站站点、计算机、服务器或数据中心通过有线(例如同轴电缆、光纤、数字用户线(DSL))或无线(例如红外、无线、微波等)方式向另一个网站站点、计算机、服务器或数据中心进行传输。所述计算机可读存储介质可以是计算机能够存取的任何可用介质或者是包含一个或多个可用介质集成的服务器、数据中心等数据存储设备。所述可用介质可以是磁 性介质,(例如,软盘、硬盘、磁带)、光介质(例如,DVD)、或者半导体介质(例如固态硬盘(Solid State Disk,SSD))等。In the above embodiments, it may be implemented in whole or in part by software, hardware, firmware or any combination thereof. When implemented by software, it can be implemented in the form of a computer program product in whole or in part. The computer program product includes one or more computer instructions. When the computer program instructions are loaded and executed on the computer, the processes or functions described in the embodiments of the present application are generated in whole or in part. The computer may be a general-purpose computer, a special-purpose computer, a computer network, or other programmable devices. The computer instructions may be stored in a computer-readable storage medium or transmitted from one computer-readable storage medium to another computer-readable storage medium. For example, the computer instructions may be transmitted from a website, computer, server, or data center. Transmission to another website site, computer, server or data center via wired (such as coaxial cable, optical fiber, digital subscriber line (DSL)) or wireless (such as infrared, wireless, microwave, etc.). The computer-readable storage medium may be any available medium that can be accessed by a computer or a data storage device such as a server or data center integrated with one or more available media. The usable medium may be a magnetic medium (for example, a floppy disk, a hard disk, and a magnetic tape), an optical medium (for example, a DVD), or a semiconductor medium (for example, a solid state disk (SSD)).
需要说明的是,在本文中,诸如第一和第二等之类的关系术语仅仅用来将一个实体或者操作与另一个实体或操作区分开来,而不一定要求或者暗示这些实体或操作之间存在任何这种实际的关系或者顺序。而且,术语“包括”、“包含”或者其任何其他变体意在涵盖非排他性的包含,从而使得包括一系列要素的过程、方法、物品或者设备不仅包括那些要素,而且还包括没有明确列出的其他要素,或者是还包括为这种过程、方法、物品或者设备所固有的要素。在没有更多限制的情况下,由语句“包括一个……”限定的要素,并不排除在包括所述要素的过程、方法、物品或者设备中还存在另外的相同要素。It should be noted that in this article, relational terms such as first and second are only used to distinguish one entity or operation from another entity or operation, and do not necessarily require or imply one of these entities or operations. There is any such actual relationship or order between. Moreover, the terms "include", "include" or any other variants thereof are intended to cover non-exclusive inclusion, so that a process, method, article, or device that includes a series of elements includes not only those elements, but also includes Other elements of, or also include elements inherent to this process, method, article or equipment. If there are no more restrictions, the element defined by the sentence "including a..." does not exclude the existence of other same elements in the process, method, article, or equipment including the element.
本说明书中的各个实施例均采用相关的方式描述,各个实施例之间相同相似的部分互相参见即可,每个实施例重点说明的都是与其他实施例的不同之处。尤其,对于视频跟踪设备、装置、计算机可读存储介质和计算机程序产品实施例而言,由于其基本相似于方法实施例,所以描述的比较简单,相关之处参见方法实施例的部分说明即可。The various embodiments in this specification are described in a related manner, and the same or similar parts between the various embodiments can be referred to each other, and each embodiment focuses on the differences from other embodiments. In particular, for the embodiments of video tracking equipment, devices, computer-readable storage media, and computer program products, since they are basically similar to the method embodiments, the description is relatively simple. For related details, please refer to the description of the method embodiments .
以上所述仅为本申请的较佳实施例而已,并不用以限制本申请,凡在本申请的精神和原则之内,所做的任何修改、等同替换、改进等,均应包含在本申请保护的范围之内。The above are only the preferred embodiments of this application and are not intended to limit this application. Any modification, equivalent replacement, improvement, etc. made within the spirit and principle of this application shall be included in this application Within the scope of protection.

Claims (16)

  1. 一种视频跟踪设备的跟踪方法,其特征在于,所述视频跟踪设备支持多种跟踪算法;所述方法包括:A tracking method for a video tracking device, characterized in that the video tracking device supports multiple tracking algorithms; the method includes:
    从成像设备获取视频图像及表征所述视频图像的场景复杂度的参数;Acquiring a video image from an imaging device and a parameter that characterizes the scene complexity of the video image;
    基于所述参数在所述多种跟踪算法中选择适用于所述视频图像的跟踪算法;Selecting a tracking algorithm suitable for the video image from the multiple tracking algorithms based on the parameter;
    根据适用于所述视频图像的跟踪算法确定所述视频图像中的跟踪目标;Determining a tracking target in the video image according to a tracking algorithm applicable to the video image;
    提取所述跟踪目标在所述视频图像中的位置信息,基于所述位置信息计算所述跟踪目标与所述视频图像的中心的距离;Extracting position information of the tracking target in the video image, and calculating the distance between the tracking target and the center of the video image based on the position information;
    根据所述距离调整所述成像设备的拍摄角度,以使所述跟踪目标处于所述成像设备拍摄的下一视频图像的中心。The shooting angle of the imaging device is adjusted according to the distance, so that the tracking target is at the center of the next video image captured by the imaging device.
  2. 根据权利要求1所述的方法,其特征在于,所述成像设备中集成了成像机芯;The method of claim 1, wherein an imaging core is integrated in the imaging device;
    所述从成像设备获取视频图像及表征所述视频图像的场景复杂度的参数的步骤,包括:The step of obtaining a video image and a parameter characterizing the scene complexity of the video image from an imaging device includes:
    接收成像设备发送的由所述成像机芯拍摄的视频图像,并接收所述成像设备发送的所述视频图像的清晰度评价参数,其中,所述清晰度评价参数为表征所述视频图像的场景复杂度的参数。Receive a video image taken by the imaging core sent by an imaging device, and receive a sharpness evaluation parameter of the video image sent by the imaging device, wherein the sharpness evaluation parameter is a scene characterizing the video image Complexity parameter.
  3. 根据权利要求1所述的方法,其特征在于,所述基于所述参数在所述多种跟踪算法中选择适用于所述视频图像的跟踪算法的步骤,包括:The method according to claim 1, wherein the step of selecting a tracking algorithm suitable for the video image from the multiple tracking algorithms based on the parameter comprises:
    比较所述参数与预先设置的每种跟踪算法对应的参数取值范围;Compare the parameter value range corresponding to each preset tracking algorithm with the parameter;
    将对应于所述参数所属参数取值范围的跟踪算法确定为适用于所述视频图像的跟踪算法。The tracking algorithm corresponding to the parameter value range to which the parameter belongs is determined as the tracking algorithm suitable for the video image.
  4. 根据权利要求1所述的方法,其特征在于,所述多种跟踪算法包括:对比度跟踪算法、二值跟踪算法和相关跟踪算法;The method according to claim 1, wherein the multiple tracking algorithms include: a contrast tracking algorithm, a binary tracking algorithm, and a correlation tracking algorithm;
    适用于所述视频图像的目标跟踪算法为对比度跟踪算法或二值跟踪算法时,所述根据适用于所述视频图像的跟踪算法确定所述视频图像中的跟踪目标的步骤,包括:When the target tracking algorithm suitable for the video image is a contrast tracking algorithm or a binary tracking algorithm, the step of determining the tracking target in the video image according to the tracking algorithm suitable for the video image includes:
    针对所述视频图像中的每一检测目标,提取该检测目标对应的用于所述 目标跟踪算法的跟踪信息;For each detection target in the video image, extract tracking information corresponding to the detection target for the target tracking algorithm;
    根据所有检测目标对应的用于所述目标跟踪算法的跟踪信息,从所有检测目标中筛选出跟踪目标;Filter out the tracking target from all the detection targets according to the tracking information corresponding to all the detection targets for the target tracking algorithm;
    所述目标跟踪算法为相关跟踪算法时,所述根据适用于所述视频图像的跟踪算法确定所述视频图像中的跟踪目标的步骤,包括:When the target tracking algorithm is a related tracking algorithm, the step of determining the tracking target in the video image according to the tracking algorithm applicable to the video image includes:
    依据预先选定的包含有跟踪目标的模板图像,从所述视频图像的所有检测目标中筛选出与所述模板图像匹配度最大的检测目标,将筛选出的检测目标确定为所述视频图像中的跟踪目标。According to the pre-selected template image containing the tracking target, the detection target with the greatest degree of matching with the template image is selected from all detection targets in the video image, and the selected detection target is determined as the video image Tracking target.
  5. 根据权利要求1-4中任一项所述的方法,其特征在于,所述根据所述距离调整所述成像设备的拍摄角度的步骤,包括:The method according to any one of claims 1-4, wherein the step of adjusting the shooting angle of the imaging device according to the distance comprises:
    将所述距离发送到搭载所述视频跟踪设备的随动控制设备,以使所述随动控制设备根据所述距离执行随动控制操作。The distance is sent to a follow-up control device equipped with the video tracking device, so that the follow-up control device performs a follow-up control operation according to the distance.
  6. 一种视频跟踪设备,其特征在于,所述视频跟踪设备支持多种跟踪算法;所述视频跟踪设备包括非瞬时性计算机可读存储介质和处理器,其中,A video tracking device, characterized in that the video tracking device supports multiple tracking algorithms; the video tracking device includes a non-transitory computer-readable storage medium and a processor, wherein,
    所述非瞬时性计算机可读存储介质,用于存储可以被所述处理器执行的指令,在所述指令由所述处理器执行时,使得所述处理器:The non-transitory computer-readable storage medium is used to store instructions that can be executed by the processor, and when the instructions are executed by the processor, the processor is caused to:
    从成像设备获取视频图像及表征所述视频图像的场景复杂度的参数;Acquiring a video image from an imaging device and a parameter that characterizes the scene complexity of the video image;
    基于所述参数在所述多种跟踪算法中选择适用于所述视频图像的跟踪算法;Selecting a tracking algorithm suitable for the video image from the multiple tracking algorithms based on the parameter;
    根据适用于所述视频图像的跟踪算法确定所述视频图像中的跟踪目标;Determining a tracking target in the video image according to a tracking algorithm applicable to the video image;
    提取所述跟踪目标在所述视频图像中的位置信息,基于所述位置信息计算所述跟踪目标与所述视频图像的中心的距离;Extracting position information of the tracking target in the video image, and calculating the distance between the tracking target and the center of the video image based on the position information;
    根据所述距离调整所述成像设备的拍摄角度,以使所述跟踪目标处于所述成像设备拍摄的下一视频图像的中心。The shooting angle of the imaging device is adjusted according to the distance, so that the tracking target is at the center of the next video image captured by the imaging device.
  7. 根据权利要求6所述的设备,其特征在于,The device according to claim 6, wherein:
    所述成像设备中集成了成像机芯;An imaging core is integrated in the imaging device;
    所述处理器,从成像设备获取视频图像及表征所述视频图像的场景复杂度的参数时,包括:When the processor acquires a video image and a parameter representing the complexity of the scene of the video image from an imaging device, it includes:
    接收成像设备发送的由所述成像机芯拍摄的视频图像,并接收所述成像 设备发送的所述视频图像的清晰度评价参数,其中,所述清晰度评价参数为表征所述视频图像的场景复杂度的参数。Receive a video image taken by the imaging core sent by an imaging device, and receive a sharpness evaluation parameter of the video image sent by the imaging device, wherein the sharpness evaluation parameter is a scene characterizing the video image Complexity parameter.
  8. 根据权利要求6所述的设备,其特征在于,The device according to claim 6, wherein:
    所述处理器,基于所述参数在所述多种跟踪算法中选择适用于所述视频图像的跟踪算法时,包括:When the processor selects a tracking algorithm suitable for the video image from the multiple tracking algorithms based on the parameter, the method includes:
    比较所述参数与预先设置的每种跟踪算法对应的参数取值范围;Compare the parameter value range corresponding to each preset tracking algorithm with the parameter;
    将对应于所述参数所属参数取值范围的跟踪算法确定为适用于所述视频图像的跟踪算法。The tracking algorithm corresponding to the parameter value range to which the parameter belongs is determined as the tracking algorithm suitable for the video image.
  9. 根据权利要求6所述的设备,其特征在于,The device according to claim 6, wherein:
    所述多种跟踪算法包括:对比度跟踪算法、二值跟踪算法、和相关跟踪算法;The multiple tracking algorithms include: contrast tracking algorithm, binary tracking algorithm, and related tracking algorithm;
    适用于所述视频图像的目标跟踪算法为对比度跟踪算法或二值跟踪算法时,所述处理器根据适用于所述视频图像的跟踪算法确定所述视频图像中的跟踪目标时,包括:When the target tracking algorithm applicable to the video image is a contrast tracking algorithm or a binary tracking algorithm, when the processor determines the tracking target in the video image according to the tracking algorithm applicable to the video image, it includes:
    针对所述视频图像中的每一检测目标,提取该检测目标对应的用于所述目标跟踪算法的跟踪信息;For each detection target in the video image, extract tracking information corresponding to the detection target for the target tracking algorithm;
    根据所有检测目标的用于所述目标跟踪算法的跟踪信息,从所有检测目标中筛选出跟踪目标;Screening out the tracking target from all the detection targets according to the tracking information of all the detection targets used in the target tracking algorithm;
    所述目标跟踪算法为相关跟踪算法时,所述处理器根据适用于所述视频图像的跟踪算法确定所述视频图像中的跟踪目标时,包括:When the target tracking algorithm is a related tracking algorithm, when the processor determines the tracking target in the video image according to the tracking algorithm applicable to the video image, the method includes:
    依据预先选定的包含有跟踪目标的模板图像,从所述视频图像的所有检测目标中筛选出与所述模板图像匹配度最大的检测目标,将筛选出的检测目标确定为所述视频图像的跟踪目标。According to the pre-selected template image containing the tracking target, the detection target with the greatest degree of matching with the template image is selected from all the detection targets of the video image, and the screened detection target is determined as the video image. Track the target.
  10. 根据权利要求6-9中任一项所述的设备,其特征在于,The device according to any one of claims 6-9, characterized in that:
    所述处理器,根据所述距离调整所述成像设备的拍摄角度时,包括:将所述距离发送到搭载所述视频跟踪设备的随动控制设备,以使所述随动控制设备根据所述距离执行随动控制操作。When the processor adjusts the shooting angle of the imaging device according to the distance, it includes: sending the distance to a follow-up control device equipped with the video tracking device, so that the follow-up control device is Perform follow-up control operations from a distance.
  11. 一种视频跟踪设备的跟踪装置,其特征在于,所述视频跟踪设备支持多种跟踪算法;所述装置包括:A tracking device for a video tracking device, wherein the video tracking device supports multiple tracking algorithms; the device includes:
    信息获取模块,用于从成像设备获取视频图像及表征所述视频图像的场景复杂度的参数;An information acquisition module for acquiring a video image from an imaging device and parameters that characterize the scene complexity of the video image;
    算法选择模块,用于基于所述参数在所述多种跟踪算法中选择适用于所述视频图像的跟踪算法;An algorithm selection module, configured to select a tracking algorithm suitable for the video image from the multiple tracking algorithms based on the parameter;
    目标确定模块,用于根据适用于所述视频图像的跟踪算法确定所述视频图像中的跟踪目标;A target determination module, configured to determine a tracking target in the video image according to a tracking algorithm applicable to the video image;
    距离计算模块,用于提取所述跟踪目标在所述视频图像中的位置信息,基于所述位置信息计算所述跟踪目标与所述视频图像的中心的距离;A distance calculation module, configured to extract position information of the tracking target in the video image, and calculate the distance between the tracking target and the center of the video image based on the position information;
    角度调整模块,用于根据所述距离调整所述成像设备的拍摄角度,以使所述跟踪目标处于所述成像设备拍摄的下一视频图像的中心。The angle adjustment module is configured to adjust the shooting angle of the imaging device according to the distance, so that the tracking target is at the center of the next video image shot by the imaging device.
  12. 根据权利要求11所述的装置,其特征在于,所述成像设备中集成了成像机芯;The apparatus according to claim 11, wherein an imaging core is integrated in the imaging device;
    所述信息获取模块,具体用于:The information acquisition module is specifically used for:
    接收成像设备发送的由所述成像机芯拍摄的视频图像,并接收所述成像设备发送的所述视频图像的清晰度评价参数,其中,所述清晰度评价参数为表征所述视频图像的场景复杂度的参数。Receive a video image taken by the imaging core sent by an imaging device, and receive a sharpness evaluation parameter of the video image sent by the imaging device, wherein the sharpness evaluation parameter is a scene characterizing the video image Complexity parameter.
  13. 根据权利要求11所述的装置,其特征在于,所述算法选择模块,具体用于:The device according to claim 11, wherein the algorithm selection module is specifically configured to:
    比较所述参数与预先设置的每种跟踪算法对应的参数取值范围;Compare the parameter value range corresponding to each preset tracking algorithm with the parameter;
    将对应于所述参数所属参数取值范围的跟踪算法确定为适用于所述视频图像的跟踪算法。The tracking algorithm corresponding to the parameter value range to which the parameter belongs is determined as the tracking algorithm suitable for the video image.
  14. 根据权利要求11所述的装置,其特征在于,所述多种跟踪算法包括:对比度跟踪算法、二值跟踪算法和相关跟踪算法;The device according to claim 11, wherein the multiple tracking algorithms include: a contrast tracking algorithm, a binary tracking algorithm, and a correlation tracking algorithm;
    适用于所述视频图像的目标跟踪算法为对比度跟踪算法或二值跟踪算法时,所述目标确定模块,具体用于:When the target tracking algorithm suitable for the video image is a contrast tracking algorithm or a binary tracking algorithm, the target determination module is specifically used for:
    针对所述视频图像中的每一检测目标,提取该检测目标对应的用于所述目标跟踪算法的跟踪信息;For each detection target in the video image, extract tracking information corresponding to the detection target for the target tracking algorithm;
    根据所有检测目标的用于所述目标跟踪算法的跟踪信息,从所有检测目标中筛选出跟踪目标;Screening out the tracking target from all the detection targets according to the tracking information of all the detection targets used in the target tracking algorithm;
    所述目标跟踪算法为相关跟踪算法时,所述目标确定模块,具体用于:When the target tracking algorithm is a related tracking algorithm, the target determination module is specifically used for:
    依据预先选定的包含有跟踪目标的模板图像,从所述视频图像的所有检测目标中筛选出与所述模板图像匹配度最大的检测目标,将筛选出的检测目标确定为所述视频图像中的跟踪目标。According to the pre-selected template image containing the tracking target, the detection target with the greatest degree of matching with the template image is selected from all detection targets in the video image, and the selected detection target is determined as the video image Tracking target.
  15. 根据权利要求11-14中任一项所述的装置,其特征在于,所述角度调整模块,具体用于:The device according to any one of claims 11-14, wherein the angle adjustment module is specifically configured to:
    将所述距离发送到搭载所述视频跟踪设备的随动控制设备,以使所述随动控制设备根据所述距离执行随动控制操作。The distance is sent to a follow-up control device equipped with the video tracking device, so that the follow-up control device performs a follow-up control operation according to the distance.
  16. 一种计算机可读存储介质,其特征在于,所述计算机可读存储介质内存储有计算机程序,所述计算机程序被处理器执行时实现权利要求1-5中任一所述的方法步骤。A computer-readable storage medium, wherein a computer program is stored in the computer-readable storage medium, and when the computer program is executed by a processor, the method steps according to any one of claims 1-5 are realized.
PCT/CN2020/075481 2019-06-25 2020-02-17 Tracking method of video tracking device, and video tracking device WO2020258889A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN201910555419.6A CN112132858A (en) 2019-06-25 2019-06-25 Tracking method of video tracking equipment and video tracking equipment
CN201910555419.6 2019-06-25

Publications (1)

Publication Number Publication Date
WO2020258889A1 true WO2020258889A1 (en) 2020-12-30

Family

ID=73849405

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2020/075481 WO2020258889A1 (en) 2019-06-25 2020-02-17 Tracking method of video tracking device, and video tracking device

Country Status (2)

Country Link
CN (1) CN112132858A (en)
WO (1) WO2020258889A1 (en)

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113949830B (en) * 2021-09-30 2023-11-24 国家能源集团广西电力有限公司 Image processing method
CN115170615B (en) * 2022-09-02 2022-12-09 环球数科集团有限公司 High-speed visual system based on intelligent camera and target tracking algorithm thereof

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104184932A (en) * 2013-05-20 2014-12-03 浙江大华技术股份有限公司 Spherical camera control method and device thereof
CN108496350A (en) * 2017-09-27 2018-09-04 深圳市大疆创新科技有限公司 A kind of focusing process method and apparatus
CN109815844A (en) * 2018-12-29 2019-05-28 西安天和防务技术股份有限公司 Object detection method and device, electronic equipment and storage medium

Family Cites Families (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR20090053295A (en) * 2007-11-23 2009-05-27 주식회사 에스원 Method and system for monitoring video
TWI482123B (en) * 2009-11-18 2015-04-21 Ind Tech Res Inst Multi-state target tracking mehtod and system
US9501701B2 (en) * 2014-01-31 2016-11-22 The Charles Stark Draper Technology, Inc. Systems and methods for detecting and tracking objects in a video stream
CN107016367B (en) * 2017-04-06 2021-02-26 北京精英路通科技有限公司 Tracking control method and tracking control system
TWI618032B (en) * 2017-10-25 2018-03-11 財團法人資訊工業策進會 Object detection and tracking method and system

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104184932A (en) * 2013-05-20 2014-12-03 浙江大华技术股份有限公司 Spherical camera control method and device thereof
CN108496350A (en) * 2017-09-27 2018-09-04 深圳市大疆创新科技有限公司 A kind of focusing process method and apparatus
CN109815844A (en) * 2018-12-29 2019-05-28 西安天和防务技术股份有限公司 Object detection method and device, electronic equipment and storage medium

Also Published As

Publication number Publication date
CN112132858A (en) 2020-12-25

Similar Documents

Publication Publication Date Title
US20230077355A1 (en) Tracker assisted image capture
US10327045B2 (en) Image processing method, image processing device and monitoring system
US9124812B2 (en) Object image capture apparatus and method
KR101958116B1 (en) Image-processing device and method for foreground mask correction for object segmentation
US10573018B2 (en) Three dimensional scene reconstruction based on contextual analysis
US9628837B2 (en) Systems and methods for providing synchronized content
CN110691193B (en) Camera switching method and device, storage medium and electronic equipment
US8508605B2 (en) Method and apparatus for image stabilization
US11823404B2 (en) Structured light depth imaging under various lighting conditions
US8913103B1 (en) Method and apparatus for focus-of-attention control
KR20150110697A (en) Systems and methods for tracking and detecting a target object
WO2020258889A1 (en) Tracking method of video tracking device, and video tracking device
WO2022057670A1 (en) Real-time focusing method, apparatus and system, and computer-readable storage medium
JP2009093644A (en) Computer-implemented method for tacking 3d position of object moving in scene
WO2021008205A1 (en) Image processing
WO2019226673A1 (en) Missing feet recovery of a human object from an image sequence based on ground plane detection
US20220309627A1 (en) Face image straight line processing method, terminal device and storage medium
CN102129692A (en) Method and system for detecting motion target in double threshold scene
US9628720B2 (en) System and method to control an imaging process
US20230368343A1 (en) Global motion detection-based image parameter control
US20240046426A1 (en) Noise removal for surveillance camera image by means of ai-based object recognition
Guilluy et al. Feature trajectories selection for video stabilization
JP2002312792A (en) Image processor and image processing method and recording medium and its program
KR101432783B1 (en) Image pre-processing apparatus for evaluating camera capability
JP2022184575A5 (en)

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 20832359

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 20832359

Country of ref document: EP

Kind code of ref document: A1