WO2019041519A1 - Target tracking device and method, and computer-readable storage medium - Google Patents

Target tracking device and method, and computer-readable storage medium Download PDF

Info

Publication number
WO2019041519A1
WO2019041519A1 PCT/CN2017/108794 CN2017108794W WO2019041519A1 WO 2019041519 A1 WO2019041519 A1 WO 2019041519A1 CN 2017108794 W CN2017108794 W CN 2017108794W WO 2019041519 A1 WO2019041519 A1 WO 2019041519A1
Authority
WO
WIPO (PCT)
Prior art keywords
tracking
frame image
target
video
sample
Prior art date
Application number
PCT/CN2017/108794
Other languages
French (fr)
Chinese (zh)
Inventor
周舒意
王建明
肖京
Original Assignee
平安科技(深圳)有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 平安科技(深圳)有限公司 filed Critical 平安科技(深圳)有限公司
Publication of WO2019041519A1 publication Critical patent/WO2019041519A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/40Scenes; Scene-specific elements in video content
    • G06V20/48Matching video sequences
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/50Context or environment of the image
    • G06V20/52Surveillance or monitoring of activities, e.g. for recognising suspicious objects

Definitions

  • the present application relates to the field of image recognition technologies, and in particular, to a target tracking device, a method, and a computer readable storage medium based on a convolutional neural network.
  • Target tracking is an important part of practical applications such as video surveillance.
  • Target tracking refers to accurately locating and tracking moving targets (such as pedestrians, vehicles, etc.) in the video, and estimating the trajectory of the target.
  • target tracking has important value in video surveillance, target recognition, and video information discovery.
  • target tracking technology has been rapidly developed, but due to actual tracking, there are many practical difficulties in target tracking tasks, such as object occlusion, viewing angle changes, target deformation, ambient illumination changes, and unpredictable
  • the complex background situation, and the existing target tracking algorithm mostly uses the difference between the target and the background to construct the classification model, separates the target from the background, and tracks the target, but this tracking algorithm is difficult to adapt to the above mentioned in the tracking process. Changes in the target and background, such as partial occlusion of the target, or similar background interference, cause the target to be tracked incorrectly, resulting in low target tracking accuracy.
  • the present application provides a target tracking device, a method and a computer readable storage medium based on a convolutional neural network, the main purpose of which is to dynamically update a model during a tracking process to adapt to changes in targets and backgrounds, and to improve target tracking. Accuracy.
  • the present application provides a target tracking device based on a convolutional neural network, the device comprising a memory and a processor, the memory storing a target tracking program executable on the processor, the target The following steps are implemented when the tracking program is executed by the processor:
  • A. Collecting a plurality of picture samples from the video frame image according to the sampling point distribution, and recording position coordinates of each picture sample;
  • step D includes:
  • the processor is further configured to execute the target tracking program, to perform the following steps after the step E:
  • G adjusting the position of the sampling point on the video frame image according to the adjusted weight to update the sampling point distribution
  • the step F includes:
  • Steps A through G are repeated until the tracking of the tracking target in all video frame images of the video is completed.
  • the step G includes:
  • sampling point within a first preset range of the sampling point corresponding to the sample whose weight is greater than the first preset weight, and decreasing the sampling point in a second preset range of the sampling point corresponding to the sample whose weight is smaller than the second preset weight,
  • the second preset weight is smaller than the first preset weight, and the number of added sampling points is equal to the number of reduced sampling points.
  • the processor is further configured to execute the target tracking program to implement the following steps:
  • the video frame image is the first frame image of the video, prompting the user to manually select a tracking target on the video frame image and receive a tracking target selected by the user based on the prompt, and after determining the tracking target
  • step A is performed.
  • the present application further provides a target tracking method based on a convolutional neural network, the method comprising:
  • A. Collecting a plurality of picture samples from the video frame image according to the sampling point distribution, and recording position coordinates of each picture sample;
  • step D includes:
  • step E the method further includes:
  • G adjusting the position of the sampling point on the video frame image according to the adjusted weight to update the sampling point distribution
  • the step F includes:
  • Steps A through G are repeated until the tracking of the tracking target in all video frame images of the video is completed.
  • the step G includes:
  • sampling point within a first preset range of the sampling point corresponding to the sample whose weight is greater than the first preset weight, and decreasing the sampling point in a second preset range of the sampling point corresponding to the sample whose weight is smaller than the second preset weight,
  • the second preset weight is smaller than the first preset weight, and the number of added sampling points is equal to the number of reduced sampling points.
  • the present application further provides a computer readable storage medium having a target tracking program stored thereon, the target tracking program being executable by one or more processors to implement Next steps:
  • A. Collecting a plurality of picture samples from the video frame image according to the sampling point distribution, and recording position coordinates of each picture sample;
  • the object tracking device, the method and the computer readable storage medium based on the convolutional neural network proposed by the present application identify the video frame image in the video frame by frame, and collect multiple image samples from the video frame image according to the sampling point distribution, and Recording position coordinates of each picture sample, extracting a plurality of sample features correspondingly from the plurality of sample pictures based on the CNN model, calculating a confidence level between each picture sample and the tracking target according to the extracted sample features, and adjusting the sample according to the confidence level Weighting, and then calculating the position coordinates of the tracking target on the video frame image according to the position coordinates and weights of the sample, and collecting positive and negative samples of the tracking target from the video frame image according to the position coordinates, and retraining the CNN using the collected samples
  • the model updates the model parameters, continues to track the next frame image using the model after updating the model parameters, and so on.
  • the model After obtaining the tracking result of each frame image, the model is updated according to the tracking result, so that the tracking target changes.
  • the updated model can adapt to changes in goals and backgrounds. Even when the phenomenon of partial occlusion, background interference appear in the image, it can also be successfully tracking the target, improve the accuracy of target tracking.
  • FIG. 1 is a schematic diagram of a preferred embodiment of a target tracking device based on a convolutional neural network
  • FIG. 2 is a schematic diagram of a program module of a target tracking program in an embodiment of a target tracking device based on a convolutional neural network;
  • FIG. 3 is a flow chart of a preferred embodiment of a target tracking method based on a convolutional neural network.
  • the application provides a target tracking device based on a convolutional neural network.
  • a schematic diagram of a preferred embodiment of a target tracking device based on a convolutional neural network is provided.
  • the target tracking device based on the convolutional neural network may be a PC (Personal Computer), or may be a smart phone, a tablet computer, an e-book reader, or a portable device.
  • a terminal device having a display function such as a computer.
  • the convolutional neural network based target tracking device includes a memory 11, a processor 12, a communication bus 13, and a network interface 14.
  • the memory 11 includes at least one type of readable storage medium including a flash memory, a hard disk, a multimedia card, a card type memory (for example, an SD or DX memory, etc.), a magnetic memory, a magnetic disk, an optical disk, and the like.
  • Memory 11 may in some embodiments be an internal storage unit of a target tracking device based on a convolutional neural network, such as a hard disk of a target tracking device based on a convolutional neural network.
  • the memory 11 may also be an external storage device of a target tracking device based on a convolutional neural network in other embodiments, such as a plug-in hard disk equipped on a target tracking device based on a convolutional neural network, a smart memory card (Smart Media Card) , SMC), Secure Digital (SD) card, Flash Card, etc. Further, the memory 11 may also include both an internal storage unit of the target tracking device based on the convolutional neural network and an external storage device. The memory 11 can be used not only for storing application software and various types of data installed on a target tracking device based on a convolutional neural network, such as code of a target tracking program, but also for temporarily storing data that has been output or is to be output.
  • the processor 12 may be a Central Processing Unit (CPU), controller, microcontroller, microprocessor or other data processing chip for running program code or processing stored in the memory 11. Data, such as executing a target tracking program.
  • CPU Central Processing Unit
  • controller microcontroller
  • microprocessor or other data processing chip for running program code or processing stored in the memory 11.
  • Data such as executing a target tracking program.
  • Communication bus 13 is used to implement connection communication between these components.
  • the network interface 14 can optionally include a standard wired interface, a wireless interface (such as a WI-FI interface), and is typically used to establish a communication connection between the device and other electronic devices.
  • a standard wired interface such as a WI-FI interface
  • Figure 1 shows only a convolutional neural network based target tracking device with components 11-14 and a target tracking program, but it should be understood that not all illustrated components may be implemented, alternative implementations may be more or more Less components.
  • the device may further include a user interface
  • the user interface may include a display
  • an input unit such as a keyboard
  • the optional user interface may further include a standard wired interface and a wireless interface.
  • the display may be an LED display, a liquid crystal display, a touch-sensitive liquid crystal display, an OLED (Organic Light-Emitting Diode) touch sensor, or the like.
  • the display may also be suitably referred to as a display screen or display unit for displaying information processed in a target tracking device based on a convolutional neural network and a user interface for displaying visualization.
  • the device may also include a touch sensor.
  • the area provided by the touch sensor for the user to perform a touch operation is referred to as a touch area.
  • the touch sensor described herein may be a resistive touch sensor, a capacitive touch sensor, or the like.
  • the touch sensor includes not only a contact type touch sensor but also a proximity type touch sensor or the like.
  • the touch sensor may be a single sensor or a plurality of sensors arranged, for example, in an array.
  • the area of the display of the device may be the same as or different from the area of the touch sensor.
  • a display is stacked with the touch sensor to form a touch display. The device is based on a touch display The screen detects the touch operation triggered by the user.
  • the device may further include a camera, an RF (Radio Frequency) circuit, a sensor, an audio circuit, a WiFi module, and the like.
  • sensors such as light sensors, motion sensors, and other sensors.
  • the light sensor may include an ambient light sensor and a proximity sensor, wherein if the device is a mobile terminal, the ambient light sensor may adjust the brightness of the display screen according to the brightness of the ambient light, and the proximity sensor may move when the mobile terminal moves to the ear. , turn off the display and / or backlight.
  • the gravity acceleration sensor can detect the magnitude of acceleration in each direction (usually three axes), and can detect the magnitude and direction of gravity when stationary, and can be used to identify the posture of the mobile terminal (such as horizontal and vertical screen switching, Related games, magnetometer attitude calibration), vibration recognition related functions (such as pedometer, tapping), etc.; of course, the mobile terminal can also be equipped with other sensors such as gyroscope, barometer, hygrometer, thermometer, infrared sensor, etc. No longer.
  • a target tracking program is stored in the memory 11; when the processor 12 executes the target tracking program stored in the memory 11, the following steps are implemented:
  • A. Collecting a plurality of picture samples from the video frame image according to a sampling point distribution.
  • a convolutional neural network is used to perform offline training on a massive picture to obtain a CNN (Convolutional Neural Network) model, which may be a two-class model, and the model can be extracted from the image.
  • CNN Convolutional Neural Network
  • the video image is tracked frame by frame.
  • a video to be subjected to target tracking is input to the device, and the device processes each video frame image in the video in accordance with the following operation.
  • the image samples are collected from the video frame image according to the sampling point distribution, wherein the number of sampling points can be preset by the user, for example, 100 image samples are collected, wherein when the first frame image is started to be recognized, the user can manually
  • the tracking target is selected in the image.
  • the tracking target is selected by a frame selection manner, and the sampling point distribution is initialized based on the position of the tracking target selected by the user.
  • the video frame image when the video frame image is received, determining whether the video frame image is the first frame image of the video; if the video frame image is the first frame image of the video, prompting the user to Manually selecting a tracking target on the video frame image and receiving a tracking target selected by the user based on the prompt; after determining the tracking target, initializing a sampling point distribution and a training sample set of the CNN model and receiving a second frame image; The video image is not the first frame image of the video, and the step A is performed.
  • the user sets the target to be tracked in advance and stores it, the tracking target is directly acquired after starting the tracking, and the user is not required to manually select from the first frame image.
  • the color histogram of the tracking target area is calculated and used as the target feature of the tracking target, and the target feature can be represented as a vector of N*1.
  • the collected sample image is input into the trained CNN model for feature extraction, and the sample feature is extracted, and the sample feature can also be represented as an N*1 vector.
  • a sample feature is extracted corresponding to each sample picture, and the confidence between each sample feature and the target feature is calculated separately.
  • the confidence of the sample features reflects the similarity between the image sample and the tracking target.
  • the similarity between the two N*1 vectors is calculated as a sample of the image. Confidence with tracking targets.
  • the weight of each picture sample is adjusted according to the confidence level, and for the sample with small confidence, the weight is reduced, and for the sample with high confidence, the weight is increased, and then for all the weights, then for all
  • the weights of the image samples are normalized such that the sum of the weights of all samples is equal to one.
  • the position coordinates of the tracking target on the video frame image are calculated based on the weight value of the picture sample and its position coordinates on the video frame image. Specifically, it is assumed that a total of k picture samples are collected, wherein the position coordinates of the sample P i are (x i , y i ), and the confidence degree with the tracking target is S i . Then, the position coordinates (x, y) of the tracking target can be predicted according to the following formula.
  • D. Collect positive and negative samples of the tracking target from the video frame image according to the position coordinates.
  • Collecting a positive sample and a negative sample of the tracking target from the video frame image according to the position coordinate specifically, acquiring a first preset number of picture samples located in a peripheral area of the position coordinate as a positive sample, wherein the periphery The area is an area formed by a point having a distance from the position coordinate that is smaller than a first preset threshold; and a second predetermined number of picture samples located in a distant area of the position coordinate are collected as a negative sample, wherein the The remote preset area is an area formed by a point that is greater than a second preset threshold value, and the second preset threshold is greater than the first preset threshold.
  • the image samples are collected from the region closer to the tracking target, and the difference between the samples and the tracking target is small, and can be used as a positive sample from the video frame image. Capture image samples from areas farther away from the tracking target.
  • the difference between the tracking targets is large, and can be added as a negative sample to the training sample set of the CNN model, and the CNN model is trained to update the model parameters and improve the accuracy of the model to identify the features of the moving target from the image samples. To enable the model to adapt to changes in the target and background in the video frame image.
  • the CNN model is continuously updated, and even if the tracking target is partially occluded or the background interferes with the tracking target, it does not interfere with the correct tracking of the target.
  • the next frame image is continuously tracked, and the updated CNN model is used for feature extraction.
  • Target tracking is performed for each frame of image according to steps A through E, and after the tracking is completed, the CNN model is trained until the tracking of the target in all frame images of the video is completed.
  • the first preset threshold, the second preset threshold, the first preset number, and the second preset number may be preset by use.
  • step E the following steps are further implemented:
  • G adjusting the position of the sampling point on the video frame image according to the adjusted weight to update the sampling point distribution
  • the step F includes repeating steps A to G until the tracking of the tracking target in all video frame images of the video is completed.
  • the distribution of the sampling points is adjusted according to the adjusted weights. Specifically, the sampling points are added in the first preset range of the sampling points corresponding to the samples whose weight is greater than the first preset weight, that is, the pictures with the weights are significant. Adding more sampling points near the sampling point corresponding to the sample, and reducing the sampling point in the second preset range of the sampling point corresponding to the sample whose weight is smaller than the second preset weight, wherein the second preset weight is smaller than the first preset
  • the weight that is, the sampling point near the sampling point corresponding to the picture sample with small weight reduction, wherein the number of added sampling points is equal to or greater than the number of reduced sampling points, or, when the weight is very small, the corresponding sampling point can be deleted. For example, the sampling point corresponding to the sample whose weight is smaller than the third preset weight is deleted, wherein the third preset weight is smaller than the fourth preset weight.
  • the target tracking device based on the convolutional neural network proposed in this embodiment performs frame-by-frame recognition on the video frame image in the video, collects multiple picture samples from the video frame image according to the sampling point distribution, and records the position coordinates of each picture sample. And extracting a plurality of sample features correspondingly from the plurality of sample pictures based on the CNN model, calculating a confidence degree between each picture sample and the tracking target according to the extracted sample features, adjusting the weight of the sample according to the confidence degree, and further according to the position of the sample Coordinates and weights calculate the position coordinates of the tracking target on the video frame image, and collect positive and negative samples of the tracking target from the video frame image according to the position coordinates, and retrain the CNN model to update the model parameters using the collected samples, using After updating the model parameters, the model continues to track the next frame image, and so on.
  • the model After obtaining the tracking result of each frame image, the model is updated according to the tracking result, so that the updated model can be changed when the tracking target changes.
  • Adapt to changes in goals and background, even if partial obscuration occurs in the image When the phenomenon of background interference, but also able to successfully track the target, improve the accuracy of target tracking.
  • the target tracking program may also be divided into one or more modules, one or more modules being stored in the memory 11 and being processed by one or more processors (this Embodiments are executed by processor 12) to accomplish the present application, and a module referred to herein refers to a series of computer program instructions that are capable of performing a particular function.
  • FIG. 2 it is a schematic diagram of a program module of a target tracking program in an embodiment of a target tracking device based on a convolutional neural network according to the present application.
  • the target tracking program may be divided into a preprocessing module 10, Tracking module 20, sampling module 30 and update module 40, illustratively,
  • the collecting module 10 is configured to: collect a plurality of picture samples from the video frame image according to the sampling point distribution, and record position coordinates of each picture sample;
  • the pre-processing module 20 is configured to: correspondingly extract a plurality of sample features from the plurality of picture samples based on a convolutional neural network CNN model, and respectively calculate a confidence between each picture sample and the tracking target according to the extracted sample features respectively degree;
  • the tracking module 30 is configured to: adjust weights of the corresponding picture samples according to the calculated confidence, and calculate position coordinates of the tracking target on the video frame image according to position coordinates and weights of all picture samples;
  • the sampling module 40 is configured to: collect positive and negative samples of the tracking target from the video frame image according to the position coordinates;
  • the updating module 50 is configured to: update the training sample set of the CNN model according to the positive sample and the negative sample, and train the CNN model to update model parameters of the CNN model by using the updated training sample set;
  • the acquisition module 10, the pre-processing module 20, the tracking module 30, the sampling module 40, and the update module 50 perform the above steps to track the target in the order of the video frame images in the video until the tracking target in all the video frame images in the video is completed. Tracking.
  • the present application also provides a target tracking method based on a convolutional neural network.
  • FIG. 3 it is a flowchart of a preferred embodiment of a target tracking method based on a convolutional neural network. The method can be performed by a device that can be implemented by software and/or hardware.
  • the target tracking method based on the convolutional neural network includes:
  • Step S10 Collect a plurality of picture samples from the video frame image according to the sampling point distribution, and record position coordinates of each picture sample.
  • a convolutional neural network is used to perform offline training on a massive picture to obtain a CNN (Convolutional Neural Network) model, which may be a two-class model, and the model can be extracted from the image.
  • CNN Convolutional Neural Network
  • the video image is tracked frame by frame.
  • a video to be subjected to target tracking is input to the device, and the device processes each video frame image in the video in accordance with the following operation.
  • the image samples are collected from the video frame image according to the sampling point distribution, wherein the number of sampling points can be preset by the user, for example, 100 image samples are collected, wherein when the first frame image is started to be recognized, the user can manually
  • the tracking target is selected in the image.
  • the tracking target is selected by a frame selection manner, and the sampling point distribution is initialized based on the position of the tracking target selected by the user.
  • the video frame image when the video frame image is received, determining whether the video frame image is the first frame image of the video; if the video frame image is the first frame image of the video, prompting the user to Manually selecting a tracking target on the video frame image and receiving a tracking target selected by the user based on the prompt; after determining the tracking target, initializing a sampling point distribution and a training sample set of the CNN model and receiving a second frame image; The video image is not the first frame image of the video, and the step S10 is performed.
  • the user sets the target to be tracked in advance and stores it, the tracking target is directly acquired after starting the tracking, and the user is not required to manually select from the first frame image.
  • the color histogram of the tracking target area is calculated and used as the target feature of the tracking target, and the target feature can be represented as a vector of N*1.
  • Step S20 correspondingly extracting a plurality of sample features from the plurality of picture samples based on the convolutional neural network CNN model, and respectively calculating a confidence level between each picture sample and the tracking target according to the extracted sample features.
  • Step S30 adjusting the weight of the corresponding picture sample according to the calculated confidence, and calculating the position coordinates of the tracking target on the video frame image according to the position coordinates and weights of all the picture samples.
  • the collected sample image is input into the trained CNN model for feature extraction, and the sample feature is extracted, and the sample feature can also be represented as an N*1 vector.
  • a sample feature is extracted corresponding to each sample picture, and the confidence between each sample feature and the target feature is calculated separately.
  • the confidence of the sample features reflects the similarity between the image sample and the tracking target.
  • the similarity between the two N*1 vectors is calculated as a sample of the image. Confidence with tracking targets.
  • the weight of each picture sample is adjusted according to the confidence level, and for the sample with small confidence, the weight is reduced, and for the sample with high confidence, the weight is increased, and then for all the weights, then for all
  • the weights of the image samples are normalized such that the sum of the weights of all samples is equal to one.
  • the position coordinates of the tracking target on the video frame image are calculated based on the weight value of the picture sample and its position coordinates on the video frame image. Specifically, it is assumed that a total of k picture samples are collected, wherein the position coordinates of the sample P i are (x i , y i ), and the confidence degree with the tracking target is S i . Then, the position coordinates (x, y) of the tracking target can be predicted according to the following formula.
  • Step S40 collecting positive and negative samples of the tracking target from the video frame image according to the position coordinates.
  • Step S50 updating the training sample set of the CNN model according to the positive sample and the negative sample, and training the CNN model to update the model parameters of the CNN model by using the updated training sample set.
  • step S60 steps S10 to S50 are repeatedly performed until the tracking of the tracking target in all the video frame images of the video is completed.
  • Collecting a positive sample and a negative sample of the tracking target from the video frame image according to the position coordinate specifically, acquiring a first preset number of picture samples located in a peripheral area of the position coordinate as a positive sample, wherein the periphery The area is an area formed by a point having a distance from the position coordinate that is smaller than a first preset threshold; and a second predetermined number of picture samples located in a distant area of the position coordinate are collected as a negative sample, wherein the The remote preset area is an area formed by a point that is greater than a second preset threshold value, and the second preset threshold is greater than the first preset threshold.
  • the image samples are collected from the region closer to the tracking target, and the difference between the samples and the tracking target is small, and can be used as a positive sample from the video frame image.
  • Image samples are taken from areas farther away from the tracking target. These samples have a large difference from the tracking targets. They can be added as negative samples to the training sample set of the CNN model, and the CNN model is trained to update the model parameters.
  • the improved model identifies the accuracy of the features of the moving object from the image samples so that the model can adapt to changes in the target and background in the video frame image.
  • the CNN model is continuously updated, and even if the tracking target is partially occluded or the background interferes with the tracking target, it does not interfere with the correct tracking of the target.
  • the next frame image is continuously tracked, and the updated CNN model is used for feature extraction.
  • Target tracking is performed for each frame of image in accordance with steps S10 through S40, and after the tracking is completed, the CNN model is trained until all tracking of the target in all frame images of the video is completed.
  • the first preset threshold, the second preset threshold, the first preset number, and the second preset number may be preset by use.
  • the method further includes the following steps: adjusting the distribution of the sampling points according to the adjusted weights, specifically, the sampling corresponding to the samples having the weight greater than the first preset weight Adding a sampling point within a first preset range of the point, that is, adding more sampling points near the sampling point corresponding to the image sample with a large weight, and second sampling of the sampling point corresponding to the sample whose weight is smaller than the second preset weight
  • the sampling point is reduced in the range, wherein the second preset weight is smaller than the first preset weight, that is, the sampling point near the sampling point corresponding to the image sample with small weight is reduced, wherein
  • the number of added sampling points is equal to or greater than the number of reduced sampling points, or, when the weight is very small, the corresponding sampling points may be deleted, for example, the sampling points corresponding to the samples whose weights are smaller than the third preset weight are deleted, wherein The third preset weight is smaller than the fourth preset weight.
  • the target tracking method based on the convolutional neural network proposed in this embodiment performs frame-by-frame recognition on the video frame image in the video, collects multiple picture samples from the video frame image according to the sampling point distribution, and records the position coordinates of each picture sample. And extracting a plurality of sample features correspondingly from the plurality of sample pictures based on the CNN model, calculating a confidence degree between each picture sample and the tracking target according to the extracted sample features, adjusting the weight of the sample according to the confidence degree, and further according to the position of the sample Coordinates and weights calculate the position coordinates of the tracking target on the video frame image, and collect positive and negative samples of the tracking target from the video frame image according to the position coordinates, and retrain the CNN model to update the model parameters using the collected samples, using After updating the model parameters, the model continues to track the next frame image, and so on.
  • the model After obtaining the tracking result of each frame image, the model is updated according to the tracking result, so that the updated model can be changed when the tracking target changes.
  • Adapt to changes in goals and background, even if partial obscuration occurs in the image When the phenomenon of background interference, but also able to successfully track the target, improve the accuracy of target tracking.
  • the embodiment of the present application further provides a computer readable storage medium, where the target readable program is stored on the computer readable storage medium, and the target tracking program can be executed by one or more processors to implement the following operations:
  • A. Collecting a plurality of picture samples from the video frame image according to the sampling point distribution, and recording position coordinates of each picture sample;
  • the position of the sampling point on the video frame image is adjusted according to the adjusted weight to update the sampling point distribution.
  • sampling point within a first preset range of the sampling point corresponding to the sample whose weight is greater than the first preset weight, and decreasing the sampling point in a second preset range of the sampling point corresponding to the sample whose weight is smaller than the second preset weight,
  • the second preset weight is smaller than the first preset weight, and the number of added sampling points is equal to the number of reduced sampling points.
  • the specific embodiment of the computer readable storage medium of the present application is substantially the same as the foregoing embodiments of the target tracking apparatus and method based on the convolutional neural network, and is not described herein.
  • the technical solution of the present application which is essential or contributes to the prior art, may be embodied in the form of a software product stored in a storage medium (such as ROM/RAM as described above). , a disk, an optical disk, including a number of instructions for causing a terminal device (which may be a mobile phone, a computer, a server, or a network device, etc.) to perform the methods described in the various embodiments of the present application.
  • a terminal device which may be a mobile phone, a computer, a server, or a network device, etc.

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Multimedia (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Computation (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Artificial Intelligence (AREA)
  • General Engineering & Computer Science (AREA)
  • Biomedical Technology (AREA)
  • Computational Linguistics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Health & Medical Sciences (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Biophysics (AREA)
  • Evolutionary Biology (AREA)
  • General Health & Medical Sciences (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Image Analysis (AREA)

Abstract

A target tracking device and method based on a convolutional neural network, and a computer-readable storage medium, these being able to improve the accuracy of target tracking. The device comprises a memory and a processor, wherein a target tracking program capable of running on the processor is stored in the memory, and when the program is executed by the processor, the following steps are implemented: collecting picture samples from a video frame image according to a sampling point distribution, and recording the position coordinates of the picture samples (S10); extracting, based on a CNN model, sample features from the picture samples, and calculating a degree of confidence of the picture samples and a tracked target according to the sample features (S20); adjusting the weight of the picture samples according to the degree of confidence, and calculating the position coordinates of the tracked target according to the position coordinates and the weight (S30); collecting a positive sample and a negative sample from the video frame image according to the position coordinates to train a CNN model with a training sample set (S40); then updating model parameters of the CNN model (S50); and repeating the above-mentioned steps until the tracking of a video is complete (S60).

Description

目标跟踪装置、方法及计算机可读存储介质Target tracking device, method and computer readable storage medium
优先权申明Priority claim
本申请基于巴黎公约申明享有2017年08月29日递交的申请号为201710754313.X、名称为“目标跟踪装置、方法及计算机可读存储介质”的中国专利申请的优先权,该中国专利申请的整体内容以参考的方式结合在本申请中。The present application is based on the priority of the Chinese Patent Application entitled "Target Tracking Apparatus, Method and Computer Readable Storage Medium", filed on Aug. 29, 2017, filed on Aug. 29, 2017, entitled The entire content is incorporated herein by reference.
技术领域Technical field
本申请涉及图像识别技术领域,尤其涉及一种基于卷积神经网络的目标跟踪装置、方法及计算机可读存储介质。The present application relates to the field of image recognition technologies, and in particular, to a target tracking device, a method, and a computer readable storage medium based on a convolutional neural network.
背景技术Background technique
计算机目标跟踪是视频监控等实际应用中的重要组成部分,目标跟踪是指对视频中的运动目标(例如行人、车辆等)的进行准确定位、跟踪,并且推测目标的轨迹。目标跟踪作为计算机视觉领域中的一个重要课题,在视频监控、目标识别、视频信息发现等方面有重要的价值。Computer target tracking is an important part of practical applications such as video surveillance. Target tracking refers to accurately locating and tracking moving targets (such as pedestrians, vehicles, etc.) in the video, and estimating the trajectory of the target. As an important topic in the field of computer vision, target tracking has important value in video surveillance, target recognition, and video information discovery.
随着大量目标跟踪算法的提出,目标跟踪技术得到了快速地发展,但是由于在实际跟踪中,目标跟踪任务存在很多现实困难,例如物体遮挡、视角变化、目标形变、周围光照变化以及难以预料的复杂的背景情况,而现有的目标跟踪算法多是利用目标与背景的差异构建分类模型,把目标从背景中分离出来,对目标进行跟踪,但是这种跟踪算法在跟踪过程中难以适应上述提到的目标及背景的变化,例如目标被部分遮挡,或者相似背景干扰等问题,造成目标的错误跟踪,导致目标跟踪准确度低。With the introduction of a large number of target tracking algorithms, target tracking technology has been rapidly developed, but due to actual tracking, there are many practical difficulties in target tracking tasks, such as object occlusion, viewing angle changes, target deformation, ambient illumination changes, and unpredictable The complex background situation, and the existing target tracking algorithm mostly uses the difference between the target and the background to construct the classification model, separates the target from the background, and tracks the target, but this tracking algorithm is difficult to adapt to the above mentioned in the tracking process. Changes in the target and background, such as partial occlusion of the target, or similar background interference, cause the target to be tracked incorrectly, resulting in low target tracking accuracy.
发明内容Summary of the invention
本申请提供一种基于卷积神经网络的目标跟踪装置、方法及计算机可读存储介质,其主要目的在于在跟踪过程中对模型进行动态的更新,以适应目标和背景的变化,提高目标跟踪的准确度。The present application provides a target tracking device, a method and a computer readable storage medium based on a convolutional neural network, the main purpose of which is to dynamically update a model during a tracking process to adapt to changes in targets and backgrounds, and to improve target tracking. Accuracy.
为实现上述目的,本申请提供一种基于卷积神经网络的目标跟踪装置,该装置包括存储器和处理器,所述存储器上存储有可在所述处理器上运行的目标跟踪程序,所述目标跟踪程序被所述处理器执行时实现如下步骤:To achieve the above object, the present application provides a target tracking device based on a convolutional neural network, the device comprising a memory and a processor, the memory storing a target tracking program executable on the processor, the target The following steps are implemented when the tracking program is executed by the processor:
A、按照采样点分布从视频帧图像上采集多个图片样本,并记录各个图片样本的位置坐标;A. Collecting a plurality of picture samples from the video frame image according to the sampling point distribution, and recording position coordinates of each picture sample;
B、基于卷积神经网络CNN模型从所述多个图片样本中对应地提取多个样本特征,并分别根据提取的样本特征分别计算每一图片样本与跟踪目标之间的置信度;B. Extracting a plurality of sample features from the plurality of picture samples based on the convolutional neural network CNN model, and respectively calculating a confidence level between each picture sample and the tracking target according to the extracted sample features;
C、根据计算得出的置信度调整对应图片样本的权重,并根据所有图片样本的位置坐标和调整后的权重计算所述跟踪目标在所述视频帧图像上的位置 坐标;C. Adjusting the weight of the corresponding picture sample according to the calculated confidence, and calculating the position of the tracking target on the video frame image according to the position coordinates of all the picture samples and the adjusted weight coordinate;
D、根据所述位置坐标从所述视频帧图像上采集所述跟踪目标的正样本和负样本;D. Collecting positive and negative samples of the tracking target from the video frame image according to the position coordinates;
E、根据所述正样本和负样本更新所述CNN模型的训练样本集,并使用更新后的训练样本集训练所述CNN模型以更新所述CNN模型的模型参数;E. Updating a training sample set of the CNN model according to the positive sample and the negative sample, and training the CNN model with the updated training sample set to update model parameters of the CNN model;
F、重复执行步骤A至E,直至完成对视频的所有视频帧图像中跟踪目标的跟踪。F. Repeat steps A through E until the tracking of the tracking target in all video frame images of the video is completed.
可选地,所述步骤D包括:Optionally, the step D includes:
采集位于所述位置坐标的周边区域内的第一预设数量的图片样本作为正样本,其中,所述周边区域为与所述位置坐标之间的距离小于第一预设阈值的点构成的区域;And acquiring a first preset number of picture samples located in a peripheral area of the position coordinate as a positive sample, wherein the peripheral area is an area formed by a point that is smaller than a first preset threshold between the position coordinates ;
采集位于所述位置坐标的远离区域内的第二预设数量的图片样本作为负样本,其中,所述远离区域为与所述位置坐标之间的距离大于第二预设阈值的点构成的区域,所述第二预设阈值大于所述第一预设阈值。And acquiring a second preset number of picture samples located in a distant area of the position coordinate as a negative sample, wherein the distant area is an area formed by a point that is greater than a second preset threshold between the position coordinates The second preset threshold is greater than the first preset threshold.
可选地,所述处理器还用于执行所述目标跟踪程序,以在步骤E之后,还实现如下步骤:Optionally, the processor is further configured to execute the target tracking program, to perform the following steps after the step E:
G、根据调整后的权重调整采样点在视频帧图像上的位置,以更新采样点分布;G, adjusting the position of the sampling point on the video frame image according to the adjusted weight to update the sampling point distribution;
所述步骤F包括:The step F includes:
重复执行步骤A至G,直至完成对视频的所有视频帧图像中的跟踪目标的跟踪。Steps A through G are repeated until the tracking of the tracking target in all video frame images of the video is completed.
可选地,所述步骤G包括:Optionally, the step G includes:
在权重大于第一预设权重的样本对应的采样点的第一预设范围内增加采样点,在权重小于第二预设权重的样本对应的采样点的第二预设范围内减少采样点,其中,所述第二预设权重小于所述第一预设权重,增加的采样点的数量等于减少的采样点的数量。Adding a sampling point within a first preset range of the sampling point corresponding to the sample whose weight is greater than the first preset weight, and decreasing the sampling point in a second preset range of the sampling point corresponding to the sample whose weight is smaller than the second preset weight, The second preset weight is smaller than the first preset weight, and the number of added sampling points is equal to the number of reduced sampling points.
可选地,所述处理器还用于执行所述目标跟踪程序,以实现如下步骤:Optionally, the processor is further configured to execute the target tracking program to implement the following steps:
判断所述视频帧图像是否为所述视频的第一帧图像;Determining whether the video frame image is the first frame image of the video;
若所述视频帧图像为所述视频的第一帧图像,则提示用户在所述视频帧图像上手动选择跟踪目标并接收用户基于所述提示选择的跟踪目标,并在确定所述跟踪目标后,初始化采样点分布和所述CNN模型的训练样本集并接收第二帧图像;If the video frame image is the first frame image of the video, prompting the user to manually select a tracking target on the video frame image and receive a tracking target selected by the user based on the prompt, and after determining the tracking target Initializing a sample point distribution and a training sample set of the CNN model and receiving a second frame image;
若所述视频图像不是所述视频的第一帧图像,则执行所述步骤A。If the video image is not the first frame image of the video, the step A is performed.
此外,为实现上述目的,本申请还提供一种基于卷积神经网络的目标跟踪方法,该方法包括:In addition, to achieve the above object, the present application further provides a target tracking method based on a convolutional neural network, the method comprising:
A、按照采样点分布从视频帧图像上采集多个图片样本,并记录各个图片样本的位置坐标; A. Collecting a plurality of picture samples from the video frame image according to the sampling point distribution, and recording position coordinates of each picture sample;
B、基于卷积神经网络CNN模型从所述多个图片样本中对应地提取多个样本特征,并分别根据提取的样本特征分别计算每一图片样本与跟踪目标之间的置信度;B. Extracting a plurality of sample features from the plurality of picture samples based on the convolutional neural network CNN model, and respectively calculating a confidence level between each picture sample and the tracking target according to the extracted sample features;
C、根据计算得出的置信度调整对应图片样本的权重,并根据所有图片样本的位置坐标和权重计算所述跟踪目标在所述视频帧图像上的位置坐标;C. Adjusting weights of the corresponding picture samples according to the calculated confidence, and calculating position coordinates of the tracking target on the video frame image according to position coordinates and weights of all picture samples;
D、根据所述位置坐标从所述视频帧图像上采集所述跟踪目标的正样本和负样本;D. Collecting positive and negative samples of the tracking target from the video frame image according to the position coordinates;
E、根据所述正样本和负样本更新所述CNN模型的训练样本集,并使用更新后的训练样本集训练所述CNN模型以更新所述CNN模型的模型参数;E. Updating a training sample set of the CNN model according to the positive sample and the negative sample, and training the CNN model with the updated training sample set to update model parameters of the CNN model;
F、重复执行步骤A至E,直至完成对视频的所有视频帧图像中跟踪目标的跟踪。F. Repeat steps A through E until the tracking of the tracking target in all video frame images of the video is completed.
可选地,所述步骤D包括:Optionally, the step D includes:
采集位于所述位置坐标的周边区域内的第一预设数量的图片样本作为正样本,其中,所述周边区域为与所述位置坐标之间的距离小于第一预设阈值的点构成的区域;And acquiring a first preset number of picture samples located in a peripheral area of the position coordinate as a positive sample, wherein the peripheral area is an area formed by a point that is smaller than a first preset threshold between the position coordinates ;
采集位于所述位置坐标的远离区域内的第二预设数量的图片样本作为负样本,其中,所述远离区域为与所述位置坐标之间的距离大于第二预设阈值的点构成的区域,所述第二预设阈值大于所述第一预设阈值。And acquiring a second preset number of picture samples located in a distant area of the position coordinate as a negative sample, wherein the distant area is an area formed by a point that is greater than a second preset threshold between the position coordinates The second preset threshold is greater than the first preset threshold.
可选地,在步骤E之后,该方法还包括:Optionally, after step E, the method further includes:
G、根据调整后的权重调整采样点在视频帧图像上的位置,以更新采样点分布;G, adjusting the position of the sampling point on the video frame image according to the adjusted weight to update the sampling point distribution;
所述步骤F包括:The step F includes:
重复执行步骤A至G,,直至完成对视频的所有视频帧图像中的跟踪目标的跟踪。Steps A through G are repeated until the tracking of the tracking target in all video frame images of the video is completed.
可选地,所述步骤G包括:Optionally, the step G includes:
在权重大于第一预设权重的样本对应的采样点的第一预设范围内增加采样点,在权重小于第二预设权重的样本对应的采样点的第二预设范围内减少采样点,其中,所述第二预设权重小于所述第一预设权重,增加的采样点的数量等于减少的采样点的数量。Adding a sampling point within a first preset range of the sampling point corresponding to the sample whose weight is greater than the first preset weight, and decreasing the sampling point in a second preset range of the sampling point corresponding to the sample whose weight is smaller than the second preset weight, The second preset weight is smaller than the first preset weight, and the number of added sampling points is equal to the number of reduced sampling points.
此外,为实现上述目的,本申请还提供一种计算机可读存储介质,所述计算机可读存储介质上存储有目标跟踪程序,所述目标跟踪程序可被一个或多个处理器执行,以实现下步骤:In addition, in order to achieve the above object, the present application further provides a computer readable storage medium having a target tracking program stored thereon, the target tracking program being executable by one or more processors to implement Next steps:
A、按照采样点分布从视频帧图像上采集多个图片样本,并记录各个图片样本的位置坐标;A. Collecting a plurality of picture samples from the video frame image according to the sampling point distribution, and recording position coordinates of each picture sample;
B、基于卷积神经网络CNN模型从所述多个图片样本中对应地提取多个样本特征,并分别根据提取的样本特征分别计算每一图片样本与跟踪目标之间的置信度; B. Extracting a plurality of sample features from the plurality of picture samples based on the convolutional neural network CNN model, and respectively calculating a confidence level between each picture sample and the tracking target according to the extracted sample features;
C、根据计算得出的置信度调整对应图片样本的权重,并根据所有图片样本的位置坐标和调整后的权重计算所述跟踪目标在所述视频帧图像上的位置坐标;C. Adjust the weight of the corresponding picture sample according to the calculated confidence, and calculate the position coordinates of the tracking target on the video frame image according to the position coordinates of all the picture samples and the adjusted weight;
D、根据所述位置坐标从所述视频帧图像上采集所述跟踪目标的正样本和负样本;D. Collecting positive and negative samples of the tracking target from the video frame image according to the position coordinates;
E、根据所述正样本和负样本更新所述CNN模型的训练样本集,并使用更新后的训练样本集训练所述CNN模型以更新所述CNN模型的模型参数;E. Updating a training sample set of the CNN model according to the positive sample and the negative sample, and training the CNN model with the updated training sample set to update model parameters of the CNN model;
F、重复执行步骤A至E,直至完成对视频的所有视频帧图像中跟踪目标的跟踪。F. Repeat steps A through E until the tracking of the tracking target in all video frame images of the video is completed.
本申请提出的基于卷积神经网络的目标跟踪装置、方法及计算机可读存储介质,对视频中的视频帧图像进行逐帧识别,按照采样点分布从视频帧图像上采集多个图片样本,并记录各个图片样本的位置坐标,基于CNN模型从多个样本图片对应地提取多个样本特征,根据提取的样本特征计算各个图片样本与跟踪目标之间的置信度,根据置信度对应的调整样本的权重,进而根据样本的位置坐标和权重计算跟踪目标在该视频帧图像上的位置坐标,并且根据该位置坐标从视频帧图像上采集跟踪目标的正样本和负样本,使用采集的样本重新训练CNN模型以更新模型参数,使用更新模型参数后的模型继续对下一帧图像跟踪,以此类推,在得到每一帧图像的跟踪结果后,根据跟踪结果对模型进行更新,使得在跟踪目标发生变化时,更新后的模型能够适应目标及背景的变化,即使图像中出现部分遮挡、背景干扰等现象时,也能够成功的进行目标的跟踪,提高目标跟踪的准确度。The object tracking device, the method and the computer readable storage medium based on the convolutional neural network proposed by the present application identify the video frame image in the video frame by frame, and collect multiple image samples from the video frame image according to the sampling point distribution, and Recording position coordinates of each picture sample, extracting a plurality of sample features correspondingly from the plurality of sample pictures based on the CNN model, calculating a confidence level between each picture sample and the tracking target according to the extracted sample features, and adjusting the sample according to the confidence level Weighting, and then calculating the position coordinates of the tracking target on the video frame image according to the position coordinates and weights of the sample, and collecting positive and negative samples of the tracking target from the video frame image according to the position coordinates, and retraining the CNN using the collected samples The model updates the model parameters, continues to track the next frame image using the model after updating the model parameters, and so on. After obtaining the tracking result of each frame image, the model is updated according to the tracking result, so that the tracking target changes. The updated model can adapt to changes in goals and backgrounds. Even when the phenomenon of partial occlusion, background interference appear in the image, it can also be successfully tracking the target, improve the accuracy of target tracking.
附图说明DRAWINGS
图1为本申请基于卷积神经网络的目标跟踪装置较佳实施例的示意图;1 is a schematic diagram of a preferred embodiment of a target tracking device based on a convolutional neural network;
图2为本申请基于卷积神经网络的目标跟踪装置一实施例中目标跟踪程序的程序模块示意图;2 is a schematic diagram of a program module of a target tracking program in an embodiment of a target tracking device based on a convolutional neural network;
图3为本申请基于卷积神经网络的目标跟踪方法较佳实施例的流程图。3 is a flow chart of a preferred embodiment of a target tracking method based on a convolutional neural network.
本申请目的的实现、功能特点及优点将结合实施例,参照附图做进一步说明。The implementation, functional features and advantages of the present application will be further described with reference to the accompanying drawings.
具体实施方式Detailed ways
应当理解,此处所描述的具体实施例仅仅用以解释本申请,并不用于限定本申请。It is understood that the specific embodiments described herein are merely illustrative of the application and are not intended to be limiting.
本申请提供一种基于卷积神经网络的目标跟踪装置。参照图1所示,为本申请基于卷积神经网络的目标跟踪装置较佳实施例的示意图。The application provides a target tracking device based on a convolutional neural network. Referring to FIG. 1, a schematic diagram of a preferred embodiment of a target tracking device based on a convolutional neural network is provided.
在本实施例中,基于卷积神经网络的目标跟踪装置可以是PC(Personal Computer,个人电脑),也可以是智能手机、平板电脑、电子书阅读器、便携 计算机等具有显示功能的终端设备。In this embodiment, the target tracking device based on the convolutional neural network may be a PC (Personal Computer), or may be a smart phone, a tablet computer, an e-book reader, or a portable device. A terminal device having a display function such as a computer.
该基于卷积神经网络的目标跟踪装置包括存储器11、处理器12,通信总线13,以及网络接口14。The convolutional neural network based target tracking device includes a memory 11, a processor 12, a communication bus 13, and a network interface 14.
其中,存储器11至少包括一种类型的可读存储介质,所述可读存储介质包括闪存、硬盘、多媒体卡、卡型存储器(例如,SD或DX存储器等)、磁性存储器、磁盘、光盘等。存储器11在一些实施例中可以是基于卷积神经网络的目标跟踪装置的内部存储单元,例如该基于卷积神经网络的目标跟踪装置的硬盘。存储器11在另一些实施例中也可以是基于卷积神经网络的目标跟踪装置的外部存储设备,例如基于卷积神经网络的目标跟踪装置上配备的插接式硬盘,智能存储卡(Smart Media Card,SMC),安全数字(Secure Digital,SD)卡,闪存卡(Flash Card)等。进一步地,存储器11还可以既包括基于卷积神经网络的目标跟踪装置的内部存储单元也包括外部存储设备。存储器11不仅可以用于存储安装于基于卷积神经网络的目标跟踪装置的应用软件及各类数据,例如目标跟踪程序的代码等,还可以用于暂时地存储已经输出或者将要输出的数据。The memory 11 includes at least one type of readable storage medium including a flash memory, a hard disk, a multimedia card, a card type memory (for example, an SD or DX memory, etc.), a magnetic memory, a magnetic disk, an optical disk, and the like. Memory 11 may in some embodiments be an internal storage unit of a target tracking device based on a convolutional neural network, such as a hard disk of a target tracking device based on a convolutional neural network. The memory 11 may also be an external storage device of a target tracking device based on a convolutional neural network in other embodiments, such as a plug-in hard disk equipped on a target tracking device based on a convolutional neural network, a smart memory card (Smart Media Card) , SMC), Secure Digital (SD) card, Flash Card, etc. Further, the memory 11 may also include both an internal storage unit of the target tracking device based on the convolutional neural network and an external storage device. The memory 11 can be used not only for storing application software and various types of data installed on a target tracking device based on a convolutional neural network, such as code of a target tracking program, but also for temporarily storing data that has been output or is to be output.
处理器12在一些实施例中可以是一中央处理器(Central Processing Unit,CPU)、控制器、微控制器、微处理器或其他数据处理芯片,用于运行存储器11中存储的程序代码或处理数据,例如执行目标跟踪程序等。The processor 12, in some embodiments, may be a Central Processing Unit (CPU), controller, microcontroller, microprocessor or other data processing chip for running program code or processing stored in the memory 11. Data, such as executing a target tracking program.
通信总线13用于实现这些组件之间的连接通信。 Communication bus 13 is used to implement connection communication between these components.
网络接口14可选的可以包括标准的有线接口、无线接口(如WI-FI接口),通常用于在该装置与其他电子设备之间建立通信连接。The network interface 14 can optionally include a standard wired interface, a wireless interface (such as a WI-FI interface), and is typically used to establish a communication connection between the device and other electronic devices.
图1仅示出了具有组件11-14以及目标跟踪程序的基于卷积神经网络的目标跟踪装置,但是应理解的是,并不要求实施所有示出的组件,可以替代的实施更多或者更少的组件。Figure 1 shows only a convolutional neural network based target tracking device with components 11-14 and a target tracking program, but it should be understood that not all illustrated components may be implemented, alternative implementations may be more or more Less components.
可选地,该装置还可以包括用户接口,用户接口可以包括显示器(Display)、输入单元比如键盘(Keyboard),可选的用户接口还可以包括标准的有线接口、无线接口。可选地,在一些实施例中,显示器可以是LED显示器、液晶显示器、触控式液晶显示器以及OLED(Organic Light-Emitting Diode,有机发光二极管)触摸器等。其中,显示器也可以适当的称为显示屏或显示单元,用于显示在基于卷积神经网络的目标跟踪装置中处理的信息以及用于显示可视化的用户界面。Optionally, the device may further include a user interface, the user interface may include a display, an input unit such as a keyboard, and the optional user interface may further include a standard wired interface and a wireless interface. Optionally, in some embodiments, the display may be an LED display, a liquid crystal display, a touch-sensitive liquid crystal display, an OLED (Organic Light-Emitting Diode) touch sensor, or the like. The display may also be suitably referred to as a display screen or display unit for displaying information processed in a target tracking device based on a convolutional neural network and a user interface for displaying visualization.
可选地,该装置还可以包括触摸传感器。所述触摸传感器所提供的供用户进行触摸操作的区域称为触控区域。此外,这里所述的触摸传感器可以为电阻式触摸传感器、电容式触摸传感器等。而且,所述触摸传感器不仅包括接触式的触摸传感器,也可包括接近式的触摸传感器等。此外,所述触摸传感器可以为单个传感器,也可以为例如阵列布置的多个传感器。该装置的显示器的面积可以与所述触摸传感器的面积相同,也可以不同。可选地,将显示器与所述触摸传感器层叠设置,以形成触摸显示屏。该装置基于触摸显示 屏侦测用户触发的触控操作。Optionally, the device may also include a touch sensor. The area provided by the touch sensor for the user to perform a touch operation is referred to as a touch area. Further, the touch sensor described herein may be a resistive touch sensor, a capacitive touch sensor, or the like. Moreover, the touch sensor includes not only a contact type touch sensor but also a proximity type touch sensor or the like. Furthermore, the touch sensor may be a single sensor or a plurality of sensors arranged, for example, in an array. The area of the display of the device may be the same as or different from the area of the touch sensor. Optionally, a display is stacked with the touch sensor to form a touch display. The device is based on a touch display The screen detects the touch operation triggered by the user.
可选地,该装置还可以包括摄像头、RF(Radio Frequency,射频)电路,传感器、音频电路、WiFi模块等。其中,传感器比如光传感器、运动传感器以及其他传感器。具体地,光传感器可包括环境光传感器及接近传感器,其中,若该装置为移动终端,环境光传感器可根据环境光线的明暗来调节显示屏的亮度,接近传感器可在移动终端移动到耳边时,关闭显示屏和/或背光。作为运动传感器的一种,重力加速度传感器可检测各个方向上(一般为三轴)加速度的大小,静止时可检测出重力的大小及方向,可用于识别移动终端姿态的应用(比如横竖屏切换、相关游戏、磁力计姿态校准)、振动识别相关功能(比如计步器、敲击)等;当然,移动终端还可配置陀螺仪、气压计、湿度计、温度计、红外线传感器等其他传感器,在此不再赘述。Optionally, the device may further include a camera, an RF (Radio Frequency) circuit, a sensor, an audio circuit, a WiFi module, and the like. Among them, sensors such as light sensors, motion sensors, and other sensors. Specifically, the light sensor may include an ambient light sensor and a proximity sensor, wherein if the device is a mobile terminal, the ambient light sensor may adjust the brightness of the display screen according to the brightness of the ambient light, and the proximity sensor may move when the mobile terminal moves to the ear. , turn off the display and / or backlight. As a kind of motion sensor, the gravity acceleration sensor can detect the magnitude of acceleration in each direction (usually three axes), and can detect the magnitude and direction of gravity when stationary, and can be used to identify the posture of the mobile terminal (such as horizontal and vertical screen switching, Related games, magnetometer attitude calibration), vibration recognition related functions (such as pedometer, tapping), etc.; of course, the mobile terminal can also be equipped with other sensors such as gyroscope, barometer, hygrometer, thermometer, infrared sensor, etc. No longer.
在图1所示的装置实施例中,存储器11中存储有目标跟踪程序;处理器12执行存储器11中存储的目标跟踪程序时实现如下步骤:In the apparatus embodiment shown in FIG. 1, a target tracking program is stored in the memory 11; when the processor 12 executes the target tracking program stored in the memory 11, the following steps are implemented:
A、按照采样点分布从所述视频帧图像上采集多个图片样本。A. Collecting a plurality of picture samples from the video frame image according to a sampling point distribution.
在本申请实施例中,基于卷积神经网络对海量图片进行离线训练得到一个CNN(Convolutional Neural Network,卷积神经网络)模型,该模型可以是一个二分类模型,利用该模型能够从图像中提取出深层次的语义化的运动目标的特征和目标的背景特征。In the embodiment of the present application, a convolutional neural network is used to perform offline training on a massive picture to obtain a CNN (Convolutional Neural Network) model, which may be a two-class model, and the model can be extracted from the image. A deeper semantic feature of the moving target and the background feature of the target.
在对视频中的运动目标进行跟踪时,逐帧对视频图像进行跟踪。具体地,将要进行目标跟踪的视频输入到该装置,该装置按照下述操作对视频中的每一视频帧图像进行处理。When tracking a moving target in a video, the video image is tracked frame by frame. Specifically, a video to be subjected to target tracking is input to the device, and the device processes each video frame image in the video in accordance with the following operation.
按照采样点分布从视频帧图像上采集图片样本,其中,采样点的数量可以由用户预先设置,例如采集100个图片样本,其中,在开始对第一帧图像进行识别时,可以由用户手动从图像中选择跟踪目标,例如,通过框选的形式选择跟踪目标,基于用户选择的跟踪目标的位置对采样点分布进行初始化。具体地,可以在接收到视频帧图像时,判断所述视频帧图像是否为所述视频的第一帧图像;若所述视频帧图像为所述视频的第一帧图像,则提示用户在所述视频帧图像上手动选择跟踪目标并接收用户基于所述提示选择的跟踪目标;在确定所述跟踪目标后,初始化采样点分布和所述CNN模型的训练样本集并接收第二帧图像;若所述视频图像不是所述视频的第一帧图像,则执行所述步骤A。或者,在其他的实施例中,用户预先设置好要跟踪的目标并存储,则在开始跟踪后直接获取跟踪目标,不需要用户再手动的从第一帧图像上选取。The image samples are collected from the video frame image according to the sampling point distribution, wherein the number of sampling points can be preset by the user, for example, 100 image samples are collected, wherein when the first frame image is started to be recognized, the user can manually The tracking target is selected in the image. For example, the tracking target is selected by a frame selection manner, and the sampling point distribution is initialized based on the position of the tracking target selected by the user. Specifically, when the video frame image is received, determining whether the video frame image is the first frame image of the video; if the video frame image is the first frame image of the video, prompting the user to Manually selecting a tracking target on the video frame image and receiving a tracking target selected by the user based on the prompt; after determining the tracking target, initializing a sampling point distribution and a training sample set of the CNN model and receiving a second frame image; The video image is not the first frame image of the video, and the step A is performed. Alternatively, in other embodiments, if the user sets the target to be tracked in advance and stores it, the tracking target is directly acquired after starting the tracking, and the user is not required to manually select from the first frame image.
在得到用户选择的跟踪目标后,计算跟踪目标区域的色彩直方图并将其作为跟踪目标的目标特征,该目标特征可以表示为一个N*1的向量。After obtaining the tracking target selected by the user, the color histogram of the tracking target area is calculated and used as the target feature of the tracking target, and the target feature can be represented as a vector of N*1.
B、基于卷积神经网络CNN模型从所述多个图片样本中对应地提取多个样本特征,并根据提取的样本特征分别计算每一图片样本与跟踪目标之间的 置信度。B. Extracting a plurality of sample features from the plurality of picture samples based on the convolutional neural network CNN model, and separately calculating each picture sample from the tracking target according to the extracted sample features Confidence.
C、根据计算得出的置信度调整对应图片样本的权重,并根据所有图片样本的位置坐标和调整后的权重计算所述跟踪目标在所述视频帧图像上的位置坐标。C. Adjust the weight of the corresponding picture sample according to the calculated confidence, and calculate the position coordinates of the tracking target on the video frame image according to the position coordinates of all the picture samples and the adjusted weight.
在采集到样本图片后,将采集的样本图片输入到上述训练好的CNN模型中进行特征提取,提取出样本特征,样本特征同样的可以表示为一个N*1的向量。每一个样本图片对应的提取一个样本特征,分别计算每一个样本特征与目标特征之间的置信度。样本特征的置信度体现出图片样本与跟踪目标之间的相似度,通过计算样本特征与目标特征之间的相似度,即计算上述两个N*1的向量之间的相似度,作为图片样本与跟踪目标之间的置信度。After the sample image is collected, the collected sample image is input into the trained CNN model for feature extraction, and the sample feature is extracted, and the sample feature can also be represented as an N*1 vector. A sample feature is extracted corresponding to each sample picture, and the confidence between each sample feature and the target feature is calculated separately. The confidence of the sample features reflects the similarity between the image sample and the tracking target. By calculating the similarity between the sample feature and the target feature, the similarity between the two N*1 vectors is calculated as a sample of the image. Confidence with tracking targets.
在得到每一个图片样本的置信度之后,根据置信度调整每一个图片样本的权重,对于置信度小的样本,减小其权重,对于置信度大的样本,则增大其权重,然后对于所有的图片样本的权重进行归一化处理,使得所有样本的权重之和等于1。根据图片样本的权重值和其在视频帧图像上的位置坐标计算跟踪目标在该视频帧图像上的位置坐标。具体地,假设一共采集了k个图片样本,其中样本Pi的位置坐标为(xi,yi),其与跟踪目标之间的置信度为Si。则根据以下公式可以预测出跟踪目标的位置坐标(x,y)。After obtaining the confidence of each picture sample, the weight of each picture sample is adjusted according to the confidence level, and for the sample with small confidence, the weight is reduced, and for the sample with high confidence, the weight is increased, and then for all the weights, then for all The weights of the image samples are normalized such that the sum of the weights of all samples is equal to one. The position coordinates of the tracking target on the video frame image are calculated based on the weight value of the picture sample and its position coordinates on the video frame image. Specifically, it is assumed that a total of k picture samples are collected, wherein the position coordinates of the sample P i are (x i , y i ), and the confidence degree with the tracking target is S i . Then, the position coordinates (x, y) of the tracking target can be predicted according to the following formula.
Figure PCTCN2017108794-appb-000001
Figure PCTCN2017108794-appb-000001
D、根据所述位置坐标从所述视频帧图像上采集所述跟踪目标的正样本和负样本。D. Collect positive and negative samples of the tracking target from the video frame image according to the position coordinates.
E、根据所述正样本和负样本更新所述CNN模型的训练样本集,并使用更新后的训练样本集训练所述CNN模型以更新所述CNN模型的模型参数。E. Updating a training sample set of the CNN model according to the positive sample and the negative sample, and training the CNN model with the updated training sample set to update model parameters of the CNN model.
F、重复执行步骤A至E,直至完成对视频的所有视频帧图像中跟踪目标的跟踪。F. Repeat steps A through E until the tracking of the tracking target in all video frame images of the video is completed.
根据该位置坐标从视频帧图像上采集跟踪目标的正样本和负样本,具体地,采集位于所述位置坐标的周边区域内的第一预设数量的图片样本作为正样本,其中,所述周边区域为与所述位置坐标之间的距离小于第一预设阈值的点构成的区域;采集位于所述位置坐标的远离区域内的第二预设数量的图片样本作为负样本,其中,所述远离区域为与所述位置坐标之间的距离大于第二预设阈值的点构成的区域,所述第二预设阈值大于所述第一预设阈值。Collecting a positive sample and a negative sample of the tracking target from the video frame image according to the position coordinate, specifically, acquiring a first preset number of picture samples located in a peripheral area of the position coordinate as a positive sample, wherein the periphery The area is an area formed by a point having a distance from the position coordinate that is smaller than a first preset threshold; and a second predetermined number of picture samples located in a distant area of the position coordinate are collected as a negative sample, wherein the The remote preset area is an area formed by a point that is greater than a second preset threshold value, and the second preset threshold is greater than the first preset threshold.
也就是说,在预测到跟踪目标在图像上的位置后,从距离跟踪目标较近的区域内采集图片样本,这些样本与跟踪目标之间的差距较小,可以作为正样本,从视频帧图像距离跟踪目标较远的区域中采集图片样本,这些样本与 跟踪目标之间的差别较大,可以作为负样本,添加到CNN模型的训练样本集中,并且使用对CNN模型进行训练,更新模型参数,提高模型从图片样本中识别出运动目标的特征的准确度,以使该模型能够适应视频帧图像中目标和背景的变化。通过这样的方式,在跟踪过程中,不断地对CNN模型更新,即使出现跟踪目标有部分遮挡或者有背景对跟踪目标造成干扰,也不会干扰到对目标的正确跟踪。在完成对该视频帧图像的跟踪后,继续对下一帧图像进行跟踪,采用更新后的CNN模型进行特征提取。按照步骤A至步骤E对每一帧图像进行目标跟踪,并且在跟踪完成后,对CNN模型进行训练,直至完成对视频的所有帧图像中目标的跟踪。可以理解的是,上述第一预设阈值、第二预设阈值、第一预设数量以及第二预设数量均可以由用预先设置。That is to say, after predicting the position of the tracking target on the image, the image samples are collected from the region closer to the tracking target, and the difference between the samples and the tracking target is small, and can be used as a positive sample from the video frame image. Capture image samples from areas farther away from the tracking target. The difference between the tracking targets is large, and can be added as a negative sample to the training sample set of the CNN model, and the CNN model is trained to update the model parameters and improve the accuracy of the model to identify the features of the moving target from the image samples. To enable the model to adapt to changes in the target and background in the video frame image. In this way, during the tracking process, the CNN model is continuously updated, and even if the tracking target is partially occluded or the background interferes with the tracking target, it does not interfere with the correct tracking of the target. After the tracking of the video frame image is completed, the next frame image is continuously tracked, and the updated CNN model is used for feature extraction. Target tracking is performed for each frame of image according to steps A through E, and after the tracking is completed, the CNN model is trained until the tracking of the target in all frame images of the video is completed. It can be understood that the first preset threshold, the second preset threshold, the first preset number, and the second preset number may be preset by use.
进一步地,在其他实施例中,在步骤E之后,还实现如下步骤:Further, in other embodiments, after step E, the following steps are further implemented:
G、根据调整后的权重调整采样点在视频帧图像上的位置,以更新采样点分布;G, adjusting the position of the sampling point on the video frame image according to the adjusted weight to update the sampling point distribution;
所述步骤F包括:重复执行步骤A至G,直至完成对视频的所有视频帧图像中的跟踪目标的跟踪。The step F includes repeating steps A to G until the tracking of the tracking target in all video frame images of the video is completed.
具体地,根据调整后的权重对采样点的分布进行调整,具体地,在权重大于第一预设权重的样本对应的采样点的第一预设范围内增加采样点,即在权重大的图片样本对应的采样点附近增加更多的采样点,在权重小于第二预设权重的样本对应的采样点的第二预设范围内减少采样点,其中,第二预设权重小于第一预设权重,即减少权重小的图片样本对应的采样点附近的采样点,其中,增加的采样点的数量等于或者大于减少的采样点的数量,或者,当权重非常小时,可以将对应的采样点删除,例如,将权重小于第三预设权重的样本对应的采样点删除,其中,第三预设权重小于所述第四预设权重。Specifically, the distribution of the sampling points is adjusted according to the adjusted weights. Specifically, the sampling points are added in the first preset range of the sampling points corresponding to the samples whose weight is greater than the first preset weight, that is, the pictures with the weights are significant. Adding more sampling points near the sampling point corresponding to the sample, and reducing the sampling point in the second preset range of the sampling point corresponding to the sample whose weight is smaller than the second preset weight, wherein the second preset weight is smaller than the first preset The weight, that is, the sampling point near the sampling point corresponding to the picture sample with small weight reduction, wherein the number of added sampling points is equal to or greater than the number of reduced sampling points, or, when the weight is very small, the corresponding sampling point can be deleted. For example, the sampling point corresponding to the sample whose weight is smaller than the third preset weight is deleted, wherein the third preset weight is smaller than the fourth preset weight.
本实施例提出的基于卷积神经网络的目标跟踪装置,对视频中的视频帧图像进行逐帧识别,按照采样点分布从视频帧图像上采集多个图片样本,并记录各个图片样本的位置坐标,基于CNN模型从多个样本图片对应地提取多个样本特征,根据提取的样本特征计算各个图片样本与跟踪目标之间的置信度,根据置信度对应的调整样本的权重,进而根据样本的位置坐标和权重计算跟踪目标在该视频帧图像上的位置坐标,并且根据该位置坐标从视频帧图像上采集跟踪目标的正样本和负样本,使用采集的样本重新训练CNN模型以更新模型参数,使用更新模型参数后的模型继续对下一帧图像跟踪,以此类推,在得到每一帧图像的跟踪结果后,根据跟踪结果对模型进行更新,使得在跟踪目标发生变化时,更新后的模型能够适应目标及背景的变化,即使图像中出现部分遮挡、背景干扰等现象时,也能够成功的进行目标的跟踪,提高目标跟踪的准确度。The target tracking device based on the convolutional neural network proposed in this embodiment performs frame-by-frame recognition on the video frame image in the video, collects multiple picture samples from the video frame image according to the sampling point distribution, and records the position coordinates of each picture sample. And extracting a plurality of sample features correspondingly from the plurality of sample pictures based on the CNN model, calculating a confidence degree between each picture sample and the tracking target according to the extracted sample features, adjusting the weight of the sample according to the confidence degree, and further according to the position of the sample Coordinates and weights calculate the position coordinates of the tracking target on the video frame image, and collect positive and negative samples of the tracking target from the video frame image according to the position coordinates, and retrain the CNN model to update the model parameters using the collected samples, using After updating the model parameters, the model continues to track the next frame image, and so on. After obtaining the tracking result of each frame image, the model is updated according to the tracking result, so that the updated model can be changed when the tracking target changes. Adapt to changes in goals and background, even if partial obscuration occurs in the image When the phenomenon of background interference, but also able to successfully track the target, improve the accuracy of target tracking.
可选地,在其他的实施例中,目标跟踪程序还可以被分割为一个或者多个模块,一个或者多个模块被存储于存储器11中,并由一个或多个处理器(本 实施例为处理器12)所执行以完成本申请,本申请所称的模块是指能够完成特定功能的一系列计算机程序指令段。Optionally, in other embodiments, the target tracking program may also be divided into one or more modules, one or more modules being stored in the memory 11 and being processed by one or more processors (this Embodiments are executed by processor 12) to accomplish the present application, and a module referred to herein refers to a series of computer program instructions that are capable of performing a particular function.
例如,参照图2所示,为本申请基于卷积神经网络的目标跟踪装置一实施例中的目标跟踪程序的程序模块示意图,该实施例中,目标跟踪程序可以被分割为预处理模块10、跟踪模块20、采样模块30和更新模块40,示例性地,For example, referring to FIG. 2, it is a schematic diagram of a program module of a target tracking program in an embodiment of a target tracking device based on a convolutional neural network according to the present application. In this embodiment, the target tracking program may be divided into a preprocessing module 10, Tracking module 20, sampling module 30 and update module 40, illustratively,
采集模块10用于:按照采样点分布从视频帧图像上采集多个图片样本,并记录各个图片样本的位置坐标;The collecting module 10 is configured to: collect a plurality of picture samples from the video frame image according to the sampling point distribution, and record position coordinates of each picture sample;
预处理模块20用于:基于卷积神经网络CNN模型从所述多个图片样本中对应地提取多个样本特征,并分别根据提取的样本特征分别计算每一图片样本与跟踪目标之间的置信度;The pre-processing module 20 is configured to: correspondingly extract a plurality of sample features from the plurality of picture samples based on a convolutional neural network CNN model, and respectively calculate a confidence between each picture sample and the tracking target according to the extracted sample features respectively degree;
跟踪模块30用于:根据计算得出的置信度调整对应图片样本的权重,并根据所有图片样本的位置坐标和权重计算所述跟踪目标在所述视频帧图像上的位置坐标;The tracking module 30 is configured to: adjust weights of the corresponding picture samples according to the calculated confidence, and calculate position coordinates of the tracking target on the video frame image according to position coordinates and weights of all picture samples;
采样模块40用于:根据所述位置坐标从所述视频帧图像上采集所述跟踪目标的正样本和负样本;The sampling module 40 is configured to: collect positive and negative samples of the tracking target from the video frame image according to the position coordinates;
更新模块50用于:根据所述正样本和负样本更新所述CNN模型的训练样本集,并使用更新后的训练样本集训练所述CNN模型以更新所述CNN模型的模型参数;The updating module 50 is configured to: update the training sample set of the CNN model according to the positive sample and the negative sample, and train the CNN model to update model parameters of the CNN model by using the updated training sample set;
采集模块10、预处理模块20、跟踪模块30、采样模块40和更新模块50按照视频中的视频帧图像的顺序执行上述步骤对目标进行跟踪,直至完成对视频中所有视频帧图像中的跟踪目标的跟踪。The acquisition module 10, the pre-processing module 20, the tracking module 30, the sampling module 40, and the update module 50 perform the above steps to track the target in the order of the video frame images in the video until the tracking target in all the video frame images in the video is completed. Tracking.
上述采集模块10、预处理模块20、跟踪模块30、采样模块40和更新模块50被执行所实现的功能或操作步骤与上述实施例大体相同,在此不再赘述。The functions or operation steps performed by the above-mentioned collection module 10, the pre-processing module 20, the tracking module 30, the sampling module 40, and the update module 50 are substantially the same as those of the foregoing embodiment, and are not described herein again.
此外,本申请还提供一种基于卷积神经网络的目标跟踪方法。参照图3所示,为本申请基于卷积神经网络的目标跟踪方法较佳实施例的流程图。该方法可以由一个装置执行,该装置可以由软件和/或硬件实现。In addition, the present application also provides a target tracking method based on a convolutional neural network. Referring to FIG. 3, it is a flowchart of a preferred embodiment of a target tracking method based on a convolutional neural network. The method can be performed by a device that can be implemented by software and/or hardware.
在本实施例中,基于卷积神经网络的目标跟踪方法包括:In this embodiment, the target tracking method based on the convolutional neural network includes:
步骤S10,按照采样点分布从视频帧图像上采集多个图片样本,并记录各个图片样本的位置坐标。Step S10: Collect a plurality of picture samples from the video frame image according to the sampling point distribution, and record position coordinates of each picture sample.
在本申请实施例中,基于卷积神经网络对海量图片进行离线训练得到一个CNN(Convolutional Neural Network,卷积神经网络)模型,该模型可以是一个二分类模型,利用该模型能够从图像中提取出深层次的语义化的运动目标的特征和目标的背景特征。In the embodiment of the present application, a convolutional neural network is used to perform offline training on a massive picture to obtain a CNN (Convolutional Neural Network) model, which may be a two-class model, and the model can be extracted from the image. A deeper semantic feature of the moving target and the background feature of the target.
在对视频中的运动目标进行跟踪时,逐帧对视频图像进行跟踪。具体地,将要进行目标跟踪的视频输入到该装置,该装置按照下述操作对视频中的每一视频帧图像进行处理。 When tracking a moving target in a video, the video image is tracked frame by frame. Specifically, a video to be subjected to target tracking is input to the device, and the device processes each video frame image in the video in accordance with the following operation.
按照采样点分布从视频帧图像上采集图片样本,其中,采样点的数量可以由用户预先设置,例如采集100个图片样本,其中,在开始对第一帧图像进行识别时,可以由用户手动从图像中选择跟踪目标,例如,通过框选的形式选择跟踪目标,基于用户选择的跟踪目标的位置对采样点分布进行初始化。具体地,可以在接收到视频帧图像时,判断所述视频帧图像是否为所述视频的第一帧图像;若所述视频帧图像为所述视频的第一帧图像,则提示用户在所述视频帧图像上手动选择跟踪目标并接收用户基于所述提示选择的跟踪目标;在确定所述跟踪目标后,初始化采样点分布和所述CNN模型的训练样本集并接收第二帧图像;若所述视频图像不是所述视频的第一帧图像,则执行所述步骤S10。或者,在其他的实施例中,用户预先设置好要跟踪的目标并存储,则在开始跟踪后直接获取跟踪目标,不需要用户再手动的从第一帧图像上选取。The image samples are collected from the video frame image according to the sampling point distribution, wherein the number of sampling points can be preset by the user, for example, 100 image samples are collected, wherein when the first frame image is started to be recognized, the user can manually The tracking target is selected in the image. For example, the tracking target is selected by a frame selection manner, and the sampling point distribution is initialized based on the position of the tracking target selected by the user. Specifically, when the video frame image is received, determining whether the video frame image is the first frame image of the video; if the video frame image is the first frame image of the video, prompting the user to Manually selecting a tracking target on the video frame image and receiving a tracking target selected by the user based on the prompt; after determining the tracking target, initializing a sampling point distribution and a training sample set of the CNN model and receiving a second frame image; The video image is not the first frame image of the video, and the step S10 is performed. Alternatively, in other embodiments, if the user sets the target to be tracked in advance and stores it, the tracking target is directly acquired after starting the tracking, and the user is not required to manually select from the first frame image.
在得到用户选择的跟踪目标后,计算跟踪目标区域的色彩直方图并将其作为跟踪目标的目标特征,该目标特征可以表示为一个N*1的向量。After obtaining the tracking target selected by the user, the color histogram of the tracking target area is calculated and used as the target feature of the tracking target, and the target feature can be represented as a vector of N*1.
步骤S20,基于卷积神经网络CNN模型从所述多个图片样本中对应地提取多个样本特征,并分别根据提取的样本特征分别计算每一图片样本与跟踪目标之间的置信度。Step S20, correspondingly extracting a plurality of sample features from the plurality of picture samples based on the convolutional neural network CNN model, and respectively calculating a confidence level between each picture sample and the tracking target according to the extracted sample features.
步骤S30,根据计算得出的置信度调整对应图片样本的权重,并根据所有图片样本的位置坐标和权重计算所述跟踪目标在所述视频帧图像上的位置坐标。Step S30, adjusting the weight of the corresponding picture sample according to the calculated confidence, and calculating the position coordinates of the tracking target on the video frame image according to the position coordinates and weights of all the picture samples.
在采集到样本图片后,将采集的样本图片输入到上述训练好的CNN模型中进行特征提取,提取出样本特征,样本特征同样的可以表示为一个N*1的向量。每一个样本图片对应的提取一个样本特征,分别计算每一个样本特征与目标特征之间的置信度。样本特征的置信度体现出图片样本与跟踪目标之间的相似度,通过计算样本特征与目标特征之间的相似度,即计算上述两个N*1的向量之间的相似度,作为图片样本与跟踪目标之间的置信度。After the sample image is collected, the collected sample image is input into the trained CNN model for feature extraction, and the sample feature is extracted, and the sample feature can also be represented as an N*1 vector. A sample feature is extracted corresponding to each sample picture, and the confidence between each sample feature and the target feature is calculated separately. The confidence of the sample features reflects the similarity between the image sample and the tracking target. By calculating the similarity between the sample feature and the target feature, the similarity between the two N*1 vectors is calculated as a sample of the image. Confidence with tracking targets.
在得到每一个图片样本的置信度之后,根据置信度调整每一个图片样本的权重,对于置信度小的样本,减小其权重,对于置信度大的样本,则增大其权重,然后对于所有的图片样本的权重进行归一化处理,使得所有样本的权重之和等于1。根据图片样本的权重值和其在视频帧图像上的位置坐标计算跟踪目标在该视频帧图像上的位置坐标。具体地,假设一共采集了k个图片样本,其中样本Pi的位置坐标为(xi,yi),其与跟踪目标之间的置信度为Si。则根据以下公式可以预测出跟踪目标的位置坐标(x,y)。 After obtaining the confidence of each picture sample, the weight of each picture sample is adjusted according to the confidence level, and for the sample with small confidence, the weight is reduced, and for the sample with high confidence, the weight is increased, and then for all the weights, then for all The weights of the image samples are normalized such that the sum of the weights of all samples is equal to one. The position coordinates of the tracking target on the video frame image are calculated based on the weight value of the picture sample and its position coordinates on the video frame image. Specifically, it is assumed that a total of k picture samples are collected, wherein the position coordinates of the sample P i are (x i , y i ), and the confidence degree with the tracking target is S i . Then, the position coordinates (x, y) of the tracking target can be predicted according to the following formula.
Figure PCTCN2017108794-appb-000002
Figure PCTCN2017108794-appb-000002
步骤S40,根据所述位置坐标从所述视频帧图像上采集所述跟踪目标的正样本和负样本。Step S40, collecting positive and negative samples of the tracking target from the video frame image according to the position coordinates.
步骤S50,根据所述正样本和负样本更新所述CNN模型的训练样本集,并使用更新后的训练样本集训练所述CNN模型以更新所述CNN模型的模型参数。Step S50, updating the training sample set of the CNN model according to the positive sample and the negative sample, and training the CNN model to update the model parameters of the CNN model by using the updated training sample set.
步骤S60,重复执行步骤S10至S50,直至完成对视频的所有视频帧图像中跟踪目标的跟踪。In step S60, steps S10 to S50 are repeatedly performed until the tracking of the tracking target in all the video frame images of the video is completed.
根据该位置坐标从视频帧图像上采集跟踪目标的正样本和负样本,具体地,采集位于所述位置坐标的周边区域内的第一预设数量的图片样本作为正样本,其中,所述周边区域为与所述位置坐标之间的距离小于第一预设阈值的点构成的区域;采集位于所述位置坐标的远离区域内的第二预设数量的图片样本作为负样本,其中,所述远离区域为与所述位置坐标之间的距离大于第二预设阈值的点构成的区域,所述第二预设阈值大于所述第一预设阈值。Collecting a positive sample and a negative sample of the tracking target from the video frame image according to the position coordinate, specifically, acquiring a first preset number of picture samples located in a peripheral area of the position coordinate as a positive sample, wherein the periphery The area is an area formed by a point having a distance from the position coordinate that is smaller than a first preset threshold; and a second predetermined number of picture samples located in a distant area of the position coordinate are collected as a negative sample, wherein the The remote preset area is an area formed by a point that is greater than a second preset threshold value, and the second preset threshold is greater than the first preset threshold.
也就是说,在预测到跟踪目标在图像上的位置后,从距离跟踪目标较近的区域内采集图片样本,这些样本与跟踪目标之间的差距较小,可以作为正样本,从视频帧图像距离跟踪目标较远的区域中采集图片样本,这些样本与跟踪目标之间的差别较大,可以作为负样本,添加到CNN模型的训练样本集中,并且使用对CNN模型进行训练,更新模型参数,提高模型从图片样本中识别出运动目标的特征的准确度,以使该模型能够适应视频帧图像中目标和背景的变化。通过这样的方式,在跟踪过程中,不断地对CNN模型更新,即使出现跟踪目标有部分遮挡或者有背景对跟踪目标造成干扰,也不会干扰到对目标的正确跟踪。在完成对该视频帧图像的跟踪后,继续对下一帧图像进行跟踪,采用更新后的CNN模型进行特征提取。按照步骤S10至步骤S40对每一帧图像进行目标跟踪,并且在跟踪完成后,对CNN模型进行训练,直至完成对视频的所有帧图像中目标的全部跟踪。可以理解的是,上述第一预设阈值、第二预设阈值、第一预设数量以及第二预设数量均可以由用预先设置。That is to say, after predicting the position of the tracking target on the image, the image samples are collected from the region closer to the tracking target, and the difference between the samples and the tracking target is small, and can be used as a positive sample from the video frame image. Image samples are taken from areas farther away from the tracking target. These samples have a large difference from the tracking targets. They can be added as negative samples to the training sample set of the CNN model, and the CNN model is trained to update the model parameters. The improved model identifies the accuracy of the features of the moving object from the image samples so that the model can adapt to changes in the target and background in the video frame image. In this way, during the tracking process, the CNN model is continuously updated, and even if the tracking target is partially occluded or the background interferes with the tracking target, it does not interfere with the correct tracking of the target. After the tracking of the video frame image is completed, the next frame image is continuously tracked, and the updated CNN model is used for feature extraction. Target tracking is performed for each frame of image in accordance with steps S10 through S40, and after the tracking is completed, the CNN model is trained until all tracking of the target in all frame images of the video is completed. It can be understood that the first preset threshold, the second preset threshold, the first preset number, and the second preset number may be preset by use.
进一步地,在其他实施例中,在步骤S50之后,该方法还包括如下步骤:根据调整后的权重对采样点的分布进行调整,具体地,在权重大于第一预设权重的样本对应的采样点的第一预设范围内增加采样点,即在权重大的图片样本对应的采样点附近增加更多的采样点,在权重小于第二预设权重的样本对应的采样点的第二预设范围内减少采样点,其中,第二预设权重小于第一预设权重,即减少权重小的图片样本对应的采样点附近的采样点,其中,增 加的采样点的数量等于或者大于减少的采样点的数量,或者,当权重非常小时,可以将对应的采样点删除,例如,将权重小于第三预设权重的样本对应的采样点删除,其中,第三预设权重小于所述第四预设权重。Further, in other embodiments, after step S50, the method further includes the following steps: adjusting the distribution of the sampling points according to the adjusted weights, specifically, the sampling corresponding to the samples having the weight greater than the first preset weight Adding a sampling point within a first preset range of the point, that is, adding more sampling points near the sampling point corresponding to the image sample with a large weight, and second sampling of the sampling point corresponding to the sample whose weight is smaller than the second preset weight The sampling point is reduced in the range, wherein the second preset weight is smaller than the first preset weight, that is, the sampling point near the sampling point corresponding to the image sample with small weight is reduced, wherein The number of added sampling points is equal to or greater than the number of reduced sampling points, or, when the weight is very small, the corresponding sampling points may be deleted, for example, the sampling points corresponding to the samples whose weights are smaller than the third preset weight are deleted, wherein The third preset weight is smaller than the fourth preset weight.
本实施例提出的基于卷积神经网络的目标跟踪方法,对视频中的视频帧图像进行逐帧识别,按照采样点分布从视频帧图像上采集多个图片样本,并记录各个图片样本的位置坐标,基于CNN模型从多个样本图片对应地提取多个样本特征,根据提取的样本特征计算各个图片样本与跟踪目标之间的置信度,根据置信度对应的调整样本的权重,进而根据样本的位置坐标和权重计算跟踪目标在该视频帧图像上的位置坐标,并且根据该位置坐标从视频帧图像上采集跟踪目标的正样本和负样本,使用采集的样本重新训练CNN模型以更新模型参数,使用更新模型参数后的模型继续对下一帧图像跟踪,以此类推,在得到每一帧图像的跟踪结果后,根据跟踪结果对模型进行更新,使得在跟踪目标发生变化时,更新后的模型能够适应目标及背景的变化,即使图像中出现部分遮挡、背景干扰等现象时,也能够成功的进行目标的跟踪,提高目标跟踪的准确度。The target tracking method based on the convolutional neural network proposed in this embodiment performs frame-by-frame recognition on the video frame image in the video, collects multiple picture samples from the video frame image according to the sampling point distribution, and records the position coordinates of each picture sample. And extracting a plurality of sample features correspondingly from the plurality of sample pictures based on the CNN model, calculating a confidence degree between each picture sample and the tracking target according to the extracted sample features, adjusting the weight of the sample according to the confidence degree, and further according to the position of the sample Coordinates and weights calculate the position coordinates of the tracking target on the video frame image, and collect positive and negative samples of the tracking target from the video frame image according to the position coordinates, and retrain the CNN model to update the model parameters using the collected samples, using After updating the model parameters, the model continues to track the next frame image, and so on. After obtaining the tracking result of each frame image, the model is updated according to the tracking result, so that the updated model can be changed when the tracking target changes. Adapt to changes in goals and background, even if partial obscuration occurs in the image When the phenomenon of background interference, but also able to successfully track the target, improve the accuracy of target tracking.
此外,本申请实施例还提出一种计算机可读存储介质,所述计算机可读存储介质上存储有目标跟踪程序,所述目标跟踪程序可被一个或多个处理器执行,以实现如下操作:In addition, the embodiment of the present application further provides a computer readable storage medium, where the target readable program is stored on the computer readable storage medium, and the target tracking program can be executed by one or more processors to implement the following operations:
A、按照采样点分布从视频帧图像上采集多个图片样本,并记录各个图片样本的位置坐标;A. Collecting a plurality of picture samples from the video frame image according to the sampling point distribution, and recording position coordinates of each picture sample;
B、基于卷积神经网络CNN模型从所述多个图片样本中对应地提取多个样本特征,并分别根据提取的样本特征分别计算每一图片样本与跟踪目标之间的置信度;B. Extracting a plurality of sample features from the plurality of picture samples based on the convolutional neural network CNN model, and respectively calculating a confidence level between each picture sample and the tracking target according to the extracted sample features;
C、根据计算得出的置信度调整对应图片样本的权重,并根据所有图片样本的位置坐标和调整后的权重计算所述跟踪目标在所述视频帧图像上的位置坐标;C. Adjust the weight of the corresponding picture sample according to the calculated confidence, and calculate the position coordinates of the tracking target on the video frame image according to the position coordinates of all the picture samples and the adjusted weight;
D、根据所述位置坐标从所述视频帧图像上采集所述跟踪目标的正样本和负样本;D. Collecting positive and negative samples of the tracking target from the video frame image according to the position coordinates;
E、根据所述正样本和负样本更新所述CNN模型的训练样本集,并使用更新后的训练样本集训练所述CNN模型以更新所述CNN模型的模型参数;E. Updating a training sample set of the CNN model according to the positive sample and the negative sample, and training the CNN model with the updated training sample set to update model parameters of the CNN model;
F、重复执行步骤A至E,直至完成对视频的所有视频帧图像中跟踪目标的跟踪。F. Repeat steps A through E until the tracking of the tracking target in all video frame images of the video is completed.
进一步地,所述目标跟踪程序被处理器执行时还实现如下操作:Further, when the target tracking program is executed by the processor, the following operations are also implemented:
采集位于所述位置坐标的周边区域内的第一预设数量的图片样本作为正样本,其中,所述周边区域为与所述位置坐标之间的距离小于第一预设阈值的点构成的区域;And acquiring a first preset number of picture samples located in a peripheral area of the position coordinate as a positive sample, wherein the peripheral area is an area formed by a point that is smaller than a first preset threshold between the position coordinates ;
采集位于所述位置坐标的远离区域内的第二预设数量的图片样本作为负 样本,其中,所述远离区域为与所述位置坐标之间的距离大于第二预设阈值的点构成的区域,所述第二预设阈值大于所述第一预设阈值。Collecting a second predetermined number of picture samples located in a distant area of the position coordinates as a negative And a sample, wherein the remote area is an area formed by a point that is greater than a second preset threshold, and the second preset threshold is greater than the first preset threshold.
进一步地,所述目标跟踪程序被处理器执行时还实现如下操作:Further, when the target tracking program is executed by the processor, the following operations are also implemented:
根据调整后的权重调整采样点在视频帧图像上的位置,以更新采样点分布。The position of the sampling point on the video frame image is adjusted according to the adjusted weight to update the sampling point distribution.
进一步地,所述目标跟踪程序被处理器执行时还实现如下操作:Further, when the target tracking program is executed by the processor, the following operations are also implemented:
在权重大于第一预设权重的样本对应的采样点的第一预设范围内增加采样点,在权重小于第二预设权重的样本对应的采样点的第二预设范围内减少采样点,其中,所述第二预设权重小于所述第一预设权重,增加的采样点的数量等于减少的采样点的数量。Adding a sampling point within a first preset range of the sampling point corresponding to the sample whose weight is greater than the first preset weight, and decreasing the sampling point in a second preset range of the sampling point corresponding to the sample whose weight is smaller than the second preset weight, The second preset weight is smaller than the first preset weight, and the number of added sampling points is equal to the number of reduced sampling points.
本申请计算机可读存储介质具体实施方式与上述基于卷积神经网络的目标跟踪装置和方法各实施例基本相同,在此不作累述。The specific embodiment of the computer readable storage medium of the present application is substantially the same as the foregoing embodiments of the target tracking apparatus and method based on the convolutional neural network, and is not described herein.
需要说明的是,上述本申请实施例序号仅仅为了描述,不代表实施例的优劣。并且本文中的术语“包括”、“包含”或者其任何其他变体意在涵盖非排他性的包含,从而使得包括一系列要素的过程、装置、物品或者方法不仅包括那些要素,而且还包括没有明确列出的其他要素,或者是还包括为这种过程、装置、物品或者方法所固有的要素。在没有更多限制的情况下,由语句“包括一个……”限定的要素,并不排除在包括该要素的过程、装置、物品或者方法中还存在另外的相同要素。It should be noted that the foregoing serial numbers of the embodiments of the present application are merely for the description, and do not represent the advantages and disadvantages of the embodiments. And the terms "including", "comprising", or any other variations thereof are intended to encompass a non-exclusive inclusion, such that a process, apparatus, article, or method that comprises a plurality of elements includes not only those elements but also Other elements listed, or elements that are inherent to such a process, device, item, or method. An element that is defined by the phrase "comprising a ..." does not exclude the presence of additional equivalent elements in the process, the device, the item, or the method that comprises the element.
通过以上的实施方式的描述,本领域的技术人员可以清楚地了解到上述实施例方法可借助软件加必需的通用硬件平台的方式来实现,当然也可以通过硬件,但很多情况下前者是更佳的实施方式。基于这样的理解,本申请的技术方案本质上或者说对现有技术做出贡献的部分可以以软件产品的形式体现出来,该计算机软件产品存储在如上所述的一个存储介质(如ROM/RAM、磁碟、光盘)中,包括若干指令用以使得一台终端设备(可以是手机,计算机,服务器,或者网络设备等)执行本申请各个实施例所述的方法。Through the description of the above embodiments, those skilled in the art can clearly understand that the foregoing embodiment method can be implemented by means of software plus a necessary general hardware platform, and of course, can also be through hardware, but in many cases, the former is better. Implementation. Based on such understanding, the technical solution of the present application, which is essential or contributes to the prior art, may be embodied in the form of a software product stored in a storage medium (such as ROM/RAM as described above). , a disk, an optical disk, including a number of instructions for causing a terminal device (which may be a mobile phone, a computer, a server, or a network device, etc.) to perform the methods described in the various embodiments of the present application.
以上仅为本申请的优选实施例,并非因此限制本申请的专利范围,凡是利用本申请说明书及附图内容所作的等效结构或等效流程变换,或直接或间接运用在其他相关的技术领域,均同理包括在本申请的专利保护范围内。 The above is only a preferred embodiment of the present application, and is not intended to limit the scope of the patent application, and the equivalent structure or equivalent process transformations made by the specification and the drawings of the present application, or directly or indirectly applied to other related technical fields. The same is included in the scope of patent protection of this application.

Claims (20)

  1. 一种基于卷积神经网络的目标跟踪装置,其特征在于,所述装置包括存储器和处理器,所述存储器上存储有可在所述处理器上运行的目标跟踪程序,所述目标跟踪程序被所述处理器执行时实现如下步骤:A target tracking device based on a convolutional neural network, the device comprising a memory and a processor, wherein the memory stores a target tracking program executable on the processor, the target tracking program is The processor implements the following steps when executed:
    A、按照采样点分布从视频帧图像上采集多个图片样本,并记录各个图片样本的位置坐标;A. Collecting a plurality of picture samples from the video frame image according to the sampling point distribution, and recording position coordinates of each picture sample;
    B、基于卷积神经网络CNN模型从所述多个图片样本中对应地提取多个样本特征,并分别根据提取的样本特征分别计算每一图片样本与跟踪目标之间的置信度;B. Extracting a plurality of sample features from the plurality of picture samples based on the convolutional neural network CNN model, and respectively calculating a confidence level between each picture sample and the tracking target according to the extracted sample features;
    C、根据计算得出的置信度调整对应图片样本的权重,并根据所有图片样本的位置坐标和调整后的权重计算所述跟踪目标在所述视频帧图像上的位置坐标;C. Adjust the weight of the corresponding picture sample according to the calculated confidence, and calculate the position coordinates of the tracking target on the video frame image according to the position coordinates of all the picture samples and the adjusted weight;
    D、根据所述位置坐标从所述视频帧图像上采集所述跟踪目标的正样本和负样本;D. Collecting positive and negative samples of the tracking target from the video frame image according to the position coordinates;
    E、根据所述正样本和负样本更新所述CNN模型的训练样本集,并使用更新后的训练样本集训练所述CNN模型以更新所述CNN模型的模型参数;E. Updating a training sample set of the CNN model according to the positive sample and the negative sample, and training the CNN model with the updated training sample set to update model parameters of the CNN model;
    F、重复执行步骤A至E,直至完成对视频的所有视频帧图像中跟踪目标的跟踪。F. Repeat steps A through E until the tracking of the tracking target in all video frame images of the video is completed.
  2. 根据权利要求1所述的基于卷积神经网络的目标跟踪装置,其特征在于,所述步骤D包括:The object tracking device based on a convolutional neural network according to claim 1, wherein the step D comprises:
    采集位于所述位置坐标的周边区域内的第一预设数量的图片样本作为正样本,其中,所述周边区域为与所述位置坐标之间的距离小于第一预设阈值的点构成的区域;And acquiring a first preset number of picture samples located in a peripheral area of the position coordinate as a positive sample, wherein the peripheral area is an area formed by a point that is smaller than a first preset threshold between the position coordinates ;
    采集位于所述位置坐标的远离区域内的第二预设数量的图片样本作为负样本,其中,所述远离区域为与所述位置坐标之间的距离大于第二预设阈值的点构成的区域,所述第二预设阈值大于所述第一预设阈值。And acquiring a second preset number of picture samples located in a distant area of the position coordinate as a negative sample, wherein the distant area is an area formed by a point that is greater than a second preset threshold between the position coordinates The second preset threshold is greater than the first preset threshold.
  3. 根据权利要求1所述的基于卷积神经网络的目标跟踪装置,其特征在于,所述处理器还用于执行所述目标跟踪程序,以在步骤E之后,还实现如下步骤:The convolutional neural network-based target tracking device according to claim 1, wherein the processor is further configured to execute the target tracking program, to further implement the following steps after the step E:
    G、根据调整后的权重调整采样点在视频帧图像上的位置,以更新采样点分布;G, adjusting the position of the sampling point on the video frame image according to the adjusted weight to update the sampling point distribution;
    所述步骤F包括:The step F includes:
    重复执行步骤A至G,直至完成对视频的所有视频帧图像中的跟踪目标的跟踪。Steps A through G are repeated until the tracking of the tracking target in all video frame images of the video is completed.
  4. 根据权利要求2所述的基于卷积神经网络的目标跟踪装置,其特征在于,所述处理器还用于执行所述目标跟踪程序,以在步骤E之后,还实现如 下步骤:The convolutional neural network-based target tracking device according to claim 2, wherein the processor is further configured to execute the target tracking program to further implement, after step E, Next steps:
    G、根据调整后的权重调整采样点在视频帧图像上的位置,以更新采样点分布;G, adjusting the position of the sampling point on the video frame image according to the adjusted weight to update the sampling point distribution;
    所述步骤F包括:The step F includes:
    重复执行步骤A至G,直至完成对视频的所有视频帧图像中的跟踪目标的跟踪。Steps A through G are repeated until the tracking of the tracking target in all video frame images of the video is completed.
  5. 根据权利要求3所述的基于卷积神经网络的目标跟踪装置,其特征在于,所述步骤G包括:The object tracking device based on a convolutional neural network according to claim 3, wherein the step G comprises:
    在权重大于第一预设权重的样本对应的采样点的第一预设范围内增加采样点,在权重小于第二预设权重的样本对应的采样点的第二预设范围内减少采样点,其中,所述第二预设权重小于所述第一预设权重,增加的采样点的数量等于减少的采样点的数量。Adding a sampling point within a first preset range of the sampling point corresponding to the sample whose weight is greater than the first preset weight, and decreasing the sampling point in a second preset range of the sampling point corresponding to the sample whose weight is smaller than the second preset weight, The second preset weight is smaller than the first preset weight, and the number of added sampling points is equal to the number of reduced sampling points.
  6. 根据权利要求1所述的基于卷积神经网络的目标跟踪装置,其特征在于,所述处理器还用于执行所述目标跟踪程序,以在步骤A之前,还实现如下步骤:The convolutional neural network-based target tracking device according to claim 1, wherein the processor is further configured to execute the target tracking program to implement the following steps before the step A:
    判断所述视频帧图像是否为所述视频的第一帧图像;Determining whether the video frame image is the first frame image of the video;
    若所述视频帧图像为所述视频的第一帧图像,则提示用户在所述视频帧图像上手动选择跟踪目标并接收用户基于所述提示选择的跟踪目标,并在确定所述跟踪目标后,初始化采样点分布和所述CNN模型的训练样本集并接收第二帧图像;If the video frame image is the first frame image of the video, prompting the user to manually select a tracking target on the video frame image and receive a tracking target selected by the user based on the prompt, and after determining the tracking target Initializing a sample point distribution and a training sample set of the CNN model and receiving a second frame image;
    若所述视频图像不是所述视频的第一帧图像,则执行所述步骤A。If the video image is not the first frame image of the video, the step A is performed.
  7. 根据权利要求2所述的基于卷积神经网络的目标跟踪装置,其特征在于,所述处理器还用于执行所述目标跟踪程序,以在步骤A之前,还实现如下步骤:The convolutional neural network-based target tracking device according to claim 2, wherein the processor is further configured to execute the target tracking program to implement the following steps before the step A:
    判断所述视频帧图像是否为所述视频的第一帧图像;Determining whether the video frame image is the first frame image of the video;
    若所述视频帧图像为所述视频的第一帧图像,则提示用户在所述视频帧图像上手动选择跟踪目标并接收用户基于所述提示选择的跟踪目标,并在确定所述跟踪目标后,初始化采样点分布和所述CNN模型的训练样本集并接收第二帧图像;If the video frame image is the first frame image of the video, prompting the user to manually select a tracking target on the video frame image and receive a tracking target selected by the user based on the prompt, and after determining the tracking target Initializing a sample point distribution and a training sample set of the CNN model and receiving a second frame image;
    若所述视频图像不是所述视频的第一帧图像,则执行所述步骤A。If the video image is not the first frame image of the video, the step A is performed.
  8. 一种基于卷积神经网络的目标跟踪方法,其特征在于,所述方法包括:A target tracking method based on a convolutional neural network, characterized in that the method comprises:
    A、按照采样点分布从视频帧图像上采集多个图片样本,并记录各个图片样本的位置坐标;A. Collecting a plurality of picture samples from the video frame image according to the sampling point distribution, and recording position coordinates of each picture sample;
    B、基于卷积神经网络CNN模型从所述多个图片样本中对应地提取多个样本特征,并分别根据提取的样本特征分别计算每一图片样本与跟踪目标之 间的置信度;B. Extracting a plurality of sample features from the plurality of picture samples based on the convolutional neural network CNN model, and respectively calculating each picture sample and the tracking target according to the extracted sample features Confidence between
    C、根据计算得出的置信度调整对应图片样本的权重,并根据所有图片样本的位置坐标和调整后的权重计算所述跟踪目标在所述视频帧图像上的位置坐标;C. Adjust the weight of the corresponding picture sample according to the calculated confidence, and calculate the position coordinates of the tracking target on the video frame image according to the position coordinates of all the picture samples and the adjusted weight;
    D、根据所述位置坐标从所述视频帧图像上采集所述跟踪目标的正样本和负样本;D. Collecting positive and negative samples of the tracking target from the video frame image according to the position coordinates;
    E、根据所述正样本和负样本更新所述CNN模型的训练样本集,并使用更新后的训练样本集训练所述CNN模型以更新所述CNN模型的模型参数;E. Updating a training sample set of the CNN model according to the positive sample and the negative sample, and training the CNN model with the updated training sample set to update model parameters of the CNN model;
    F、重复执行步骤A至E,直至完成对视频的所有视频帧图像中跟踪目标的跟踪。F. Repeat steps A through E until the tracking of the tracking target in all video frame images of the video is completed.
  9. 根据权利要求8所述的基于卷积神经网络的目标跟踪方法,其特征在于,所述步骤D包括:The method for tracking a target based on a convolutional neural network according to claim 8, wherein the step D comprises:
    采集位于所述位置坐标的周边区域内的第一预设数量的图片样本作为正样本,其中,所述周边区域为与所述位置坐标之间的距离小于第一预设阈值的点构成的区域;And acquiring a first preset number of picture samples located in a peripheral area of the position coordinate as a positive sample, wherein the peripheral area is an area formed by a point that is smaller than a first preset threshold between the position coordinates ;
    采集位于所述位置坐标的远离区域内的第二预设数量的图片样本作为负样本,其中,所述远离区域为与所述位置坐标之间的距离大于第二预设阈值的点构成的区域,所述第二预设阈值大于所述第一预设阈值。And acquiring a second preset number of picture samples located in a distant area of the position coordinate as a negative sample, wherein the distant area is an area formed by a point that is greater than a second preset threshold between the position coordinates The second preset threshold is greater than the first preset threshold.
  10. 根据权利要求8所述的基于卷积神经网络的目标跟踪方法,其特征在于,在步骤E之后,该方法还包括:The convolutional neural network-based target tracking method according to claim 8, wherein after the step E, the method further comprises:
    G、根据调整后的权重调整采样点在视频帧图像上的位置,以更新采样点分布;G, adjusting the position of the sampling point on the video frame image according to the adjusted weight to update the sampling point distribution;
    所述步骤F包括:The step F includes:
    重复执行步骤A至G,直至完成对视频的所有视频帧图像中的跟踪目标的跟踪。Steps A through G are repeated until the tracking of the tracking target in all video frame images of the video is completed.
  11. 根据权利要求9所述的基于卷积神经网络的目标跟踪方法,其特征在于,在步骤E之后,该方法还包括:The convolutional neural network-based target tracking method according to claim 9, wherein after the step E, the method further comprises:
    G、根据调整后的权重调整采样点在视频帧图像上的位置,以更新采样点分布;G, adjusting the position of the sampling point on the video frame image according to the adjusted weight to update the sampling point distribution;
    所述步骤F包括:The step F includes:
    重复执行步骤A至G,直至完成对视频的所有视频帧图像中的跟踪目标的跟踪。Steps A through G are repeated until the tracking of the tracking target in all video frame images of the video is completed.
  12. 根据权利要求10所述的基于卷积神经网络的目标跟踪方法,其特征在于,所述步骤G包括:The method for tracking a target based on a convolutional neural network according to claim 10, wherein the step G comprises:
    在权重大于第一预设权重的样本对应的采样点的第一预设范围内增加采 样点,在权重小于第二预设权重的样本对应的采样点的第二预设范围内减少采样点,其中,所述第二预设权重小于所述第一预设权重,增加的采样点的数量等于减少的采样点的数量。Adding a first preset range of sampling points corresponding to samples whose weight is greater than the first preset weight a sampling point, where the sampling point is decreased in a second preset range of the sampling point corresponding to the sample whose weight is smaller than the second preset weight, wherein the second preset weight is smaller than the first preset weight, and the added sampling point The number is equal to the number of reduced sampling points.
  13. 根据权利要求8所述的基于卷积神经网络的目标跟踪方法,其特征在于,在步骤A之前,所述方法还包括如下步骤:The convolutional neural network-based target tracking method according to claim 8, wherein before the step A, the method further comprises the following steps:
    判断所述视频帧图像是否为所述视频的第一帧图像;Determining whether the video frame image is the first frame image of the video;
    若所述视频帧图像为所述视频的第一帧图像,则提示用户在所述视频帧图像上手动选择跟踪目标并接收用户基于所述提示选择的跟踪目标,并在确定所述跟踪目标后,初始化采样点分布和所述CNN模型的训练样本集并接收第二帧图像;If the video frame image is the first frame image of the video, prompting the user to manually select a tracking target on the video frame image and receive a tracking target selected by the user based on the prompt, and after determining the tracking target Initializing a sample point distribution and a training sample set of the CNN model and receiving a second frame image;
    若所述视频图像不是所述视频的第一帧图像,则执行所述步骤A。If the video image is not the first frame image of the video, the step A is performed.
  14. 根据权利要求9所述的基于卷积神经网络的目标跟踪方法,其特征在于,在步骤A之前,所述方法还包括如下步骤:The convolutional neural network-based target tracking method according to claim 9, wherein before the step A, the method further comprises the following steps:
    判断所述视频帧图像是否为所述视频的第一帧图像;Determining whether the video frame image is the first frame image of the video;
    若所述视频帧图像为所述视频的第一帧图像,则提示用户在所述视频帧图像上手动选择跟踪目标并接收用户基于所述提示选择的跟踪目标,并在确定所述跟踪目标后,初始化采样点分布和所述CNN模型的训练样本集并接收第二帧图像;If the video frame image is the first frame image of the video, prompting the user to manually select a tracking target on the video frame image and receive a tracking target selected by the user based on the prompt, and after determining the tracking target Initializing a sample point distribution and a training sample set of the CNN model and receiving a second frame image;
    若所述视频图像不是所述视频的第一帧图像,则执行所述步骤A。If the video image is not the first frame image of the video, the step A is performed.
  15. 一种计算机可读存储介质,其特征在于,所述计算机可读存储介质上存储有目标跟踪程序,所述目标跟踪程序可被一个或多个处理器执行,以实现如下步骤:A computer readable storage medium, characterized in that the computer readable storage medium stores a target tracking program, and the target tracking program can be executed by one or more processors to implement the following steps:
    A、按照采样点分布从视频帧图像上采集多个图片样本,并记录各个图片样本的位置坐标;A. Collecting a plurality of picture samples from the video frame image according to the sampling point distribution, and recording position coordinates of each picture sample;
    B、基于卷积神经网络CNN模型从所述多个图片样本中对应地提取多个样本特征,并分别根据提取的样本特征分别计算每一图片样本与跟踪目标之间的置信度;B. Extracting a plurality of sample features from the plurality of picture samples based on the convolutional neural network CNN model, and respectively calculating a confidence level between each picture sample and the tracking target according to the extracted sample features;
    C、根据计算得出的置信度调整对应图片样本的权重,并根据所有图片样本的位置坐标和调整后的权重计算所述跟踪目标在所述视频帧图像上的位置坐标;C. Adjust the weight of the corresponding picture sample according to the calculated confidence, and calculate the position coordinates of the tracking target on the video frame image according to the position coordinates of all the picture samples and the adjusted weight;
    D、根据所述位置坐标从所述视频帧图像上采集所述跟踪目标的正样本和负样本;D. Collecting positive and negative samples of the tracking target from the video frame image according to the position coordinates;
    E、根据所述正样本和负样本更新所述CNN模型的训练样本集,并使用更新后的训练样本集训练所述CNN模型以更新所述CNN模型的模型参数;E. Updating a training sample set of the CNN model according to the positive sample and the negative sample, and training the CNN model with the updated training sample set to update model parameters of the CNN model;
    F、重复执行步骤A至E,直至完成对视频的所有视频帧图像中跟踪目标的跟踪。 F. Repeat steps A through E until the tracking of the tracking target in all video frame images of the video is completed.
  16. 根据权利要求15所述的计算机可读存储介质,其特征在于,所述步骤D包括:The computer readable storage medium of claim 15, wherein the step D comprises:
    采集位于所述位置坐标的周边区域内的第一预设数量的图片样本作为正样本,其中,所述周边区域为与所述位置坐标之间的距离小于第一预设阈值的点构成的区域;And acquiring a first preset number of picture samples located in a peripheral area of the position coordinate as a positive sample, wherein the peripheral area is an area formed by a point that is smaller than a first preset threshold between the position coordinates ;
    采集位于所述位置坐标的远离区域内的第二预设数量的图片样本作为负样本,其中,所述远离区域为与所述位置坐标之间的距离大于第二预设阈值的点构成的区域,所述第二预设阈值大于所述第一预设阈值。And acquiring a second preset number of picture samples located in a distant area of the position coordinate as a negative sample, wherein the distant area is an area formed by a point that is greater than a second preset threshold between the position coordinates The second preset threshold is greater than the first preset threshold.
  17. 根据权利要求15所述的计算机可读存储介质,其特征在于,所述目标跟踪程序还可被一个或多个处理器执行,以在步骤E之后,还实现如下步骤:The computer readable storage medium of claim 15, wherein the target tracking program is further executable by one or more processors to further implement the following steps after step E:
    G、根据调整后的权重调整采样点在视频帧图像上的位置,以更新采样点分布;G, adjusting the position of the sampling point on the video frame image according to the adjusted weight to update the sampling point distribution;
    所述步骤F包括:The step F includes:
    重复执行步骤A至G,直至完成对视频的所有视频帧图像中的跟踪目标的跟踪。Steps A through G are repeated until the tracking of the tracking target in all video frame images of the video is completed.
  18. 根据权利要求16所述的计算机可读存储介质,其特征在于,所述目标跟踪程序还可被一个或多个处理器执行,以在步骤E之后,还实现如下步骤:The computer readable storage medium of claim 16, wherein the target tracking program is further executable by one or more processors to further implement the following steps after step E:
    G、根据调整后的权重调整采样点在视频帧图像上的位置,以更新采样点分布;G, adjusting the position of the sampling point on the video frame image according to the adjusted weight to update the sampling point distribution;
    所述步骤F包括:The step F includes:
    重复执行步骤A至G,直至完成对视频的所有视频帧图像中的跟踪目标的跟踪。Steps A through G are repeated until the tracking of the tracking target in all video frame images of the video is completed.
  19. 根据权利要求17所述的计算机可读存储介质,其特征在于,所述步骤G包括:The computer readable storage medium of claim 17, wherein the step G comprises:
    在权重大于第一预设权重的样本对应的采样点的第一预设范围内增加采样点,在权重小于第二预设权重的样本对应的采样点的第二预设范围内减少采样点,其中,所述第二预设权重小于所述第一预设权重,增加的采样点的数量等于减少的采样点的数量。Adding a sampling point within a first preset range of the sampling point corresponding to the sample whose weight is greater than the first preset weight, and decreasing the sampling point in a second preset range of the sampling point corresponding to the sample whose weight is smaller than the second preset weight, The second preset weight is smaller than the first preset weight, and the number of added sampling points is equal to the number of reduced sampling points.
  20. 根据权利要求15所述的计算机可读存储介质,其特征在于,所述目标跟踪程序还可被一个或多个处理器执行,以在步骤A之前,还实现如下步骤:The computer readable storage medium of claim 15, wherein the target tracking program is further executable by one or more processors to implement the following steps prior to step A:
    判断所述视频帧图像是否为所述视频的第一帧图像; Determining whether the video frame image is the first frame image of the video;
    若所述视频帧图像为所述视频的第一帧图像,则提示用户在所述视频帧图像上手动选择跟踪目标并接收用户基于所述提示选择的跟踪目标,并在确定所述跟踪目标后,初始化采样点分布和所述CNN模型的训练样本集并接收第二帧图像;If the video frame image is the first frame image of the video, prompting the user to manually select a tracking target on the video frame image and receive a tracking target selected by the user based on the prompt, and after determining the tracking target Initializing a sample point distribution and a training sample set of the CNN model and receiving a second frame image;
    若所述视频图像不是所述视频的第一帧图像,则执行所述步骤A。 If the video image is not the first frame image of the video, the step A is performed.
PCT/CN2017/108794 2017-08-29 2017-10-31 Target tracking device and method, and computer-readable storage medium WO2019041519A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN201710754313.XA CN107679455A (en) 2017-08-29 2017-08-29 Target tracker, method and computer-readable recording medium
CN201710754313.X 2017-08-29

Publications (1)

Publication Number Publication Date
WO2019041519A1 true WO2019041519A1 (en) 2019-03-07

Family

ID=61134784

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2017/108794 WO2019041519A1 (en) 2017-08-29 2017-10-31 Target tracking device and method, and computer-readable storage medium

Country Status (2)

Country Link
CN (1) CN107679455A (en)
WO (1) WO2019041519A1 (en)

Cited By (31)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109858472A (en) * 2019-04-09 2019-06-07 武汉领普科技有限公司 A kind of embedded humanoid detection method and device in real time
CN110148052A (en) * 2019-04-17 2019-08-20 深圳壹账通智能科技有限公司 Management-control method, device, computer equipment and storage medium after businessman borrows
CN110176027A (en) * 2019-05-27 2019-08-27 腾讯科技(深圳)有限公司 Video target tracking method, device, equipment and storage medium
CN110443827A (en) * 2019-07-22 2019-11-12 浙江大学 A kind of UAV Video single goal long-term follow method based on the twin network of improvement
CN110516556A (en) * 2019-07-31 2019-11-29 平安科技(深圳)有限公司 Multi-target tracking detection method, device and storage medium based on Darkflow-DeepSort
CN110532883A (en) * 2019-07-30 2019-12-03 平安科技(深圳)有限公司 On-line tracking is improved using off-line tracking algorithm
CN110543818A (en) * 2019-07-25 2019-12-06 长沙行深智能科技有限公司 Traffic light tracking method, device, medium and equipment based on weight graph matching
CN110738101A (en) * 2019-09-04 2020-01-31 平安科技(深圳)有限公司 Behavior recognition method and device and computer readable storage medium
CN110929605A (en) * 2019-11-11 2020-03-27 中国建设银行股份有限公司 Video key frame storage method, device, equipment and storage medium
CN110956131A (en) * 2019-11-27 2020-04-03 北京迈格威科技有限公司 Single-target tracking method, device and system
CN110991458A (en) * 2019-11-25 2020-04-10 创新奇智(北京)科技有限公司 Artificial intelligence recognition result sampling system and sampling method based on image characteristics
CN111091098A (en) * 2019-12-20 2020-05-01 浙江大华技术股份有限公司 Training method and detection method of detection model and related device
CN111241931A (en) * 2019-12-30 2020-06-05 沈阳理工大学 Aerial unmanned aerial vehicle target identification and tracking method based on YOLOv3
CN111242977A (en) * 2020-01-09 2020-06-05 影石创新科技股份有限公司 Target tracking method of panoramic video, readable storage medium and computer equipment
CN111275011A (en) * 2020-02-25 2020-06-12 北京百度网讯科技有限公司 Mobile traffic light detection method and device, electronic equipment and storage medium
CN111311635A (en) * 2020-02-08 2020-06-19 腾讯科技(深圳)有限公司 Target positioning method, device and system
CN111401172A (en) * 2020-03-06 2020-07-10 大连海事大学 Port hoisting material bag automatic counting method based on video
CN111652080A (en) * 2020-05-12 2020-09-11 合肥的卢深视科技有限公司 Target tracking method and device based on RGB-D image
CN111694423A (en) * 2019-03-12 2020-09-22 阿里巴巴集团控股有限公司 Positioning, capturing, data processing and display method and equipment for augmented reality
CN111815670A (en) * 2019-04-10 2020-10-23 曜科智能科技(上海)有限公司 Multi-view target tracking method, device and system, electronic terminal and storage medium
CN111862158A (en) * 2020-07-21 2020-10-30 湖南师范大学 Staged target tracking method and device, terminal and readable storage medium
CN112183235A (en) * 2020-09-07 2021-01-05 根尖体育科技(北京)有限公司 Automatic control method for video acquisition aiming at sport places
CN112270661A (en) * 2020-10-19 2021-01-26 北京宇航系统工程研究所 Space environment monitoring method based on rocket telemetry video
CN112655021A (en) * 2020-04-09 2021-04-13 深圳市大疆创新科技有限公司 Image processing method, image processing device, electronic equipment and storage medium
CN112733690A (en) * 2020-12-31 2021-04-30 北京易华录信息技术股份有限公司 High-altitude parabolic detection method and device and electronic equipment
CN113066108A (en) * 2021-04-14 2021-07-02 武汉卓目科技有限公司 Anti-occlusion visual target tracking method and device based on ECO algorithm
CN113573137A (en) * 2021-07-01 2021-10-29 厦门美图之家科技有限公司 Video canvas boundary detection method, system, terminal equipment and storage medium
CN113609971A (en) * 2021-08-04 2021-11-05 广州威拓电子科技有限公司 Method, device and equipment for inspecting microseism observation equipment and storage medium
CN113642360A (en) * 2020-04-27 2021-11-12 杭州海康威视数字技术股份有限公司 Behavior timing method and device, electronic equipment and storage medium
CN114511793A (en) * 2020-11-17 2022-05-17 中国人民解放军军事科学院国防科技创新研究院 Unmanned aerial vehicle ground detection method and system based on synchronous detection and tracking
US11538246B2 (en) 2018-07-27 2022-12-27 Tencent Technology (Shenzhen) Company Limited Method and apparatus for training feature extraction model, computer device, and computer-readable storage medium

Families Citing this family (24)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108320297B (en) * 2018-03-09 2020-06-19 湖北工业大学 Video target real-time tracking method and system
WO2019213820A1 (en) * 2018-05-07 2019-11-14 合刃科技(武汉)有限公司 Photographing control method and electronic device
CN108960046A (en) 2018-05-23 2018-12-07 北京图森未来科技有限公司 A kind of training data method of sampling and its device, computer server
CN108765455B (en) * 2018-05-24 2021-09-21 中国科学院光电技术研究所 Target stable tracking method based on TLD algorithm
CN108846850B (en) * 2018-05-24 2022-06-10 中国科学院光电技术研究所 Target tracking method based on TLD algorithm
CN110569690B (en) * 2018-06-06 2022-05-13 浙江宇视科技有限公司 Target information acquisition method and device
CN109448018B (en) * 2018-09-18 2023-08-01 平安科技(深圳)有限公司 Tracking target positioning method, device, equipment and storage medium
CN111127509B (en) * 2018-10-31 2023-09-01 杭州海康威视数字技术股份有限公司 Target tracking method, apparatus and computer readable storage medium
CN109711248A (en) * 2018-11-08 2019-05-03 平安科技(深圳)有限公司 A kind of Identify Environment and terminal device based on video
CN109635657B (en) * 2018-11-12 2023-01-06 平安科技(深圳)有限公司 Target tracking method, device, equipment and storage medium
CN109600627B (en) * 2018-12-11 2021-12-24 国信优易数据股份有限公司 Video identification method and device
CN111310526B (en) * 2018-12-12 2023-10-20 杭州海康威视数字技术股份有限公司 Parameter determination method and device for target tracking model and storage medium
CN109598885B (en) * 2018-12-21 2021-06-11 广东中安金狮科创有限公司 Monitoring system and alarm method thereof
CN109886998A (en) * 2019-01-23 2019-06-14 平安科技(深圳)有限公司 Multi-object tracking method, device, computer installation and computer storage medium
CN110288082B (en) * 2019-06-05 2022-04-05 北京字节跳动网络技术有限公司 Convolutional neural network model training method and device and computer readable storage medium
CN111462194B (en) * 2020-03-30 2023-08-11 苏州科达科技股份有限公司 Training method, device and storage medium of object tracking model
CN111653103A (en) * 2020-05-07 2020-09-11 浙江大华技术股份有限公司 Target object identification method and device
CN112132088B (en) * 2020-09-29 2024-01-12 动联(山东)电子科技有限公司 Inspection point missing inspection identification method
CN112702194B (en) * 2020-12-16 2023-04-07 中国联合网络通信集团有限公司 Indoor cell fault positioning method and device and electronic equipment
CN112802050B (en) * 2021-01-25 2024-04-16 商汤集团有限公司 Network training method, target tracking device, electronic equipment and storage medium
CN112884040B (en) * 2021-02-19 2024-04-30 北京小米松果电子有限公司 Training sample data optimization method, system, storage medium and electronic equipment
CN113052874B (en) * 2021-03-18 2022-01-25 上海商汤智能科技有限公司 Target tracking method and device, electronic equipment and storage medium
CN112735535B (en) * 2021-04-01 2021-06-25 腾讯科技(深圳)有限公司 Prediction model training method, prediction model training device, data prediction method, data prediction device and storage medium
CN114596337B (en) * 2022-03-03 2022-11-25 捻果科技(深圳)有限公司 Self-recognition target tracking method and system based on linkage of multiple camera positions

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106056628A (en) * 2016-05-30 2016-10-26 中国科学院计算技术研究所 Target tracking method and system based on deep convolution nerve network feature fusion
WO2017015947A1 (en) * 2015-07-30 2017-02-02 Xiaogang Wang A system and a method for object tracking
CN106709936A (en) * 2016-12-14 2017-05-24 北京工业大学 Single target tracking method based on convolution neural network
CN106846364A (en) * 2016-12-30 2017-06-13 明见(厦门)技术有限公司 A kind of method for tracking target and device based on convolutional neural networks
CN107066990A (en) * 2017-05-04 2017-08-18 厦门美图之家科技有限公司 A kind of method for tracking target and mobile device

Family Cites Families (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9669300B2 (en) * 2013-12-27 2017-06-06 Ballcraft, Llc Motion detection for existing portable devices
CN104616318B (en) * 2015-01-22 2017-06-16 重庆邮电大学 A kind of motion target tracking method in video sequence image
CN105741316B (en) * 2016-01-20 2018-10-16 西北工业大学 Robust method for tracking target based on deep learning and multiple dimensioned correlation filtering
CN105911518A (en) * 2016-03-31 2016-08-31 山东大学 Robot positioning method
CN106296734B (en) * 2016-08-05 2018-08-28 合肥工业大学 Method for tracking target based on extreme learning machine and boosting Multiple Kernel Learnings
CN106326924A (en) * 2016-08-23 2017-01-11 武汉大学 Object tracking method and object tracking system based on local classification
CN106920248A (en) * 2017-01-19 2017-07-04 博康智能信息技术有限公司上海分公司 A kind of method for tracking target and device

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2017015947A1 (en) * 2015-07-30 2017-02-02 Xiaogang Wang A system and a method for object tracking
CN106056628A (en) * 2016-05-30 2016-10-26 中国科学院计算技术研究所 Target tracking method and system based on deep convolution nerve network feature fusion
CN106709936A (en) * 2016-12-14 2017-05-24 北京工业大学 Single target tracking method based on convolution neural network
CN106846364A (en) * 2016-12-30 2017-06-13 明见(厦门)技术有限公司 A kind of method for tracking target and device based on convolutional neural networks
CN107066990A (en) * 2017-05-04 2017-08-18 厦门美图之家科技有限公司 A kind of method for tracking target and mobile device

Cited By (52)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11538246B2 (en) 2018-07-27 2022-12-27 Tencent Technology (Shenzhen) Company Limited Method and apparatus for training feature extraction model, computer device, and computer-readable storage medium
CN111694423B (en) * 2019-03-12 2023-05-26 阿里巴巴集团控股有限公司 Positioning, grabbing, data processing and display method and device for augmented reality
CN111694423A (en) * 2019-03-12 2020-09-22 阿里巴巴集团控股有限公司 Positioning, capturing, data processing and display method and equipment for augmented reality
CN109858472A (en) * 2019-04-09 2019-06-07 武汉领普科技有限公司 A kind of embedded humanoid detection method and device in real time
CN109858472B (en) * 2019-04-09 2023-08-04 武汉领普科技有限公司 Embedded real-time humanoid detection method and device
CN111815670A (en) * 2019-04-10 2020-10-23 曜科智能科技(上海)有限公司 Multi-view target tracking method, device and system, electronic terminal and storage medium
CN110148052A (en) * 2019-04-17 2019-08-20 深圳壹账通智能科技有限公司 Management-control method, device, computer equipment and storage medium after businessman borrows
CN110176027A (en) * 2019-05-27 2019-08-27 腾讯科技(深圳)有限公司 Video target tracking method, device, equipment and storage medium
CN110176027B (en) * 2019-05-27 2023-03-14 腾讯科技(深圳)有限公司 Video target tracking method, device, equipment and storage medium
CN110443827A (en) * 2019-07-22 2019-11-12 浙江大学 A kind of UAV Video single goal long-term follow method based on the twin network of improvement
CN110443827B (en) * 2019-07-22 2022-12-20 浙江大学 Unmanned aerial vehicle video single-target long-term tracking method based on improved twin network
CN110543818A (en) * 2019-07-25 2019-12-06 长沙行深智能科技有限公司 Traffic light tracking method, device, medium and equipment based on weight graph matching
CN110543818B (en) * 2019-07-25 2023-01-17 长沙行深智能科技有限公司 Traffic light tracking method, device, medium and equipment based on weight graph matching
CN110532883A (en) * 2019-07-30 2019-12-03 平安科技(深圳)有限公司 On-line tracking is improved using off-line tracking algorithm
CN110532883B (en) * 2019-07-30 2023-09-01 平安科技(深圳)有限公司 Improvement of on-line tracking algorithm by off-line tracking algorithm
CN110516556A (en) * 2019-07-31 2019-11-29 平安科技(深圳)有限公司 Multi-target tracking detection method, device and storage medium based on Darkflow-DeepSort
CN110516556B (en) * 2019-07-31 2023-10-31 平安科技(深圳)有限公司 Multi-target tracking detection method and device based on Darkflow-deep Sort and storage medium
CN110738101B (en) * 2019-09-04 2023-07-25 平安科技(深圳)有限公司 Behavior recognition method, behavior recognition device and computer-readable storage medium
CN110738101A (en) * 2019-09-04 2020-01-31 平安科技(深圳)有限公司 Behavior recognition method and device and computer readable storage medium
CN110929605A (en) * 2019-11-11 2020-03-27 中国建设银行股份有限公司 Video key frame storage method, device, equipment and storage medium
CN110991458A (en) * 2019-11-25 2020-04-10 创新奇智(北京)科技有限公司 Artificial intelligence recognition result sampling system and sampling method based on image characteristics
CN110956131A (en) * 2019-11-27 2020-04-03 北京迈格威科技有限公司 Single-target tracking method, device and system
CN110956131B (en) * 2019-11-27 2024-01-05 北京迈格威科技有限公司 Single-target tracking method, device and system
CN111091098B (en) * 2019-12-20 2023-08-15 浙江大华技术股份有限公司 Training method of detection model, detection method and related device
CN111091098A (en) * 2019-12-20 2020-05-01 浙江大华技术股份有限公司 Training method and detection method of detection model and related device
CN111241931A (en) * 2019-12-30 2020-06-05 沈阳理工大学 Aerial unmanned aerial vehicle target identification and tracking method based on YOLOv3
CN111241931B (en) * 2019-12-30 2023-04-18 沈阳理工大学 Aerial unmanned aerial vehicle target identification and tracking method based on YOLOv3
CN111242977A (en) * 2020-01-09 2020-06-05 影石创新科技股份有限公司 Target tracking method of panoramic video, readable storage medium and computer equipment
CN111242977B (en) * 2020-01-09 2023-04-25 影石创新科技股份有限公司 Target tracking method of panoramic video, readable storage medium and computer equipment
CN111311635A (en) * 2020-02-08 2020-06-19 腾讯科技(深圳)有限公司 Target positioning method, device and system
CN111275011A (en) * 2020-02-25 2020-06-12 北京百度网讯科技有限公司 Mobile traffic light detection method and device, electronic equipment and storage medium
CN111275011B (en) * 2020-02-25 2023-12-19 阿波罗智能技术(北京)有限公司 Mobile traffic light detection method and device, electronic equipment and storage medium
CN111401172B (en) * 2020-03-06 2023-10-27 大连海事大学 Port crane ladle automatic counting method based on video
CN111401172A (en) * 2020-03-06 2020-07-10 大连海事大学 Port hoisting material bag automatic counting method based on video
CN112655021A (en) * 2020-04-09 2021-04-13 深圳市大疆创新科技有限公司 Image processing method, image processing device, electronic equipment and storage medium
CN113642360A (en) * 2020-04-27 2021-11-12 杭州海康威视数字技术股份有限公司 Behavior timing method and device, electronic equipment and storage medium
CN113642360B (en) * 2020-04-27 2024-06-04 杭州海康威视数字技术股份有限公司 Behavior timing method and device, electronic equipment and storage medium
CN111652080A (en) * 2020-05-12 2020-09-11 合肥的卢深视科技有限公司 Target tracking method and device based on RGB-D image
CN111652080B (en) * 2020-05-12 2023-10-17 合肥的卢深视科技有限公司 Target tracking method and device based on RGB-D image
CN111862158A (en) * 2020-07-21 2020-10-30 湖南师范大学 Staged target tracking method and device, terminal and readable storage medium
CN111862158B (en) * 2020-07-21 2023-08-29 湖南师范大学 Staged target tracking method, device, terminal and readable storage medium
CN112183235A (en) * 2020-09-07 2021-01-05 根尖体育科技(北京)有限公司 Automatic control method for video acquisition aiming at sport places
CN112270661A (en) * 2020-10-19 2021-01-26 北京宇航系统工程研究所 Space environment monitoring method based on rocket telemetry video
CN112270661B (en) * 2020-10-19 2024-05-07 北京宇航系统工程研究所 Rocket telemetry video-based space environment monitoring method
CN114511793A (en) * 2020-11-17 2022-05-17 中国人民解放军军事科学院国防科技创新研究院 Unmanned aerial vehicle ground detection method and system based on synchronous detection and tracking
CN114511793B (en) * 2020-11-17 2024-04-05 中国人民解放军军事科学院国防科技创新研究院 Unmanned aerial vehicle ground detection method and system based on synchronous detection tracking
CN112733690B (en) * 2020-12-31 2024-02-20 北京易华录信息技术股份有限公司 High-altitude parabolic detection method and device and electronic equipment
CN112733690A (en) * 2020-12-31 2021-04-30 北京易华录信息技术股份有限公司 High-altitude parabolic detection method and device and electronic equipment
CN113066108A (en) * 2021-04-14 2021-07-02 武汉卓目科技有限公司 Anti-occlusion visual target tracking method and device based on ECO algorithm
CN113573137A (en) * 2021-07-01 2021-10-29 厦门美图之家科技有限公司 Video canvas boundary detection method, system, terminal equipment and storage medium
CN113573137B (en) * 2021-07-01 2023-08-08 厦门美图之家科技有限公司 Video canvas boundary detection method, system, terminal equipment and storage medium
CN113609971A (en) * 2021-08-04 2021-11-05 广州威拓电子科技有限公司 Method, device and equipment for inspecting microseism observation equipment and storage medium

Also Published As

Publication number Publication date
CN107679455A (en) 2018-02-09

Similar Documents

Publication Publication Date Title
WO2019041519A1 (en) Target tracking device and method, and computer-readable storage medium
CN109961009B (en) Pedestrian detection method, system, device and storage medium based on deep learning
US10534957B2 (en) Eyeball movement analysis method and device, and storage medium
US9390340B2 (en) Image-based character recognition
US8792722B2 (en) Hand gesture detection
CN104350509B (en) Quick attitude detector
US8750573B2 (en) Hand gesture detection
CN109325456B (en) Target identification method, target identification device, target identification equipment and storage medium
US9436883B2 (en) Collaborative text detection and recognition
US9710698B2 (en) Method, apparatus and computer program product for human-face features extraction
US20180211104A1 (en) Method and device for target tracking
WO2019033572A1 (en) Method for detecting whether face is blocked, device and storage medium
US10650234B2 (en) Eyeball movement capturing method and device, and storage medium
US9058536B1 (en) Image-based character recognition
US9269009B1 (en) Using a front-facing camera to improve OCR with a rear-facing camera
CN110598559B (en) Method and device for detecting motion direction, computer equipment and storage medium
US9569679B1 (en) Adaptive image sampling for text detection
WO2019033570A1 (en) Lip movement analysis method, apparatus and storage medium
WO2019033568A1 (en) Lip movement capturing method, apparatus and storage medium
US20210281744A1 (en) Action recognition method and device for target object, and electronic apparatus
WO2015003606A1 (en) Method and apparatus for recognizing pornographic image
KR20120044484A (en) Apparatus and method for tracking object in image processing system
CN110910445A (en) Object size detection method and device, detection equipment and storage medium
US9760177B1 (en) Color maps for object tracking
CN111199169A (en) Image processing method and device

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 17923072

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

32PN Ep: public notification in the ep bulletin as address of the adressee cannot be established

Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205A DATED 25/09/2020)

122 Ep: pct application non-entry in european phase

Ref document number: 17923072

Country of ref document: EP

Kind code of ref document: A1