WO2019041519A1 - Procédé et dispositif de suivi de cible, et support de stockage lisible par ordinateur - Google Patents

Procédé et dispositif de suivi de cible, et support de stockage lisible par ordinateur Download PDF

Info

Publication number
WO2019041519A1
WO2019041519A1 PCT/CN2017/108794 CN2017108794W WO2019041519A1 WO 2019041519 A1 WO2019041519 A1 WO 2019041519A1 CN 2017108794 W CN2017108794 W CN 2017108794W WO 2019041519 A1 WO2019041519 A1 WO 2019041519A1
Authority
WO
WIPO (PCT)
Prior art keywords
tracking
frame image
target
video
sample
Prior art date
Application number
PCT/CN2017/108794
Other languages
English (en)
Chinese (zh)
Inventor
周舒意
王建明
肖京
Original Assignee
平安科技(深圳)有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 平安科技(深圳)有限公司 filed Critical 平安科技(深圳)有限公司
Publication of WO2019041519A1 publication Critical patent/WO2019041519A1/fr

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/40Scenes; Scene-specific elements in video content
    • G06V20/48Matching video sequences
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/50Context or environment of the image
    • G06V20/52Surveillance or monitoring of activities, e.g. for recognising suspicious objects

Definitions

  • the present application relates to the field of image recognition technologies, and in particular, to a target tracking device, a method, and a computer readable storage medium based on a convolutional neural network.
  • Target tracking is an important part of practical applications such as video surveillance.
  • Target tracking refers to accurately locating and tracking moving targets (such as pedestrians, vehicles, etc.) in the video, and estimating the trajectory of the target.
  • target tracking has important value in video surveillance, target recognition, and video information discovery.
  • target tracking technology has been rapidly developed, but due to actual tracking, there are many practical difficulties in target tracking tasks, such as object occlusion, viewing angle changes, target deformation, ambient illumination changes, and unpredictable
  • the complex background situation, and the existing target tracking algorithm mostly uses the difference between the target and the background to construct the classification model, separates the target from the background, and tracks the target, but this tracking algorithm is difficult to adapt to the above mentioned in the tracking process. Changes in the target and background, such as partial occlusion of the target, or similar background interference, cause the target to be tracked incorrectly, resulting in low target tracking accuracy.
  • the present application provides a target tracking device, a method and a computer readable storage medium based on a convolutional neural network, the main purpose of which is to dynamically update a model during a tracking process to adapt to changes in targets and backgrounds, and to improve target tracking. Accuracy.
  • the present application provides a target tracking device based on a convolutional neural network, the device comprising a memory and a processor, the memory storing a target tracking program executable on the processor, the target The following steps are implemented when the tracking program is executed by the processor:
  • A. Collecting a plurality of picture samples from the video frame image according to the sampling point distribution, and recording position coordinates of each picture sample;
  • step D includes:
  • the processor is further configured to execute the target tracking program, to perform the following steps after the step E:
  • G adjusting the position of the sampling point on the video frame image according to the adjusted weight to update the sampling point distribution
  • the step F includes:
  • Steps A through G are repeated until the tracking of the tracking target in all video frame images of the video is completed.
  • the step G includes:
  • sampling point within a first preset range of the sampling point corresponding to the sample whose weight is greater than the first preset weight, and decreasing the sampling point in a second preset range of the sampling point corresponding to the sample whose weight is smaller than the second preset weight,
  • the second preset weight is smaller than the first preset weight, and the number of added sampling points is equal to the number of reduced sampling points.
  • the processor is further configured to execute the target tracking program to implement the following steps:
  • the video frame image is the first frame image of the video, prompting the user to manually select a tracking target on the video frame image and receive a tracking target selected by the user based on the prompt, and after determining the tracking target
  • step A is performed.
  • the present application further provides a target tracking method based on a convolutional neural network, the method comprising:
  • A. Collecting a plurality of picture samples from the video frame image according to the sampling point distribution, and recording position coordinates of each picture sample;
  • step D includes:
  • step E the method further includes:
  • G adjusting the position of the sampling point on the video frame image according to the adjusted weight to update the sampling point distribution
  • the step F includes:
  • Steps A through G are repeated until the tracking of the tracking target in all video frame images of the video is completed.
  • the step G includes:
  • sampling point within a first preset range of the sampling point corresponding to the sample whose weight is greater than the first preset weight, and decreasing the sampling point in a second preset range of the sampling point corresponding to the sample whose weight is smaller than the second preset weight,
  • the second preset weight is smaller than the first preset weight, and the number of added sampling points is equal to the number of reduced sampling points.
  • the present application further provides a computer readable storage medium having a target tracking program stored thereon, the target tracking program being executable by one or more processors to implement Next steps:
  • A. Collecting a plurality of picture samples from the video frame image according to the sampling point distribution, and recording position coordinates of each picture sample;
  • the object tracking device, the method and the computer readable storage medium based on the convolutional neural network proposed by the present application identify the video frame image in the video frame by frame, and collect multiple image samples from the video frame image according to the sampling point distribution, and Recording position coordinates of each picture sample, extracting a plurality of sample features correspondingly from the plurality of sample pictures based on the CNN model, calculating a confidence level between each picture sample and the tracking target according to the extracted sample features, and adjusting the sample according to the confidence level Weighting, and then calculating the position coordinates of the tracking target on the video frame image according to the position coordinates and weights of the sample, and collecting positive and negative samples of the tracking target from the video frame image according to the position coordinates, and retraining the CNN using the collected samples
  • the model updates the model parameters, continues to track the next frame image using the model after updating the model parameters, and so on.
  • the model After obtaining the tracking result of each frame image, the model is updated according to the tracking result, so that the tracking target changes.
  • the updated model can adapt to changes in goals and backgrounds. Even when the phenomenon of partial occlusion, background interference appear in the image, it can also be successfully tracking the target, improve the accuracy of target tracking.
  • FIG. 1 is a schematic diagram of a preferred embodiment of a target tracking device based on a convolutional neural network
  • FIG. 2 is a schematic diagram of a program module of a target tracking program in an embodiment of a target tracking device based on a convolutional neural network;
  • FIG. 3 is a flow chart of a preferred embodiment of a target tracking method based on a convolutional neural network.
  • the application provides a target tracking device based on a convolutional neural network.
  • a schematic diagram of a preferred embodiment of a target tracking device based on a convolutional neural network is provided.
  • the target tracking device based on the convolutional neural network may be a PC (Personal Computer), or may be a smart phone, a tablet computer, an e-book reader, or a portable device.
  • a terminal device having a display function such as a computer.
  • the convolutional neural network based target tracking device includes a memory 11, a processor 12, a communication bus 13, and a network interface 14.
  • the memory 11 includes at least one type of readable storage medium including a flash memory, a hard disk, a multimedia card, a card type memory (for example, an SD or DX memory, etc.), a magnetic memory, a magnetic disk, an optical disk, and the like.
  • Memory 11 may in some embodiments be an internal storage unit of a target tracking device based on a convolutional neural network, such as a hard disk of a target tracking device based on a convolutional neural network.
  • the memory 11 may also be an external storage device of a target tracking device based on a convolutional neural network in other embodiments, such as a plug-in hard disk equipped on a target tracking device based on a convolutional neural network, a smart memory card (Smart Media Card) , SMC), Secure Digital (SD) card, Flash Card, etc. Further, the memory 11 may also include both an internal storage unit of the target tracking device based on the convolutional neural network and an external storage device. The memory 11 can be used not only for storing application software and various types of data installed on a target tracking device based on a convolutional neural network, such as code of a target tracking program, but also for temporarily storing data that has been output or is to be output.
  • the processor 12 may be a Central Processing Unit (CPU), controller, microcontroller, microprocessor or other data processing chip for running program code or processing stored in the memory 11. Data, such as executing a target tracking program.
  • CPU Central Processing Unit
  • controller microcontroller
  • microprocessor or other data processing chip for running program code or processing stored in the memory 11.
  • Data such as executing a target tracking program.
  • Communication bus 13 is used to implement connection communication between these components.
  • the network interface 14 can optionally include a standard wired interface, a wireless interface (such as a WI-FI interface), and is typically used to establish a communication connection between the device and other electronic devices.
  • a standard wired interface such as a WI-FI interface
  • Figure 1 shows only a convolutional neural network based target tracking device with components 11-14 and a target tracking program, but it should be understood that not all illustrated components may be implemented, alternative implementations may be more or more Less components.
  • the device may further include a user interface
  • the user interface may include a display
  • an input unit such as a keyboard
  • the optional user interface may further include a standard wired interface and a wireless interface.
  • the display may be an LED display, a liquid crystal display, a touch-sensitive liquid crystal display, an OLED (Organic Light-Emitting Diode) touch sensor, or the like.
  • the display may also be suitably referred to as a display screen or display unit for displaying information processed in a target tracking device based on a convolutional neural network and a user interface for displaying visualization.
  • the device may also include a touch sensor.
  • the area provided by the touch sensor for the user to perform a touch operation is referred to as a touch area.
  • the touch sensor described herein may be a resistive touch sensor, a capacitive touch sensor, or the like.
  • the touch sensor includes not only a contact type touch sensor but also a proximity type touch sensor or the like.
  • the touch sensor may be a single sensor or a plurality of sensors arranged, for example, in an array.
  • the area of the display of the device may be the same as or different from the area of the touch sensor.
  • a display is stacked with the touch sensor to form a touch display. The device is based on a touch display The screen detects the touch operation triggered by the user.
  • the device may further include a camera, an RF (Radio Frequency) circuit, a sensor, an audio circuit, a WiFi module, and the like.
  • sensors such as light sensors, motion sensors, and other sensors.
  • the light sensor may include an ambient light sensor and a proximity sensor, wherein if the device is a mobile terminal, the ambient light sensor may adjust the brightness of the display screen according to the brightness of the ambient light, and the proximity sensor may move when the mobile terminal moves to the ear. , turn off the display and / or backlight.
  • the gravity acceleration sensor can detect the magnitude of acceleration in each direction (usually three axes), and can detect the magnitude and direction of gravity when stationary, and can be used to identify the posture of the mobile terminal (such as horizontal and vertical screen switching, Related games, magnetometer attitude calibration), vibration recognition related functions (such as pedometer, tapping), etc.; of course, the mobile terminal can also be equipped with other sensors such as gyroscope, barometer, hygrometer, thermometer, infrared sensor, etc. No longer.
  • a target tracking program is stored in the memory 11; when the processor 12 executes the target tracking program stored in the memory 11, the following steps are implemented:
  • A. Collecting a plurality of picture samples from the video frame image according to a sampling point distribution.
  • a convolutional neural network is used to perform offline training on a massive picture to obtain a CNN (Convolutional Neural Network) model, which may be a two-class model, and the model can be extracted from the image.
  • CNN Convolutional Neural Network
  • the video image is tracked frame by frame.
  • a video to be subjected to target tracking is input to the device, and the device processes each video frame image in the video in accordance with the following operation.
  • the image samples are collected from the video frame image according to the sampling point distribution, wherein the number of sampling points can be preset by the user, for example, 100 image samples are collected, wherein when the first frame image is started to be recognized, the user can manually
  • the tracking target is selected in the image.
  • the tracking target is selected by a frame selection manner, and the sampling point distribution is initialized based on the position of the tracking target selected by the user.
  • the video frame image when the video frame image is received, determining whether the video frame image is the first frame image of the video; if the video frame image is the first frame image of the video, prompting the user to Manually selecting a tracking target on the video frame image and receiving a tracking target selected by the user based on the prompt; after determining the tracking target, initializing a sampling point distribution and a training sample set of the CNN model and receiving a second frame image; The video image is not the first frame image of the video, and the step A is performed.
  • the user sets the target to be tracked in advance and stores it, the tracking target is directly acquired after starting the tracking, and the user is not required to manually select from the first frame image.
  • the color histogram of the tracking target area is calculated and used as the target feature of the tracking target, and the target feature can be represented as a vector of N*1.
  • the collected sample image is input into the trained CNN model for feature extraction, and the sample feature is extracted, and the sample feature can also be represented as an N*1 vector.
  • a sample feature is extracted corresponding to each sample picture, and the confidence between each sample feature and the target feature is calculated separately.
  • the confidence of the sample features reflects the similarity between the image sample and the tracking target.
  • the similarity between the two N*1 vectors is calculated as a sample of the image. Confidence with tracking targets.
  • the weight of each picture sample is adjusted according to the confidence level, and for the sample with small confidence, the weight is reduced, and for the sample with high confidence, the weight is increased, and then for all the weights, then for all
  • the weights of the image samples are normalized such that the sum of the weights of all samples is equal to one.
  • the position coordinates of the tracking target on the video frame image are calculated based on the weight value of the picture sample and its position coordinates on the video frame image. Specifically, it is assumed that a total of k picture samples are collected, wherein the position coordinates of the sample P i are (x i , y i ), and the confidence degree with the tracking target is S i . Then, the position coordinates (x, y) of the tracking target can be predicted according to the following formula.
  • D. Collect positive and negative samples of the tracking target from the video frame image according to the position coordinates.
  • Collecting a positive sample and a negative sample of the tracking target from the video frame image according to the position coordinate specifically, acquiring a first preset number of picture samples located in a peripheral area of the position coordinate as a positive sample, wherein the periphery The area is an area formed by a point having a distance from the position coordinate that is smaller than a first preset threshold; and a second predetermined number of picture samples located in a distant area of the position coordinate are collected as a negative sample, wherein the The remote preset area is an area formed by a point that is greater than a second preset threshold value, and the second preset threshold is greater than the first preset threshold.
  • the image samples are collected from the region closer to the tracking target, and the difference between the samples and the tracking target is small, and can be used as a positive sample from the video frame image. Capture image samples from areas farther away from the tracking target.
  • the difference between the tracking targets is large, and can be added as a negative sample to the training sample set of the CNN model, and the CNN model is trained to update the model parameters and improve the accuracy of the model to identify the features of the moving target from the image samples. To enable the model to adapt to changes in the target and background in the video frame image.
  • the CNN model is continuously updated, and even if the tracking target is partially occluded or the background interferes with the tracking target, it does not interfere with the correct tracking of the target.
  • the next frame image is continuously tracked, and the updated CNN model is used for feature extraction.
  • Target tracking is performed for each frame of image according to steps A through E, and after the tracking is completed, the CNN model is trained until the tracking of the target in all frame images of the video is completed.
  • the first preset threshold, the second preset threshold, the first preset number, and the second preset number may be preset by use.
  • step E the following steps are further implemented:
  • G adjusting the position of the sampling point on the video frame image according to the adjusted weight to update the sampling point distribution
  • the step F includes repeating steps A to G until the tracking of the tracking target in all video frame images of the video is completed.
  • the distribution of the sampling points is adjusted according to the adjusted weights. Specifically, the sampling points are added in the first preset range of the sampling points corresponding to the samples whose weight is greater than the first preset weight, that is, the pictures with the weights are significant. Adding more sampling points near the sampling point corresponding to the sample, and reducing the sampling point in the second preset range of the sampling point corresponding to the sample whose weight is smaller than the second preset weight, wherein the second preset weight is smaller than the first preset
  • the weight that is, the sampling point near the sampling point corresponding to the picture sample with small weight reduction, wherein the number of added sampling points is equal to or greater than the number of reduced sampling points, or, when the weight is very small, the corresponding sampling point can be deleted. For example, the sampling point corresponding to the sample whose weight is smaller than the third preset weight is deleted, wherein the third preset weight is smaller than the fourth preset weight.
  • the target tracking device based on the convolutional neural network proposed in this embodiment performs frame-by-frame recognition on the video frame image in the video, collects multiple picture samples from the video frame image according to the sampling point distribution, and records the position coordinates of each picture sample. And extracting a plurality of sample features correspondingly from the plurality of sample pictures based on the CNN model, calculating a confidence degree between each picture sample and the tracking target according to the extracted sample features, adjusting the weight of the sample according to the confidence degree, and further according to the position of the sample Coordinates and weights calculate the position coordinates of the tracking target on the video frame image, and collect positive and negative samples of the tracking target from the video frame image according to the position coordinates, and retrain the CNN model to update the model parameters using the collected samples, using After updating the model parameters, the model continues to track the next frame image, and so on.
  • the model After obtaining the tracking result of each frame image, the model is updated according to the tracking result, so that the updated model can be changed when the tracking target changes.
  • Adapt to changes in goals and background, even if partial obscuration occurs in the image When the phenomenon of background interference, but also able to successfully track the target, improve the accuracy of target tracking.
  • the target tracking program may also be divided into one or more modules, one or more modules being stored in the memory 11 and being processed by one or more processors (this Embodiments are executed by processor 12) to accomplish the present application, and a module referred to herein refers to a series of computer program instructions that are capable of performing a particular function.
  • FIG. 2 it is a schematic diagram of a program module of a target tracking program in an embodiment of a target tracking device based on a convolutional neural network according to the present application.
  • the target tracking program may be divided into a preprocessing module 10, Tracking module 20, sampling module 30 and update module 40, illustratively,
  • the collecting module 10 is configured to: collect a plurality of picture samples from the video frame image according to the sampling point distribution, and record position coordinates of each picture sample;
  • the pre-processing module 20 is configured to: correspondingly extract a plurality of sample features from the plurality of picture samples based on a convolutional neural network CNN model, and respectively calculate a confidence between each picture sample and the tracking target according to the extracted sample features respectively degree;
  • the tracking module 30 is configured to: adjust weights of the corresponding picture samples according to the calculated confidence, and calculate position coordinates of the tracking target on the video frame image according to position coordinates and weights of all picture samples;
  • the sampling module 40 is configured to: collect positive and negative samples of the tracking target from the video frame image according to the position coordinates;
  • the updating module 50 is configured to: update the training sample set of the CNN model according to the positive sample and the negative sample, and train the CNN model to update model parameters of the CNN model by using the updated training sample set;
  • the acquisition module 10, the pre-processing module 20, the tracking module 30, the sampling module 40, and the update module 50 perform the above steps to track the target in the order of the video frame images in the video until the tracking target in all the video frame images in the video is completed. Tracking.
  • the present application also provides a target tracking method based on a convolutional neural network.
  • FIG. 3 it is a flowchart of a preferred embodiment of a target tracking method based on a convolutional neural network. The method can be performed by a device that can be implemented by software and/or hardware.
  • the target tracking method based on the convolutional neural network includes:
  • Step S10 Collect a plurality of picture samples from the video frame image according to the sampling point distribution, and record position coordinates of each picture sample.
  • a convolutional neural network is used to perform offline training on a massive picture to obtain a CNN (Convolutional Neural Network) model, which may be a two-class model, and the model can be extracted from the image.
  • CNN Convolutional Neural Network
  • the video image is tracked frame by frame.
  • a video to be subjected to target tracking is input to the device, and the device processes each video frame image in the video in accordance with the following operation.
  • the image samples are collected from the video frame image according to the sampling point distribution, wherein the number of sampling points can be preset by the user, for example, 100 image samples are collected, wherein when the first frame image is started to be recognized, the user can manually
  • the tracking target is selected in the image.
  • the tracking target is selected by a frame selection manner, and the sampling point distribution is initialized based on the position of the tracking target selected by the user.
  • the video frame image when the video frame image is received, determining whether the video frame image is the first frame image of the video; if the video frame image is the first frame image of the video, prompting the user to Manually selecting a tracking target on the video frame image and receiving a tracking target selected by the user based on the prompt; after determining the tracking target, initializing a sampling point distribution and a training sample set of the CNN model and receiving a second frame image; The video image is not the first frame image of the video, and the step S10 is performed.
  • the user sets the target to be tracked in advance and stores it, the tracking target is directly acquired after starting the tracking, and the user is not required to manually select from the first frame image.
  • the color histogram of the tracking target area is calculated and used as the target feature of the tracking target, and the target feature can be represented as a vector of N*1.
  • Step S20 correspondingly extracting a plurality of sample features from the plurality of picture samples based on the convolutional neural network CNN model, and respectively calculating a confidence level between each picture sample and the tracking target according to the extracted sample features.
  • Step S30 adjusting the weight of the corresponding picture sample according to the calculated confidence, and calculating the position coordinates of the tracking target on the video frame image according to the position coordinates and weights of all the picture samples.
  • the collected sample image is input into the trained CNN model for feature extraction, and the sample feature is extracted, and the sample feature can also be represented as an N*1 vector.
  • a sample feature is extracted corresponding to each sample picture, and the confidence between each sample feature and the target feature is calculated separately.
  • the confidence of the sample features reflects the similarity between the image sample and the tracking target.
  • the similarity between the two N*1 vectors is calculated as a sample of the image. Confidence with tracking targets.
  • the weight of each picture sample is adjusted according to the confidence level, and for the sample with small confidence, the weight is reduced, and for the sample with high confidence, the weight is increased, and then for all the weights, then for all
  • the weights of the image samples are normalized such that the sum of the weights of all samples is equal to one.
  • the position coordinates of the tracking target on the video frame image are calculated based on the weight value of the picture sample and its position coordinates on the video frame image. Specifically, it is assumed that a total of k picture samples are collected, wherein the position coordinates of the sample P i are (x i , y i ), and the confidence degree with the tracking target is S i . Then, the position coordinates (x, y) of the tracking target can be predicted according to the following formula.
  • Step S40 collecting positive and negative samples of the tracking target from the video frame image according to the position coordinates.
  • Step S50 updating the training sample set of the CNN model according to the positive sample and the negative sample, and training the CNN model to update the model parameters of the CNN model by using the updated training sample set.
  • step S60 steps S10 to S50 are repeatedly performed until the tracking of the tracking target in all the video frame images of the video is completed.
  • Collecting a positive sample and a negative sample of the tracking target from the video frame image according to the position coordinate specifically, acquiring a first preset number of picture samples located in a peripheral area of the position coordinate as a positive sample, wherein the periphery The area is an area formed by a point having a distance from the position coordinate that is smaller than a first preset threshold; and a second predetermined number of picture samples located in a distant area of the position coordinate are collected as a negative sample, wherein the The remote preset area is an area formed by a point that is greater than a second preset threshold value, and the second preset threshold is greater than the first preset threshold.
  • the image samples are collected from the region closer to the tracking target, and the difference between the samples and the tracking target is small, and can be used as a positive sample from the video frame image.
  • Image samples are taken from areas farther away from the tracking target. These samples have a large difference from the tracking targets. They can be added as negative samples to the training sample set of the CNN model, and the CNN model is trained to update the model parameters.
  • the improved model identifies the accuracy of the features of the moving object from the image samples so that the model can adapt to changes in the target and background in the video frame image.
  • the CNN model is continuously updated, and even if the tracking target is partially occluded or the background interferes with the tracking target, it does not interfere with the correct tracking of the target.
  • the next frame image is continuously tracked, and the updated CNN model is used for feature extraction.
  • Target tracking is performed for each frame of image in accordance with steps S10 through S40, and after the tracking is completed, the CNN model is trained until all tracking of the target in all frame images of the video is completed.
  • the first preset threshold, the second preset threshold, the first preset number, and the second preset number may be preset by use.
  • the method further includes the following steps: adjusting the distribution of the sampling points according to the adjusted weights, specifically, the sampling corresponding to the samples having the weight greater than the first preset weight Adding a sampling point within a first preset range of the point, that is, adding more sampling points near the sampling point corresponding to the image sample with a large weight, and second sampling of the sampling point corresponding to the sample whose weight is smaller than the second preset weight
  • the sampling point is reduced in the range, wherein the second preset weight is smaller than the first preset weight, that is, the sampling point near the sampling point corresponding to the image sample with small weight is reduced, wherein
  • the number of added sampling points is equal to or greater than the number of reduced sampling points, or, when the weight is very small, the corresponding sampling points may be deleted, for example, the sampling points corresponding to the samples whose weights are smaller than the third preset weight are deleted, wherein The third preset weight is smaller than the fourth preset weight.
  • the target tracking method based on the convolutional neural network proposed in this embodiment performs frame-by-frame recognition on the video frame image in the video, collects multiple picture samples from the video frame image according to the sampling point distribution, and records the position coordinates of each picture sample. And extracting a plurality of sample features correspondingly from the plurality of sample pictures based on the CNN model, calculating a confidence degree between each picture sample and the tracking target according to the extracted sample features, adjusting the weight of the sample according to the confidence degree, and further according to the position of the sample Coordinates and weights calculate the position coordinates of the tracking target on the video frame image, and collect positive and negative samples of the tracking target from the video frame image according to the position coordinates, and retrain the CNN model to update the model parameters using the collected samples, using After updating the model parameters, the model continues to track the next frame image, and so on.
  • the model After obtaining the tracking result of each frame image, the model is updated according to the tracking result, so that the updated model can be changed when the tracking target changes.
  • Adapt to changes in goals and background, even if partial obscuration occurs in the image When the phenomenon of background interference, but also able to successfully track the target, improve the accuracy of target tracking.
  • the embodiment of the present application further provides a computer readable storage medium, where the target readable program is stored on the computer readable storage medium, and the target tracking program can be executed by one or more processors to implement the following operations:
  • A. Collecting a plurality of picture samples from the video frame image according to the sampling point distribution, and recording position coordinates of each picture sample;
  • the position of the sampling point on the video frame image is adjusted according to the adjusted weight to update the sampling point distribution.
  • sampling point within a first preset range of the sampling point corresponding to the sample whose weight is greater than the first preset weight, and decreasing the sampling point in a second preset range of the sampling point corresponding to the sample whose weight is smaller than the second preset weight,
  • the second preset weight is smaller than the first preset weight, and the number of added sampling points is equal to the number of reduced sampling points.
  • the specific embodiment of the computer readable storage medium of the present application is substantially the same as the foregoing embodiments of the target tracking apparatus and method based on the convolutional neural network, and is not described herein.
  • the technical solution of the present application which is essential or contributes to the prior art, may be embodied in the form of a software product stored in a storage medium (such as ROM/RAM as described above). , a disk, an optical disk, including a number of instructions for causing a terminal device (which may be a mobile phone, a computer, a server, or a network device, etc.) to perform the methods described in the various embodiments of the present application.
  • a terminal device which may be a mobile phone, a computer, a server, or a network device, etc.

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Multimedia (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Computation (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Artificial Intelligence (AREA)
  • General Engineering & Computer Science (AREA)
  • Biomedical Technology (AREA)
  • Computational Linguistics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Health & Medical Sciences (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Biophysics (AREA)
  • Evolutionary Biology (AREA)
  • General Health & Medical Sciences (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Image Analysis (AREA)

Abstract

L'invention concerne un dispositif et un procédé de suivi de cible basés sur un réseau neuronal convolutionnel, et un support de stockage lisible par ordinateur, ceux-ci pouvant améliorer la précision de suivi de cible. Le dispositif comprend une mémoire et un processeur, un programme de suivi cible pouvant être exécuté sur le processeur étant stocké dans la mémoire, et lorsque le programme est exécuté par le processeur, les étapes suivantes sont mises en œuvre : collecter des échantillons d'image à partir d'une image de trame vidéo selon une distribution de point d'échantillonnage, et enregistrer des coordonnées de position des échantillons d'image (S10); extraire, sur la base d'un modèle de CNN, des caractéristiques d'échantillon à partir des échantillons d'image, et calculer un degré de confiance des échantillons d'image et une cible suivie selon les caractéristiques d'échantillon (S20); ajuster le poids des échantillons d'image selon le degré de confiance, et calculer les coordonnées de position de la cible suivie selon les coordonnées de position et le poids (S30); collecter un échantillon positif et un échantillon négatif à partir de l'image de trame vidéo selon les coordonnées de position pour entraîner un modèle de CNN avec un ensemble d'échantillons d'apprentissage (S40); puis mettre à jour des paramètres de modèle du modèle de CNN (S50); et répéter les étapes mentionnées ci-dessus jusqu'à ce que le suivi d'une vidéo soit achevé (S60).
PCT/CN2017/108794 2017-08-29 2017-10-31 Procédé et dispositif de suivi de cible, et support de stockage lisible par ordinateur WO2019041519A1 (fr)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN201710754313.XA CN107679455A (zh) 2017-08-29 2017-08-29 目标跟踪装置、方法及计算机可读存储介质
CN201710754313.X 2017-08-29

Publications (1)

Publication Number Publication Date
WO2019041519A1 true WO2019041519A1 (fr) 2019-03-07

Family

ID=61134784

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2017/108794 WO2019041519A1 (fr) 2017-08-29 2017-10-31 Procédé et dispositif de suivi de cible, et support de stockage lisible par ordinateur

Country Status (2)

Country Link
CN (1) CN107679455A (fr)
WO (1) WO2019041519A1 (fr)

Cited By (31)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109858472A (zh) * 2019-04-09 2019-06-07 武汉领普科技有限公司 一种嵌入式实时人形检测方法和装置
CN110148052A (zh) * 2019-04-17 2019-08-20 深圳壹账通智能科技有限公司 商家贷后管控方法、装置、计算机设备及存储介质
CN110176027A (zh) * 2019-05-27 2019-08-27 腾讯科技(深圳)有限公司 视频目标跟踪方法、装置、设备及存储介质
CN110443827A (zh) * 2019-07-22 2019-11-12 浙江大学 一种基于改进孪生网络的无人机视频单目标长期跟踪方法
CN110516556A (zh) * 2019-07-31 2019-11-29 平安科技(深圳)有限公司 基于Darkflow-DeepSort的多目标追踪检测方法、装置及存储介质
CN110532883A (zh) * 2019-07-30 2019-12-03 平安科技(深圳)有限公司 应用离线跟踪算法对在线跟踪算法进行改进
CN110543818A (zh) * 2019-07-25 2019-12-06 长沙行深智能科技有限公司 基于权图匹配的红绿灯跟踪方法、装置、介质及设备
CN110738101A (zh) * 2019-09-04 2020-01-31 平安科技(深圳)有限公司 行为识别方法、装置及计算机可读存储介质
CN110929605A (zh) * 2019-11-11 2020-03-27 中国建设银行股份有限公司 视频关键帧的保存方法、装置、设备及存储介质
CN110956131A (zh) * 2019-11-27 2020-04-03 北京迈格威科技有限公司 单目标追踪方法、装置及系统
CN110991458A (zh) * 2019-11-25 2020-04-10 创新奇智(北京)科技有限公司 基于图像特征的人工智能识别结果抽样系统及抽样方法
CN111091098A (zh) * 2019-12-20 2020-05-01 浙江大华技术股份有限公司 检测模型的训练方法、检测方法及相关装置
CN111241931A (zh) * 2019-12-30 2020-06-05 沈阳理工大学 一种基于YOLOv3的空中无人机目标识别和跟踪方法
CN111242977A (zh) * 2020-01-09 2020-06-05 影石创新科技股份有限公司 全景视频的目标跟踪方法、可读存储介质及计算机设备
CN111275011A (zh) * 2020-02-25 2020-06-12 北京百度网讯科技有限公司 移动红绿灯检测方法、装置、电子设备和存储介质
CN111311635A (zh) * 2020-02-08 2020-06-19 腾讯科技(深圳)有限公司 一种目标定位方法、装置及系统
CN111401172A (zh) * 2020-03-06 2020-07-10 大连海事大学 一种基于视频的港口吊装料包自动计数方法
CN111652080A (zh) * 2020-05-12 2020-09-11 合肥的卢深视科技有限公司 基于rgb-d图像的目标跟踪方法和装置
CN111694423A (zh) * 2019-03-12 2020-09-22 阿里巴巴集团控股有限公司 定位、抓取、数据处理与用于增强现实的显示方法及设备
CN111815670A (zh) * 2019-04-10 2020-10-23 曜科智能科技(上海)有限公司 多视图目标跟踪方法、装置、系统、电子终端、及存储介质
CN111862158A (zh) * 2020-07-21 2020-10-30 湖南师范大学 一种分阶段目标跟踪方法、装置、终端及可读存储介质
CN112183235A (zh) * 2020-09-07 2021-01-05 根尖体育科技(北京)有限公司 针对运动场所进行视频采集的自动控制方法
CN112270661A (zh) * 2020-10-19 2021-01-26 北京宇航系统工程研究所 一种基于火箭遥测视频的空间环境监测方法
CN112655021A (zh) * 2020-04-09 2021-04-13 深圳市大疆创新科技有限公司 图像处理方法、装置、电子设备和存储介质
CN112733690A (zh) * 2020-12-31 2021-04-30 北京易华录信息技术股份有限公司 一种高空抛物检测方法、装置及电子设备
CN113066108A (zh) * 2021-04-14 2021-07-02 武汉卓目科技有限公司 基于eco算法的抗遮挡视觉目标跟踪方法及装置
CN113573137A (zh) * 2021-07-01 2021-10-29 厦门美图之家科技有限公司 视频画布边界检测方法、系统、终端设备及存储介质
CN113609971A (zh) * 2021-08-04 2021-11-05 广州威拓电子科技有限公司 微地震观测设备的检查方法、装置、设备及存储介质
CN113642360A (zh) * 2020-04-27 2021-11-12 杭州海康威视数字技术股份有限公司 一种行为计时方法、装置、电子设备及存储介质
CN114511793A (zh) * 2020-11-17 2022-05-17 中国人民解放军军事科学院国防科技创新研究院 一种基于同步检测跟踪的无人机对地探测方法及系统
US11538246B2 (en) 2018-07-27 2022-12-27 Tencent Technology (Shenzhen) Company Limited Method and apparatus for training feature extraction model, computer device, and computer-readable storage medium

Families Citing this family (24)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108320297B (zh) * 2018-03-09 2020-06-19 湖北工业大学 一种视频目标实时跟踪方法及系统
WO2019213820A1 (fr) * 2018-05-07 2019-11-14 合刃科技(武汉)有限公司 Procédé de commande de photographie et dispositif électronique
CN108960046A (zh) 2018-05-23 2018-12-07 北京图森未来科技有限公司 一种训练数据采样方法及其装置、计算机服务器
CN108765455B (zh) * 2018-05-24 2021-09-21 中国科学院光电技术研究所 一种基于tld算法的目标稳定跟踪方法
CN108846850B (zh) * 2018-05-24 2022-06-10 中国科学院光电技术研究所 一种基于tld算法的目标跟踪方法
CN110569690B (zh) * 2018-06-06 2022-05-13 浙江宇视科技有限公司 目标信息获取方法及装置
CN109448018B (zh) * 2018-09-18 2023-08-01 平安科技(深圳)有限公司 跟踪目标的定位方法、装置、设备及存储介质
CN111127509B (zh) * 2018-10-31 2023-09-01 杭州海康威视数字技术股份有限公司 目标跟踪方法、装置和计算机可读存储介质
CN109711248A (zh) * 2018-11-08 2019-05-03 平安科技(深圳)有限公司 一种基于视频的环境识别方法及终端设备
CN109635657B (zh) * 2018-11-12 2023-01-06 平安科技(深圳)有限公司 目标跟踪方法、装置、设备及存储介质
CN109600627B (zh) * 2018-12-11 2021-12-24 国信优易数据股份有限公司 一种视频识别方法和装置
CN111310526B (zh) * 2018-12-12 2023-10-20 杭州海康威视数字技术股份有限公司 目标跟踪模型的参数确定方法、装置及存储介质
CN109598885B (zh) * 2018-12-21 2021-06-11 广东中安金狮科创有限公司 监控系统及其报警方法
CN109886998A (zh) * 2019-01-23 2019-06-14 平安科技(深圳)有限公司 多目标跟踪方法、装置、计算机装置及计算机存储介质
CN110288082B (zh) * 2019-06-05 2022-04-05 北京字节跳动网络技术有限公司 卷积神经网络模型训练方法、装置和计算机可读存储介质
CN111462194B (zh) * 2020-03-30 2023-08-11 苏州科达科技股份有限公司 对象跟踪模型的训练方法、装置及存储介质
CN111653103A (zh) * 2020-05-07 2020-09-11 浙江大华技术股份有限公司 一种目标对象的识别方法及装置
CN112132088B (zh) * 2020-09-29 2024-01-12 动联(山东)电子科技有限公司 一种巡检点位漏巡识别方法
CN112702194B (zh) * 2020-12-16 2023-04-07 中国联合网络通信集团有限公司 室分小区故障定位方法、装置及电子设备
CN112802050B (zh) * 2021-01-25 2024-04-16 商汤集团有限公司 网络训练、目标跟踪方法及装置、电子设备和存储介质
CN112884040B (zh) * 2021-02-19 2024-04-30 北京小米松果电子有限公司 训练样本数据的优化方法、系统、存储介质及电子设备
CN113052874B (zh) * 2021-03-18 2022-01-25 上海商汤智能科技有限公司 目标跟踪方法及装置、电子设备和存储介质
CN112735535B (zh) * 2021-04-01 2021-06-25 腾讯科技(深圳)有限公司 预测模型训练、数据预测方法、装置和存储介质
CN114596337B (zh) * 2022-03-03 2022-11-25 捻果科技(深圳)有限公司 一种基于多摄像机位联动性的自识别目标追踪方法及系统

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106056628A (zh) * 2016-05-30 2016-10-26 中国科学院计算技术研究所 基于深度卷积神经网络特征融合的目标跟踪方法及系统
WO2017015947A1 (fr) * 2015-07-30 2017-02-02 Xiaogang Wang Système et procédé pour un suivi d'objet
CN106709936A (zh) * 2016-12-14 2017-05-24 北京工业大学 一种基于卷积神经网络的单目标跟踪方法
CN106846364A (zh) * 2016-12-30 2017-06-13 明见(厦门)技术有限公司 一种基于卷积神经网络的目标跟踪方法及装置
CN107066990A (zh) * 2017-05-04 2017-08-18 厦门美图之家科技有限公司 一种目标跟踪方法及移动设备

Family Cites Families (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9669300B2 (en) * 2013-12-27 2017-06-06 Ballcraft, Llc Motion detection for existing portable devices
CN104616318B (zh) * 2015-01-22 2017-06-16 重庆邮电大学 一种视频序列图像中的运动目标跟踪方法
CN105741316B (zh) * 2016-01-20 2018-10-16 西北工业大学 基于深度学习和多尺度相关滤波的鲁棒目标跟踪方法
CN105911518A (zh) * 2016-03-31 2016-08-31 山东大学 机器人定位方法
CN106296734B (zh) * 2016-08-05 2018-08-28 合肥工业大学 基于极限学习机和boosting多核学习的目标跟踪方法
CN106326924A (zh) * 2016-08-23 2017-01-11 武汉大学 一种基于局部分类的目标跟踪方法及系统
CN106920248A (zh) * 2017-01-19 2017-07-04 博康智能信息技术有限公司上海分公司 一种目标跟踪方法及装置

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2017015947A1 (fr) * 2015-07-30 2017-02-02 Xiaogang Wang Système et procédé pour un suivi d'objet
CN106056628A (zh) * 2016-05-30 2016-10-26 中国科学院计算技术研究所 基于深度卷积神经网络特征融合的目标跟踪方法及系统
CN106709936A (zh) * 2016-12-14 2017-05-24 北京工业大学 一种基于卷积神经网络的单目标跟踪方法
CN106846364A (zh) * 2016-12-30 2017-06-13 明见(厦门)技术有限公司 一种基于卷积神经网络的目标跟踪方法及装置
CN107066990A (zh) * 2017-05-04 2017-08-18 厦门美图之家科技有限公司 一种目标跟踪方法及移动设备

Cited By (52)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11538246B2 (en) 2018-07-27 2022-12-27 Tencent Technology (Shenzhen) Company Limited Method and apparatus for training feature extraction model, computer device, and computer-readable storage medium
CN111694423B (zh) * 2019-03-12 2023-05-26 阿里巴巴集团控股有限公司 定位、抓取、数据处理与用于增强现实的显示方法及设备
CN111694423A (zh) * 2019-03-12 2020-09-22 阿里巴巴集团控股有限公司 定位、抓取、数据处理与用于增强现实的显示方法及设备
CN109858472A (zh) * 2019-04-09 2019-06-07 武汉领普科技有限公司 一种嵌入式实时人形检测方法和装置
CN109858472B (zh) * 2019-04-09 2023-08-04 武汉领普科技有限公司 一种嵌入式实时人形检测方法和装置
CN111815670A (zh) * 2019-04-10 2020-10-23 曜科智能科技(上海)有限公司 多视图目标跟踪方法、装置、系统、电子终端、及存储介质
CN110148052A (zh) * 2019-04-17 2019-08-20 深圳壹账通智能科技有限公司 商家贷后管控方法、装置、计算机设备及存储介质
CN110176027A (zh) * 2019-05-27 2019-08-27 腾讯科技(深圳)有限公司 视频目标跟踪方法、装置、设备及存储介质
CN110176027B (zh) * 2019-05-27 2023-03-14 腾讯科技(深圳)有限公司 视频目标跟踪方法、装置、设备及存储介质
CN110443827A (zh) * 2019-07-22 2019-11-12 浙江大学 一种基于改进孪生网络的无人机视频单目标长期跟踪方法
CN110443827B (zh) * 2019-07-22 2022-12-20 浙江大学 一种基于改进孪生网络的无人机视频单目标长期跟踪方法
CN110543818A (zh) * 2019-07-25 2019-12-06 长沙行深智能科技有限公司 基于权图匹配的红绿灯跟踪方法、装置、介质及设备
CN110543818B (zh) * 2019-07-25 2023-01-17 长沙行深智能科技有限公司 基于权图匹配的红绿灯跟踪方法、装置、介质及设备
CN110532883A (zh) * 2019-07-30 2019-12-03 平安科技(深圳)有限公司 应用离线跟踪算法对在线跟踪算法进行改进
CN110532883B (zh) * 2019-07-30 2023-09-01 平安科技(深圳)有限公司 应用离线跟踪算法对在线跟踪算法进行改进
CN110516556A (zh) * 2019-07-31 2019-11-29 平安科技(深圳)有限公司 基于Darkflow-DeepSort的多目标追踪检测方法、装置及存储介质
CN110516556B (zh) * 2019-07-31 2023-10-31 平安科技(深圳)有限公司 基于Darkflow-DeepSort的多目标追踪检测方法、装置及存储介质
CN110738101B (zh) * 2019-09-04 2023-07-25 平安科技(深圳)有限公司 行为识别方法、装置及计算机可读存储介质
CN110738101A (zh) * 2019-09-04 2020-01-31 平安科技(深圳)有限公司 行为识别方法、装置及计算机可读存储介质
CN110929605A (zh) * 2019-11-11 2020-03-27 中国建设银行股份有限公司 视频关键帧的保存方法、装置、设备及存储介质
CN110991458A (zh) * 2019-11-25 2020-04-10 创新奇智(北京)科技有限公司 基于图像特征的人工智能识别结果抽样系统及抽样方法
CN110956131A (zh) * 2019-11-27 2020-04-03 北京迈格威科技有限公司 单目标追踪方法、装置及系统
CN110956131B (zh) * 2019-11-27 2024-01-05 北京迈格威科技有限公司 单目标追踪方法、装置及系统
CN111091098B (zh) * 2019-12-20 2023-08-15 浙江大华技术股份有限公司 检测模型的训练方法、检测方法及相关装置
CN111091098A (zh) * 2019-12-20 2020-05-01 浙江大华技术股份有限公司 检测模型的训练方法、检测方法及相关装置
CN111241931A (zh) * 2019-12-30 2020-06-05 沈阳理工大学 一种基于YOLOv3的空中无人机目标识别和跟踪方法
CN111241931B (zh) * 2019-12-30 2023-04-18 沈阳理工大学 一种基于YOLOv3的空中无人机目标识别和跟踪方法
CN111242977A (zh) * 2020-01-09 2020-06-05 影石创新科技股份有限公司 全景视频的目标跟踪方法、可读存储介质及计算机设备
CN111242977B (zh) * 2020-01-09 2023-04-25 影石创新科技股份有限公司 全景视频的目标跟踪方法、可读存储介质及计算机设备
CN111311635A (zh) * 2020-02-08 2020-06-19 腾讯科技(深圳)有限公司 一种目标定位方法、装置及系统
CN111275011A (zh) * 2020-02-25 2020-06-12 北京百度网讯科技有限公司 移动红绿灯检测方法、装置、电子设备和存储介质
CN111275011B (zh) * 2020-02-25 2023-12-19 阿波罗智能技术(北京)有限公司 移动红绿灯检测方法、装置、电子设备和存储介质
CN111401172B (zh) * 2020-03-06 2023-10-27 大连海事大学 一种基于视频的港口吊装料包自动计数方法
CN111401172A (zh) * 2020-03-06 2020-07-10 大连海事大学 一种基于视频的港口吊装料包自动计数方法
CN112655021A (zh) * 2020-04-09 2021-04-13 深圳市大疆创新科技有限公司 图像处理方法、装置、电子设备和存储介质
CN113642360A (zh) * 2020-04-27 2021-11-12 杭州海康威视数字技术股份有限公司 一种行为计时方法、装置、电子设备及存储介质
CN113642360B (zh) * 2020-04-27 2024-06-04 杭州海康威视数字技术股份有限公司 一种行为计时方法、装置、电子设备及存储介质
CN111652080A (zh) * 2020-05-12 2020-09-11 合肥的卢深视科技有限公司 基于rgb-d图像的目标跟踪方法和装置
CN111652080B (zh) * 2020-05-12 2023-10-17 合肥的卢深视科技有限公司 基于rgb-d图像的目标跟踪方法和装置
CN111862158A (zh) * 2020-07-21 2020-10-30 湖南师范大学 一种分阶段目标跟踪方法、装置、终端及可读存储介质
CN111862158B (zh) * 2020-07-21 2023-08-29 湖南师范大学 一种分阶段目标跟踪方法、装置、终端及可读存储介质
CN112183235A (zh) * 2020-09-07 2021-01-05 根尖体育科技(北京)有限公司 针对运动场所进行视频采集的自动控制方法
CN112270661A (zh) * 2020-10-19 2021-01-26 北京宇航系统工程研究所 一种基于火箭遥测视频的空间环境监测方法
CN112270661B (zh) * 2020-10-19 2024-05-07 北京宇航系统工程研究所 一种基于火箭遥测视频的空间环境监测方法
CN114511793A (zh) * 2020-11-17 2022-05-17 中国人民解放军军事科学院国防科技创新研究院 一种基于同步检测跟踪的无人机对地探测方法及系统
CN114511793B (zh) * 2020-11-17 2024-04-05 中国人民解放军军事科学院国防科技创新研究院 一种基于同步检测跟踪的无人机对地探测方法及系统
CN112733690B (zh) * 2020-12-31 2024-02-20 北京易华录信息技术股份有限公司 一种高空抛物检测方法、装置及电子设备
CN112733690A (zh) * 2020-12-31 2021-04-30 北京易华录信息技术股份有限公司 一种高空抛物检测方法、装置及电子设备
CN113066108A (zh) * 2021-04-14 2021-07-02 武汉卓目科技有限公司 基于eco算法的抗遮挡视觉目标跟踪方法及装置
CN113573137A (zh) * 2021-07-01 2021-10-29 厦门美图之家科技有限公司 视频画布边界检测方法、系统、终端设备及存储介质
CN113573137B (zh) * 2021-07-01 2023-08-08 厦门美图之家科技有限公司 视频画布边界检测方法、系统、终端设备及存储介质
CN113609971A (zh) * 2021-08-04 2021-11-05 广州威拓电子科技有限公司 微地震观测设备的检查方法、装置、设备及存储介质

Also Published As

Publication number Publication date
CN107679455A (zh) 2018-02-09

Similar Documents

Publication Publication Date Title
WO2019041519A1 (fr) Procédé et dispositif de suivi de cible, et support de stockage lisible par ordinateur
CN109961009B (zh) 基于深度学习的行人检测方法、系统、装置及存储介质
US10534957B2 (en) Eyeball movement analysis method and device, and storage medium
US9390340B2 (en) Image-based character recognition
US8792722B2 (en) Hand gesture detection
CN104350509B (zh) 快速姿势检测器
US8750573B2 (en) Hand gesture detection
CN109325456B (zh) 目标识别方法、装置、目标识别设备及存储介质
US9436883B2 (en) Collaborative text detection and recognition
US9710698B2 (en) Method, apparatus and computer program product for human-face features extraction
US20180211104A1 (en) Method and device for target tracking
WO2019033572A1 (fr) Procédé de détection de situation de visage bloqué, dispositif et support d'informations
US10650234B2 (en) Eyeball movement capturing method and device, and storage medium
US9058536B1 (en) Image-based character recognition
US9269009B1 (en) Using a front-facing camera to improve OCR with a rear-facing camera
CN110598559B (zh) 检测运动方向的方法、装置、计算机设备和存储介质
US9569679B1 (en) Adaptive image sampling for text detection
WO2019033570A1 (fr) Procédé d'analyse de mouvement labial, appareil et support d'informations
WO2019033568A1 (fr) Procédé de saisie de mouvement labial, appareil et support d'informations
US20210281744A1 (en) Action recognition method and device for target object, and electronic apparatus
WO2015003606A1 (fr) Procédé et appareil de reconnaissance d'une image pornographique
KR20120044484A (ko) 이미지 처리 시스템에서 물체 추적 장치 및 방법
CN110910445A (zh) 一种物件尺寸检测方法、装置、检测设备及存储介质
US9760177B1 (en) Color maps for object tracking
CN111199169A (zh) 图像处理方法和装置

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 17923072

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

32PN Ep: public notification in the ep bulletin as address of the adressee cannot be established

Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205A DATED 25/09/2020)

122 Ep: pct application non-entry in european phase

Ref document number: 17923072

Country of ref document: EP

Kind code of ref document: A1