CN113470073A - Animal center tracking method based on deep learning - Google Patents

Animal center tracking method based on deep learning Download PDF

Info

Publication number
CN113470073A
CN113470073A CN202110765039.2A CN202110765039A CN113470073A CN 113470073 A CN113470073 A CN 113470073A CN 202110765039 A CN202110765039 A CN 202110765039A CN 113470073 A CN113470073 A CN 113470073A
Authority
CN
China
Prior art keywords
animal
tracking
video
training
model
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202110765039.2A
Other languages
Chinese (zh)
Inventor
李新建
孙广龙
高利霞
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Zhejiang University ZJU
Original Assignee
Zhejiang University ZJU
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Zhejiang University ZJU filed Critical Zhejiang University ZJU
Priority to CN202110765039.2A priority Critical patent/CN113470073A/en
Publication of CN113470073A publication Critical patent/CN113470073A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/20Analysis of motion
    • G06T7/246Analysis of motion using feature-based methods, e.g. the tracking of corners or segments
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • G06N3/084Backpropagation, e.g. using gradient descent
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/13Edge detection
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/60Analysis of geometric attributes
    • G06T7/66Analysis of geometric attributes of image moments or centre of gravity
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10016Video; Image sequence
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10024Color image
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20081Training; Learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20084Artificial neural networks [ANN]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30241Trajectory

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • General Health & Medical Sciences (AREA)
  • Artificial Intelligence (AREA)
  • Computational Linguistics (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Computation (AREA)
  • Biomedical Technology (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • General Engineering & Computer Science (AREA)
  • Biophysics (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Health & Medical Sciences (AREA)
  • Multimedia (AREA)
  • Geometry (AREA)
  • Image Analysis (AREA)

Abstract

The invention discloses an animal center tracking method based on deep learning, which comprises model training and animal tracking, wherein during the model training, a pre-trained neural network is adopted for carrying out the characteristic extraction of a picture, a target detection layer of a YOLO algorithm is adopted for recognizing a boundary frame of a target animal, and a trained model is finally obtained; when the animal is tracked, all the frame pictures in the video are tracked by using the trained model, the boundary frame of the animal is calculated, and finally the trajectory graph of the animal motion is obtained. The animal center tracking method based on deep learning provides a new experimental animal behavioristics tracking and positioning method, the method can accurately position the animal center point in various experimental paradigms and complex experimental environments, overcomes the interference of in-vivo recording wires and the entrance of experimenters into scenes, and can be suitable for different animal models.

Description

Animal center tracking method based on deep learning
Technical Field
The invention relates to the field of image processing methods, in particular to an animal center tracking method based on deep learning.
Background
In behavioral research in the fields of biology, medicine, neuroscience and the like, the method has an important role in accurately obtaining the position of an animal in an experimental scene through video data. With the development of in vivo recording technology and the diversification of experimental paradigms, researchers often encounter the problem that experimental animals are difficult to accurately position.
Conventional methods (e.g., Limelight, ANY-maze, Lima, and Lima, respectively,
Figure BDA0003150515500000011
XT, TopScan, etc.) extracts the outline of an experimental animal by using a background noise reduction method, and then determines the position of the animal by calculating the center of the outline, although the method has high calculation speed, the method is only suitable for simple scenes with clean background, high image signal-to-noise ratio and small interference, and is easily influenced by the wires of in-vivo recording equipment and the necessary operation of experimenters entering the scenes in the experimental process, and cannot adapt to the diversification of experimental paradigms.
The development of computer vision technology in recent years provides a large number of image data processing algorithms and provides a new solution for tracking experimental animals. However, most of these algorithms are currently based on the evaluation of animal pose by characteristic point detection, such as DeepLabCut (1. Mathis, A., et al., DeepLabCut: market testing of user-defined body parts with depth testing. Nat. Neurosci, 2018.21 (9): p.1281-1289.[2] Nath, T., et al., use DeepLabCut for 3D market testing of space protocols and devices, 2019.14 (7): p.2152-2176., LEAP.1. Wang, Z.R., et al., leap.1. basic vision testing of animal pose, repair, soil, repair, 2019.8.), Trex (Walter, T.and I.D.Couzin, TRex, a fast multi-animal tracking system with Markesses identification, and 2D evaluation of location and visual fields, Elife, 2021.10.), etc. Although the method can track a plurality of characteristic points of the animal, the behavior analysis of the animal is richer. But because the exact animal center point cannot be objectively evaluated when creating training data, it takes longer and less effective to track the animal center position.
Disclosure of Invention
In order to overcome the defect that the existing method is difficult to accurately track the experimental animal in a diversified experimental paradigm, the invention provides an animal center tracking method based on deep learning based on a deep learning algorithm.
An animal center tracking method based on deep learning comprises the following steps:
(1) model training
(1.1) randomly extracting pictures from a scene video to be tested, then manually creating a data set of a bounding box for extracting the pictures for model training,
(1.2) dividing the data set into a training set, a cross validation set and a test set, wherein the data of the training set is used for training the parameters of the model, the cross validation set and the test set are used for validating and testing the effect of the model,
during training, the pre-trained neural network is adopted to extract the characteristics of the picture, the target detection layer of the YOLO algorithm is adopted to identify the boundary frame of the target animal, and finally the trained model is obtained,
(2) animal tracking
(2.1) manually defining the tracked area before each tracking, and setting the part of each frame image in the video outside the area as a color similar to the background,
(2.2) tracking all the frame pictures in the video by using the trained model, calculating the boundary frame of the animal,
(2.3) converting the format of each frame in the video into a gray scale format, calculating the average value of all the frames of the video as the background noise of the video,
(2.4) subtracting the background noise corresponding to the bounding box area from each frame of picture in the video by using a background noise reduction method to obtain the outline of the animal in the bounding box,
and (2.5) finally, tracking the animal by calculating the centroid of the animal outline in all the frames of the video to obtain a trajectory diagram of the animal movement.
Preferably, in step (1.1), the picture is in RGB format.
Preferably, in step (1.2), the pre-trained neural network is resnet18, mobilenetv2 or resnet 50.
Preferably, in step (1.2), the target detection layer of the YOLO-specific recognition bounding box is connected after the pre-trained neural network.
Preferably, in step (1.2), a mini-batch gradient descent method is adopted during training, and parameters in the network are adjusted through back propagation in an iterative process.
Preferably, in step (2.2), if the YOLO calculates a plurality of possible bounding boxes, the bounding box with the largest predicted p-value is selected for subsequent processing.
Preferably, in step (2.2), the predicted bounding box is enlarged in a ratio of 1: 1.5.
Preferably, when the animal is tracked, the animal tracking of a single video file is realized by dbt _ singleTracking; animal tracking of multiple video files is achieved by dbt _ batch tracking; the tracking completed video is created by dbt _ createlabedvideo.
Preferably, the animal center tracking method corrects the failed tracking frame manually through dbt _ manual tracking, or derives the failed tracking frame through dbt _ optimum, and combines the failed tracking frame with the data set in the step (1.1), and retrains a new model through the step (1.2), so as to serve as a final trained model for the animal tracking in the step (2).
The animal center tracking method based on deep learning provides a new experimental animal behavioristics tracking and positioning method, the method can accurately position the animal center point in various experimental paradigms and complex experimental environments, overcomes the interference of in-vivo recording wires and the entrance of experimenters into scenes, and can be suitable for different animal models.
Drawings
Fig. 1 is a technical flowchart of the present invention, wherein a: model training and optimizing processes; b: animal tracking process; c: schematic diagram of the procedure for tracking black mice in open field experiment.
Fig. 2 is a graph comparing the performance of three pre-trained neural networks, wherein a: comparing training time of the YOLO models of the three neural networks under different picture sizes; b: comparing the detection speed and the detection accuracy of the YOLO model of the three neural networks under different picture sizes, wherein the dots represent the picture sizes of 224, 320, 416 and 512 from small to large in sequence.
FIG. 3 is a diagram showing the comparison result between DeepBhvTracking and three other methods, wherein a-d are a schematic tracking and positioning diagram and an animal track diagram of an L-shaped labyrinth by a background noise reduction method, a YOLO method, a DeepLabcut method and a DeepBhvTracking method respectively; e-h are respectively the animal track schematic diagrams of a background noise reduction method, a YOLO method, a DeepLabCut method and a DeepBhvTracking method in a three-box maze; i: comparing the training time of the DeepBhvTracking model and the DeepLabCut model; j: comparing detection time of DeepBhvTracking with three other methods; k: comparing the pixel variation of each frame of four tracking methods under different experimental paradigms; l: and comparing the error of the four tracking methods in different experimental paradigms with the error of the real value. P < 0.05, p < 0.01, p < 0.001, and significant results are the Bonferroni corrected rank sum test.
Fig. 4 is a graph of the effect of deep bhvtracking in various experimental paradigms of various animals, wherein a: black mouse treadmill experiment; b: the black mouse inverted V-shaped maze; c: black mouse elevated plus maze; d: three-box experiment of white mice; e: marmoset monkey cage.
Fig. 5 is a graph of an application example result of deep bhvtracking in the medical field, wherein a: schematic trace of wild type C57BL/6 under open field; b: indicating the residence time of the mouse at each position of the open field; c: indicating the movement speed of the mouse at each position of the open field; d: average speed comparison of three mice; e: comparing the residence time of the three mice in the center and corner areas of the open field; f: the average speed of movement of the three mice in the center and corner areas of the open field was compared. P < 0.05, p < 0.01, p < 0.001, and significant results are the Bonferroni corrected rank sum test.
Detailed Description
The method is a tool kit developed based on MATLAB software, and is named as DeepBhvTracking, the detailed code list of the DeepBhvTracking is shown in Table 1, and when the method is used, a deep learning tool kit, a computer vision tool kit and a pre-trained neural network (such as a net18, a mobilenetv2, a net50 or other pre-trained neural networks) need to be installed in advance. The testing part related to the application is completed on dell computer (CPU Intel (R) core (TM) i7-10700 CPU @2.90GHz, RAM 64GB, GPU Inter (R) UHD Graphics 6308 GB).
TABLE 1 DeepBhvTracking detailed code List
Figure BDA0003150515500000041
Example 1
The method mainly comprises the steps of carrying out transfer learning through a pre-trained neural network and calculating a boundary box of an animal through a You Only Look One (YOLO) algorithm, then extracting the outline of the animal in the boundary box by using a background noise reduction method, and finally calculating the central locus of the animal according to the outline of the animal. The method is divided into two parts of model training (figure 1a) and animal tracking (figures 1b and c), and comprises the following steps:
model training
Firstly, randomly extracting pictures (RGB format) from a scene video to be detected as a data set, and then manually creating a bounding box of the data set pictures through an Image Label. The acquisition of the data set is realized by dbt _ dataset.
The data set is divided into a training set (70%), a cross validation set (10%) and a test set (20%) according to the number of pictures, data of the training set is used for training parameters of a YOLO model, and the cross validation set and the test set are used for verifying and testing the effect of the model. In order to make the generalization capability of the model stronger, data enhancement (including flipping, rotating and adding noise) is carried out on the data of the training set. The pre-trained neural network is adopted to extract the features of the pictures of the training set, and the parameters on the network train a large number of images in the ImageNet large database, so that the capability of extracting rich information features from the images is achieved. And connecting a target detection layer of a YOLO specific recognition bounding box after the pre-trained neural network. During training, a mini-batch gradient descent method is adopted, and parameters in the network are adjusted through back propagation in the iterative process. Finally, the trained YOLO model is saved for later animal tracking. The training and verification of the model is realized by dbt _ training.
(II) animal tracking
To avoid the effect of factors outside the experimental field on tracking animals, a tracking area is manually defined before each tracking, and the part outside the area of each frame of image in the video is set to be similar color of the background. Then, tracking all the frame pictures in the video by using the trained YOLO model, and calculating the boundary frame of the animal. Since YOLO accounts work out a number of possible bounding boxes, we select the box with the largest predicted p-value for later processing. To avoid that the predicted bounding box only covers the influence of the animal part body on the tracking effect, we enlarge the bounding box in a ratio of 1: 1.5. Then the video is converted from a color format to a gray scale format, and then the average value of each pixel in the video in time is calculated to be used as the background noise of the video. At the moment, a background noise reduction method is used, and the corresponding background noise is subtracted from each frame of the boundary frame area picture in the video, so that the outline of the animal in the boundary frame can be obtained. And finally, tracking the animal by calculating the centroid of the contour of the animal in the boundary box in all frames of the video to obtain a trajectory diagram of the animal movement. And then analyzing the motion related parameters of the animal, such as the motion speed, the acceleration, the motion time and the like through the trajectory graph. Animal tracking of a single video file can be realized through dbt _ singleTracking, and batch tracking of animals in a plurality of video files can also be realized through dbt _ batchTracking. Finally, the tracking effect can also be detected by dbt _ createlabedvideo creating a tracking completed video.
To avoid the impact of the tracking failed frames (including frames where YOLO failed to detect animals and detected errors) on the experimental analysis, dbt _ manual tracking was provided to correct the tracking failed frames manually. At the same time, dbt _ optimum is provided to derive images of failed tracking, which are merged with the previous training data set and the new model is trained to optimize the model previously trained.
(III) Pre-trained neural network comparison
Fig. 2 shows a comparison of the performance of 3 common pre-trained neural networks (resnet18, mobilenetv2, and resnet50) at different picture sizes. In the resnet18 network, the part before the 'res 5b _ branch2a _ relu' layer is selected as a feature extraction layer, and the layer after the feature extraction layer is replaced by a YOLO target detection layer; in the mobilenetv2 network, selecting a part before a 'block _16_ expanded _ relu' layer as a feature extraction layer, and replacing the layer after the layer with a YOLO target detection layer; in the 'rescenet 50' network, the part before the 'activation _40_ relu' layer is selected as a feature extraction layer, and the layer after the layer is replaced with a YOLO target detection layer.
The results show that the training time and detection accuracy of the YOLO model increase with increasing picture size, but the detection speed decreases with increasing picture size (fig. 2a, b). Under the same picture size, the training time of the resnet18 network is shortest, the detection speed is fastest, and the detection accuracy is minimum; the training time of the resnet50 network is longest, the detection speed is slowest, but the detection accuracy is highest (fig. 2a and b). For the trade-off of detection speed and accuracy, the resnet50 network was chosen as the pre-trained neural network for training the YOLO model.
Example 2
DeepBhvTracking was compared in performance with three commonly used tracking and localization methods in three typical experimental paradigms. The three experimental paradigms are respectively: (1) black mouse open field experiment (n ═ 6): the paradigm scene is simple, no interference of wires and experimenters exists, and the signal-to-noise ratio is high; (2) black mouse L-maze experiment (n ═ 6): the model mouse head is provided with a calcium imaging recording device, and an experimenter needs to enter a scene to perform necessary operation due to the interference of electric wires; (3) white mouse three-box experiment (n ═ 6): the model mouse head is provided with a calcium imaging recording device, the interference of electric wires exists, the color of the mouse is similar to that of the background, and the signal-to-noise ratio is low. The three tracking methods are as follows: the method comprises a background noise reduction method, a YOLO algorithm and a DeepLabCut algorithm, wherein the YOLO, the DeepLabCut and the DeepBhvTracking are all methods developed based on deep learning, and a data set is required for model training. The three methods adopt the same data set pictures for training, and the data set is formed by extracting 300 pictures from each of the three normal forms (6 videos in each normal form and 50 pictures in each video). When comparing DeepBhvTracking with other methods, the significance level was corrected by Bonferroni method using the rank sum test. The results of the four methods of tracking and locating are not manually corrected for later comparative analysis.
The results show that the background noise reduction method cannot accurately track animals in complex scenes such as an L-shaped maze and a three-box experiment (fig. 3a and e), and the variation per frame and the error from the true value calculated in the two paradigms are both significantly larger than those of the DeepBhvtracking method (fig. 3k and 1). But the background noise reduction method has the advantages of high detection speed and no need of training a model (fig. 3j), and has better tracking effect in a simple scene such as an open field experiment (fig. 3k and 1). Although the YOLO algorithm effectively solves the interference of wires and human factors, since YOLO can only locate the animal through the bounding box, the center position of the animal cannot be accurately obtained, so that the tracked track jumps, resulting in track abnormality (fig. 3b, f, k, l). The DeepLabCut is used as an attitude estimation method based on feature point detection, and can also effectively solve the interference of wires and human factors (figures 3c and g). However, the central position of the animal cannot be accurately marked manually when the training data is created, and partial frame identification fails, so that the tracking effect in the L-shaped maze is poor, and the calculated variation of each frame and the error from the true value are both significantly larger than those of the deep bvtracking method (fig. 3k and L). Meanwhile, the detection speed of the deep LabCut is also slower (FIG. 2 j). In summary, deep bhvtracking can guarantee relatively good tracking effect at a faster processing speed in various paradigms (fig. 2).
Example 3
To verify the availability and generalization ability of DeepBhvTracking, it was applied in tracking animals of different paradigms: black mouse treadmill experiment (fig. 4a), black mouse inverted V-maze (fig. 4b), black mouse elevated plus maze (fig. 4c), white mouse triple box experiment (fig. 4d) and marmoset movement in monkey cages (fig. 4 e). The animal relatively smooth motion track can be obtained in different paradigms, which shows that the DeepBhvTracking can be widely applied to different animal models and different motion scenes.
Example 4
Fig. 5 is an application example of deep bhvtracking in the field of neuromedical research, and 3 mice were used for open field experiments, which are respectively: wild type C57BL/6 mice (n ═ 6), mutant PRRT2 mice (n ═ 6), and mutant FMRl mice (n ═ 6). Open field experiments are one of the common paradigms for assessing rodent locomotor status and anxiety levels. First, we calculated the trajectory of each mouse moving within 8 minutes of open field using DeepBhvTracking (figure 5 a). Then, the moving time and the moving speed of the mouse at each position of the open field can be calculated through the moving track (fig. 5b and c). All mice tend to stay more in the corners and the speed of movement at the corners is less than the speed of movement at the center. Also, the mean locomotor speed was greater for both mutant mice than for wild-type mice, but PRRT2 mice stayed more centrally than for wild-type mice, indicating a lower level of anxiety.

Claims (9)

1. An animal center tracking method based on deep learning is characterized by comprising the following steps:
(1) model training
(1.1) randomly extracting pictures from a scene video to be tested, then manually creating a data set of a bounding box for extracting the pictures for model training,
(1.2) dividing the data set into a training set, a cross validation set and a test set, wherein the data of the training set is used for training the parameters of the model, the cross validation set and the test set are used for validating and testing the effect of the model,
during training, the pre-trained neural network is adopted to extract the characteristics of the picture, the target detection layer of the YOLO algorithm is adopted to identify the boundary frame of the target animal, and finally the trained model is obtained,
(2) animal tracking
(2.1) manually defining the tracked area before each tracking, and setting the part of each frame image in the video outside the area as a color similar to the background,
(2.2) tracking all frame pictures in the video by using a trained model, calculating a boundary frame of an animal, (2.3) converting the format of each frame in the video into a gray format, calculating the average value of all frames of the video as the background noise of the video,
(2.4) subtracting the background noise corresponding to the bounding box area from each frame of picture in the video by using a background noise reduction method to obtain the outline of the animal in the bounding box,
and (2.5) finally, tracking the animal by calculating the centroid of the animal outline in all the frames of the video to obtain a trajectory diagram of the animal movement.
2. The animal centric tracking method according to claim 1, characterized in that in step (1.1), the picture is in RGB format.
3. The animal centric tracking method according to claim 1, characterized in that in step (1.2), the pre-trained neural network is resnet18, mobilenetv2 or resnet 50.
4. The animal centric tracking method according to claim 1, characterized in that in step (1.2) an object detection layer of a YOLO-specific recognition bounding box is connected after the pre-trained neural network.
5. The animal centric tracking method according to claim 1, characterized in that in step (1.2), the mini-batch gradient descent method is used during training, and parameters in the network are adjusted by back propagation in an iterative process.
6. The animal centric tracking method according to claim 1, characterized in that in step (2.2), if YOLO calculates a plurality of possible bounding boxes, the bounding box with the largest predicted p-value is selected for later processing.
7. The method for animal center tracking according to claim 1, wherein in step (2.2), the predicted bounding box is enlarged in a ratio of 1: 1.5.
8. The animal centric tracking method according to claim 1, characterized in that in animal tracking, animal tracking of a single video file is realized by dbt _ singleTracking; animal tracking of multiple video files is achieved by dbt _ batch tracking; the tracking completed video is created by dbt _ createlabedvideo.
9. The animal centric tracking method according to claim 1, characterized in that the failed tracking frame is corrected manually by dbt _ manual tracking or derived by dbt _ optimization, and is merged with the data set in step (1.1), and then retrained by step (1.2) again to be used as the final trained model for animal tracking in step (2).
CN202110765039.2A 2021-07-06 2021-07-06 Animal center tracking method based on deep learning Pending CN113470073A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202110765039.2A CN113470073A (en) 2021-07-06 2021-07-06 Animal center tracking method based on deep learning

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202110765039.2A CN113470073A (en) 2021-07-06 2021-07-06 Animal center tracking method based on deep learning

Publications (1)

Publication Number Publication Date
CN113470073A true CN113470073A (en) 2021-10-01

Family

ID=77878699

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202110765039.2A Pending CN113470073A (en) 2021-07-06 2021-07-06 Animal center tracking method based on deep learning

Country Status (1)

Country Link
CN (1) CN113470073A (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN116309727A (en) * 2023-05-26 2023-06-23 中南大学 Unmanned aerial vehicle target tracking method and simulation system based on deep learning algorithm

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108550147A (en) * 2018-04-13 2018-09-18 华中科技大学苏州脑空间信息研究院 Full cranial nerve beam projected path method for automatic tracking and system
CN109584266A (en) * 2018-11-15 2019-04-05 腾讯科技(深圳)有限公司 A kind of object detection method and device
CN110781806A (en) * 2019-10-23 2020-02-11 浙江工业大学 Pedestrian detection tracking method based on YOLO
CN111612002A (en) * 2020-06-04 2020-09-01 广州市锲致智能技术有限公司 Multi-target object motion tracking method based on neural network

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108550147A (en) * 2018-04-13 2018-09-18 华中科技大学苏州脑空间信息研究院 Full cranial nerve beam projected path method for automatic tracking and system
CN109584266A (en) * 2018-11-15 2019-04-05 腾讯科技(深圳)有限公司 A kind of object detection method and device
CN110781806A (en) * 2019-10-23 2020-02-11 浙江工业大学 Pedestrian detection tracking method based on YOLO
CN111612002A (en) * 2020-06-04 2020-09-01 广州市锲致智能技术有限公司 Multi-target object motion tracking method based on neural network

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
于红: "水产动物目标探测与追踪技术及应用研究进展", 《大连海洋大学学报》, vol. 35, no. 6, pages 793 - 804 *
王春林 等: "《人工智能》", 30 September 2020, 西安电子科技大学出版社, pages: 181 *

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN116309727A (en) * 2023-05-26 2023-06-23 中南大学 Unmanned aerial vehicle target tracking method and simulation system based on deep learning algorithm

Similar Documents

Publication Publication Date Title
Liu et al. Attribute-aware face aging with wavelet-based generative adversarial networks
US11908244B2 (en) Human posture detection utilizing posture reference maps
CN110070074B (en) Method for constructing pedestrian detection model
CN107424171B (en) Block-based anti-occlusion target tracking method
CN104392468B (en) Based on the moving target detecting method for improving visual background extraction
CN107516316B (en) Method for segmenting static human body image by introducing focusing mechanism into FCN
CN107220596B (en) Human body posture estimation method based on cascade error correction mechanism
CN105160310A (en) 3D (three-dimensional) convolutional neural network based human body behavior recognition method
CN106203423B (en) Weak structure perception visual target tracking method fusing context detection
CN106648078B (en) Multi-mode interaction method and system applied to intelligent robot
CN105426929B (en) Object shapes alignment device, object handles devices and methods therefor
KR20130013122A (en) Apparatus and method for detecting object pose
CN101986348A (en) Visual target identification and tracking method
CN110991278A (en) Human body action recognition method and device in video of computer vision system
CN103902989A (en) Human body motion video recognition method based on non-negative matrix factorization
JP2021503139A (en) Image processing equipment, image processing method and image processing program
Hu et al. Depth sensor based human detection for indoor surveillance
KR20200101521A (en) Semantic matchaing apparatus and method
CN113470073A (en) Animal center tracking method based on deep learning
CN114708645A (en) Object identification device and object identification method
CN117541994A (en) Abnormal behavior detection model and detection method in dense multi-person scene
CN116994049A (en) Full-automatic flat knitting machine and method thereof
US20230005162A1 (en) Image processing system, image processing method, and storage medium
CN113627255B (en) Method, device and equipment for quantitatively analyzing mouse behaviors and readable storage medium
Kourbane et al. Skeleton-aware multi-scale heatmap regression for 2D hand pose estimation

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination