CN110569719B - Animal head posture estimation method and system - Google Patents

Animal head posture estimation method and system Download PDF

Info

Publication number
CN110569719B
CN110569719B CN201910698158.3A CN201910698158A CN110569719B CN 110569719 B CN110569719 B CN 110569719B CN 201910698158 A CN201910698158 A CN 201910698158A CN 110569719 B CN110569719 B CN 110569719B
Authority
CN
China
Prior art keywords
detection frame
animal
head
picture
coordinates
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201910698158.3A
Other languages
Chinese (zh)
Other versions
CN110569719A (en
Inventor
黄章进
汪方军
贺翔翔
邹露
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
University of Science and Technology of China USTC
Original Assignee
University of Science and Technology of China USTC
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by University of Science and Technology of China USTC filed Critical University of Science and Technology of China USTC
Priority to CN201910698158.3A priority Critical patent/CN110569719B/en
Publication of CN110569719A publication Critical patent/CN110569719A/en
Application granted granted Critical
Publication of CN110569719B publication Critical patent/CN110569719B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/20Image preprocessing
    • G06V10/25Determination of region of interest [ROI] or a volume of interest [VOI]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/20Movements or behaviour, e.g. gesture recognition

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Multimedia (AREA)
  • General Health & Medical Sciences (AREA)
  • Health & Medical Sciences (AREA)
  • Human Computer Interaction (AREA)
  • Biophysics (AREA)
  • Mathematical Physics (AREA)
  • Data Mining & Analysis (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • General Engineering & Computer Science (AREA)
  • Computational Linguistics (AREA)
  • Evolutionary Computation (AREA)
  • Software Systems (AREA)
  • Biomedical Technology (AREA)
  • Artificial Intelligence (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Psychiatry (AREA)
  • Social Psychology (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Image Analysis (AREA)

Abstract

A method and a system for estimating the head posture of an animal are provided, the method comprises the following steps: carrying out target detection processing on the animal picture, and marking a first detection frame of the head of the animal on the animal picture; marking a second detection frame of the head of the animal according to the first detection frame, and recording the position coordinates of the second detection frame, wherein the second detection frame comprises the whole head of the animal; extracting the characteristics of the animal head within the range defined by the second detection frame to obtain a plurality of key point position coordinates corresponding to the animal head; and obtaining the posture of the animal head according to the position coordinates of the second detection frame and the position coordinates of the plurality of key points. The method and the system perform attitude estimation in the head region instead of the whole picture, can effectively reduce the calculated amount and input information redundancy, and improve the accuracy and speed of zebra fish head attitude estimation.

Description

Animal head posture estimation method and system
Technical Field
The invention relates to the field of image processing, in particular to an animal head posture estimation method and system.
Background
The method controls and records the activity of the whole brain neurons under the free behavior of the animal, and has important significance for analyzing the corresponding relation between the behavior of the animal and the brain activity. Such as zebrafish, whose whole genome has 87% similarity to the human genome, are often used to explore the relationship between neuronal populations and behavior. In order to realize the optogenetic experiment under the free behavior of animals (zebra fish), the head images of the animals need to be subjected to posture estimation, so that the heads of the animals are quantized.
At present, the method for quantifying the head of the zebra fish mainly comprises a template matching-based method, an HOG feature-based method and a deep learning-based method.
The template matching method is based on the main idea that a region template is designed to locate a specific region, left eye, right eye and central template images required by template matching are firstly set, then reasonable threshold values are set, binarization is carried out on the images, and finally a target is tracked in an ROI region according to the set template so as to achieve the purpose of quantization. The algorithm has high requirements on the setting of the template, no universality, particular sensitivity to environment and noise, and poor robustness, flexibility and stability.
The method based on the HOG features mainly comprises the steps of designing a head feature description according to head features and imaging features of the head feature description, then combining a global matching algorithm, different from other template-based matching algorithms, extracting candidate points through a fast algorithm, and then performing filtering processing, so that the head of the zebra fish can be quickly tracked. The method belongs to the process of artificially designing features, and is used for counting bottom-layer features, only effectively expressing features such as edges and the like, and not expressing the features of high-layer semantic information.
The deep learning-based method currently comprises DeepLabCut, which quantifies animal behaviors by performing posture estimation on a user-defined animal body part. And adopting ResNet as a base frame by the DeepLabcut, extracting high-level semantic features, performing deconvolution on the features to obtain thermodynamic diagrams and offsets, and predicting the positions of key points by combining the thermodynamic diagrams and the offsets. But it uses the entire image as input, resulting in computational and information redundancy throughout the process, causing bottlenecks in speed and accuracy.
Disclosure of Invention
Technical problem to be solved
In view of the prior art, the present invention provides a method and a system for estimating a head pose of an animal, which are used to at least partially solve the above technical problems.
(II) technical scheme
The invention provides an animal head posture estimation method, which comprises the following steps: s1, carrying out target detection processing on the animal picture, and marking a first detection frame of the animal head on the animal picture; s2, marking a second detection frame of the animal head according to the first detection frame, and recording a position coordinate of the second detection frame, wherein the second detection frame comprises the whole head of the animal; s3, extracting the features of the animal head within the range defined by the second detection frame to obtain a plurality of key point position coordinates corresponding to the animal head; and S4, obtaining the posture of the animal head according to the position coordinates of the second detection frame and the position coordinates of the plurality of key points.
Optionally, the first detection frame for marking the head of the animal on the animal picture comprises: and recording the coordinates, the length and the width of the two pairs of corner points of the first detection frame and the confidence coefficient of the head of the animal.
Optionally, the second detection frame marking the head of the animal according to the first detection frame comprises: mapping the animal picture marked with the first detection frame back to the original animal picture, extracting a corresponding region of the first detection frame, and inputting the corresponding region of the first detection frame into a regression optimization neural network to obtain an offset required when the first detection frame contains the whole head of the animal; calculating the coordinates, the length and the width of the two pairs of corner points of the second detection frame according to the offset and the coordinates, the length and the width of the two pairs of corner points of the first detection frame; and marking the second detection frame of the animal head on the animal picture according to the coordinates, the length and the width of the two pairs of corner points of the second detection frame.
Optionally, the animal picture is input into a Micro-YOLO neural network for target detection processing, wherein the Micro-YOLO neural network at least includes a convolutional layer.
Optionally, within the range defined by the second detection frame, the performing feature extraction on the animal head comprises: and mapping the animal picture marked with the second detection frame back to the original animal picture, extracting a region corresponding to the second detection frame, inputting the region corresponding to the second detection frame into an hourglass neural network, and extracting the characteristics of the head of the animal, wherein the hourglass neural network at least comprises a convolution layer and a pooling layer.
Optionally, the formula for calculating the coordinates, the length, and the width of the two pairs of corner points of the second detection frame is as follows:
xc=(x1+x2)/2,yc=(y1+y2)/2
bw=x2-x1,bh=y2-y1
xc′=xc+tx·bw,yc′=yc+yy·bh
Figure BDA0002148846410000031
wherein (x)1,y1) And (x)2,y2) Two pairs of angular coordinates for the first detection box, (x)c,yc) As the center coordinates of the first detection frame, bwIs the length of the first detection frame, bhWidth of the first detection frame, (x)c′,yc') center coordinates of the second detection frame, bw' is the length of the second detection frame, bh' is the width of the second detection frame, (t)x,ty,tw,th) Is an offset.
Optionally, inputting the corresponding region of the second detection box into the hourglass neural network, and performing feature extraction on the animal head includes: downsampling the area corresponding to the second detection frame to obtain a first feature map corresponding to the head of the animal; upsampling the first characteristic diagram to obtain a second characteristic diagram; fusing the second characteristic diagram in a jumping connection mode to obtain a thermal point diagram corresponding to the head of the animal; and (4) obtaining the position coordinates of the maximum activation value of each heat point in the heat point diagram to obtain the position coordinates of a plurality of key points corresponding to the head of the animal.
Optionally, the upsampling employs a bilinear interpolation method.
In another aspect, the present invention provides a system for estimating the head pose of an animal, comprising: the first processing module is used for carrying out target detection processing on the animal picture and marking a first detection frame of the head of the animal on the animal picture; the second processing module is used for marking a second detection frame of the head of the animal according to the first detection frame and recording the position coordinates of the second detection frame, wherein the second detection frame comprises the whole head of the animal; the characteristic extraction module is used for extracting the characteristics of the animal head within the range defined by the second detection frame to obtain a plurality of key point position coordinates corresponding to the animal head; and the calculation module is used for obtaining the posture of the animal head according to the position coordinates of the second detection frame and the position coordinates of the plurality of key points.
Optionally, the first processing module performs target detection processing on the animal picture by using a Micro-YOLO neural network; the second processing module marks a second detection frame of the animal head according to the first detection frame by adopting a regression optimization neural network; the feature extraction module adopts a hourglass neural network to extract features of the animal head.
(III) advantageous effects
The invention provides an animal head posture estimation method and system, which are characterized in that a zebra fish head region is detected firstly, and then posture estimation is carried out on the head region instead of the whole picture, so that the calculated amount and the input information redundancy can be effectively reduced, and the accuracy and the speed of zebra fish head posture estimation are improved.
Drawings
Fig. 1 schematically shows a flowchart of an animal head pose estimation method provided by an embodiment of the present invention.
Fig. 2 schematically shows a detection block diagram of a zebra fish head marker according to an embodiment of the present invention.
FIG. 3 schematically shows a Fast regressNet network architecture diagram provided by an embodiment of the invention.
FIG. 4 is a schematic diagram of a network architecture of the Tiny Hourglass provided by the embodiment of the present invention
Fig. 5 schematically shows a block diagram of an animal head pose estimation system provided by an embodiment of the invention.
Fig. 6 schematically shows a result diagram of posture estimation and quantization of the zebra fish head by using the method and system for estimating the animal head posture of the embodiment of the invention.
Detailed Description
In order to make the objects, technical solutions and advantages of the present invention more apparent, the present invention is further described in detail with reference to the following embodiments and the accompanying drawings.
Referring to fig. 1, fig. 1 schematically shows a flowchart of a method for estimating a head pose of an animal according to an embodiment of the present invention, which is described in detail below by taking zebra fish as an example. As shown in fig. 1, the method includes:
and S1, carrying out target detection processing on the animal picture, and marking a first detection frame of the animal head on the animal picture.
The purpose of operation S1 is to quickly obtain a zebrafish head region that is not particularly precise and likely does not fully contain the zebrafish head. The method adopts a first-order detection neural network to carry out target detection processing on the whole zebra fish picture. In this embodiment, a Micro-YOLO neural network is adopted, the Micro-YOLO neural network at least includes a convolutional layer, and the invention is not limited to the specific detection of the neural network type.
Specifically, the zebra picture is input into a Micro-YOLO neural network, and a first detection frame (x) of the zebra fish head is finally obtained according to network output after sequential forward propagation1,y1,x2,y2) Wherein (x)1,y1) Represents the coordinates of the upper left corner of the first detection box, (x)2,y2) Representing the coordinates of the lower right corner of the first detection box.
The specific network architecture of Micro-YOLO is shown in table 1, the input picture is 224 × 224 × 1, and is divided into 7 × 7 grids, and each grid predicts 2 first detection boxes. Each first detection frame comprises 5 predicted values which are respectively a center point coordinate (coordinates of two pairs of corner points), a length and a width of the first detection frame and a confidence coefficient comprising the head of the animal, and meanwhile, each grid needs to predict a category confidence coefficient. Therefore, the input image passes through 6 convolutional layers and 3 fully-connected layers, and finally a vector of dimension 7 × 7 × (2 × 5+1) ═ 539 is obtained. The location of the first detection frame of the head can be obtained by a 539-dimensional vector. The detection result of Micro-YOLO is shown in fig. 2, where the frame indicated by the reference numeral 1 in fig. 2 represents the first detection frame, and the difference is large compared to the real frame (the frame indicated by the reference numeral 3) that completely surrounds the head of the zebra fish, and the zebra fish head is not completely contained.
Figure BDA0002148846410000051
TABLE 1
And S2, marking a second detection frame of the animal head according to the first detection frame, and recording the position coordinates of the second detection frame, wherein the second detection frame comprises the whole head of the animal.
The purpose of operation S2 is to perform regression optimization on the zebra fish head region (first detection box) obtained in the first stage to obtain position regression coordinate values, and optimize the region (first detection box) obtained in the first stage into a more accurate zebra fish head region (second detection box). The method used is to input the animal picture marked with the first detection frame into a regression optimization neural network for elaboration.
Specifically, this embodiment adopts Fast RegressNet to detect the first detection frame (x)1,y1, x2,y2) And (6) optimizing. The general idea is to map the detection frame obtained in operation S1 back to the original image, extract the corresponding area, and scale the size to the image block R1 of 56 × 56 × 1. Then, the image block R1 is input into Fast regressNet, so that a more accurate second detection frame (x) is obtained1′,y1′,x2′,y2') the second detection frame contains the entire area of the zebrafish head.
The Fast RegressNet network architecture is shown in fig. 4, with an input size of 56 × 56 × 1. 4 convolution layers and 2 full-connection layers are passed through to finally obtain 6-dimensional vector (t)x,ty,tw,th,c1,c2)。 (tx,ty,tw,th) Represents the amount of offset required for regression of the first test frame (so that the first test frame encompasses the entire head of the animal), (c)1,c2) Representing a picture background probability of c1Probability of zebrafish being c2. Calculating the coordinates, the length and the width of the two pairs of corner points of the second detection frame according to the offset and the coordinates, the length and the width of the two pairs of corner points of the first detection frame; and marking the second detection frame of the animal head on the animal picture according to the coordinates, the length and the width of the two pairs of corner points of the second detection frame. The coordinates (x) of the center point of the optimized second detection framec′,yc') and length and width (b)w′,bh') can be calculated by the following formula:
xc=(x1+x2)/2,yc=(y1+y2)/2
bw=x2-x1,bh=y2-y1
xc′=xc+tx·bw,yc′=yc+yv·bh
Figure BDA0002148846410000061
wherein (x)1,y1) And (x)2,y2) Two diagonal coordinates of the first detection frame, (x)c,yc) As the center coordinates of the first detection frame, bwIs the length of the first detection frame, bhWidth of the first detection frame, (x)c′,yc') center coordinates of the second detection frame, bw' is the length of the second detection frame, bh' is the width of the second detection frame, (t)x,ty,tw,th) Is an offset. Similarly, by the center coordinate (x) of the second detection framec′,yc') and length and width (b)w′,bh') determine coordinates (x) of the upper left corner of the second detection frame1′,y1) And the coordinates of the lower right corner (x)2′,y2′)。
The regression optimization result of Fast RegressNet is shown in fig. 2, and the frame indicated by the reference numeral 2 in fig. 2 represents a second detection frame, and it can be seen that the difference between the second detection frame and a real green rectangular frame (the frame indicated by the reference numeral 3) which completely surrounds the head of the zebra fish is small, and the head of the zebra fish is completely surrounded.
And S3, extracting the features of the animal head within the range defined by the second detection frame to obtain the position coordinates of a plurality of key points corresponding to the animal head.
The purpose of operation S3 is to obtain the position of the key point of the animal head for pose estimation by performing feature extraction on the animal head local area contained in the second detection frame.
Specifically, the Hourglass neural network (Tiny Hourglass) used in this embodiment is a network structure of the zebra fish head region obtained in operation S2, and the network structure is shown in fig. 4. The general idea is to map the second detection frame obtained in operation S2 back to the original image, extract the corresponding area, and scale the size to image block R2 of 56 × 56 × 1. Then, R2 is input into the Tiny Hourglass to obtain the position coordinates (P) of a plurality of key points0,P1,…, Pn). The number of the specific key points is determined according to actual requirements, the invention is not limited, and the positions of the key points of 21 zebra fish heads are obtained in the embodiment.
To speed up the reasoning process, the input resolution is reduced to 56 × 56. The Tiny Hourglass uses a convolutional layer and a max pooling layer to process the features, down-samples the input to 7 x 7 to obtain a first feature map, then up-samples the first feature map to the original input size using a fully symmetric structure to obtain a second feature map, and the up-sampling method is bilinear interpolation. In order to better extract the features, the second feature maps obtained by upsampling are fused through jump connection, and finally 21 thermodynamic point maps are obtained, which represent 21 key point positions. All convolution operations in the Tiny Hourglass use convolution kernels of size 3X 64. And respectively taking the position coordinates of the maximum activation value of the 21 heat point diagrams to obtain the position coordinates of 21 key points.
And S4, obtaining the posture of the animal head according to the position coordinates of the second detection frame and the position coordinates of the plurality of key points.
Finally, combining the position coordinates of each key point with a second detection frame (x)l′,y1′,x2′,y2') obtaining the position coordinates of the zebra fish head in the original image, and obtaining the posture of the zebra fish head.
The animal head posture estimation method provided by the embodiment carries out posture estimation on the zebra fish head instead of the whole picture, can effectively reduce the calculated amount and input information redundancy, and improves the accuracy and speed of zebra fish head posture estimation.
Referring to fig. 5, fig. 5 schematically shows a block diagram of an animal head pose estimation system provided by an embodiment of the present invention, including:
the first processing module 510 is configured to perform target detection processing on the animal picture, and mark a first detection frame of the head of the animal on the animal picture;
the second processing module 520 is configured to mark a second detection frame of the head of the animal according to the first detection frame, and record a position coordinate of the second detection frame, where the second detection frame includes the entire head of the animal;
the feature extraction module 530 is configured to perform feature extraction on the animal head within a range defined by the second detection frame to obtain a plurality of key point position coordinates corresponding to the animal head;
and the calculating module 540 is configured to obtain the posture of the head of the animal according to the position coordinates of the second detection frame and the position coordinates of the plurality of key points.
The first processing module 510 performs target detection processing on the animal picture by using a Micro-YOLO neural network; the second processing module 520 marks a second detection frame of the animal head according to the first detection frame by adopting a regression optimization neural network; the feature extraction module 530 performs feature extraction on the head of the animal by using the hourglass neural network.
Please refer to the first embodiment of the present invention for details of the present embodiment.
The network model framework built by the animal head posture estimation system provided by the embodiment is smaller, the characteristic extraction capability is strong, and the accuracy of zebra fish head posture estimation is improved.
Referring to fig. 6, fig. 6 schematically shows a result diagram of posture estimation and quantization of the zebra fish head by using the method and system for estimating the animal head posture of the embodiment of the invention. Picture a represents a zebra fish original picture, wherein roi (region of interest) represents the zebra fish head region; the picture b represents the definition of the zebra fish head pose estimation, and the zebra fish head pose estimation totally comprises 21 key points; the picture c represents the result of ellipse fitting and straight line fitting of the key points, and the picture d represents the quantized result of the zebra fish head, including the sight line direction of both eyes and the head median line direction.
The posture of the zebra fish is estimated by adopting the animal head posture estimation system and the currently best user-defined posture estimation method DeepLabCut, and the comparison result is shown in Table 2.
Figure BDA0002148846410000081
TABLE 2
As can be seen from Table 2, the method and the system provided by the invention are superior to DeepLabCut in the speed and the accuracy of attitude estimation.
In conclusion, the invention firstly detects the head area of the zebra fish, then carries out attitude estimation on the zebra fish head instead of the whole picture, has smaller built network model frame and strong capability of extracting characteristics, can effectively reduce the calculated amount and input information redundancy, improves the accuracy and speed of the attitude estimation of the zebra fish head, and is suitable for the quantification of the zebra fish head.
The above-mentioned embodiments are intended to illustrate the objects, technical solutions and advantages of the present invention in further detail, and it should be understood that the above-mentioned embodiments are only exemplary embodiments of the present invention, and are not intended to limit the present invention, and any modifications, equivalent substitutions, improvements and the like made within the spirit and principle of the present invention should be included in the protection scope of the present invention.

Claims (7)

1. A method for estimating the head pose of an animal, comprising:
s1, carrying out target detection processing on the animal picture, marking a first detection frame of the animal head on the animal picture, and recording the coordinates, the length and the width of two pairs of corner points of the first detection frame and the confidence coefficient containing the animal head;
s2, marking a second detection frame of the animal head according to the first detection frame, and recording the position coordinates of the second detection frame, wherein the animal picture marked with the first detection frame is mapped back to the original animal picture, the corresponding region of the first detection frame is extracted, the corresponding region of the first detection frame is reduced and then input into a regression optimization neural network, and the offset required when the first detection frame comprises the whole head of the animal is obtained; calculating the coordinates, the length and the width of the two pairs of corner points of the second detection frame according to the offset and the coordinates, the length and the width of the two pairs of corner points of the first detection frame; marking a second detection frame of the animal head on the animal picture according to the coordinates, the length and the width of the two pairs of corner points of the second detection frame; the second detection frame comprises the entire head of the animal;
s3, extracting the features of the animal head within the range defined by the second detection frame to obtain a plurality of key point position coordinates corresponding to the animal head;
s4, obtaining the posture of the animal head according to the position coordinates of the second detection frame and the position coordinates of the plurality of key points;
the formula for calculating the coordinates, the length and the width of the two pairs of corner points of the second detection frame is as follows:
Figure DEST_PATH_IMAGE002
Figure DEST_PATH_IMAGE004
Figure DEST_PATH_IMAGE006
wherein (A), (B), (C), (D), (C), (B), (C)x 1y 1) And (a)x 2y 2) For two diagonal coordinates of said first detection frame, ((s))x c y c ) Is the center coordinate of the first detection frame,b w is the first detectionThe length of the frame is such that,b h the width of the first detection frame, ((ii))x c ′,y c ') is the center coordinates of the second detection frame,b w ' is the length of the second detection frame,b h ' is the width of the second detection frame,
Figure DEST_PATH_IMAGE008
is the offset.
2. The method of claim 1, wherein the animal picture is input into a Micro-YOLO neural network for target detection processing, wherein the Micro-YOLO neural network comprises at least a convolutional layer.
3. The method for estimating the head pose of the animal according to claim 1, wherein said extracting the features of the head of the animal within the range defined by the second detection frame comprises:
and mapping the animal picture marked with the second detection frame back to the original animal picture, extracting a region corresponding to the second detection frame, inputting the region corresponding to the second detection frame into an hourglass neural network, and extracting the characteristics of the head of the animal, wherein the hourglass neural network at least comprises a convolution layer and a pooling layer.
4. The method for estimating the head pose of the animal according to the claim 3, wherein the inputting the corresponding area of the second detection frame into the hourglass neural network, and the extracting the features of the head of the animal comprises:
downsampling the area corresponding to the second detection frame to obtain a first feature map corresponding to the head of the animal;
the first characteristic diagram is up-sampled to obtain a second characteristic diagram;
fusing the second characteristic diagram in a jumping connection mode to obtain a thermal point diagram corresponding to the head of the animal;
and obtaining the position coordinates of the maximum activation value of each heat point in the heat point diagram to obtain the position coordinates of a plurality of key points corresponding to the animal head.
5. The animal head pose estimation method of claim 4, wherein the upsampling employs a bilinear interpolation method.
6. An animal head pose estimation system, comprising:
the first processing module is used for carrying out target detection processing on the animal picture and marking a first detection frame of the head of the animal on the animal picture;
the second processing module is used for marking a second detection frame of the head of the animal according to the first detection frame and recording the position coordinates of the second detection frame, wherein the second detection frame comprises the whole head of the animal, and formulas for calculating the coordinates, the length and the width of two pairs of corner points of the second detection frame are as follows:
Figure DEST_PATH_IMAGE002A
Figure DEST_PATH_IMAGE004A
Figure DEST_PATH_IMAGE006A
wherein (A), (B), (C), (D), (C), (B), (C)x 1y 1) And (a)x 2y 2) For two diagonal coordinates of said first detection frame, ((s))x c y c ) Is the center coordinate of the first detection frame,b w is the length of the first detection frame,b h the width of the first detection frame, ((ii))x c ′,y c ′)Is the center coordinate of the second detection frame,b w ' is the length of the second detection frame,b h ' is the width of the second detection frame,
Figure DEST_PATH_IMAGE008A
is the offset;
the feature extraction module is used for extracting features of the animal head within the range defined by the second detection frame to obtain a plurality of key point position coordinates corresponding to the animal head;
and the calculation module is used for obtaining the posture of the animal head according to the position coordinates of the second detection frame and the position coordinates of the plurality of key points.
7. The system for estimating the head pose of the animal according to claim 6, wherein the first processing module performs a target detection process on the animal picture by using a Micro-YOLO neural network;
the second processing module marks a second detection frame of the animal head according to the first detection frame by adopting a regression optimization neural network;
the feature extraction module adopts a hourglass neural network to extract features of the animal head.
CN201910698158.3A 2019-07-30 2019-07-30 Animal head posture estimation method and system Active CN110569719B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201910698158.3A CN110569719B (en) 2019-07-30 2019-07-30 Animal head posture estimation method and system

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910698158.3A CN110569719B (en) 2019-07-30 2019-07-30 Animal head posture estimation method and system

Publications (2)

Publication Number Publication Date
CN110569719A CN110569719A (en) 2019-12-13
CN110569719B true CN110569719B (en) 2022-05-17

Family

ID=68773235

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910698158.3A Active CN110569719B (en) 2019-07-30 2019-07-30 Animal head posture estimation method and system

Country Status (1)

Country Link
CN (1) CN110569719B (en)

Families Citing this family (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110569719B (en) * 2019-07-30 2022-05-17 中国科学技术大学 Animal head posture estimation method and system
CN111931764B (en) * 2020-06-30 2024-04-16 华为云计算技术有限公司 Target detection method, target detection frame and related equipment
CN112101259A (en) * 2020-09-21 2020-12-18 中国农业大学 Single pig body posture recognition system and method based on stacked hourglass network
CN112016527B (en) * 2020-10-19 2022-02-01 成都大熊猫繁育研究基地 Panda behavior recognition method, system, terminal and medium based on deep learning
CN112464744A (en) * 2020-11-09 2021-03-09 湖北省农业科学院农产品加工与核农技术研究所 Fish posture identification method
CN116883458B (en) * 2023-09-06 2024-01-09 中国科学技术大学 Transformer-based multi-target tracking system fusing motion characteristics with observation as center

Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107909005A (en) * 2017-10-26 2018-04-13 西安电子科技大学 Personage's gesture recognition method under monitoring scene based on deep learning
CN108447090A (en) * 2016-12-09 2018-08-24 株式会社理光 The method, apparatus and electronic equipment of object gesture estimation
CN108875492A (en) * 2017-10-11 2018-11-23 北京旷视科技有限公司 Face datection and crucial independent positioning method, device, system and storage medium
CN108985148A (en) * 2018-05-31 2018-12-11 成都通甲优博科技有限责任公司 A kind of hand critical point detection method and device
CN109064514A (en) * 2018-07-03 2018-12-21 北京航空航天大学 A kind of six-freedom degree pose algorithm for estimating returned based on subpoint coordinate
CN109190537A (en) * 2018-08-23 2019-01-11 浙江工商大学 A kind of more personage's Attitude estimation methods based on mask perceived depth intensified learning
CN109376571A (en) * 2018-08-03 2019-02-22 西安电子科技大学 Estimation method of human posture based on deformation convolution
CN109685013A (en) * 2018-12-25 2019-04-26 上海智臻智能网络科技股份有限公司 The detection method and device of header key point in human body attitude identification
CN110569719A (en) * 2019-07-30 2019-12-13 中国科学技术大学 animal head posture estimation method and system

Patent Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108447090A (en) * 2016-12-09 2018-08-24 株式会社理光 The method, apparatus and electronic equipment of object gesture estimation
CN108875492A (en) * 2017-10-11 2018-11-23 北京旷视科技有限公司 Face datection and crucial independent positioning method, device, system and storage medium
CN107909005A (en) * 2017-10-26 2018-04-13 西安电子科技大学 Personage's gesture recognition method under monitoring scene based on deep learning
CN108985148A (en) * 2018-05-31 2018-12-11 成都通甲优博科技有限责任公司 A kind of hand critical point detection method and device
CN109064514A (en) * 2018-07-03 2018-12-21 北京航空航天大学 A kind of six-freedom degree pose algorithm for estimating returned based on subpoint coordinate
CN109376571A (en) * 2018-08-03 2019-02-22 西安电子科技大学 Estimation method of human posture based on deformation convolution
CN109190537A (en) * 2018-08-23 2019-01-11 浙江工商大学 A kind of more personage's Attitude estimation methods based on mask perceived depth intensified learning
CN109685013A (en) * 2018-12-25 2019-04-26 上海智臻智能网络科技股份有限公司 The detection method and device of header key point in human body attitude identification
CN110569719A (en) * 2019-07-30 2019-12-13 中国科学技术大学 animal head posture estimation method and system

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
"Cascade R-CNN: Delving into High Quality Object Detection";Zhaowei Cai等;《arXiv》;20171203;第1-9页 *

Also Published As

Publication number Publication date
CN110569719A (en) 2019-12-13

Similar Documents

Publication Publication Date Title
CN110569719B (en) Animal head posture estimation method and system
CN109934115B (en) Face recognition model construction method, face recognition method and electronic equipment
CN114424250A (en) Structural modeling
CN110879982B (en) Crowd counting system and method
CN111627050B (en) Training method and device for target tracking model
CN110555481A (en) Portrait style identification method and device and computer readable storage medium
CN112862824A (en) Novel coronavirus pneumonia focus detection method, system, device and storage medium
CN110909618B (en) Method and device for identifying identity of pet
CN111626295B (en) Training method and device for license plate detection model
CN113705583B (en) Target detection and identification method based on convolutional neural network model
CN110738650B (en) Infectious disease infection identification method, terminal device and storage medium
CN112926696A (en) Interpretable local migration mutual learning method based on attention diagram
CN115457492A (en) Target detection method and device, computer equipment and storage medium
CN111626379A (en) X-ray image detection method for pneumonia
CN113420648B (en) Target detection method and system with rotation adaptability
CN112750155B (en) Panoramic depth estimation method based on convolutional neural network
CN118196885A (en) Mouse gesture estimation method and system based on deep learning
CN114972492A (en) Position and pose determination method and device based on aerial view and computer storage medium
Oh et al. Local selective vision transformer for depth estimation using a compound eye camera
CN111709269A (en) Human hand segmentation method and device based on two-dimensional joint information in depth image
CN116523957A (en) Multi-target tracking method, system, electronic equipment and storage medium
CN116310899A (en) YOLOv 5-based improved target detection method and device and training method
CN114724190A (en) Mood recognition method based on pet posture
CN115544190A (en) Semantic map updating method and device, computer equipment and storage medium
KR20180012638A (en) Method and apparatus for detecting object in vision recognition with aggregate channel features

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant