CN106485735A - Human body target recognition and tracking method based on stereovision technique - Google Patents

Human body target recognition and tracking method based on stereovision technique Download PDF

Info

Publication number
CN106485735A
CN106485735A CN201510550955.9A CN201510550955A CN106485735A CN 106485735 A CN106485735 A CN 106485735A CN 201510550955 A CN201510550955 A CN 201510550955A CN 106485735 A CN106485735 A CN 106485735A
Authority
CN
China
Prior art keywords
target
image
camera
coordinate system
point
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201510550955.9A
Other languages
Chinese (zh)
Inventor
吕芳
任侃
潘佳惠
韶阿俊
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Nanjing University of Science and Technology
Original Assignee
Nanjing University of Science and Technology
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Nanjing University of Science and Technology filed Critical Nanjing University of Science and Technology
Priority to CN201510550955.9A priority Critical patent/CN106485735A/en
Publication of CN106485735A publication Critical patent/CN106485735A/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/20Movements or behaviour, e.g. gesture recognition
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10016Video; Image sequence
    • G06T2207/10021Stereoscopic video; Stereoscopic image sequence

Landscapes

  • Engineering & Computer Science (AREA)
  • Health & Medical Sciences (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • General Health & Medical Sciences (AREA)
  • Psychiatry (AREA)
  • Social Psychology (AREA)
  • Human Computer Interaction (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Multimedia (AREA)
  • Theoretical Computer Science (AREA)
  • Image Processing (AREA)

Abstract

The invention discloses a kind of human body target recognition and tracking method based on stereovision technique, including:Same width scene picture is obtained by two cameras simultaneously from two different angles, form stereo pairs;The inside and outside parameter of video camera is determined by camera calibration, establishes imaging model;Using the matching algorithm based on window, window is created centered on the point to be matched of wherein piece image, identical sliding window is created on another piece image, sliding window is moved in units of pixel successively along EP point, calculation window matching degree, optimal match point is found, the three-dimensional geometric information of target is obtained by principle of parallax, generate depth image;Head and shoulder information are distinguished with gray feature using One-Dimensional Maximum-Entropy thresholding method;Targetpath is tracked and obtains using Adaptive window tracking to human body target.Present invention achieves the accurate identification of human body target and count tracking, and effectively remove ambient noise, it is to avoid the interference of environment.

Description

Human body target identification and tracking method based on stereoscopic vision technology
Technical Field
The invention relates to a human body target tracking method, in particular to a human body target identification and tracking method based on a stereoscopic vision technology.
Background
With the rapid improvement of performances in the aspects of computer storage, operation and the like, people gradually apply computers to realize complex functions such as scene reconstruction, target recognition, man-machine interaction and the like, so that the scale and the research direction of the computer application field are exploited, and the rapid development of related subjects is promoted. As an active research field at present, the essence of computer vision is to use a camera to replace human eyes, use a computer to replace a brain, recognize and track a target, perform corresponding graphic analysis and processing, and generate an image suitable for instrument detection or human eye observation. The video technology can continuously transmit images in a period of time, contains more detailed information, and has the advantages of intuition, concrete performance, easy processing and the like. Identification of video objects has become an important issue in the fields of image processing, pattern recognition, human-computer interaction, and the like, and is widely used in various intelligent systems in the fields of manufacturing, medical diagnosis, military, and the like.
The traditional APC system mainly comprises a pressure sensing system and an infrared shielding system, and with the rapid development of laser infrared technology, a proper heat release infrared probe is adopted to detect signals sent by a human body for identification and counting. When the target moves, the infrared sensor detects the change caused by the infrared spectrum of the human body, namely the process of the motion of the human body target is obtained, the target can be distinguished through signal processing, and the method is low in cost and simple to operate, but has the problems of inaccurate identification and statistical result, limited application place and the like.
Meanwhile, the image processing method can also be used for solving the human body recognition problem. However, most methods only use some recognition algorithms for two-dimensional images, such as selecting some parts of the human body as features, and try to obtain a match in the images, thereby achieving the purpose of recognition. At present, there are many methods commonly used for human body recognition: the method based on the human body model and the structural elements has higher requirements on extracting image information of the whole person, is difficult to process moving and deforming objects, and has high requirements on the real-time property of image acquisition; the method is mainly based on the principle of a wavelet template, the whole image needs to be searched according to different sizes, and the calculation amount is large.
Disclosure of Invention
The invention aims to provide a human body target recognition and tracking method based on a stereoscopic vision technology.
The technical solution for realizing the purpose of the invention is as follows: a human body target recognition and tracking method based on a stereoscopic vision technology is characterized by comprising the following steps:
step 1, two cameras simultaneously acquire pictures of the same scene from two different angles to form a stereo image pair;
step 2, determining internal and external parameters of the camera through camera calibration, and establishing an imaging model;
3, adopting a window-based matching algorithm, creating a window by taking a point to be matched of one image as a center, creating the same sliding window on the other image, sequentially moving the sliding windows along an epipolar line by taking a pixel point as a unit, calculating window matching measure, finding the optimal matching point, obtaining three-dimensional geometric information of a target by a parallax principle, and generating a depth image;
step 4, distinguishing head and shoulder information by adopting a one-dimensional maximum entropy threshold segmentation method and combining with gray features, and identifying a human body target;
and 5, tracking the human body target by adopting a self-adaptive wave gate tracking method and obtaining a target track.
Compared with the prior art, the invention has the following remarkable advantages:
(1) the invention has small calculated amount, can quickly and accurately identify the human body target by using simple images,
(2) under the condition of crowding, the depth information of the image can be used for identification, interference is effectively eliminated, and a moving target is distinguished;
(3) the invention can effectively realize human target tracking and counting and avoid repetition.
Drawings
Fig. 1 is a flow chart of a human body target recognition and tracking method based on a stereoscopic vision technology.
FIG. 2 is an original image according to an embodiment of the present invention.
FIG. 3 is a segmented image of the shoulders of a human target according to an embodiment of the present invention.
Fig. 4 is a motion trail diagram of a human target according to an embodiment of the present invention.
Detailed Description
With reference to fig. 1, the human body target recognition and tracking method based on the stereoscopic vision technology of the present invention includes the following steps:
step 1, obtaining a stereo image pair:
two MTV-1881EX-3 cameras are placed in parallel, and pictures of the same scene are simultaneously obtained from two different angles to form a stereo image pair;
step 2, determining internal and external parameters of the camera through camera calibration, and establishing an imaging model, specifically:
step 2-1, calibrating the coordinates of the camera, wherein the calibration graph is a checkerboard, and the calibration principle is as follows:
assuming that the world coordinate system plane with z equal to 0 is the template plane, [ r ]1r2r3]Is a rotation matrix of the camera coordinate system relative to the world coordinate system, and t is a translation vector of the camera coordinate system relative to the world coordinate system, [ X Y1 ]]THomogeneous coordinates of points on the template, [ u v 1 ]]TProjecting points on the template plane to homogeneous coordinates on the image plane, wherein K represents a camera internal reference matrix;
step 2-2, setting a camera coordinate system OxcyczcIs a rectangular coordinate system fixed on the camera, and the origin O is defined as the optical center, x, of the camerac,ycThe axes are respectively parallel to the x, y and z axes of the image physical coordinate systemcThe axis coinciding with the optical axis, i.e. zcThe axis being perpendicular to the imaging plane of the camera, the distance OO of the optical centre to the image plane1Is the effective focal length f of the camera;
step 2-3, setting (x)w,yw,zw) Is the three-dimensional coordinate of a certain P point in a three-dimensional world coordinate system, (x)c,yc,zc) For the three-dimensional coordinates of the same point p in the camera coordinate system, the transformation of the point in the world coordinate system into the camera coordinate system is represented by an orthogonal rotation matrix R and a translation transformation matrix T as:
where R is a 3 × 3 rotation matrix and a translation matrix
The orthogonal matrix R is a direction cosine combination of the optical axis with respect to the coordinate axes of the world coordinate system, and includes three independent angle variables (euler angles): rotation by an angle ψ (yaw) about the x-axis; rotation by an angle θ (pitch) about the y-axis; rotating the phi angle (rolling) around the z axis, and adding three variables of T to form six parameters which are called external parameters of the camera;
step 2-4, simplifying rigid transformation of the world coordinate system and the camera coordinate system into a uniform coordinate and matrix form:
thus, a matrix M may be used between the world coordinate system and the camera coordinate system2Is expressed as long as M is known2The transformation of coordinates between two coordinate systems can be performed;
the transformation from the camera coordinate system to the ideal image physical coordinate system, i.e. the ideal perspective projection transformation under the pinhole model, has the following formula:
x=f·xc/zcy=f·yc/zc
x and y are respectively the abscissa and the ordinate of an ideal image physical coordinate system;
the above formula is also expressed in terms of homogeneous coordinates and matrices as:
the transformation of the ideal image coordinate system to the image pixel coordinate system, expressed in homogeneous coordinates, is:
wherein u is0、v0Coordinates representing a camera coordinate system;
the inverse relationship is:
by substituting the above relational expressions, the relation between the coordinates of the point P represented by the world coordinate system and the coordinates (u, v) of the projection P' thereof can be obtained:
wherein α f/dx f sx/dy,β=f/dy;M1For an array of internal parameters, M2The method is characterized in that the method is an external parameter matrix, wherein M is a matrix 3 × 4, called a projection matrix, and represents the basic relationship between two-dimensional image coordinates and three-dimensional world coordinates, the world coordinates of known object points can be used for solving corresponding ideal image coordinates, and conversely, if the image coordinates of the M matrix and the image points are known, a space ray corresponding to the optical center of a camera can be solved;
obtaining the basic relation between the two-dimensional image coordinate and the three-dimensional world coordinate, namely completing the calibration of the camera; the coordinates of the image obtained by the camera can be uniformly converted into coordinates on a three-dimensional world coordinate system through the imaging model, namely, the imaging model of the image obtained by the camera on the three-dimensional world coordinate system is determined.
3, adopting a window-based matching algorithm, creating a window by taking a point to be matched of one image as a center, creating the same sliding window on the other image, sequentially moving the sliding windows along an epipolar line by taking a pixel point as a unit, calculating window matching measure, finding the optimal matching point, obtaining three-dimensional geometric information of a target by a parallax principle, and generating a depth image; the method specifically comprises the following steps:
step 3-1, the right image is assumed as a reference, and a difference is made between the right image and the background to obtain a foreground image;
step 3-2, determining parallax:
the first step, in the foreground image, assuming that the right image is taken as a reference, calculating the gray difference value of each pixel at a given parallax with the corresponding point of the left image;
secondly, on each parallax, a narrow-strip-shaped window perpendicular to the base line direction is used instead, and the sum of the gray level difference of the window taking each pixel as the center is calculated by adopting a window-based matching algorithm, wherein the expression is as follows:
wherein m x n is the size of the template window, γ is the unit length of the template window, and is the unit width of the template window, and Iright[xe+γ,ye+]Is a right view [ xe+γ,ye+]Gray value at coordinate, Ileft[xe+γ+d,ye+]Is a left view [ xe+γ+d,ye+]Gray value at coordinates, d is parallax;
thirdly, in the set parallax range, taking the maximum parallax from the minimum parallax of the d, sequentially comparing the values of the expressions, wherein the point corresponding to the minimum value is the optimal matching point, and the corresponding parallax value is used as the parallax value of the pixel point;
step 3-3, determining depth information of the target:
the binocular distance measurement mainly utilizes the relationship that the difference (namely parallax) directly existing in the transverse coordinates of the imaging of a target point on a left view and a right view and the distance Z from the target point to an imaging plane have inverse proportion, and under the condition that the focal length of a camera is known, the depth information of any point, namely the coordinate value of the point on the Z axis under a camera coordinate system, and b is set as the optical center distance of two cameras; the vertical distance from the target Q to the camera is H; same focal length of f, Q1、Q2An imaging point of a target Q in a camera; d isParallax, assuming that the optical axes of the two cameras are parallel to each other, is derived from a similar triangle:
H=(b×f)/d
the obtained vertical distance from the target Q to the camera is the depth information of the target;
therefore, the stereoscopic vision counting can utilize two or more than two cameras with position offset to obtain the depth information of the scene by triangulation, if the point in the scene is required to have an image point in the left and right images; in the left view and the right view, the positions of image points are different, namely parallax, the distance between the points in the scene and the camera is different, the parallax is also different, and the parallax is smaller and larger along with the distance from the camera; binocular stereo vision is based on this disparity and uses trigonometric calculations to determine the distance of an object from the camera.
Step 4, distinguishing head and shoulder information by adopting a one-dimensional maximum entropy threshold segmentation method and combining with gray features, and identifying a human body target, wherein the method specifically comprises the following steps:
step 4-1, dividing the depth image into small lattices with L x L pixels, wherein L is a positive integer, one nine-square lattice is taken as a unit, the small lattices with L x L pixels are moved once every comparison in the sequence from left to right and from top to bottom when the small lattices are moved, and if the average gray scale of the middle lattice is higher than that of the surrounding eight neighborhoods, the middle lattice is determined to be a head target area;
step 4-2, setting a threshold value for the head target area, carrying out binarization, and segmenting the head target; the method specifically comprises the following steps:
adopting one-dimensional maximum entropy threshold segmentation method to segment head and non-head regions, and making p beiRepresenting the proportion of pixels with gray value i in the image, dividing the head and shoulder regions by taking the gray level t as a threshold, forming a head region by pixel points higher than the gray level t in the region, and forming a non-head region by pixel points lower than the gray level t, wherein the entropies of the non-head region and the head region are respectively defined as:
HO=-Σi[pi/(1-pt)]lg[pi/(1-pt)]
wherein:where i represents the gray value of the pixel (0. ltoreq. i. ltoreq.255), Ht=-Σtpilgpi,HE=-ΣipilgpiWhen the value of the entropy function is summedWhen the maximum is obtained, the gray level t can be used as a threshold value for dividing the image:
t=arg{max{HB+HO}}
step 4-3, determining the average gray level and the gray level variance of the segmented head region:
m, N respectively represents the row and column number of each region, epsilon respectively represents the row and column number of the unit, f (,) represents the gray value of the point, and when the gray variance is larger than a set threshold value, the pixel point is filtered;
4-4, filtering out a false target with a long and narrow appearance according to whether the ratio of the width of the total pixel to the height of the field of view of the head of the human body under different heights of the field of view meets a set range;
the geometric characteristics of the head mainly comprise ellipticity, head area, length and width and the like; the range of the width of the head total pixel under a certain field height can be known through continuous simulation tests, and meanwhile, the range of w/h can be obtained through simulation to be [0.65, 1.5], and a false target with a long and narrow appearance can be effectively filtered through the threshold discrimination.
Step 5, tracking the human body target by adopting a self-adaptive wave gate tracking method to obtain a target track, which specifically comprises the following steps:
step 5-1, respectively expanding 8 pixel values in four directions by setting a wave gate size as an external rectangular frame of a human head target, wherein the moving direction is from left to right and from top to bottom;
step 5-2, in the process of moving the wave gate, correcting the wave gate rectangular frame by using the single chip microcomputer, predicting the target position of the next field by using the target position quantity of the first five fields obtained by the camera, wherein the difference between the predicted value and the accurate value of the target of the actual next field is the wave gate tracking error;
and 5-3, in the moving process of the gate, improving the real-time performance and tracking accuracy of the system by adopting a 2-point linear prediction algorithm:
in the formula, tkIs the gate information of the k-th field,is a predicted value of the X coordinate of the wave gate center,for the x coordinate of the kth field wave gate information, the predicted value of the Y coordinate of the wave gate center can be obtained in the same way
And 5-4, after the target tracking is realized by adopting the self-adaptive wave gate, acquiring the flight path as follows:
firstly, initializing a flight path;
step two, the zeta frame appears the target for the first time, and the wave gate, the center coordinate and the frame number information of the target are determined;
thirdly, in the zeta +1 th frame, a target with the highest similarity is searched according to the gate coordinate prediction of the zeta frame target, a temporary track is generated, and the position of the target in the zeta +2 th frame is predicted;
and fourthly, determining the target track according to the information of the previous two frames in the zeta +2 th frame, and entering a stable tracking stage.
Step 5-5, counting human body targets and recording images:
the number of people entering and exiting is recorded at intervals, so that monitoring personnel can conveniently check the passenger flow at each time interval; and meanwhile, the generated depth image is stored in a computer hard disk, and the camera is read out from the hard disk when in a non-working state. On one hand, the system is used for the monitor to selectively watch the passenger flow video and restore the scene data; on the other hand, the experiment video can be repeatedly read to modify the threshold parameter so as to improve the identification precision.
The present invention will be further described with reference to the following specific examples.
Examples
And in combination with the original depth image of fig. 2, the human motion object which is dark in human hair and colored in dark clothes or is easily confused with the background when the human hair is light in color is processed by using a stereoscopic vision-based algorithm under the condition of visible light.
Fig. 3 is a segmented image of the head and shoulder of the human body, and it can be clearly seen that, as a result of processing the moving object under visible light, when the human body is dark and is painted with dark clothes or the human body is light and is easily confused with the background, the depth image output by the stereoscopic vision technology has high recognition accuracy, the head object of the human body is marked by a circle and is separated from the shoulder object, which is not easily affected by light and the background, and the object and the background are clearly distinguished.
With the combination of fig. 4, the human moving target which is crowded and is too close to contact with people is identified and track-traced, the three-dimensional information of the human body can be displayed, the human head target is marked by a square frame, and the curve of the head represents the track of the motion of the human body, so that the limitation of two-dimensional image human body identification is effectively broken through, the accurate identification and tracking of the human body target are realized, the same target is distinguished and counted, the background noise can be effectively removed, and the interference of the environment is avoided.

Claims (5)

1. A human body target recognition and tracking method based on a stereoscopic vision technology is characterized by comprising the following steps:
step 1, two cameras simultaneously acquire pictures of the same scene from two different angles to form a stereo image pair;
step 2, determining internal and external parameters of the camera through camera calibration, and establishing an imaging model;
3, adopting a window-based matching algorithm, creating a window by taking a point to be matched of one image as a center, creating the same sliding window on the other image, sequentially moving the sliding windows along an epipolar line by taking a pixel point as a unit, calculating window matching measure, finding the optimal matching point, obtaining three-dimensional geometric information of a target by a parallax principle, and generating a depth image;
step 4, distinguishing head and shoulder information by adopting a one-dimensional maximum entropy threshold segmentation method and combining with gray features, and identifying a human body target;
and 5, tracking the human body target by adopting a self-adaptive wave gate tracking method and obtaining a target track.
2. The human body target recognition and tracking method based on the stereoscopic vision technology as claimed in claim 1, wherein the step 2 is specifically as follows:
step 2-1, calibrating the coordinates of the camera, wherein the calibration graph is a checkerboard, and the calibration principle is as follows:
u v 1 = K r 1 r 2 r 3 t X Y 0 1 = K r 1 r 2 t X Y 1
assuming that the world coordinate system plane with z equal to 0 is the template plane, [ r ]1r2r3]Is a rotation matrix of the camera coordinate system relative to the world coordinate system, and t is a translation vector of the camera coordinate system relative to the world coordinate system, [ X Y1 ]]THomogeneous coordinates of points on the template, [ u v 1 ]]TProjecting points on the template plane to homogeneous coordinates on the image plane, wherein K represents a camera internal reference matrix;
step 2-2, setting a camera coordinate system OxcyczcIs a rectangular coordinate system fixed on the camera, and the origin O is defined as the optical center, x, of the camerac,ycThe axes are respectively parallel to the x, y and z axes of the image physical coordinate systemcThe axis coinciding with the optical axis, i.e. zcThe axis being perpendicular to the imaging plane of the camera, the distance OO of the optical centre to the image plane1Is the effective focal length f of the camera;
step 2-3, setting (x)w,yw,zw) Is the three-dimensional coordinate of a certain P point in a three-dimensional world coordinate system, (x)c,yc,zc) For the three-dimensional coordinates of the same point P in the camera coordinate system, the transformation of the point in the world coordinate system into the camera coordinate system is represented by an orthogonal rotation matrix R and a translation transformation matrix T as:
x c y c z c = R x w y w z w + T
where R is a 3 × 3 rotation matrix and a translation matrix T = t x t y t z ;
The orthogonal matrix R is a direction cosine combination of an optical axis relative to the coordinate axis of the world coordinate system, and comprises three independent angle variables: rotating the phi angle around the x axis, rotating the theta angle around the y axis and rotating the phi angle around the z axis, wherein three variables of the phi angle and the T are collectively called as external parameters of the camera;
step 2-4, simplifying rigid transformation of the world coordinate system and the camera coordinate system into a uniform coordinate and matrix form:
x c y c z c 1 = R T 0 T 1 x w y w z w 1 = M 2 x w y w z w 1
x c ~ = R T 0 T 1 x w ~ ⇒ x c ~ = M 2 x w ~ ⇒ x w ~ = M 2 - 1 x c ~
the transformation from the camera coordinate system to the ideal image physical coordinate system, i.e. the ideal perspective projection transformation under the pinhole model, has the following formula:
x=f·xc/zcy=f·yc/zc
x and y are respectively the abscissa and the ordinate of an ideal image physical coordinate system;
the above formula is also expressed in terms of homogeneous coordinates and matrices as:
z c x y 1 = f 0 0 0 0 f 0 0 0 0 1 0 x c y c z c 1
the transformation of the ideal image coordinate system to the image pixel coordinate system, expressed in homogeneous coordinates, is:
u v 1 = 1 / d x 0 u 0 0 1 / d x v 0 0 0 1 x y 1 = s x / d y 0 u 0 0 1 / d y v 0 0 0 1 x y 1
the inverse relationship is:
x y 1 = d y / s x 0 - u 0 d y / s x 0 d y - v 0 d y 0 0 1 u v 1
obtaining the relation between the coordinates of the point P represented by the world coordinate system and the coordinates (u, v) of the projection P' of the point P:
z c u v 1 = s x / d y 0 u 0 0 1 / d y v 0 0 0 1 f 0 0 0 0 f 0 0 0 0 1 0 R T 0 T 1 x w y w z w 1 = α 0 u 0 0 0 β v 0 0 0 0 1 0 R T 0 T 1 x w y w z w 1 = M 1 M 2 x w ~ = M x w ~
wherein α f/dx fsx/dy,β=f/dy;M1For an array of internal parameters, M2M is a projection matrix of 3 × 4, and represents the basic relationship between the two-dimensional image coordinates and the three-dimensional world coordinates.
3. The human body target recognition and tracking method based on the stereoscopic vision technology as claimed in claim 1, wherein the step 3 is specifically as follows:
step 3-1, the right image is assumed as a reference, and a difference is made between the right image and the background to obtain a foreground image;
step 3-2, determining parallax:
the first step, in the foreground image, assuming that the right image is taken as a reference, calculating the gray difference value of each pixel at a given parallax with the corresponding point of the left image;
secondly, on each parallax, a narrow-strip-shaped window perpendicular to the base line direction is used instead, and the sum of the gray level difference of the window taking each pixel as the center is calculated by adopting a window-based matching algorithm, wherein the expression is as follows:
Σ γ = - m 2 m 2 Σ δ = - n 2 n 2 | I r i g h t [ x e + γ , y e + δ ] - I l e f t [ x e + γ + d , y e + δ ] |
wherein m x n is the size of the template window, γ is the unit length of the template window, and is the unit width of the template window, and Iright[xe+γ,ye+]Is a right view [ xe+γ,ye+]Gray value at coordinate, Ileft[xe+γ+d,ye+]Is a left view [ xe+γ+d,ye+]Gray value at coordinates, d is parallax;
thirdly, in the set parallax range, taking the maximum parallax from the minimum parallax of the d, sequentially comparing the values of the expressions, wherein the point corresponding to the minimum value is the optimal matching point, and the corresponding parallax value is used as the parallax value of the pixel point;
step 3-3, determining depth information of the target:
knowing the focal length of the camera, setting b as the optical center distance of the two cameras, wherein the depth information of any point, namely the coordinate value of the point on the Z axis in the camera coordinate system; the vertical distance from the target Q to the camera is H; same focal length of f, Q1、Q2Respectively imaging points of the target Q in the two cameras; d is parallax, and assuming that the optical axes of the two cameras are parallel to each other, the derivation is given by the similar triangle:
H=(b×f)/d
and the obtained vertical distance from the target Q to the camera is the depth information of the target.
4. The human body target recognition and tracking method based on the stereoscopic vision technology as claimed in claim 1, wherein the step 4 is specifically as follows:
step 4-1, dividing the depth image into small lattices with L x L pixels, wherein L is a positive integer, one nine-square lattice is taken as a unit, the small lattices with L x L pixels are moved once every comparison in the sequence from left to right and from top to bottom when the small lattices are moved, and if the average gray scale of the middle lattice is higher than that of the surrounding eight neighborhoods, the middle lattice is determined to be a head target area;
step 4-2, setting a threshold value for the head target area, carrying out binarization, and segmenting the head target; the method specifically comprises the following steps:
adopting one-dimensional maximum entropy threshold segmentation method to segment head and non-head regions, and making p beiRepresenting the proportion of pixels with gray value i in the image, dividing the head and shoulder regions by taking the gray level t as a threshold value, forming the head region by the pixel points with gray level higher than t in the region,the pixel points lower than the gray level t form a non-head region, and the entropies of the non-head region and the head region are respectively defined as:
H B = - Σ i ( p i p t ) lg ( p i p t )
HO=-Σi[pi/(1-pt)]lg[pi/(1-pt)]
wherein,i represents the gray value of the pixel (0 ≦ i ≦ 255), Ht=-Σtpilgpi,HE=-ΣipilgpiWhen the value of the entropy function is summedTaking the maximum, the gray level t as the threshold for segmenting the image:
t=arg{max{HB+HO}}
step 4-3, determining the average gray level and the gray level variance of the segmented head region:
g ‾ = Σ ϵ = 0 M - 1 Σ ∈ = 0 N - 1 f ( ϵ , ∈ ) M * N
var = Σ ϵ = 0 M - 1 Σ ∈ = 0 N - 1 ( f ( ϵ , ∈ ) - g ‾ ) 2 M * N
m, N respectively represents the row and column number of each region, epsilon respectively represents the row and column number of the unit, f (,) represents the gray value of the point, and when the gray variance is larger than a set threshold value, the pixel point is filtered;
and 4-4, filtering out a false target with a long and narrow appearance according to whether the ratio of the width of the total pixel to the height of the field of view of the head of the human body under different field of view heights meets a set range, so as to obtain the human body target.
5. The human body target recognition and tracking method based on the stereoscopic vision technology as claimed in claim 1, wherein the step 5 is specifically as follows:
step 5-1, respectively expanding 8 pixel values in four directions by setting a wave gate size as an external rectangular frame of a human head target, wherein the moving direction is from left to right and from top to bottom;
step 5-2, in the process of moving the wave gate, correcting the wave gate rectangular frame by using the single chip microcomputer, predicting the target position of the next field by using the target position quantity of the first five fields obtained by the camera, wherein the difference between the predicted value and the accurate value of the target of the actual next field is the wave gate tracking error;
and 5-3, in the moving process of the gate, improving the real-time performance and tracking accuracy of the system by adopting a 2-point linear prediction algorithm:
X ^ ( t k + 1 t k ) = 2 x t k - x t k - 1
in the formula, tkIs the gate information of the k-th field,is a predicted value of the X coordinate of the wave gate center,for the x coordinate of the kth field wave gate information, the predicted value of the Y coordinate of the wave gate center can be obtained in the same way
And 5-4, after the target tracking is realized by adopting the self-adaptive wave gate, acquiring the flight path as follows:
firstly, initializing a flight path;
step two, the zeta frame appears the target for the first time, and the wave gate, the center coordinate and the frame number information of the target are determined;
thirdly, in the zeta +1 th frame, a target with the highest similarity is searched according to the gate coordinate prediction of the zeta frame target, a temporary track is generated, and the position of the target in the zeta +2 th frame is predicted;
and fourthly, determining the target track according to the information of the previous two frames in the zeta +2 th frame, and entering a stable tracking stage.
CN201510550955.9A 2015-09-01 2015-09-01 Human body target recognition and tracking method based on stereovision technique Pending CN106485735A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201510550955.9A CN106485735A (en) 2015-09-01 2015-09-01 Human body target recognition and tracking method based on stereovision technique

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201510550955.9A CN106485735A (en) 2015-09-01 2015-09-01 Human body target recognition and tracking method based on stereovision technique

Publications (1)

Publication Number Publication Date
CN106485735A true CN106485735A (en) 2017-03-08

Family

ID=58235758

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201510550955.9A Pending CN106485735A (en) 2015-09-01 2015-09-01 Human body target recognition and tracking method based on stereovision technique

Country Status (1)

Country Link
CN (1) CN106485735A (en)

Cited By (24)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108413584A (en) * 2018-02-11 2018-08-17 四川虹美智能科技有限公司 A kind of air-conditioning and its control method
CN108960014A (en) * 2017-05-23 2018-12-07 北京旷视科技有限公司 Image processing method, device and system and storage medium
CN109104574A (en) * 2018-10-09 2018-12-28 中国科学院深圳先进技术研究院 A kind of method and system of cradle head camera masking privacy area
CN109155067A (en) * 2018-01-23 2019-01-04 深圳市大疆创新科技有限公司 The control method of moveable platform, equipment, computer readable storage medium
CN109274871A (en) * 2018-09-27 2019-01-25 维沃移动通信有限公司 A kind of image imaging method and device of mobile terminal
CN109308718A (en) * 2018-08-09 2019-02-05 上海青识智能科技有限公司 A kind of space personnel positioning apparatus and method based on more depth cameras
CN109352654A (en) * 2018-11-23 2019-02-19 武汉科技大学 A kind of intelligent robot system for tracking and method based on ROS
CN109446999A (en) * 2018-10-31 2019-03-08 中电科新型智慧城市研究院有限公司 Quick sensing system and method based on the dynamic human body movement that statistics calculates
CN109684991A (en) * 2018-12-24 2019-04-26 北京旷视科技有限公司 Image processing method, device, electronic equipment and storage medium
CN109753858A (en) * 2017-11-07 2019-05-14 北京中科慧眼科技有限公司 A kind of road barricade object detecting method and device based on binocular vision
CN109961455A (en) * 2017-12-22 2019-07-02 杭州萤石软件有限公司 Target detection method and device
CN110342134A (en) * 2019-07-23 2019-10-18 珠海市一微半导体有限公司 A kind of garbage classification identifying system and its method based on binocular vision
CN110443228A (en) * 2019-08-20 2019-11-12 图谱未来(南京)人工智能研究院有限公司 A kind of method for pedestrian matching, device, electronic equipment and storage medium
CN110900611A (en) * 2019-12-13 2020-03-24 合肥工业大学 Novel mechanical arm target positioning and path planning method
CN111028271A (en) * 2019-12-06 2020-04-17 浩云科技股份有限公司 Multi-camera personnel three-dimensional positioning and tracking system based on human skeleton detection
CN111080712A (en) * 2019-12-06 2020-04-28 浩云科技股份有限公司 Multi-camera personnel positioning, tracking and displaying method based on human body skeleton detection
CN111145220A (en) * 2019-12-31 2020-05-12 东南大学 Tunnel target track tracking method based on visual information
CN111243274A (en) * 2020-01-20 2020-06-05 陈俊言 Road collision early warning system and method for non-internet traffic individuals
CN112396611A (en) * 2020-10-27 2021-02-23 武汉理工大学 Point-line visual odometer self-adaptive optimization method and device and storage medium
CN112396633A (en) * 2020-10-19 2021-02-23 北京理工大学 Target tracking and track three-dimensional reproduction method and device based on single camera
CN112580426A (en) * 2020-10-30 2021-03-30 江苏集萃未来城市应用技术研究所有限公司 Monocular vision-based outdoor personnel dynamic tracking and positioning method
WO2021098081A1 (en) * 2019-11-22 2021-05-27 大连理工大学 Trajectory feature alignment-based multispectral stereo camera self-calibration algorithm
WO2021139176A1 (en) * 2020-07-30 2021-07-15 平安科技(深圳)有限公司 Pedestrian trajectory tracking method and apparatus based on binocular camera calibration, computer device, and storage medium
CN113835074A (en) * 2021-08-04 2021-12-24 南京常格科技发展有限公司 People flow dynamic monitoring method based on millimeter wave radar

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102222265A (en) * 2010-04-13 2011-10-19 上海申腾三盛信息技术工程有限公司 Binocular vision and laterally mounted video camera-based passenger flow counting method
US20130236058A1 (en) * 2007-07-03 2013-09-12 Shoppertrak Rct Corporation System And Process For Detecting, Tracking And Counting Human Objects Of Interest
CN103455792A (en) * 2013-08-20 2013-12-18 深圳市飞瑞斯科技有限公司 Guest flow statistics method and system
CN103839038A (en) * 2012-11-23 2014-06-04 浙江大华技术股份有限公司 People counting method and device
CN104504688A (en) * 2014-12-10 2015-04-08 上海大学 Method and system based on binocular stereoscopic vision for passenger flow density estimation

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20130236058A1 (en) * 2007-07-03 2013-09-12 Shoppertrak Rct Corporation System And Process For Detecting, Tracking And Counting Human Objects Of Interest
CN102222265A (en) * 2010-04-13 2011-10-19 上海申腾三盛信息技术工程有限公司 Binocular vision and laterally mounted video camera-based passenger flow counting method
CN103839038A (en) * 2012-11-23 2014-06-04 浙江大华技术股份有限公司 People counting method and device
CN103455792A (en) * 2013-08-20 2013-12-18 深圳市飞瑞斯科技有限公司 Guest flow statistics method and system
CN104504688A (en) * 2014-12-10 2015-04-08 上海大学 Method and system based on binocular stereoscopic vision for passenger flow density estimation

Non-Patent Citations (6)

* Cited by examiner, † Cited by third party
Title
刘欣: "基于立体视觉的公交客流统计方法与实现", 《中国优秀硕士论文全文数据库 工程科技Ⅱ辑》 *
孙中旭: "基于双目视觉的运动物体测距测速研究", 《中国优秀硕士学位论文全文数据库 信息科技辑》 *
尹章芹: "三维自动客流计数系统设计", 《中国优秀硕士学位论文全文数据库 信息科技辑》 *
李秋竹: "基于DSP的多目标跟踪系统算法研究", 《中国优秀硕士学位论文全文数据库 信息科技辑》 *
杨帆: "《图像数字处理与分析》", 31 May 2015, 北京航空航天大学出版社 *
赛干内克,希伯特: "《三维计算机视觉技术和算法导论》", 1 October 2014, 国防工业出版社 *

Cited By (37)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108960014A (en) * 2017-05-23 2018-12-07 北京旷视科技有限公司 Image processing method, device and system and storage medium
CN108960014B (en) * 2017-05-23 2021-05-11 北京旷视科技有限公司 Image processing method, device and system and storage medium
CN109753858A (en) * 2017-11-07 2019-05-14 北京中科慧眼科技有限公司 A kind of road barricade object detecting method and device based on binocular vision
US11367276B2 (en) 2017-12-22 2022-06-21 Hangzhou Ezviz Software Co., Ltd. Target detection method and apparatus
CN109961455A (en) * 2017-12-22 2019-07-02 杭州萤石软件有限公司 Target detection method and device
CN109155067A (en) * 2018-01-23 2019-01-04 深圳市大疆创新科技有限公司 The control method of moveable platform, equipment, computer readable storage medium
US12002221B2 (en) 2018-01-23 2024-06-04 SZ DJI Technology Co., Ltd. Control method and device for mobile platform, and computer readable storage medium
US11227388B2 (en) 2018-01-23 2022-01-18 SZ DJI Technology Co., Ltd. Control method and device for mobile platform, and computer readable storage medium
CN108413584A (en) * 2018-02-11 2018-08-17 四川虹美智能科技有限公司 A kind of air-conditioning and its control method
CN109308718B (en) * 2018-08-09 2022-09-23 上海青识智能科技有限公司 Space personnel positioning device and method based on multiple depth cameras
CN109308718A (en) * 2018-08-09 2019-02-05 上海青识智能科技有限公司 A kind of space personnel positioning apparatus and method based on more depth cameras
CN109274871A (en) * 2018-09-27 2019-01-25 维沃移动通信有限公司 A kind of image imaging method and device of mobile terminal
CN109104574A (en) * 2018-10-09 2018-12-28 中国科学院深圳先进技术研究院 A kind of method and system of cradle head camera masking privacy area
CN109446999A (en) * 2018-10-31 2019-03-08 中电科新型智慧城市研究院有限公司 Quick sensing system and method based on the dynamic human body movement that statistics calculates
CN109446999B (en) * 2018-10-31 2021-08-31 中电科新型智慧城市研究院有限公司 Rapid sensing system and method for dynamic human body movement based on statistical calculation
CN109352654A (en) * 2018-11-23 2019-02-19 武汉科技大学 A kind of intelligent robot system for tracking and method based on ROS
CN109684991A (en) * 2018-12-24 2019-04-26 北京旷视科技有限公司 Image processing method, device, electronic equipment and storage medium
CN110342134A (en) * 2019-07-23 2019-10-18 珠海市一微半导体有限公司 A kind of garbage classification identifying system and its method based on binocular vision
CN110443228A (en) * 2019-08-20 2019-11-12 图谱未来(南京)人工智能研究院有限公司 A kind of method for pedestrian matching, device, electronic equipment and storage medium
WO2021098081A1 (en) * 2019-11-22 2021-05-27 大连理工大学 Trajectory feature alignment-based multispectral stereo camera self-calibration algorithm
US11575873B2 (en) 2019-11-22 2023-02-07 Dalian University Of Technology Multispectral stereo camera self-calibration algorithm based on track feature registration
CN111028271B (en) * 2019-12-06 2023-04-14 浩云科技股份有限公司 Multi-camera personnel three-dimensional positioning and tracking system based on human skeleton detection
CN111080712A (en) * 2019-12-06 2020-04-28 浩云科技股份有限公司 Multi-camera personnel positioning, tracking and displaying method based on human body skeleton detection
CN111080712B (en) * 2019-12-06 2023-04-18 浩云科技股份有限公司 Multi-camera personnel positioning, tracking and displaying method based on human body skeleton detection
CN111028271A (en) * 2019-12-06 2020-04-17 浩云科技股份有限公司 Multi-camera personnel three-dimensional positioning and tracking system based on human skeleton detection
CN110900611A (en) * 2019-12-13 2020-03-24 合肥工业大学 Novel mechanical arm target positioning and path planning method
CN111145220B (en) * 2019-12-31 2022-11-18 东南大学 Tunnel target track tracking method based on visual information
CN111145220A (en) * 2019-12-31 2020-05-12 东南大学 Tunnel target track tracking method based on visual information
CN111243274A (en) * 2020-01-20 2020-06-05 陈俊言 Road collision early warning system and method for non-internet traffic individuals
WO2021139176A1 (en) * 2020-07-30 2021-07-15 平安科技(深圳)有限公司 Pedestrian trajectory tracking method and apparatus based on binocular camera calibration, computer device, and storage medium
CN112396633A (en) * 2020-10-19 2021-02-23 北京理工大学 Target tracking and track three-dimensional reproduction method and device based on single camera
CN112396633B (en) * 2020-10-19 2023-02-28 北京理工大学 Target tracking and track three-dimensional reproduction method and device based on single camera
CN112396611A (en) * 2020-10-27 2021-02-23 武汉理工大学 Point-line visual odometer self-adaptive optimization method and device and storage medium
CN112396611B (en) * 2020-10-27 2024-02-13 武汉理工大学 Self-adaptive optimization method, device and storage medium for point-line visual odometer
CN112580426A (en) * 2020-10-30 2021-03-30 江苏集萃未来城市应用技术研究所有限公司 Monocular vision-based outdoor personnel dynamic tracking and positioning method
CN113835074A (en) * 2021-08-04 2021-12-24 南京常格科技发展有限公司 People flow dynamic monitoring method based on millimeter wave radar
CN113835074B (en) * 2021-08-04 2024-01-16 南京常格科技发展有限公司 Dynamic people flow monitoring method based on millimeter wave radar

Similar Documents

Publication Publication Date Title
CN106485735A (en) Human body target recognition and tracking method based on stereovision technique
CN106503605A (en) Human body target recognition methods based on stereovision technique
CN108549873B (en) Three-dimensional face recognition method and three-dimensional face recognition system
Zaarane et al. Distance measurement system for autonomous vehicles using stereo camera
CN108416791B (en) Binocular vision-based parallel mechanism moving platform pose monitoring and tracking method
CN108629946B (en) Human body falling detection method based on RGBD sensor
WO2020172783A1 (en) Head posture tracking system used for transcranial magnetic stimulation diagnosis and treatment
CN108628306B (en) Robot walking obstacle detection method and device, computer equipment and storage medium
CN110910421B (en) Weak and small moving object detection method based on block characterization and variable neighborhood clustering
CN109035307B (en) Set area target tracking method and system based on natural light binocular vision
ES2764484T3 (en) Asynchronous signal processing procedure
CN105913013A (en) Binocular vision face recognition algorithm
CN102831427A (en) Texture feature extraction method fused with visual significance and gray level co-occurrence matrix (GLCM)
CN105869166A (en) Human body action identification method and system based on binocular vision
CN107203743B (en) Face depth tracking device and implementation method
CN109961092B (en) Binocular vision stereo matching method and system based on parallax anchor point
CN112257641A (en) Face recognition living body detection method
Byeong-Ho A Review on Image and Video processing
CN102510437B (en) Method for detecting background of video image based on distribution of red, green and blue (RGB) components
CN109443319A (en) Barrier range-measurement system and its distance measuring method based on monocular vision
CN107145820B (en) Binocular positioning method based on HOG characteristics and FAST algorithm
Wang et al. LBP-based edge detection method for depth images with low resolutions
CN105138979A (en) Method for detecting the head of moving human body based on stereo visual sense
CN117113284A (en) Multi-sensor fusion data processing method and device and multi-sensor fusion method
CN111489384B (en) Method, device, system and medium for evaluating shielding based on mutual viewing angle

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication
RJ01 Rejection of invention patent application after publication

Application publication date: 20170308