CN105550667A - Stereo camera based framework information action feature extraction method - Google Patents

Stereo camera based framework information action feature extraction method Download PDF

Info

Publication number
CN105550667A
CN105550667A CN201610047866.7A CN201610047866A CN105550667A CN 105550667 A CN105550667 A CN 105550667A CN 201610047866 A CN201610047866 A CN 201610047866A CN 105550667 A CN105550667 A CN 105550667A
Authority
CN
China
Prior art keywords
human body
coordinate system
matrix
stereo camera
rightarrow
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201610047866.7A
Other languages
Chinese (zh)
Other versions
CN105550667B (en
Inventor
陈启军
尹晓川
韩云
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Suzhou Tongqi Artificial Intelligence Technology Co ltd
Original Assignee
Tongji University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Tongji University filed Critical Tongji University
Priority to CN201610047866.7A priority Critical patent/CN105550667B/en
Publication of CN105550667A publication Critical patent/CN105550667A/en
Application granted granted Critical
Publication of CN105550667B publication Critical patent/CN105550667B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/20Movements or behaviour, e.g. gesture recognition
    • G06V40/23Recognition of whole body movements, e.g. for sport training
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/42Global feature extraction by analysis of the whole pattern, e.g. using frequency domain transformations or autocorrelation
    • G06V10/422Global feature extraction by analysis of the whole pattern, e.g. using frequency domain transformations or autocorrelation for representing the structure of the pattern or shape of an object therefor
    • G06V10/426Graphical representations

Landscapes

  • Engineering & Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Multimedia (AREA)
  • Theoretical Computer Science (AREA)
  • Health & Medical Sciences (AREA)
  • General Health & Medical Sciences (AREA)
  • Psychiatry (AREA)
  • Social Psychology (AREA)
  • Human Computer Interaction (AREA)
  • Measurement Of The Respiration, Hearing Ability, Form, And Blood Characteristics Of Living Organisms (AREA)
  • Image Analysis (AREA)
  • Image Processing (AREA)

Abstract

The invention relates to a stereo camera based framework information action feature extraction method. The method comprises the following steps: (1) obtaining coordinates of each joint point of a human body by utilizing a stereo camera; (2) determining a coordinate system of the human body; (3) calculating a first position matrix from a coordinate system of the camera to the coordinate system of the human body, wherein the first position matrix includes a rotary part matrix and a translation part matrix; (4) calculating a second position matrix from each joint point of the human body to the coordinate system of the camera; (5) calculating a relative position of each joint point of the human body relative to the coordinate system of the human body; (6) performing smoothing treatment on a rotary translation matrix between two adjacent frames, and accumulating variations between the two adjacent frames to obtain a translation variation and a rotary variation of the coordinate system of the human body from the beginning of an action to a current moment; and (7) generating an eigenvector of the current moment. Compared with the prior art, the method has the advantages of high precision, convenience for application and the like.

Description

Skeleton information action feature extraction method based on stereo camera
Technical Field
The invention relates to a skeleton information action feature extraction method, in particular to a skeleton information action feature extraction method based on a stereo camera.
Background
Human motion characteristics are being studied in biomedical engineering, physiotherapy, medical diagnosis and rehabilitation. The detection of the human motion characteristics has wide requirements in places such as nursing homes, hospitals and the like, and also has many applications in the fields of safety protection, battlefield reconnaissance and the like. Under the impetus of the development of motion performance analysis, visual monitoring and biometry, the method for extracting and analyzing different human motion is widely regarded.
Currently, the most common method for detecting motion characteristics of a human body is to use a sequence of visual images. However, the visual perception of human body movement is affected by distance, light changes, clothing changes and the shielding of various parts of the human body on the appearance, and the detection performance is reduced. Radar is an electromagnetic sensor, which can work in daytime and at night due to its long range of action and has the ability to penetrate through walls and the ground, and is also commonly used for detecting human motion characteristics. However, the traditional radar has lower working frequency, the influence of the micro Doppler effect of human motion is very small, and the human motion characteristics are more difficult to detect with high resolution.
Disclosure of Invention
The invention aims to overcome the defects of the prior art and provide a method for representing the action characteristics of skeleton information extracted by a stereo camera, which has high precision and convenient application.
The purpose of the invention can be realized by the following technical scheme: a skeleton information action feature extraction method based on a stereo camera comprises the following steps:
(1) acquiring coordinates of each joint point of the human body by using a stereo camera;
(2) determining a human body coordinate system;
(3) calculating a first position matrix from a camera coordinate system to a human body coordinate system, wherein the first position matrix comprises a rotating part matrix and a translating part matrix;
(4) calculating a second position matrix from each joint point of the human body to a camera coordinate system;
(5) calculating the relative position of each joint point of the human body relative to a human body coordinate system, and taking the relative position as a part of the required characteristics of motion recognition;
(6) smoothing the rotation translation matrix between two adjacent frames, and accumulating the variation between two adjacent needles to obtain the translation variation and the rotation variation of the human body coordinate system from the beginning of the action to the current moment;
(7) and generating a feature vector of the current moment.
The human body coordinate system is characterized in that an origin O of the human body coordinate system is an origin which is an intersection point of a connecting line of a left shoulder L and a right shoulder R of a human body and a human body symmetry axis, an x axis of the human body coordinate system is a ray from the origin O to the left shoulder L, a y axis of the human body coordinate system is a ray from the origin O to a human body center T, and a z axis of the human body coordinate system is perpendicular to a plane where the x axis and the y axis are located and meets the right-hand rule.
The step (3) is specifically as follows: let the coordinate of the left shoulder L of the human body acquired by the stereo camera be (CXL,CYL,CZL) The coordinate of the right shoulder R of the human body is (CXR,CYR,CZR) The coordinate of the center T of the human body is (CXT,CYT,CZT) The rotating parts are in matrix ofSaid first position matrix P O = R C O P O + P O C O R G , Wherein,OPCORGin order to translate a portion of the matrix,OPCORG=[COx,COy,COz]T O c x = X C L + X C R 2 , O c y = Y C L + Y C R 2 , O c z = Z C L + Z C R + Z C T 3 .
the rotating part matrix isThe calculation is carried out according to the formula (1),
d i a g | R L → | | T O → | | R L → × R T → | = R C O R L → T O → O Z → - - - ( 1 )
wherein,representing the vector from the right shoulder to the left shoulder of the human body,a vector representing the center of the human body to the origin of the human body coordinate system,represents the vector from the right shoulder of the human body to the center of the human body,
the step (4) is specifically as follows: orthogonalizing the rotating part matrix in the first position matrix to obtain a second position matrixComprises the following steps:
R C O ′ = U T · d i a g λ 1 - 1 2 λ 2 - 1 2 λ 3 - 1 2 · U · R C O
wherein,for a rotating partial matrix, U and λ satisfy R C O · R C O T = U T · d i a g λ 1 λ 2 λ 3 · U . Since the rotating part matrices calculated from the acquired point cloud data and obtained based on the three left and right shoulders and the body center are not necessarily orthogonal, it is necessary to orthogonalize the rotating part matrices,
the step (5) is specifically as follows: the coordinates of each joint point of the human body are converted into a human body coordinate system through coordinate transformation, so that the relative position of each joint point of the human body with respect to the human body coordinate system is obtained, and the influence of visual angle change in action recognition is solved; however, for the motion sequence with the indistinct four-limb movement, the change of the human body coordinate system in the motion needs to be added for distinguishing the motion such as rotation, jumping and the like.
The smoothing treatment specifically comprises the following steps: and performing arithmetic average on the obtained rotation and translation matrixes between two adjacent frames. Because the human body motion is equivalent to a rigid body, the change of the human body motion between two adjacent frames is smooth, and sudden change does not occur, so that the obtained change quantity between two adjacent frames is smoothed.
The step (7) is specifically as follows: and (5) taking the relative position in the step (5) and the translation variation and the rotation variation in the step (6) as the feature vector of the current moment.
Compared with the prior art, the invention has the following advantages:
(1) according to the invention, a human body coordinate system is established, so that the scene information does not need to be modeled, and the extra error caused by calculating a horizontal or vertical reference coordinate system is avoided;
(2) after the human body is identified, the method solves the influence of visual angle change in action identification by calculating the relative position of the human body joint point relative to the human body coordinate system;
(3) in the actual test, the method reduces the error and improves the accuracy of action identification by accumulating the variable quantity between two adjacent frames;
(4) the stereo camera used by the invention makes up the defect of the loss of space information of the traditional monocular camera, has more and more extensive application due to the price advantage compared with a binocular camera and a TOF camera, and improves the identification precision and capability of the skeleton.
Drawings
Fig. 1 is a flowchart of a skeleton information action feature extraction method based on a stereo camera according to the present application;
FIG. 2 is a schematic diagram of a human coordinate system;
fig. 3 is a schematic diagram illustrating the transformation from point a to point B of the human body at time t.
Detailed Description
The invention is described in detail below with reference to the figures and specific embodiments.
As shown in fig. 1, a skeleton information action feature extraction method based on a stereo camera includes the following steps:
(1) acquiring coordinates of each joint point of the human body by using a stereo camera;
(2) and determining a human body coordinate system, wherein an origin O of the human body coordinate system is an origin at an intersection point of a connecting line of a left shoulder L and a right shoulder R of the human body and a human body symmetry axis, an x-axis of the human body coordinate system is a ray from the origin O to the left shoulder L, a y-axis of the human body coordinate system is a ray from the origin O to a human body center T, and a z-axis of the human body coordinate system is perpendicular to a plane where the x-axis and the y-axis are located and meets the right-hand rule, as shown in fig. 2.
(3) Calculating a first position matrix from a camera coordinate system to a human body coordinate system, wherein the first position matrix comprises a rotating part matrix and a translating part matrix; let the coordinate of the left shoulder L of the human body acquired by the stereo camera be (CXL,CYL,CZL) The coordinate of the right shoulder R of the human body is (CXR,CYR,CZR) The coordinate of the center T of the human body is (CXT,CYT,CZT) The rotating parts are in matrix ofThen the first position matrix P O = R C O P O + P O C O R G , Wherein,OPCORGin order to translate a portion of the matrix,OPCORG=[COx,COy,COz]T O c x = X C L + X C R 2 , O c y = Y C L + Y C R 2 , O c z = Z C L + Z C R + Z C T 3 .
rotating parts are in matrix ofThe calculation is carried out according to the formula (1),
d i a g | R L → | | T O → | | R L → × R T → | = R C O R L → T O → O Z → - - - ( 1 )
wherein,representing the vector from the right shoulder to the left shoulder of the human body,a vector representing the center of the human body to the origin of the human body coordinate system,represents the vector from the right shoulder of the human body to the center of the human body,
(4) since the rotating part matrices calculated from the acquired point cloud data and obtained based on the three left shoulders, right shoulders and body center are not necessarily orthogonal, it is necessary to orthogonalize the rotating part matrices to obtain a second position matrix, which is a position matrix with a high accuracyComprises the following steps:
R C O ′ = U T · d i a g λ 1 - 1 2 λ 2 - 1 2 λ 3 - 1 2 · U · R C O
wherein,for a rotating partial matrix, U and λ satisfy R C O · R C O T = U T · d i a g λ 1 λ 2 λ 3 · U .
(5) The coordinates of each joint point of the human body are converted into a human body coordinate system through coordinate transformation to obtain the relative position of each joint point of the human body relative to the human body coordinate system, and the relative position is used as a part of the required characteristics of action identification, so that the influence of visual angle change in the action identification is solved; however, for the motion sequence with the indistinct four-limb movement, the change of the human body coordinate system in the motion needs to be added for distinguishing the motion such as rotation, jumping and the like.
(6) Because human motion equivalence becomes a rigid body, the motion of human motion is comparatively gentle between two adjacent frames, can not take place the sudden change, consequently carries out the smoothing to the rotational translation matrix between two adjacent frames that obtain, and the smoothing process specifically is: and performing arithmetic average on the obtained rotation and translation matrixes between two adjacent frames. After the smoothing treatment, accumulating the variation between two adjacent needles to obtain the translation variation and the rotation variation of the human body coordinate system from the beginning of the action to the current moment;
assuming that the human body is located at point A at time t, the human body is located at point B at time t +1, and α, β and gamma are the rotation angles of the human body around the x-axis, the y-axis and the z-axis, respectively, as shown in FIG. 3, the rotation matrix is rotated R A B = r 11 r 12 r 13 r 21 r 22 r 23 r 31 r 32 r 33 , The Euler angle can be obtained by calculation α = tan - 1 ( r 21 r 11 ) , β = tan - 1 ( - r 31 r 11 2 + r 21 2 ) , γ = tan - 1 ( r 32 r 33 ) , The obtained variation is a change between frames and is small, and in the action recognition, the variation needs to be accumulated to calculate the translation and rotation changes of the human coordinate system from the action start to the current time.
(7) And (5) taking the relative position in the step (5) and the translation variation and the rotation variation in the step (6) as the feature vector of the current moment.

Claims (8)

1. A skeleton information action feature extraction method based on a stereo camera is characterized by comprising the following steps:
(1) acquiring coordinates of each joint point of the human body by using a stereo camera;
(2) determining a human body coordinate system;
(3) calculating a first position matrix from a camera coordinate system to a human body coordinate system, wherein the first position matrix comprises a rotating part matrix and a translating part matrix;
(4) calculating a second position matrix from each joint point of the human body to a camera coordinate system;
(5) calculating the relative position of each joint point of the human body relative to a human body coordinate system;
(6) smoothing the rotation translation matrix between two adjacent frames, and accumulating the variation between two adjacent needles to obtain the translation variation and the rotation variation of the human body coordinate system from the beginning of the action to the current moment;
(7) and generating a feature vector of the current moment.
2. The method as claimed in claim 1, wherein an origin O of the human body coordinate system is an intersection point of a connecting line of a left shoulder L and a right shoulder R of the human body and a human body symmetry axis, an x-axis of the human body coordinate system is a ray from the origin O to the left shoulder L, a y-axis of the human body coordinate system is a ray from the origin O to a human body center T, and a z-axis of the human body coordinate system is perpendicular to a plane where the x-axis and the y-axis are located and satisfies a right-hand rule.
3. The method for extracting skeleton information action features based on a stereo camera according to claim 1, wherein the step (3) specifically comprises: let the coordinate of the left shoulder L of the human body acquired by the stereo camera be (CXL,CYL,CZL) The coordinate of the right shoulder R of the human body is (CXR,CYR,CZR) The coordinate of the center T of the human body is (CXT,CYT,CZT) The rotating parts are in matrix ofSaid first position matrixWherein,OPCORGin order to translate a portion of the matrix,OPCORG=[COx,COy,COz]T Cand P is the coordinate of each joint point of the human body in the human body coordinate system.
4. The method according to claim 1, wherein the matrix of rotating parts isThe calculation is carried out according to the formula (1),
d i a g | R L → | | T O → | | R L → × R T → | = R C O R L → T O → O Z → - - - ( 1 )
wherein,representing the vector from the right shoulder to the left shoulder of the human body,a vector representing the center of the human body to the origin of the human body coordinate system,represents the vector from the right shoulder of the human body to the center of the human body,
5. the method for extracting skeleton information action features based on a stereo camera according to claim 1, wherein the step (4) specifically comprises: orthogonalizing the rotating part matrix in the first position matrix to obtain a second position matrixComprises the following steps:
R C O ′ = U T · d i a g λ 1 - 1 2 λ 2 - 1 2 λ 3 - 1 2 · U · R C O
wherein,for a rotating partial matrix, U and λ satisfy R C O · R C O T = U T · d i a g λ 1 λ 2 λ 3 · U .
6. The method for extracting skeleton information action features based on a stereo camera according to claim 1, wherein the step (5) specifically comprises: and converting the coordinates of each joint point of the human body into a human body coordinate system through coordinate transformation to obtain the relative position of each joint point of the human body with respect to the human body coordinate system.
7. The method for extracting skeleton information action features based on a stereo camera according to claim 1, wherein the smoothing process specifically comprises: and performing arithmetic average on the obtained rotation and translation matrixes between two adjacent frames.
8. The method for extracting skeleton information action features based on a stereo camera according to claim 1, wherein the step (7) specifically comprises: and (5) taking the relative position in the step (5) and the translation variation and the rotation variation in the step (6) as the feature vector of the current moment.
CN201610047866.7A 2016-01-25 2016-01-25 A kind of framework information motion characteristic extracting method based on stereoscopic camera Active CN105550667B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201610047866.7A CN105550667B (en) 2016-01-25 2016-01-25 A kind of framework information motion characteristic extracting method based on stereoscopic camera

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201610047866.7A CN105550667B (en) 2016-01-25 2016-01-25 A kind of framework information motion characteristic extracting method based on stereoscopic camera

Publications (2)

Publication Number Publication Date
CN105550667A true CN105550667A (en) 2016-05-04
CN105550667B CN105550667B (en) 2019-01-25

Family

ID=55829850

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201610047866.7A Active CN105550667B (en) 2016-01-25 2016-01-25 A kind of framework information motion characteristic extracting method based on stereoscopic camera

Country Status (1)

Country Link
CN (1) CN105550667B (en)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109271845A (en) * 2018-07-31 2019-01-25 浙江理工大学 Human action analysis and evaluation methods based on computer vision
WO2019196313A1 (en) * 2018-04-10 2019-10-17 平安科技(深圳)有限公司 Robot walking obstacle detection method and apparatus, computer device, and storage medium
CN110567135A (en) * 2019-10-08 2019-12-13 珠海格力电器股份有限公司 air conditioner control method and device, storage medium and household equipment

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20110279368A1 (en) * 2010-05-12 2011-11-17 Microsoft Corporation Inferring user intent to engage a motion capture system
CN103544496A (en) * 2012-07-12 2014-01-29 同济大学 Method for recognizing robot scenes on basis of space and time information fusion
CN103577793A (en) * 2012-07-27 2014-02-12 中兴通讯股份有限公司 Gesture recognition method and device

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20110279368A1 (en) * 2010-05-12 2011-11-17 Microsoft Corporation Inferring user intent to engage a motion capture system
CN103544496A (en) * 2012-07-12 2014-01-29 同济大学 Method for recognizing robot scenes on basis of space and time information fusion
CN103577793A (en) * 2012-07-27 2014-02-12 中兴通讯股份有限公司 Gesture recognition method and device

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
XINXIAO WU 等: "Action recognition feedback-based framework for human pose reconstruction from monocular images", 《PATTERN RECOGNITION LETTERS》 *
杜鹃 等: "基于共形几何代数的三维人体骨架运动序列", 《计算机工程与设计》 *

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2019196313A1 (en) * 2018-04-10 2019-10-17 平安科技(深圳)有限公司 Robot walking obstacle detection method and apparatus, computer device, and storage medium
CN109271845A (en) * 2018-07-31 2019-01-25 浙江理工大学 Human action analysis and evaluation methods based on computer vision
CN110567135A (en) * 2019-10-08 2019-12-13 珠海格力电器股份有限公司 air conditioner control method and device, storage medium and household equipment

Also Published As

Publication number Publication date
CN105550667B (en) 2019-01-25

Similar Documents

Publication Publication Date Title
Tang et al. Real-time detection of surface deformation and strain in recycled aggregate concrete-filled steel tubular columns via four-ocular vision
CN111432733B (en) Apparatus and method for determining motion of an ultrasound probe
CN102506757B (en) Self-positioning method of binocular stereo measuring system in multiple-visual angle measurement
CN105205858B (en) A kind of indoor scene three-dimensional rebuilding method based on single deep vision sensor
CN103424112B (en) A kind of motion carrier vision navigation method auxiliary based on laser plane
CN104933718A (en) Physical coordinate positioning method based on binocular vision
CN110617814A (en) Monocular vision and inertial sensor integrated remote distance measuring system and method
CN110044374B (en) Image feature-based monocular vision mileage measurement method and odometer
JP5977214B2 (en) Image processing method, apparatus, and program
CN104677330A (en) Small binocular stereoscopic vision ranging system
CN103528571A (en) Monocular stereo vision relative position/pose measuring method
CN105395196B (en) Positioning device and method of EEG the or MEG electrodes in brain MR images
CN107796370B (en) Method and device for acquiring conversion parameters and mobile mapping system
CN105550667B (en) A kind of framework information motion characteristic extracting method based on stereoscopic camera
CN107374638A (en) A kind of height measuring system and method based on binocular vision module
CN204070693U (en) A kind of three-dimensional human body measurement device of structure based light technology
CN106056121A (en) Satellite assembly workpiece fast-identification method based on SIFT image feature matching
CN105277144A (en) Land area rapid detection method based on binocular vision and detection device thereof
Horanyi et al. Multiview absolute pose using 3D-2D perspective line correspondences and vertical direction
CN116071424A (en) Fruit space coordinate positioning method based on monocular vision
Zhou et al. Point cloud registration for agriculture and forestry crops based on calibration balls using Kinect V2
CN112950527B (en) Stereo matching morphology measurement method based on limited geometric association constraint
CN106251359A (en) Based on Clifford algebraic geometry relative to the 3D rendering method for registering of invariant
JP2015058246A (en) Sight line analysis system
WO2014106747A1 (en) Methods and apparatus for image processing

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant
TR01 Transfer of patent right
TR01 Transfer of patent right

Effective date of registration: 20210225

Address after: Room 306, 30 / F, Tiancheng Times Business Plaza, 58 qinglonggang Road, high speed rail new town, Xiangcheng District, Suzhou City, Jiangsu Province, 215000

Patentee after: Suzhou tongqi Artificial Intelligence Technology Co.,Ltd.

Address before: 200092 Siping Road 1239, Shanghai, Yangpu District

Patentee before: TONGJI University

PE01 Entry into force of the registration of the contract for pledge of patent right
PE01 Entry into force of the registration of the contract for pledge of patent right

Denomination of invention: A Method for Extracting Skeleton Information Action Features Based on Stereoscopic Cameras

Granted publication date: 20190125

Pledgee: Bank of Jiangsu Co.,Ltd. Suzhou Branch

Pledgor: Suzhou tongqi Artificial Intelligence Technology Co.,Ltd.

Registration number: Y2024980015790