CN116129467A - Method for identifying gesture and behavior of transformer operation and maintenance personnel and gesture of tool - Google Patents
Method for identifying gesture and behavior of transformer operation and maintenance personnel and gesture of tool Download PDFInfo
- Publication number
- CN116129467A CN116129467A CN202211645542.5A CN202211645542A CN116129467A CN 116129467 A CN116129467 A CN 116129467A CN 202211645542 A CN202211645542 A CN 202211645542A CN 116129467 A CN116129467 A CN 116129467A
- Authority
- CN
- China
- Prior art keywords
- human body
- tool
- camera
- gesture
- coordinate system
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
- 238000012423 maintenance Methods 0.000 title claims abstract description 55
- 238000000034 method Methods 0.000 title claims abstract description 35
- 238000003066 decision tree Methods 0.000 claims abstract description 33
- 230000009471 action Effects 0.000 claims abstract description 13
- 230000036544 posture Effects 0.000 claims description 60
- 238000012795 verification Methods 0.000 claims description 28
- 230000009466 transformation Effects 0.000 claims description 25
- 238000012549 training Methods 0.000 claims description 24
- 238000011156 evaluation Methods 0.000 claims description 21
- 239000011159 matrix material Substances 0.000 claims description 20
- 210000001503 joint Anatomy 0.000 claims description 15
- 210000000689 upper leg Anatomy 0.000 claims description 8
- 210000000544 articulatio talocruralis Anatomy 0.000 claims description 4
- 210000002310 elbow joint Anatomy 0.000 claims description 4
- 210000000629 knee joint Anatomy 0.000 claims description 4
- 210000002414 leg Anatomy 0.000 claims description 4
- 238000005070 sampling Methods 0.000 claims description 4
- 210000000323 shoulder joint Anatomy 0.000 claims description 4
- 210000003857 wrist joint Anatomy 0.000 claims description 4
- 238000006243 chemical reaction Methods 0.000 claims description 3
- 230000008569 process Effects 0.000 claims description 3
- 238000012935 Averaging Methods 0.000 claims description 2
- 238000001914 filtration Methods 0.000 claims description 2
- 238000003384 imaging method Methods 0.000 claims description 2
- 238000009434 installation Methods 0.000 claims description 2
- 238000002372 labelling Methods 0.000 claims description 2
- 238000010200 validation analysis Methods 0.000 claims description 2
- 230000004044 response Effects 0.000 abstract description 2
- 238000010586 diagram Methods 0.000 description 4
- 210000003423 ankle Anatomy 0.000 description 3
- 238000012544 monitoring process Methods 0.000 description 3
- 238000012545 processing Methods 0.000 description 3
- 230000011218 segmentation Effects 0.000 description 3
- 238000004088 simulation Methods 0.000 description 3
- 210000000707 wrist Anatomy 0.000 description 3
- 238000004519 manufacturing process Methods 0.000 description 2
- 230000019771 cognition Effects 0.000 description 1
- 238000013527 convolutional neural network Methods 0.000 description 1
- 210000003414 extremity Anatomy 0.000 description 1
- 230000006870 function Effects 0.000 description 1
- 230000006872 improvement Effects 0.000 description 1
- 210000003127 knee Anatomy 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 230000003287 optical effect Effects 0.000 description 1
- 230000000717 retained effect Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/103—Static body considered as a whole, e.g. static pedestrian or occupant recognition
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F17/00—Digital computing or data processing equipment or methods, specially adapted for specific functions
- G06F17/10—Complex mathematical operations
- G06F17/16—Matrix or vector computation, e.g. matrix-matrix or matrix-vector multiplication, matrix factorization
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/20—Image preprocessing
- G06V10/25—Determination of region of interest [ROI] or a volume of interest [VOI]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/40—Extraction of image or video features
- G06V10/56—Extraction of image or video features relating to colour
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/764—Arrangements for image or video recognition or understanding using pattern recognition or machine learning using classification, e.g. of video objects
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/77—Processing image or video features in feature spaces; using data integration or data reduction, e.g. principal component analysis [PCA] or independent component analysis [ICA] or self-organising maps [SOM]; Blind source separation
- G06V10/774—Generating sets of training patterns; Bootstrap methods, e.g. bagging or boosting
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y04—INFORMATION OR COMMUNICATION TECHNOLOGIES HAVING AN IMPACT ON OTHER TECHNOLOGY AREAS
- Y04S—SYSTEMS INTEGRATING TECHNOLOGIES RELATED TO POWER NETWORK OPERATION, COMMUNICATION OR INFORMATION TECHNOLOGIES FOR IMPROVING THE ELECTRICAL POWER GENERATION, TRANSMISSION, DISTRIBUTION, MANAGEMENT OR USAGE, i.e. SMART GRIDS
- Y04S10/00—Systems supporting electrical power generation, transmission or distribution
- Y04S10/50—Systems or methods supporting the power network operation or management, involving a certain degree of interaction with the load-side end user applications
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Multimedia (AREA)
- Computing Systems (AREA)
- Mathematical Physics (AREA)
- Software Systems (AREA)
- Databases & Information Systems (AREA)
- Health & Medical Sciences (AREA)
- Mathematical Analysis (AREA)
- Evolutionary Computation (AREA)
- General Health & Medical Sciences (AREA)
- Medical Informatics (AREA)
- Artificial Intelligence (AREA)
- Computational Mathematics (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Mathematical Optimization (AREA)
- Data Mining & Analysis (AREA)
- Pure & Applied Mathematics (AREA)
- Algebra (AREA)
- Human Computer Interaction (AREA)
- General Engineering & Computer Science (AREA)
- Image Analysis (AREA)
Abstract
The invention provides a method for identifying the gesture and behavior of a transformer operation and maintenance person and the gesture of a tool, which comprises the following steps: pre-storing standard action gestures of personnel and standard shapes of tools and instruments; an RGB camera and an infrared camera are adopted to collect human body images and scene images; acquiring two-dimensional coordinates of a joint point of a human body by adopting an openpost model, and further determining three-dimensional coordinates of the joint point; constructing a cart decision tree; and identifying human behavior types by the cart decision tree, and identifying the gesture of the substation operation and maintenance personnel. Dividing and converting the scene image to obtain tool feature data; identifying a possible two-dimensional coordinate range of the tool by adopting a unet network model; and identifying the tool gesture according to the tool shape feature, the tool position feature, the tool infrared image and the tool color image. The method for identifying the gesture and the behavior of the transformer operation and maintenance personnel and the gesture of the tool has the characteristics of high precision, quick response and the like, and can be widely applied to the power industry.
Description
Technical Field
The invention relates to a recognition technology, in particular to a recognition method for the gesture and behavior of a transformer operation and maintenance person and the gesture of a tool.
Background
The electric power system as a national core industry plays a significant role in the national economy production and life. In order to ensure the normal work of the power system, the power transformation operation and maintenance work is difficult and the task amount is large. In actual power transformation operation and maintenance work, the problems that importance cognition of partial staff on the work of the staff is insufficient, the operation flow is not standard, the staff training requirement is increased due to power transformation equipment updating iteration and the like exist. Personnel training is an important measure to solve the above problems.
The personnel training mainly adopts theoretical teaching, master teaching, simulation operation and other modes, and has the problems of single theoretical learning form, hard back of the dead, careless operation and the like; moreover, the master belt freehand is that the master carries out skill teaching according to rules and experience, is difficult to standardize and has a long training period; the simulation operation training requires a large amount of resources to manufacture simulation equipment, and has poor expansibility. How to monitor the standard operation degree of the substation operation and maintenance personnel in an intelligent mode is a blank in the field of the substation operation and maintenance at present.
Therefore, in the prior art, an intelligent method for monitoring the standard operation degree of the transformer operation and maintenance personnel is not available.
Disclosure of Invention
Therefore, the main purpose of the invention is to provide a method for identifying the gesture and behavior of the transformer operation and maintenance personnel and the gesture of the tools and instruments, which has high precision, quick response and convenient use.
In order to achieve the above purpose, the technical scheme provided by the invention is as follows:
a method for identifying the gesture and behavior of a transformer operation and maintenance person and the gesture of a tool comprises the following steps:
and step 1, pre-storing standard action postures of power transformation operation and maintenance personnel and standard appearance of tools.
Step 2, respectively acquiring human body infrared images and human body color images related to the postures of the transformer operation and maintenance personnel by adopting an RGB camera and an infrared camera; simultaneously, respectively acquiring an infrared image and a color image of a scene with tools.
Step 3, acquiring a two-dimensional human body joint point by adopting an openpore model; and (3) obtaining the human body joint point in a three-dimensional form according to the human body joint point in a two-dimensional form, the human body infrared image and the human body color image acquired in the step (2).
And 4, extracting human body posture characteristic data consisting of angles of all joints and distances among all joint points according to the human body joint points in the three-dimensional form obtained in the step 3.
Step 5, constructing a cart decision tree according to the human body posture characteristic data; and identifying human behavior types by the cart decision tree, and identifying the gesture of the substation operation and maintenance personnel.
And 6, segmenting and converting the scene infrared image and the scene color image with the tool acquired in the step 2 to obtain the characteristic data of the tools.
And 7, dividing the tool feature data into a tool feature training set and a tool feature verification set.
Step 8, training the unet network model by adopting a tool feature training set; and verifying the trained unet network model by adopting the tool feature verification set to obtain a trained unet model.
And 9, identifying a possible two-dimensional coordinate range of the tool by the trained unet model.
And 10, according to the shape characteristics and the position characteristics of the tool, combining the tool infrared image and the tool color image obtained in the step 5, and identifying the tool posture.
In summary, according to the method for recognizing the gesture and behavior of the transformer operation and maintenance personnel and the gesture of the tool, the infrared camera and the RGB camera are adopted to collect the infrared image and the color image of the human body, and the infrared image and the color image of the scene with tools. For a human body color image, acquiring two-dimensional coordinates of joint points related to the human body posture by adopting an openpost model, and further determining three-dimensional coordinates of the joint points related to the human body posture according to the human body infrared image. And determining the angle of the joint point and the distance between the joint points according to the three-dimensional coordinates of the joint point of the human body posture, and constructing a cart decision tree. Human body gestures are identified by the cart decision tree. Therefore, the gesture recognition method for the transformer operation and maintenance personnel can remotely recognize the real-time working behavior and working state of the transformer operation and maintenance personnel, obtain first hand data of the first line operation of the transformer operation and maintenance personnel, and realize accurate and rapid recognition on standard training of the personnel and monitoring of the standard degree of the working behavior of the personnel. In addition, the invention also carries out segmentation processing on the scene color image with the tools and adopts the unet network model to identify the possible two-dimensional coordinate range of the tools; and combining the scene infrared image with the tools, the tool shape characteristics and the tool position characteristics to identify the gesture of the tools. The gesture of tool can further confirm the actual operation action of transformer fortune dimension personnel, further accurate, quick, pertinence personnel training provides powerful basis.
Drawings
Fig. 1 is a general flow diagram of a method for recognizing the gesture and behavior of a transformer operation and maintenance person and the gesture of a tool according to the present invention.
Fig. 2 is a schematic structural diagram of a human body joint according to the present invention.
Fig. 3 is an angular schematic view of a human body joint according to the present invention.
Fig. 4 is a schematic view of the distances between the joints according to the present invention.
Detailed Description
The present invention will be described in further detail with reference to the drawings and the embodiments, in order to make the objects, technical solutions and advantages of the present invention more apparent.
Fig. 1 is a general flow diagram of a method for recognizing the gesture and behavior of a transformer operation and maintenance person and the gesture of a tool according to the present invention. As shown in fig. 1, the method for identifying the gesture and behavior of the transformer operation and maintenance personnel and the gesture of the tool comprises the following steps:
and step 1, pre-storing standard action postures of power transformation operation and maintenance personnel and standard appearance of tools.
Step 2, respectively acquiring human body infrared images and human body color images related to the postures of the transformer operation and maintenance personnel by adopting an RGB camera and an infrared camera; simultaneously, respectively acquiring an infrared image and a color image of a scene with tools.
Step 3, acquiring a two-dimensional human body joint point by adopting an openpore model; and (3) obtaining the human body joint point in a three-dimensional form according to the human body joint point in a two-dimensional form, the human body infrared image and the human body color image acquired in the step (2).
And 4, extracting human body posture characteristic data consisting of angles of all joints and distances among all joint points according to the human body joint points in the three-dimensional form obtained in the step 3.
Step 5, constructing a cart decision tree according to the human body posture characteristic data; and identifying human behavior types by the cart decision tree, and identifying the gesture of the substation operation and maintenance personnel.
And 6, segmenting and converting the scene infrared image and the scene color image with the tool acquired in the step 2 to obtain the characteristic data of the tools.
And 7, dividing the tool feature data into a tool feature training set and a tool feature verification set.
Step 8, training the unet network model by adopting a tool feature training set; and verifying the trained unet network model by adopting the tool feature verification set to obtain a trained unet model.
And 9, identifying a possible two-dimensional coordinate range of the tool by the trained unet model.
And 10, according to the shape characteristics and the position characteristics of the tool, combining the tool infrared image and the tool color image obtained in the step 5, and identifying the tool posture.
In a word, the method for identifying the gesture and the behavior of the transformer operation and maintenance personnel and the gesture of the tool adopts the infrared camera and the RGB camera to collect the infrared image and the color image of the human body, and the infrared image and the color image of the scene with tools. For a human body color image, acquiring two-dimensional coordinates of joint points related to the human body posture by adopting an openpost model, and further determining three-dimensional coordinates of the joint points related to the human body posture according to the human body infrared image. And determining the angle of the joint point and the distance between the joint points according to the three-dimensional coordinates of the joint point of the human body posture, and constructing a cart decision tree. Human body gestures are identified by the cart decision tree. Therefore, the gesture recognition method for the transformer operation and maintenance personnel can remotely recognize the real-time working behavior and working state of the transformer operation and maintenance personnel, obtain first hand data of the first line operation of the transformer operation and maintenance personnel, and realize accurate and rapid recognition on standard training of the personnel and monitoring of the standard degree of the working behavior of the personnel. In addition, the invention also carries out segmentation processing on the scene color image with the tools and adopts the unet network model to identify the possible two-dimensional coordinate range of the tools; and combining the scene infrared image with the tools, the tool shape characteristics and the tool position characteristics to identify the gesture of the tools. The gesture of tool can further confirm the actual operation action of transformer fortune dimension personnel, further accurate, quick, pertinence personnel training provides powerful basis.
In the invention, after the step 1 and before the step 2, the method further comprises the following steps:
and A1, calibrating the RGB camera and the infrared camera to obtain the internal parameters of the RGB camera participating in the infrared camera.
And A2, acquiring an external parameter matrix and a pose transformation matrix between the RGB camera and the infrared camera according to the step A1.
In the present invention, the step A1 includes the steps of:
step A11, according to the imaging principle, the setting of a physical coordinate system and the setting of a pixel coordinate system, obtaining the relationship between the physical coordinate system and the pixel coordinate system of the RGB camera or the infrared camera, wherein the relationship is as follows:
wherein ,(XR ,Y R ) Representing coordinates of any pixel point in the image in the physical coordinate system of the RGB camera, (u) R ,v R ) Representing coordinates of any pixel point in the image in the pixel coordinate system of the RGB camera, (x) R ,y R ,z R ) Representing the coordinates of any pixel point in the image in the camera coordinate system of the RGB camera, d Rx 、d Ry Respectively representing x of any pixel point in the image in a camera coordinate system of the RGB camera R Axis, y R Actual dimension in axial direction s Rx 、s Ry Respectively representing x of any pixel point in the image in a camera coordinate system of the RGB camera R Axis, y R Sampling frequency in axial direction, (X) R ,Y R )、(u R ,v R )、(x R ,y R ,z R ) The two have one-to-one correspondence; physical coordinate system center point (X) of RGB camera R0 ,Y R0 ) The corresponding coordinates in the pixel coordinate system are (u) R0 ,v R0 );(X h ,Y h ) Representing coordinates of any pixel point in the image in a physical coordinate system of the infrared camera, (u) h ,v h ) Representing coordinates of any pixel point in the image in a pixel coordinate system of the infrared camera, (x) h ,y h ,z h ) Representing the coordinates of any pixel point in the image in the camera coordinate system of the infrared camera, d hx 、d hy Respectively representing x of any pixel point in the image in a camera coordinate system of the infrared camera h Axis, y h Actual dimension in axial direction s hx 、s hy Respectively representing x of any pixel point in the image in a camera coordinate system of the infrared camera h Axis, y h Sampling frequency in axial direction, (X) h ,Y h )、(u h ,v h )、(x h ,y h ,z h ) The two have one-to-one correspondence; physical coordinate system center point (X) of infrared camera h0 ,Y h0 ) The corresponding coordinates in the pixel coordinate system are (u) h0 ,v h0 )。
Step A12, respectively obtaining the relation between the image physical coordinate system of the RGB camera and the image camera coordinate system and the relation between the image physical coordinate system of the infrared camera and the image camera coordinate system, wherein the relation is as follows:
wherein ,fR Focal length f of RGB camera h Is the focal length of the camera of the infrared camera.
Step A13, obtaining an internal reference matrix M of the RGB camera according to the step A11 and the step A12 and the expression mode of the homogeneous equation R Internal reference matrix M of infrared camera h :
wherein ,fRx 、f Ry 、u R0 、v R0 Is an internal reference of RGB camera, f hx 、f hy 、u h0 、v h0 Is an internal reference of an infrared camera; f (f) Rx =f R /dx R Representing RGB camera x R Focal length in axial direction, f Ry =f R /dy R Representing RGB camera y R Focal length in axial direction, f hx Representing an infrared camera x h Focal length in axial direction, f hy Representing an infrared camera y h Focal length in the axial direction.
In the present invention, the step A2 includes the following steps:
step A21, determining the relation between a camera coordinate system of the infrared camera and a camera coordinate system of the RGB camera according to the installation positions of the infrared camera and the RGB camera:
wherein ,TRh Representing the extrinsic matrix between the infrared camera and the RGB camera.
Step A22, according to steps A11 to A13 and step A21, determining the relation between the pixel coordinate system of the RGB camera and the pixel coordinate system of the infrared camera, wherein the relation is as follows:
Step A23, z of RGB camera-based camera coordinate System R Z of camera coordinate system of axial and infrared camera h Axially parallel, take z R =z h The method comprises the steps of carrying out a first treatment on the surface of the Thereafter, the following is obtained:
In practical application, the pose transformation matrix H has the following functions: in the subsequent image processing or gesture recognition process, the method is used for aligning the color image shot by the RGB camera with the infrared image shot by the infrared camera so as to accurately acquire the relevant information corresponding to each pixel point in the image.
In practical application, three coordinate systems commonly used by cameras are an image coordinate system, a camera coordinate system and a world coordinate system. The world coordinate system is the basis for locating cameras and other objects in the space, and the origin position of the world coordinate system can be defined at any position according to actual needs. The origin of the camera coordinate system is the focus of the pinhole camera model, and the z-axis is the optical axis of the camera. The image coordinate system is specifically divided into an image pixel coordinate system and an image physical coordinate system, wherein the physical coordinate system is a two-dimensional rectangular coordinate system, and the x axis and the y axis of the physical coordinate system are parallel to the camera coordinate system. The pixel coordinate system is a two-dimensional coordinate system in the image, with the origin generally set to the upper left corner of the image. The transformation relationships among the world coordinate system, the camera coordinate system, the image coordinate system and the image physical coordinate system and the pixel coordinate system are the prior art, and are not repeated here.
In the present invention, the step 3 specifically includes the following steps:
and 31, identifying two-dimensional coordinates of all human body joints of the transformer operation and maintenance personnel, connection relations among all the human body joints and confidence degrees of all the human body joints by using an Openphase gesture estimation system for the human body color image.
In practical application, the openpost attitude estimation system U.S. kansui university (CMU) is based on convolutional neural network and open source library developed by taking caffe as framework, and is used for identifying limb language, and will not be described here again.
Step 32, for human body joint points with the confidence coefficient smaller than 0.2, searching for adjacent human body joint points with the confidence coefficient larger than 0.8; setting a virtual sphere by taking the adjacent human body joint point with the confidence coefficient smaller than 0.2 and the adjacent human body joint point with the confidence coefficient larger than 0.8 as a radius: if the human body joint point falls outside the virtual sphere, deleting the human body joint point; if the human body node falls within the virtual sphere, the human body node is retained.
Step 33, correcting and predicting the deleted human body joint points and the joint points lost due to shielding by adopting Kalman filtering; and then, obtaining the connection relation between the two-dimensional coordinates of the 18 human body joint points of the power transformation operation and maintenance personnel and each human body joint point in the 18 human body joint points. Fig. 2 is a schematic structural diagram of a human body joint according to the present invention. As shown in fig. 2, the 18 human body nodes include: nasal node N 1 Shoulder center joint N 2 Left shoulder joint point N 3 Left elbow joint N 4 Left wrist joint N 5 Right shoulder joint point N 6 Right elbow joint point N 7 Right wrist joint point N 8 Left crotch node N 9 Left knee joint N 10 Left ankle joint N 11 Right crotch joint point N 12 Right knee joint point N 13 Right ankle joint point N 14 Left eye joint N 15 Right eye joint N 16 Left ear joint point N 17 Joint point N of right ear 18 。
And step 34, determining three-dimensional coordinates of 18 human body joint points of the power transformation operation and maintenance personnel according to the human body infrared image conforming to the pose transformation matrix between the power transformation operation and maintenance personnel and the human body color image.
In the present invention, the step 4 specifically includes the following steps:
and step 41, obtaining 12 groups of joint angles through the rest 16 joints except for the left eye joint and the right eye joint in the 18 human joints. Fig. 3 is an angular schematic view of a human body joint according to the present invention. As shown in fig. 3, the 12 sets of joint angles include: left arm angle theta 1 Included angle theta of right arm 2 Included angle theta of left leg 3 Included angle θ of right leg 4 Included angle theta between left upper arm and left shoulder 5 Included angle θ between right upper arm and right shoulder 6 Included angle θ between left thigh and crotch 7 Included angle θ between right thigh and crotch 8 Included angle theta between left upper arm and spine 9 Included angle theta of right upper arm and spine 10 Included angle theta between left thigh and spine 11 Included angle of right thigh and spineθ 12 。
In practical application, the angle solving is the prior art, and the description is omitted here.
Step 42, distance from nose joint point to shoulder center joint pointAs the actual head height; wherein, (x) 1 ,y 1 ,z 1 ) Three-dimensional coordinates of the nasal joint point, (x) 2 ,y 2 ,z 2 ) Is the three-dimensional coordinate of the joint point of the center of the two shoulders.
Step 43, normalizing the actual head height to obtain the scaling ratioWhere d represents the standard head height and d=20 cm.
Step 44, determining the distance between any 2 joint points for the 18 human body joint points according to the scaling ratio k: wherein ,(xi ,y i ,z i ) Represents the N < th i Three-dimensional coordinates of the individual nodes, (x) i ,y i ,z i ) Represents the N < th j Three-dimensional coordinates of the individual nodes, i, j are natural numbers, and i=1, 2,..18, j=1, 2,..18, i+.j; n represents a human body joint point.
Fig. 4 is a schematic view of the distances between the joints according to the present invention. As shown in fig. 4, the distance between the joint points includes: head-to-neck distance L0, head-to-spine lower end distance L1, left shoulder-to-left ankle distance L2, right shoulder-to-right ankle distance L3, left and right double elbow distance L4, left and right double wrist distance L5, left and right double knee distance L6, and left and right double ankle distance L7.
Step 45, combining the joint angles determined in step 41 and the distances between the joint points determined in step 44 together as human body posture feature data.
In the present invention, the step 5 specifically includes the following:
and 51, dividing the human body posture feature data obtained in the step 45 into a human body posture feature training set and a human body posture feature verification set.
And 52, constructing a cart decision tree according to the human body posture feature training set.
Step 53, verifying and evaluating the cart decision tree constructed in the step 52 by adopting a human body posture feature verification set; and analyzing the human body posture feature data by adopting a cart decision tree meeting the evaluation result, identifying the human body posture and the action corresponding to the human body posture, comparing the action corresponding to the human body posture with the pre-stored standard action posture of the power transformation operation and maintenance personnel, and determining the posture and the behavior of the power transformation operation and maintenance personnel in the actual operation process.
In practical application, the art decision tree theory is the prior art, and will not be described here again.
In the present invention, in step 53, the verification and evaluation are performed on the cart decision tree constructed in step 52 by using the human body posture feature verification set, which specifically includes: randomly selecting 70% of the data from the human body posture feature data to form a human body posture feature training set, and the rest 30% of the data to form a human body posture feature verification set; the constructed cart decision tree of step 52 is validated and evaluated by the body posture feature validation set.
In the present invention, in step 53, the verification and evaluation are performed on the cart decision tree constructed in step 52 by using the human body posture feature verification set, which specifically includes:
step B1, dividing the human body posture characteristic data into m sub-data sets; wherein m is a natural number.
And B2, randomly selecting one sub-data set from the m sub-data sets as a human body posture feature verification set, and performing verification and evaluation on the cart decision tree constructed in the step 52 to obtain an evaluation value.
Step B3, repeating the step B2 until the rest m-1 sub-data sets are respectively and independently used as human body posture feature verification sets, and performing verification evaluation on the cart decision tree constructed in the step 52; (m-1) evaluation values are obtained.
And B4, averaging the m evaluation values obtained in the steps B2 to B3, wherein the average value is the evaluation result of the cart decision tree constructed in the step 52.
The two methods for verifying and evaluating the cart decision tree constructed in the step 52 are two parallel methods, and in practical application, one of the two methods can be selected according to the needs. In practical application, the verification and evaluation is the accuracy rate of motion recognition: when the action recognition accuracy, i.e., the evaluation value or the average evaluation value, is greater than 90%, the cart decision tree can be considered to satisfy the evaluation result.
In the present invention, the step 6 specifically includes the following steps:
and 61, labeling and dividing over 4000 scene infrared images with tools and scene color images by using labelme, and correspondingly obtaining tool infrared data and tool RGB data after file conversion.
In practical application, the related theory of labelme label segmentation and file conversion is the prior art, and is not repeated here.
Step 62, combining the tool infrared data with the tool RGB data as tool feature data.
In the present invention, the step 10 specifically includes the following steps:
step 101, calculating the nearest distance between each possible two-dimensional coordinate range and the left wrist articulation point or the right wrist articulation point for the possible two-dimensional coordinate ranges of the tools and instruments obtained in the step 9; comparing the nearest distances, wherein the two-dimensional coordinate range corresponding to the minimum value of the nearest distances is the actual coordinate range of the tool; and simultaneously, obtaining a tool color image.
And 102, comparing the actual coordinate range of the tool with a pre-stored standard shape of the tool to obtain two-dimensional coordinates of each point of the tool on the tool color image.
And 103, obtaining three-dimensional coordinates of each point of the tool according to the tool infrared image conforming to the pose transformation matrix between the tool color image and the tool color image, namely, recognizing the pose of the tool.
In summary, the above embodiments are only preferred embodiments of the present invention, and are not intended to limit the scope of the present invention. Any modification, equivalent replacement, improvement, etc. made within the spirit and principle of the present invention should be included in the protection scope of the present invention.
Claims (10)
1. The method for identifying the gesture and the behavior of the transformer operation and maintenance personnel and the gesture of the tool is characterized by comprising the following steps:
step 1, pre-storing standard action gestures and tool standard shapes of substation operation and maintenance personnel;
step 2, respectively acquiring human body infrared images and human body color images related to the postures of the transformer operation and maintenance personnel by adopting an RGB camera and an infrared camera; simultaneously, respectively acquiring a scene infrared image and a scene color image with tools;
step 3, acquiring a two-dimensional human body joint point by adopting an openpore model; obtaining a three-dimensional human body joint point according to the two-dimensional human body joint point, the human body infrared image and the human body color image acquired in the step 2;
step 4, extracting human body posture characteristic data formed by angles of all joints and distances among all joint points according to the human body joint points in the three-dimensional form obtained in the step 3;
step 5, constructing a cart decision tree according to the human body posture characteristic data; recognizing human behavior categories by the cart decision tree, and recognizing the gesture of a substation operation and maintenance person;
step 6, segmenting and converting the scene infrared image and the scene color image with the tool acquired in the step 2 to obtain tool feature data;
step 7, dividing the tool feature data into a tool feature training set and a tool feature verification set;
step 8, training the unet network model by adopting a tool feature training set; verifying the trained unet network model by adopting a tool feature verification set to obtain a trained unet model;
step 9, identifying a possible two-dimensional coordinate range of the tool by the trained unet model;
and 10, according to the shape characteristics and the position characteristics of the tool, combining the tool infrared image and the tool color image obtained in the step 5, and identifying the tool posture.
2. The method for identifying the gesture and behavior of the transformer operation and maintenance personnel and the gesture of the tool according to claim 1, wherein the steps after the step 1 and before the step 2 further comprise the following steps:
a1, calibrating an RGB camera and an infrared camera to obtain an internal parameter of the RGB camera participating in the infrared camera;
and A2, acquiring an external parameter matrix and a pose transformation matrix between the RGB camera and the infrared camera according to the step A1.
3. The method for recognizing the gesture and behavior of the transformer operation and maintenance personnel and the gesture of the tool according to claim 2, wherein the step A1 comprises the following steps:
step A11, according to the imaging principle, the setting of a physical coordinate system and the setting of a pixel coordinate system, obtaining the relationship between the physical coordinate system and the pixel coordinate system of the RGB camera or the infrared camera, wherein the relationship is as follows:
wherein ,(XR ,Y R ) Representing coordinates of any pixel point in the image in the physical coordinate system of the RGB camera, (u) R ,v R ) Representing coordinates of any pixel point in the image in the pixel coordinate system of the RGB camera, (x) R ,y R ,z R ) Representing the coordinates of any pixel point in the image in the camera coordinate system of the RGB camera, d Rx 、d Ry Respectively representing x of any pixel point in the image in a camera coordinate system of the RGB camera R Axis, y R Actual dimension in axial direction s Rx 、s Ry Respectively representing any pixel point in an image in a camera coordinate system of an RGB cameraX in the middle R Axis, y R Sampling frequency in axial direction, (X) R ,Y R )、(u R ,v R )、(x R ,y R ,z R ) The two have one-to-one correspondence; physical coordinate system center point (X) of RGB camera R0 ,Y R0 ) The corresponding coordinates in the pixel coordinate system are (u) R0 ,v R0 );(X h ,Y h ) Representing coordinates of any pixel point in the image in a physical coordinate system of the infrared camera, (u) h ,v h ) Representing coordinates of any pixel point in the image in a pixel coordinate system of the infrared camera, (x) h ,y h ,z h ) Representing the coordinates of any pixel point in the image in the camera coordinate system of the infrared camera, d hx 、d hy Respectively representing x of any pixel point in the image in a camera coordinate system of the infrared camera h Axis, y h Actual dimension in axial direction s hx 、s hy Respectively representing x of any pixel point in the image in a camera coordinate system of the infrared camera h Axis, y h Sampling frequency in axial direction, (X) h ,Y h )、(u h ,v h )、(x h ,y h ,z h ) The two have one-to-one correspondence; physical coordinate system center point (X) of infrared camera h0 ,Y h0 ) The corresponding coordinates in the pixel coordinate system are (u) h0 ,v h0 );
Step A12, respectively obtaining the relation between the image physical coordinate system of the RGB camera and the image camera coordinate system and the relation between the image physical coordinate system of the infrared camera and the image camera coordinate system, wherein the relation is as follows:
wherein ,fR Focal length f of RGB camera h The focal length of a camera of the infrared camera;
step A13, obtaining an internal reference matrix M of the RGB camera according to the step A11 and the step A12 and the expression mode of the homogeneous equation R Inner part of infrared cameraParameter matrix M h :
wherein ,fRx 、f Ry 、u R0 、v R0 Is an internal reference of RGB camera, f hx 、f hy 、u h0 、v h0 Is an internal reference of an infrared camera; f (f) Rx =f R /dx R Representing RGB camera x R Focal length in axial direction, f Ry =f R /d y R Representing RGB camera y R Focal length in axial direction, f hx Representing an infrared camera x h Focal length in axial direction, f hy Representing an infrared camera y h Focal length in the axial direction.
4. The method for recognizing gesture and behavior of power transformation operation and maintenance personnel and gesture of tool according to claim 3, wherein the step A2 comprises the following steps:
step A21, determining the relation between a camera coordinate system of the infrared camera and a camera coordinate system of the RGB camera according to the installation positions of the infrared camera and the RGB camera:
wherein ,TRh An external parameter matrix representing a distance between the infrared camera and the RGB camera;
step A22, according to steps A11 to A13 and step A21, determining the relation between the pixel coordinate system of the RGB camera and the pixel coordinate system of the infrared camera, wherein the relation is as follows:
wherein ,indicating infrared cameraIs an internal reference matrix M of (2) h An inverse matrix of (a);
step A23, z of RGB camera-based camera coordinate System R Z of camera coordinate system of axial and infrared camera h Axially parallel, take z R =z h The method comprises the steps of carrying out a first treatment on the surface of the Thereafter, the following is obtained:
5. The method for recognizing the gesture and behavior of the transformer operation and maintenance personnel and the gesture of the tool according to claim 4, wherein the step 3 specifically comprises the following steps:
step 31, identifying two-dimensional coordinates of all human body joints of a transformer operation and maintenance person, connection relations among all human body joints and confidence degrees of all human body joints by using an Openphase gesture estimation system for a human body color image;
step 32, for human body joint points with the confidence coefficient smaller than 0.2, searching for adjacent human body joint points with the confidence coefficient larger than 0.8; setting a virtual sphere by taking the adjacent human body joint point with the confidence coefficient smaller than 0.2 and the adjacent human body joint point with the confidence coefficient larger than 0.8 as a radius: if the human body joint point falls outside the virtual sphere, deleting the human body joint point; if the human body joint point falls in the virtual sphere, the human body joint point is reserved;
step 33, correcting and predicting the deleted human body joint points and the joint points lost due to shielding by adopting Kalman filtering; then, obtaining the connection relation between the two-dimensional coordinates of the 18 human body joint points of the power transformation operation and maintenance personnel and each human body joint point in the 18 human body joint points; the 18 human body articulation points include: a nose joint point (N1), a shoulder center joint point (N2), a left shoulder joint point, a left elbow joint point, a left wrist joint point, a right shoulder joint point, a right elbow joint point, a right wrist joint point, a left crotch joint point, a left knee joint point, a left ankle joint point, a right crotch joint point, a right knee joint point, a right ankle joint point, a left eye joint point, a right eye joint point, a left ear joint point and a right ear joint point;
and step 34, determining three-dimensional coordinates of 18 human body joint points of the power transformation operation and maintenance personnel according to the human body infrared image conforming to the pose transformation matrix between the power transformation operation and maintenance personnel and the human body color image.
6. The method for recognizing the gesture and behavior of the transformer operation and maintenance personnel and the gesture of the tool according to claim 4, wherein the step 4 specifically comprises the following steps:
step 41, obtaining 12 groups of joint angles through the rest 16 joints except for the left eye joint and the right eye joint in the 18 human joints: left arm angle (theta) 1 ) Included angle of right arm (theta) 2 ) Included angle of left leg (theta) 3 ) Included angle of right leg (theta) 4 ) Included angle (theta) between left upper arm and left shoulder 5 ) Included angle (theta) between right upper arm and right shoulder 6 ) Included angle between left thigh and crotch (θ) 7 ) Included angle between right thigh and crotch (θ) 8 ) Included angle (θ) between upper left arm and spine 9 ) Included angle (θ) between right upper arm and spine 10 ) Included angle (θ) between left thigh and spine 11 ) Included angle between right thigh and spine (θ) 12 );
Step 42, distance from the nose joint point to the central joint point of the shoulders As the actual head height; wherein, (x) 1 ,y 1 ,z 1 ) Three-dimensional coordinates of the nasal joint point, (x) 2 ,y 2 ,z 2 ) The three-dimensional coordinates of the joint points of the centers of the shoulders;
step 43, normalizing the actual head height to obtain the scaling ratioWherein d represents a standard head height, and d=20 cm;
step 44, determining the distance between any 2 joint points for the 18 human body joint points according to the scaling ratio k: wherein ,(xi ,y i ,z i ) Represents the N < th i Three-dimensional coordinates of the individual nodes, (x) i ,y i ,z i ) Represents the N < th j Three-dimensional coordinates of the individual nodes, i, j are natural numbers, and i=1, 2,..18, j=1, 2,..18, i+.j; n represents a human body joint point.
Step 45, combining the joint angles determined in step 41 and the distances between the joint points determined in step 44 together as human body posture feature data.
7. The method for recognizing the gesture and behavior of the transformer operation and maintenance personnel and the gesture of the tool according to claim 6, wherein the step 5 specifically comprises the following steps:
51, dividing the human body posture feature data obtained in the step 45 into a human body posture feature training set and a human body posture feature verification set;
step 52, constructing a cart decision tree according to the human body posture feature training set;
step 53, verifying and evaluating the cart decision tree constructed in the step 52 by adopting a human body posture feature verification set; and analyzing the human body posture feature data by adopting a cart decision tree meeting the evaluation result, identifying the human body posture and the action corresponding to the human body posture, comparing the action corresponding to the human body posture with the pre-stored standard action posture of the power transformation operation and maintenance personnel, and determining the posture and the behavior of the power transformation operation and maintenance personnel in the actual operation process.
8. The method for identifying the gesture and behavior of the transformer operation and maintenance personnel and the gesture of the tool according to claim 7, wherein in step 53, the verification and evaluation of the cart decision tree constructed in step 52 are performed by using a human gesture feature verification set, specifically: randomly selecting 70% of the data from the human body posture feature data to form a human body posture feature training set, and the rest 30% of the data to form a human body posture feature verification set; the constructed cart decision tree of step 52 is validated and evaluated by the body posture feature validation set.
9. The method for identifying the gesture and behavior of the transformer operation and maintenance personnel and the gesture of the tool according to claim 7, wherein in step 53, the verification and evaluation of the cart decision tree constructed in step 52 are performed by using a human gesture feature verification set, specifically:
step B1, dividing the human body posture characteristic data into m sub-data sets; wherein m is a natural number;
step B2, randomly selecting one sub-data set from the m sub-data sets as a human body posture feature verification set, and verifying and evaluating the cart decision tree constructed in the step 52 to obtain an evaluation value;
step B3, repeating the step B2 until the rest m-1 sub-data sets are respectively and independently used as human body posture feature verification sets, and performing verification evaluation on the cart decision tree constructed in the step 52; obtaining (m-1) evaluation values;
and B4, averaging the m evaluation values obtained in the steps B2 to B3, wherein the average value is the evaluation result of the cart decision tree constructed in the step 52.
10. The method for recognizing the gesture and behavior of the transformer operation and maintenance personnel and the gesture of the tool according to claim 1, wherein the step 6 specifically comprises the following steps:
step 61, labeling and dividing more than 4000 scene infrared images with tools and scene color images by using labelme, and correspondingly obtaining tool infrared data and tool RGB data after file conversion;
step 62, combining the tool infrared data with the tool RGB data as tool feature data.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202211645542.5A CN116129467A (en) | 2022-12-16 | 2022-12-16 | Method for identifying gesture and behavior of transformer operation and maintenance personnel and gesture of tool |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202211645542.5A CN116129467A (en) | 2022-12-16 | 2022-12-16 | Method for identifying gesture and behavior of transformer operation and maintenance personnel and gesture of tool |
Publications (1)
Publication Number | Publication Date |
---|---|
CN116129467A true CN116129467A (en) | 2023-05-16 |
Family
ID=86296545
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202211645542.5A Pending CN116129467A (en) | 2022-12-16 | 2022-12-16 | Method for identifying gesture and behavior of transformer operation and maintenance personnel and gesture of tool |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN116129467A (en) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN117830961A (en) * | 2024-03-06 | 2024-04-05 | 山东达斯特信息技术有限公司 | Environment-friendly equipment operation and maintenance behavior analysis method and system based on image analysis |
-
2022
- 2022-12-16 CN CN202211645542.5A patent/CN116129467A/en active Pending
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN117830961A (en) * | 2024-03-06 | 2024-04-05 | 山东达斯特信息技术有限公司 | Environment-friendly equipment operation and maintenance behavior analysis method and system based on image analysis |
CN117830961B (en) * | 2024-03-06 | 2024-05-10 | 山东达斯特信息技术有限公司 | Environment-friendly equipment operation and maintenance behavior analysis method and system based on image analysis |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN110222665B (en) | Human body action recognition method in monitoring based on deep learning and attitude estimation | |
CN109919141A (en) | A kind of recognition methods again of the pedestrian based on skeleton pose | |
WO2019120108A1 (en) | Image coding method, action recognition method, and computer device | |
CN110135277B (en) | Human behavior recognition method based on convolutional neural network | |
JP2019096113A (en) | Processing device, method and program relating to keypoint data | |
CN112435731B (en) | Method for judging whether real-time gesture meets preset rules | |
CN113077519B (en) | Multi-phase external parameter automatic calibration method based on human skeleton extraction | |
CN112215172A (en) | Human body prone position three-dimensional posture estimation method fusing color image and depth information | |
CN111145865A (en) | Vision-based hand fine motion training guidance system and method | |
CN113191243B (en) | Human hand three-dimensional attitude estimation model establishment method based on camera distance and application thereof | |
CN116129467A (en) | Method for identifying gesture and behavior of transformer operation and maintenance personnel and gesture of tool | |
CN108830861A (en) | A kind of hybrid optical motion capture method and system | |
JP2008065368A (en) | System for recognizing position and posture of object using stereoscopic image, method of recognizing position and posture of object, and program for executing method | |
CN112184898A (en) | Digital human body modeling method based on motion recognition | |
CN116524586A (en) | Dance scoring algorithm based on CNN and GCN gesture estimation and similarity matching | |
CN112183316B (en) | Athlete human body posture measuring method | |
CN114170686A (en) | Elbow bending behavior detection method based on human body key points | |
CN113435293A (en) | Human body posture estimation method based on joint relation | |
CN113011344A (en) | Pull-up quantity calculation method based on machine vision | |
CN110163112B (en) | Examinee posture segmentation and smoothing method | |
CN116749168A (en) | Rehabilitation track acquisition method based on gesture teaching | |
CN111241936A (en) | Human body posture estimation method based on depth and color image feature fusion | |
CN116030135A (en) | Real-time attitude measurement system in remote operation | |
CN116386137A (en) | Mobile terminal design method for lightweight recognition of Taiji boxing | |
CN114494341A (en) | Real-time completion method for optical motion capture mark points by fusing time-space constraints |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination |