CN112454365A - Human behavior recognition technology-based human-computer interaction safety monitoring system - Google Patents

Human behavior recognition technology-based human-computer interaction safety monitoring system Download PDF

Info

Publication number
CN112454365A
CN112454365A CN202011396870.7A CN202011396870A CN112454365A CN 112454365 A CN112454365 A CN 112454365A CN 202011396870 A CN202011396870 A CN 202011396870A CN 112454365 A CN112454365 A CN 112454365A
Authority
CN
China
Prior art keywords
behavior
value
robot arm
human
word
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202011396870.7A
Other languages
Chinese (zh)
Inventor
涂宏斌
李�杰
高晓飞
段军
丁莉
聂芳华
杜变霞
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Hunan Great Wall Science And Technology Information Co ltd
Original Assignee
Hunan Great Wall Science And Technology Information Co ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Hunan Great Wall Science And Technology Information Co ltd filed Critical Hunan Great Wall Science And Technology Information Co ltd
Priority to CN202011396870.7A priority Critical patent/CN112454365A/en
Publication of CN112454365A publication Critical patent/CN112454365A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • BPERFORMING OPERATIONS; TRANSPORTING
    • B25HAND TOOLS; PORTABLE POWER-DRIVEN TOOLS; MANIPULATORS
    • B25JMANIPULATORS; CHAMBERS PROVIDED WITH MANIPULATION DEVICES
    • B25J9/00Programme-controlled manipulators
    • B25J9/16Programme controls
    • B25J9/1694Programme controls characterised by use of sensors other than normal servo-feedback from position, speed or acceleration sensors, perception control, multi-sensor controlled systems, sensor fusion
    • B25J9/1697Vision controlled systems
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B25HAND TOOLS; PORTABLE POWER-DRIVEN TOOLS; MANIPULATORS
    • B25JMANIPULATORS; CHAMBERS PROVIDED WITH MANIPULATION DEVICES
    • B25J13/00Controls for manipulators
    • B25J13/08Controls for manipulators by means of sensing devices, e.g. viewing or touching devices
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B25HAND TOOLS; PORTABLE POWER-DRIVEN TOOLS; MANIPULATORS
    • B25JMANIPULATORS; CHAMBERS PROVIDED WITH MANIPULATION DEVICES
    • B25J9/00Programme-controlled manipulators
    • B25J9/16Programme controls
    • B25J9/1674Programme controls characterised by safety, monitoring, diagnostic

Abstract

The invention provides a human body behavior recognition technology-based human-computer interaction safety monitoring system. The human behavior recognition technology-based human interaction safety monitoring system intelligently recognizes the actions of the robot arm through the intelligent monitoring system and realizes human interaction through the robot control system, and the human interaction safety monitoring system specifically comprises the following steps: step S1, establishing a robot arm behavior recognition system; step S2, recognizing behaviors of the robot arm and the operator; and step S3, reasoning the behavior category of the opposite side by adopting a case reasoning method combining the front and back action association according to the recognized behaviors of the robot arm and the operator, and realizing man-machine interaction. According to the invention, the camera is arranged at the key position of the robot equipment, the monitoring is carried out at the position where the human-machine interaction exists, and the intelligent video monitoring system can identify the robot hand action of the robot, so that the human-machine interaction is realized.

Description

Human behavior recognition technology-based human-computer interaction safety monitoring system
Technical Field
The invention relates to the technical field of human-computer interaction monitoring, in particular to a human-computer interaction safety monitoring system based on a human body behavior recognition technology.
Background
At present, in a factory workshop, machine equipment and operators do not have intelligent monitoring and alarming systems, and the safety of operation completely depends on the proficiency and personal protection of the operators and passive safety equipment such as limit switches, light curtains and the like arranged on the machines. This often causes injury to the operator of the device before the alarm is given.
Disclosure of Invention
The invention provides a human body behavior recognition technology-based human-computer interaction safety monitoring system, and aims to solve the technical problem that the degree of human-computer interaction is not high in factory monitoring in the background technology.
In order to achieve the above object, the human-computer interaction security monitoring system based on the human behavior recognition technology provided in the embodiments of the present invention intelligently recognizes the actions of the robot arm through the intelligent monitoring system, and realizes human-computer interaction through the robot control system, which specifically includes the following steps:
step S1, establishing a robot arm behavior recognition system: acquiring two parallax images of the robot arm through a binocular camera, and calculating the three-dimensional coordinate and geometric information of the robot arm space through the acquired parallax images so as to reconstruct the three-dimensional shape and position of the monitored part of the robot arm;
step S2, recognizing robot arm and operator behavior: shooting the motion attitude of the robot arm and the behavior of an operator in a certain range of the station through a camera arranged on the appropriate station, and calculating the motion attitude of the robot arm and the behavior and action recognition result of the operator;
and step S3, reasoning the behavior category of the opposite side by adopting a case reasoning method combining the front and back action association according to the recognized behaviors of the robot arm and the operator, and realizing man-machine interaction.
Preferably, in step S1, the calculating the three-dimensional space coordinate and the geometric information of the robot arm specifically includes the following steps:
step S11, where the imaging point coordinate of the left camera at the point P (X, Y, Z) where the imaging object exists is L (X)L,yL) The coordinate of the imaging point of the right camera is R (x)R,yR);
Step S12, according to the focal length f of the left and right camerasLAnd fRObtaining coordinate values (X) of the imaged object point P (X, Y, Z) with respect to the right cameraR,yR,zR);
Step S13, coordinate value (x) for right cameraR,yR,zR) Rotating and translating to obtain the value of the imaging object point P (X, Y, Z) in the coordinate system (X ', Y ', Z ') relative to the coordinate value of the right camera;
step S14, coordinate values (X, Y, Z) of the imaged object are obtained.
Preferably, the step S12 obtains the imaging object point P (X, Y, Z) relative to the right camera (X)R,yR,zR) The coordinate values are specifically: obtaining imaged object point P (X, Y, Z) relative to right camera (X) by obtaining a model of camera perspective transformationR,yR,zR) Coordinate value of (2) ([ X ]R YR ZR]TThe model of the camera perspective transformation is:
Figure BDA0002815567140000021
wherein f isLIs the focal length of the left camera, fRIs the focal length of the right camera, kLIs the scale factor, k, of the left cameraRIs the scale factor of the left camera;
the step S13 specifically includes:
the rotation transformation matrix R of the coordinate system (X, Y, Z) and the right camera coordinate system is:
R=R(α)·R(β)·R(γ);
Figure BDA0002815567140000022
wherein alpha, beta and gamma are rotation angles around x, y and z axes respectively;
the translation transformation vector T of the coordinate system (X ', Y ', Z ') and the right camera coordinate system is:
[T]=[tx ty tz]T
the calculation formula of the value of the imaged object point P (X, Y, Z) in the coordinate system (X ', Y ', Z ') with respect to the right camera coordinate values is:
Figure BDA0002815567140000031
the obtaining of the coordinate values (X, Y, Z) of the imaging object in step S14 specifically includes:
is provided with
Figure BDA0002815567140000032
Thereby having
Figure BDA0002815567140000033
The formula in the merging step S12 can be obtained:
Figure BDA0002815567140000034
determining R by rotating the transformation matrix R1~r9And obtaining the three-dimensional coordinates of the imaging object point P (X, Y, Z) in the coordinate system (X ', Y ', Z ') according to the corresponding relation:
X=ZxL/fL;Y=ZyL/fL
Figure BDA0002815567140000035
preferably, in step S2, the calculating the result of the robot arm motion gesture and the operator behavior recognition specifically includes the following steps:
step S21, acquiring space-time characteristic points of the movement parts of the robot arm and the operator;
step S22, converting the human body behavior in each frame into corresponding word packet information, and converting the obtained space-time characteristic points of each key point of the human body into word packets; establishing a word book by using an improved k-means clustering algorithm, namely, coding behavior actions with high similarity or the same similarity into one class, namely generating the word book;
and step S23, adopting the improved PLSA algorithm to recognize and classify the human body behavior.
Preferably, the step S21 is specifically:
for spatial domain image fspGiven an observation scale
Figure BDA0002815567140000036
And a smoothing scale
Figure BDA0002815567140000037
sp represents space domain and can be obtained through the found image interest points according to the Harris angular point detection method:
Figure BDA0002815567140000038
wherein the content of the first and second substances,
Figure BDA0002815567140000041
and
Figure BDA0002815567140000042
is a gaussian derivative function defined as follows;
Figure BDA0002815567140000043
Figure BDA0002815567140000044
in the formula (I), the compound is shown in the specification,
Figure BDA0002815567140000045
is a pair of
Figure BDA0002815567140000046
The differential of x is taken to be the difference,
Figure BDA0002815567140000047
is a pair of
Figure BDA0002815567140000048
Differentiating y, gsp(x,y;δ2) And muspRespectively a smooth window function and an observation window function;
Figure BDA0002815567140000049
μsp=det(μsp)-ktrac2sp)
=λ1λ2-k(λ12)2
in the formula, delta is a correlation coefficient between pixel points, det is a formula expression in the Harris corner detection method, trac is the rank of the solution u matrix, and lambda1λ2Is det (mu)sp) A representation of the calculation of the determinant;
when the image intensity value of a pixel point in a video image has obvious changes in a space domain and a time domain, namely a space-time characteristic point, defining a function: f is R2Xr → R, and convolving it with a gaussian kernel to obtain the result, namely the scale space L, which is defined as: l is R2×R×R+→R
Figure BDA00028155671400000410
Delta is a correlation coefficient between pixel points, tau is a time coefficient of the same pixel point under a scene, and is specific time with the unit of second;
where g is a Gaussian kernel function:
Figure BDA00028155671400000411
in the formula, deltalIs the correlation coefficient between pixels, taulFor a time coefficient appearing in the same pixel under a scene, t is the time of appearing the pixel, a 2 × 2 matrix of a spatial domain is popularized to a 3 × 3 matrix of the spatial domain containing time differentiation, and the following results are obtained:
Figure BDA0002815567140000051
wherein the content of the first and second substances,
Figure BDA0002815567140000052
is shown as
Figure BDA0002815567140000053
Convolution of, LxLy、LtAre respectively as
Figure BDA0002815567140000054
The smoothing function has smoothing scales in a space domain and a time domain as follows:
Figure BDA0002815567140000055
and
Figure BDA0002815567140000056
and detecting the spots by Harris method;
to model a sequence of spatiotemporal images, functions f and L are used and the following functions are defined:
Figure BDA0002815567140000057
the spatio-temporal function with a gaussian kernel is:
Figure BDA0002815567140000058
Figure BDA0002815567140000059
wherein the fusion scale is:
Figure BDA00028155671400000510
and
Figure BDA00028155671400000511
finally, an extended Harris function is defined, and the maximum value of the function is the space-time characteristic point:
H=det(μ)-ktrace3(μ)
=λ1λ2λ3-k(λ123)3
in the formula, λ3Is the eigenvalue of the u matrix.
Preferably, the step S22 specifically includes the following steps:
step S221, establishing a digital-analog definition word packet and a word book:
parameter definition: d ═ D1,d2,....,dNDenotes a set of good word packets, and W ═ W1,w2,....,wMDenotes a word book, then D and W are merged into an N M co-occurrence matrix N, which represents WjAppears at diThe number of times of (1) is as shown in the following formula
Figure BDA0002815567140000061
Wherein d isiFor the ith word package, wjRepresenting the jth word, in the matrix N, the rows represent the number of times each word appears in a certain word package, and the columns represent the number of times a certain word appears in each document; defining a subject variable z, p (d)i) For a word appearing at diProbability in the piece document; p (w)j|zk) To be on the subject zkIn occurrence ofwjThe probability of (d); p (z)k|di) As a subject zkProbability of occurrence in the ith document;
where P (Z | d) is the probability distribution of topic Z under a given topic Z, and P (w | Z) is the probability distribution of a topic w for a given topic Z;
step S222, classifying the space-time characteristic points of the robot arm and the joint points of the operator by applying a digital model:
the category d of the robot arm behavior and the operator behavior, N, can be obtained from the clustering in the video image qvA visual word is represented as NvVector of dimensions d (q):
Figure BDA0002815567140000062
in the formula, n (p, v)i) The video image q representing the human body behavior d contains visual words viThe number of (2);
the space-time characteristic point feature set of the camera to the joint points of the robot arm and the operator is as follows:
X={x1,......xN},
Figure BDA0002815567140000063
n1.. N, wherein Y ═ Y ·1,......yN},
Figure BDA0002815567140000064
i-1, …. n. is a reduced-dimension set of the set X, i.e., a set of noise-removed points, and the cluster category is
Figure BDA0002815567140000065
Step S223, creating a word list through a k-means clustering algorithm:
Figure BDA0002815567140000066
wherein k is a cluster number, rijE {0,1} is a number of labels, e.g.Fruit yiIs class j, then rij1, otherwise rij=0;
Figure BDA0002815567140000067
Wherein the content of the first and second substances,
Figure BDA0002815567140000068
is composed of
Figure BDA0002815567140000069
The weight value after the dimension reduction is carried out,
Figure BDA00028155671400000610
the probability expectation value of the x picture at the t moment and the i frame is obtained; k is the size of the word list,
Figure BDA00028155671400000611
is composed of
Figure BDA00028155671400000612
The weight value after dimensionality reduction; x ═ X1,......xN},
Figure BDA00028155671400000613
N is divided into k classes, defining a cluster variance:
Figure RE-GDA00029232924700000613
weight ε defining variance in clusterw
Figure BDA0002815567140000072
Figure BDA0002815567140000073
0≤p<1
Wherein the index p represents the sensitivity degree of weight updating between classes; an empirical value of 0.7 was chosen, giving the following formula:
Figure BDA0002815567140000074
Figure BDA0002815567140000075
0≤p<1
firstly fixing the weight value to find a new cluster, and then calculating mk
Figure BDA0002815567140000076
When w iskWhen increasing progressively, mkThe calculation formula is as follows:
Figure BDA0002815567140000077
wherein σijFor the i frame and j frame image correlation coefficient, vkIs xiiFor the difference between the ith frame image x and the ith frame probability expectation value mu,
Figure BDA0002815567140000078
0≤p<1,
Figure BDA0002815567140000079
Figure BDA00028155671400000711
the difference value of the probability expectation value mu of the jth frame image x and the ith frame image x is obtained;
the updating weight value is defined as follows in order to enhance and improve the stability of the k-means algorithm:
Figure RE-GDA0002923292470000078
0≤β≤1
wherein the content of the first and second substances,
Figure BDA0002815567140000081
the difference value of the probability expectation value mu of the jth frame image x and the ith frame image x is obtained;
in the continuous iterative calculation process, the beta parameter can influence the updating of the control weight and can smooth the weight values of continuous iteration;
given weight matrix wkWe can obtain the following formula:
Figure BDA0002815567140000082
wherein W is a weight matrix WkThe sum, X is the image matrix, C is the cluster class
Figure BDA0002815567140000083
T is matrix transposition;
the minimum value of J is the word book, i.e., the coded classification of high similarity or same behavioral actions.
Preferably, the step S23 specifically includes the following steps:
defining a joint distribution model and a posterior probability formula:
p(w,d,z)=p(w|z,z0)p(z,z0|d)p(d)
Figure BDA0002815567140000084
in order to find the maximum log-likelihood function in the EM step of the maximum expectation algorithm, the estimation value in the EM step is as follows:
Figure BDA0002815567140000085
Figure BDA0002815567140000086
Figure BDA0002815567140000087
the parameters in the above formula are regularized by setting appropriate Lagrange multipliers alpha and beta to obtain
Figure RE-GDA0002923292470000087
Figure RE-GDA0002923292470000088
Figure RE-GDA0002923292470000089
Figure BDA0002815567140000093
Bayesian estimation of p (z | w, d) and p (z) is then used0L w, d), the specific calculation steps are as follows:
Figure BDA0002815567140000094
Figure BDA0002815567140000095
p (z) finally obtained0And | w, d) is a probability value, and the probability value is the result of motion recognition.
Preferably, the step S3 specifically includes the following steps:
step S31, defining an interactive behavior Case of the robot arm to be recognized and the operator, and dividing the Case into 3 parts: the previous frame has determined a man-machine interaction sub-behavior part, a current frame basic sub-behavior part and a basic related sub-behavior time characteristic part:
Case={{A_prior}{B_current},{B_related}}={{A1,A2,....An},{B1,B2,B3,....Bn},{t1,t2, ....,tn}};
wherein { A _ prior } identifies the behavior of the previous frame; a set denoted as a1.. An, respectively;
{ B _ current } is the current behavior to be identified; { B _ related } is a behavior that is identified by the behavior identification algorithm above as being similar to the behavior that has been identified for A;
b is represented as a set of B1.. Bn;
t 1.. tn is the time series corresponding to it;
and step S32, judging the current man-machine interaction behavior through the previous man-machine interaction action by using a case reasoning mechanism according to the behavior category determined in the previous frame.
Preferably, the step S32 specifically includes the following steps:
step S321, calculating the similarity attribute characteristics of the human-computer interaction behaviors of the front frame and the rear frame: defining a one-dimensional vector consisting of the interaction behaviors of the robot arm and the operator in the previous frame of image as { A _ prior }, and the current behavior part as { B _ current }, then calculating the similarity between the current behavior and the determined behavior in the previous frame:
Sim(A,B)=∑ni=1
wi×sim(Ai,Bi)
n is the characteristic number in each case base, sim (A, B) is the case of known determined behavior and ambiguous behavior to be tested, wiThe case is A, the associated weight of the case base B, the weight is the frequency of the case base A when the behavior B occurs, and the calculation formula is as follows: w is ai=Bi/A;
Step S322, calculating time attribute characteristics related to behaviors: the time sequence set of the basic sub-behaviors according to the sequence set of the front frame number and the back frame number is as follows: b _ duration, the total duration of the occurrence of each simple child behavior in vector B ═ T1,T2,...,TN}; wherein, Ti is the total duration of the ith sub-behavior corresponding to the vector B _ duration;
the specific algorithm is as follows:
calculating two attribute characteristics of each simple behavior of the human body for the collected video sequence;
calculating comprehensive similarity sim (A, B);
when the maximum value of the similarity is larger than a certain threshold value T, namely sim (A, B) > T, judging that the current interactive behavior belongs to the same interactive behavior, wherein the threshold value is determined according to a field test experience value, and adding the threshold value into a case library to update the case library in real time for classifying the subsequent human-computer interactive behavior; and sends deaf the results to the operator to effect the corresponding interaction.
The technical effects which can be achieved by adopting the invention are as follows: according to the invention, the camera is arranged at the key position of the machine equipment, the monitoring is carried out at the position where the man-machine interaction exists, and when the camera shoots the machine, such as an operator operating materials, the intelligent video monitoring system can identify the robot hand action of the robot, so that the man-machine interaction is realized. When the movement of the robot arm harms the operator, the robot automatically gives an alarm and performs emergency braking.
Compared with the traditional method, the intelligent monitoring method has the greatest advantages that a contactless, large-range and long-distance intelligent monitoring technology can be realized, the repeated behavior interaction and dangerous behavior emergency braking of the robot-operating personnel in the monitored area can be realized, the technical scheme can be used for the man-machine interaction monitoring of the autonomous controllable intelligent factory, and the safety of the online operating personnel can be ensured on the premise of improving the production efficiency.
Drawings
FIG. 1 is a schematic structural diagram of a human-computer interaction safety monitoring system based on human behavior recognition technology according to the present invention;
fig. 2 is a binocular camera device installation schematic diagram and an imaging schematic diagram of the human behavior recognition technology-based human interaction safety monitoring system of the invention.
Detailed Description
In order to make the technical problems, technical solutions and advantages to be solved by the present invention clearer, the following detailed description is made with reference to the accompanying drawings and specific embodiments.
The invention provides a human-computer interaction safety monitoring system based on a human behavior recognition technology, which is characterized in that as shown in figure 1, the intelligent monitoring system intelligently recognizes the actions of a robot arm and realizes human-computer interaction through a robot control system, and the human-computer interaction safety monitoring system specifically comprises the following steps:
step S1, establishing a robot arm behavior recognition system: acquiring two parallax images of the robot arm through a binocular camera, and calculating the three-dimensional coordinate and geometric information of the robot arm space through the acquired parallax images so as to reconstruct the three-dimensional shape and position of the monitored part of the robot arm;
step S2, recognizing robot arm and operator behavior: shooting the motion attitude of the robot arm and the behavior of an operator in a certain range of the station through a camera arranged on the appropriate station, and calculating the motion attitude of the robot arm and the behavior and action recognition result of the operator;
and step S3, reasoning the behavior category of the opposite side by adopting a case reasoning method combining the front and back action association according to the recognized behaviors of the robot arm and the operator, and realizing man-machine interaction.
In step S1, the behavior of the robot arm in the operation area, such as turning, grabbing, lifting, etc., is identified, and due to the actual scene, there may be human-machine, human-human and machine-machine occlusion. Behavior recognition in the case of occlusion must be employed to recognize this type of behavior.
The invention installs a binocular camera or a multi-camera on a proper station, and the binocular camera adopts two cameras with completely same performance parameters during imaging so as to shoot operators and equipment. Taking binocular as an example, two images of parallax images of the robot arm are acquired from two different angles respectively, and the spatial three-dimensional coordinates and geometric information of the robot arm can be calculated through the acquired parallax images, so that the three-dimensional shape and position of the monitored part of the robot arm are reconstructed.
The binocular camera imaging uses two cameras with identical performance parameters, arranged in a left-right structure. As shown in fig. 1, wherein OL is the optical center of the left camera, OR represents the optical center of the right camera, by establishing a three-dimensional coordinate on the object under the action of the light source (indoor OR outdoor light source), it can be known that there is a point P (X, Y, Z) on the object, connecting the optical centers of the left and right cameras and the object point P, respectively, and the connecting line must intersect the XY plane of the cameras. Under the imaging of the cameras, the imaging point coordinate of the point P on the left camera is L (xL, yL), and the imaging point coordinate of the point P on the right camera is R (xR, yR). The two imaging points are images of the object P with different view angles.
In step S1, the step of calculating the three-dimensional coordinate and the geometric information of the robot arm space specifically includes the following steps:
step S11, where the imaging point coordinate of the left camera at the point P (X, Y, Z) where the imaging object exists is L (X)L,yL) The coordinate of the imaging point of the right camera is R (x)R,yR);
Step S12, according to the focal length f of the left and right camerasLAnd fRObtaining coordinate values (X) of the imaged object point P (X, Y, Z) with respect to the right cameraR,yR,zR);
Step S13, coordinate value (x) for right cameraR,yR,zR) Rotating and translating to obtain the value of the imaging object point P (X, Y, Z) in the coordinate system (X ', Y ', Z ') relative to the coordinate value of the right camera;
step S14, coordinate values (X, Y, Z) of the imaged object are obtained.
1. The step S12 obtains the imaging object point P (X, Y, Z) with respect to the right camera (X)R,yR,zR) The coordinate values are specifically: obtaining imaged object point P (X, Y, Z) relative to right camera (X) by obtaining a model of camera perspective transformationR,yR,zR) Coordinate value of (2) ([ X ]R YR ZR]TThe model of the camera perspective transformation is:
Figure BDA0002815567140000121
wherein f isLIs the focal length of the left camera, fRIs the focal length of the right camera, kLIs the scale factor, k, of the left cameraRIs the scale factor of the left camera;
the step S13 specifically includes:
the rotation transformation matrix R of the coordinate system (X, Y, Z) and the right camera coordinate system is:
R=R(α)·R(β)·R(γ);
Figure BDA0002815567140000131
wherein alpha, beta and gamma are rotation angles around x, y and z axes respectively;
the translation transformation vector T of the coordinate system (X ', Y ', Z ') and the right camera coordinate system is:
[T]=[tx ty tz]T
the calculation formula of the value of the imaged object point P (X, Y, Z) in the coordinate system (X ', Y ', Z ') with respect to the right camera coordinate values is:
Figure BDA0002815567140000132
the obtaining of the coordinate values (X, Y, Z) of the imaging object in step S14 specifically includes:
is provided with
Figure BDA0002815567140000133
Thereby having
Figure BDA0002815567140000134
The formula in the merging step S12 can be obtained:
Figure BDA0002815567140000135
determining R by rotating the transformation matrix R1~r9And obtaining the three-dimensional coordinates of the imaging object point P (X, Y, Z) in the coordinate system (X ', Y ', Z ') according to the corresponding relation:
X=ZxL/fL;Y=ZyL/fL
Figure BDA0002815567140000136
in step S2, the calculating the result of the robot arm motion gesture and the operator behavior recognition specifically includes the following steps:
step S21, acquiring space-time characteristic points of the movement parts of the robot arm and the operator;
step S22, converting the human body behavior in each frame into corresponding word packet information, and converting the obtained space-time characteristic points of each key point of the human body into word packets, namely encoding 01, 11, 10 and the like; establishing a word book by using an improved k-means clustering algorithm, namely, classifying behavior action codes with high similarity or the same behavior action codes into one class, namely generating the word book;
and step S23, adopting the improved PLSA algorithm to recognize and classify the human body behavior.
In step S21, the camera captures the robot arm and the operator simultaneously in the field of view, and the robot arm and the operator' S joints can be regarded as angular points. Therefore, the design method finds these corner points. The method for acquiring the space-time characteristic points of the moving parts of the robot arm and the operator specifically comprises the following steps:
for spatial domain image fspGiven an observation scale
Figure BDA0002815567140000141
And a smoothing scale
Figure BDA0002815567140000142
sp represents the space domain; the method can be obtained by finding the interest points of the image according to a Harris corner detection method:
Figure BDA0002815567140000143
wherein the content of the first and second substances,
Figure BDA0002815567140000144
and
Figure BDA0002815567140000145
is a gaussian derivative function defined as follows;
Figure BDA0002815567140000146
Figure BDA0002815567140000147
in the formula (I), the compound is shown in the specification,
Figure BDA0002815567140000148
is a pair of
Figure BDA0002815567140000149
The differential of x is taken to be the difference,
Figure BDA00028155671400001410
is a pair of
Figure BDA00028155671400001411
Differentiating y, gsp(x,y;δ2) And muspRespectively a smooth window function and an observation window function;
Figure BDA00028155671400001412
μsp=det(μsp)-ktrac2sp)
=λ1λ2-k(λ12)2
in the formula, delta is a correlation coefficient between pixel points, det is a formula expression in the Harris corner detection method, trac is the rank of the solution u matrix, and lambda1λ2Is det (mu)sp) A representation of the calculation of the determinant;
when pixels in a video image are clicked outWhen the image intensity values have obvious changes in a space domain and a time domain, namely, when space-time characteristic points appear, a function is defined: f is R2Xr → R, and convolving it with a gaussian kernel to obtain the result, namely the scale space L, which is defined as: l is R2×R×R+→R
Figure BDA0002815567140000151
Delta is a correlation coefficient between pixel points, tau is a time coefficient of the same pixel point under a scene, and is specific time with the unit of second;
where g is a Gaussian kernel function:
Figure BDA0002815567140000152
in the formula, deltalIs the correlation coefficient between pixels, taulFor a time coefficient appearing in the same pixel under a scene, t is the time of appearing the pixel, a 2 × 2 matrix of a spatial domain is popularized to a 3 × 3 matrix of the spatial domain containing time differentiation, and the following results are obtained:
Figure BDA0002815567140000153
wherein the content of the first and second substances,
Figure BDA0002815567140000154
is shown as
Figure BDA0002815567140000155
Convolution of, LxLy、LtAre respectively as
Figure BDA0002815567140000156
The smoothing function has smoothing scales in a space domain and a time domain as follows:
Figure BDA0002815567140000157
and
Figure BDA0002815567140000158
and detecting the spots by Harris method;
to model a sequence of spatiotemporal images, functions f and L are used and the following functions are defined:
Figure BDA0002815567140000159
the spatio-temporal function with a gaussian kernel is:
Figure BDA00028155671400001510
Figure BDA0002815567140000161
wherein the fusion scale is:
Figure BDA0002815567140000162
and
Figure BDA0002815567140000163
finally, an extended Harris function is defined, and the maximum value of the function is the space-time characteristic point:
H=det(μ)-ktrace3(μ)
=λ1λ2λ3-k(λ123)3
in the formula, λ3Is the eigenvalue of the u matrix.
The step S22 specifically includes the following steps:
step S221, establishing a digital-analog definition word packet and a word book:
parameter definition: d ═ D1,d2,....,dNDenotes a set of good word packets, and W ═ W1,w2,....,wMDenotes a word book, then D and W are merged into an N M co-occurrence matrix N, which represents WjAppears at diThe number of times of (1) is as shown in the following formula
Figure BDA0002815567140000164
Wherein d isiFor the ith word package, wjRepresenting the jth word, in the matrix N, the rows represent the number of times each word appears in a certain word package, and the columns represent the number of times a certain word appears in each document; defining a subject variable z, p (d)i) For a word appearing at diProbability in the piece document; p (w)j|zk) To be on the subject zkIn the presence of wjThe probability of (d); p (z)k|di) As a subject zkProbability of occurrence in the ith document;
where P (Z | d) is the probability distribution of topic Z under a given topic Z, and P (w | Z) is the probability distribution of a topic w for a given topic Z;
step S222, classifying the space-time characteristic points of the robot arm and the joint points of the operator by applying a digital model:
the category d of the robot arm behavior and the operator behavior, N, can be obtained from the clustering in the video image qvA visual word is represented as NvVector of dimensions d (q):
Figure BDA0002815567140000165
in the formula, n (p, v)i) The video image q representing the human body behavior d contains visual words viThe number of (2);
the space-time characteristic point feature set of the camera to the joint points of the robot arm and the operator is as follows:
X={x1,......xN},
Figure BDA0002815567140000171
n1.. N, wherein Y ═ Y ·1,......yN},
Figure BDA0002815567140000172
i-1, …. n. is a reduced-dimension set of the set X, i.e., a set of noise-removed points, and the cluster category is
Figure BDA0002815567140000173
Step S223, creating a word list through a k-means clustering algorithm:
Figure BDA0002815567140000174
wherein k is a cluster number, rijE {0,1} is a tag number if yiIs class j, then rij1, otherwise rij=0;
Figure BDA0002815567140000175
Wherein the content of the first and second substances,
Figure RE-GDA0002923292470000168
is composed of
Figure RE-GDA0002923292470000169
The weight value after the dimension reduction is carried out,
Figure RE-GDA00029232924700001610
the probability expectation value of the x picture at the t moment and the i frame is obtained; k is the size of the word list,
Figure RE-GDA00029232924700001611
is composed of
Figure RE-GDA00029232924700001612
The weight value after dimensionality reduction; x ═ X1,......xN},
Figure RE-GDA00029232924700001613
Divided into k classes, we define an intra-cluster variance:
Figure RE-GDA0002923292470000171
weight ε defining variance in clusterw
Figure BDA00028155671400001713
Figure BDA00028155671400001714
0≤p<1
Wherein the index p represents the sensitivity degree of weight updating between classes; an empirical value of 0.7 was chosen, giving the following formula:
Figure BDA00028155671400001715
Figure BDA00028155671400001716
0≤p<1
firstly fixing the weight value to find a new cluster, and then calculating mk
Figure BDA00028155671400001717
When w iskWhen increasing progressively, mkThe calculation formula is as follows:
Figure BDA0002815567140000181
wherein σijFor the i frame and j frame image correlation coefficient, vkIs xiiFor the difference between the ith frame image x and the ith frame probability expectation value mu,
Figure BDA0002815567140000182
0≤p<1,
Figure BDA0002815567140000183
Figure BDA0002815567140000189
the difference value of the probability expectation value mu of the jth frame image x and the ith frame image x is obtained;
the updating weight value is defined as follows in order to enhance and improve the stability of the k-means algorithm:
Figure RE-GDA00029232924700001710
0≤β≤1
wherein the content of the first and second substances,
Figure BDA0002815567140000185
the difference value of the probability expectation value mu of the jth frame image x and the ith frame image x is obtained;
in the continuous iterative calculation process, the beta parameter can influence the updating of the control weight and can smooth the weight value of continuous iteration;
given weight matrix wkWe can obtain the following formula:
Figure BDA0002815567140000186
wherein W is a weight matrix WkThe sum, X is the image matrix, C is the cluster class
Figure BDA0002815567140000187
T is matrix transposition;
the minimum value of J is the word book, i.e., the coded classification of high similarity or same behavioral actions.
The step S23 specifically includes the following steps:
defining a joint distribution model and a posterior probability formula:
p(w,d,z)=p(w|z,z0)p(z,z0|d)p(d)
Figure BDA0002815567140000188
the Expectation maximization (Expectation maximization) algorithm is an iterative optimization algorithm, and the calculation method is that each iteration is divided into an Expectation (E) step and a Maximum (M) step. In order to find the maximum log-likelihood function in the EM step of the maximum expectation algorithm, the estimation value in the EM step is as follows:
Figure BDA0002815567140000191
Figure BDA0002815567140000192
Figure BDA0002815567140000193
the parameters in the above formula are regularized by setting appropriate Lagrange multipliers alpha and beta to obtain
Figure BDA0002815567140000194
Figure RE-GDA0002923292470000191
Figure RE-GDA0002923292470000192
Figure BDA0002815567140000197
Then using shellfishLeaf estimates p (z | w, d) and p (z)0L w, d), the specific calculation steps are as follows:
Figure BDA0002815567140000198
Figure BDA0002815567140000201
p (z) finally obtained0And | w, d) is a probability value, and the probability value is the result of motion recognition.
The step S3 specifically includes the following steps:
step S31, defining an interactive behavior Case of the robot arm to be recognized and the operator, and dividing the Case into 3 parts: the previous frame has determined a man-machine interaction sub-behavior part, a current frame basic sub-behavior part and a basic related sub-behavior time characteristic part:
Case={{A_prior}{B_current},{B_related}}={{A1,A2,....An},{B1,B2,B3,....Bn},{t1,t2, ....,tn}};
wherein { A _ prior } identifies the behavior of the previous frame; a set denoted as a1.. An, respectively;
{ B _ current } is the current behavior to be identified; { B _ related } is a behavior that is identified by the behavior identification algorithm above as being similar to the behavior that has been identified for A;
b is represented as a set of B1.. Bn;
t 1.. tn is the time series corresponding to it;
and step S32, judging the current man-machine interaction behavior through the previous man-machine interaction action by using a case reasoning mechanism according to the behavior category determined in the previous frame.
The step S32 specifically includes the following steps:
step S321, calculating the similarity attribute characteristics of the human-computer interaction behaviors of the front frame and the rear frame: defining a one-dimensional vector consisting of the interaction behaviors of the robot arm and the operator in the previous frame of image as { A _ prior }, and the current behavior part as { B _ current }, then calculating the similarity between the current behavior and the determined behavior in the previous frame:
Sim(A,B)=∑ni=1
wi×sim(Ai,Bi)
n is the characteristic number in each case base, sim (A, B) is the case of known determined behavior and ambiguous behavior to be tested, wiThe case is A, the associated weight of the case base B, the weight is the frequency of the case base A when the behavior B occurs, and the calculation formula is as follows: w is ai=Bi/A;
Step S322, calculating time attribute characteristics related to behaviors: the time sequence set of the basic sub-behaviors according to the sequence set of the front frame number and the back frame number is as follows: b _ duration, the total duration of the occurrence of each simple child behavior in vector B ═ T1,T2,...,TN}; wherein, Ti is the total duration of the ith sub-behavior corresponding to the vector B _ duration;
the specific algorithm is as follows:
calculating two attribute characteristics of each simple behavior of the human body for the collected video sequence;
calculating comprehensive similarity sim (A, B);
when the maximum value of the similarity is larger than a certain threshold value T, namely sim (A, B) > T, judging that the current interactive behavior belongs to the same interactive behavior, wherein the threshold value is determined according to a field test experience value, and adding the threshold value into a case library to update the case library in real time for classifying the subsequent human-computer interactive behavior; and sends deaf the results to the operator to effect corresponding interactions, such as: and delivering the workpiece to the robot arm, and automatically lifting or grabbing the workpiece by the robot arm.
The technical effects which can be achieved by adopting the invention are as follows: according to the invention, the camera is arranged at the key position of the machine equipment, the monitoring is carried out at the position where the man-machine interaction exists, and when the camera shoots the machine, such as an operator operating materials, the intelligent video monitoring system can identify the robot hand action of the robot, so that the man-machine interaction is realized. When the movement of the robot arm harms the operator, the robot automatically gives an alarm and performs emergency braking.
Compared with the traditional method, the intelligent monitoring method has the greatest advantages that a contactless, large-range and long-distance intelligent monitoring technology can be realized, the repeated behavior interaction and dangerous behavior emergency braking of the robot-operating personnel in the monitored area can be realized, the technical scheme can be used for the man-machine interaction monitoring of the autonomous controllable intelligent factory, and the safety of the online operating personnel can be ensured on the premise of improving the production efficiency.
While the foregoing is directed to the preferred embodiment of the present invention, it will be understood by those skilled in the art that various changes and modifications may be made without departing from the spirit and scope of the invention as defined in the appended claims.

Claims (9)

1. Human action recognition technology based human interaction safety monitoring system, its characterized in that through intelligent monitoring system intelligent recognition robot arm action to realize human-computer interaction through robot control system, specifically include the following step:
step S1, establishing a robot arm behavior recognition system: acquiring two parallax images of the robot arm through a binocular camera, and calculating the three-dimensional coordinate and geometric information of the robot arm space through the acquired parallax images so as to reconstruct the three-dimensional shape and position of the monitored part of the robot arm;
step S2, recognizing robot arm and operator behavior: shooting the motion attitude of the robot arm and the behavior of an operator in a certain range of the station through a camera arranged on the appropriate station, and calculating the motion attitude of the robot arm and the behavior and action recognition result of the operator;
and step S3, reasoning the behavior category of the opposite side by adopting a case reasoning method combining the front and back action association according to the recognized behaviors of the robot arm and the operator, and realizing man-machine interaction.
2. The human body behavior recognition technology-based human body interaction security monitoring system according to claim 1, wherein the step S1 of calculating the spatial three-dimensional coordinates and the geometric information of the robot arm specifically comprises the steps of:
step S11, where the imaging point coordinate of the left camera at the point P (X, Y, Z) where the imaging object exists is L (X)L,yL) The coordinate of the imaging point of the right camera is R (x)R,yR);
Step S12, according to the focal length f of the left and right camerasLAnd fRObtaining coordinate values (X) of the imaged object point P (X, Y, Z) with respect to the right cameraR,yR,zR);
Step S13, coordinate value (x) for right cameraR,yR,zR) Rotating and translating to obtain the value of the imaging object point P (X, Y, Z) in the coordinate system (X ', Y ', Z ') relative to the coordinate value of the right camera;
step S14, coordinate values (X, Y, Z) of the imaged object are obtained.
3. The human body behavior recognition technology-based human body interaction security monitoring system according to claim 2, wherein the imaging object point P (X, Y, Z) obtained in the step S12 is opposite to the right camera (X)R,yR,zR) The coordinate values are specifically: obtaining imaged object point P (X, Y, Z) relative to right camera (X) by obtaining a model of camera perspective transformationR,yR,zR) Coordinate value of (2) ([ X ]R YR ZR]TThe model of the camera perspective transformation is:
Figure FDA0002815567130000021
wherein f isLIs the focal length of the left camera, fRIs the focal length of the right camera, kLIs the scale factor, k, of the left cameraRIs the scale factor of the left camera;
the step S13 specifically includes:
the rotation transformation matrix R of the coordinate system (X, Y, Z) and the right camera coordinate system is:
R=R(α)·R(β)·R(γ);
Figure FDA0002815567130000022
wherein alpha, beta and gamma are rotation angles around x, y and z axes respectively;
the translation transformation vector T of the coordinate system (X ', Y ', Z ') and the right camera coordinate system is:
[T]=[tx ty tz]T
the calculation formula of the value of the imaged object point P (X, Y, Z) in the coordinate system (X ', Y ', Z ') with respect to the right camera coordinate values is:
Figure FDA0002815567130000023
the obtaining of the coordinate values (X, Y, Z) of the imaging object in step S14 specifically includes:
is provided with
Figure FDA0002815567130000024
Thereby having
Figure FDA0002815567130000025
The formula in the merging step S12 can be obtained:
Figure FDA0002815567130000031
determining R by rotating the transformation matrix R1~r9And obtaining the three-dimensional coordinates of the imaging object point P (X, Y, Z) in the coordinate system (X ', Y ', Z ') according to the corresponding relation:
X=ZxL/fL;Y=ZyL/fL
Figure FDA0002815567130000032
4. the human body behavior recognition technology-based human body interaction security monitoring system according to claim 1, wherein in the step S2, the calculating of the robot arm motion gesture and the operator behavior recognition result specifically includes the following steps:
step S21, acquiring space-time characteristic points of the movement parts of the robot arm and the operator;
step S22, converting the human body behavior in each frame into corresponding word packet information, and converting the obtained space-time characteristic points of each key point of the human body into word packets; establishing a word book by using an improved k-means clustering algorithm, namely, coding behavior actions with high similarity or the same similarity into one class, namely generating the word book;
and step S23, adopting the improved PLSA algorithm to recognize and classify the human body behavior.
5. The human behavior recognition technology-based human interaction security monitoring system of claim 4, wherein the step S21 specifically comprises:
for spatial domain image fspGiven an observation scale
Figure FDA0002815567130000033
And a smoothing scale
Figure FDA0002815567130000034
sp represents space domain and can be obtained through the found image interest points according to the Harris angular point detection method:
Figure FDA0002815567130000035
wherein the content of the first and second substances,
Figure FDA0002815567130000036
and
Figure FDA0002815567130000037
is a gaussian derivative function defined as follows;
Figure FDA0002815567130000038
Figure FDA0002815567130000039
in the formula (I), the compound is shown in the specification,
Figure FDA00028155671300000310
is a pair of
Figure FDA00028155671300000311
The differential of x is taken to be the difference,
Figure FDA00028155671300000312
is a pair of
Figure FDA00028155671300000313
Differentiating y, gsp(x,y;δ2) And muspRespectively a smooth window function and an observation window function;
Figure FDA00028155671300000314
μsp=det(μsp)-ktrac2sp)
=λ1λ2-k(λ12)2
in the formula, delta is a correlation coefficient between pixel points, det is a formula expression in the Harris corner detection method, trac is the rank of the solution u matrix, and lambda1λ2Is det (mu)sp) A representation of the calculation of the determinant;
when the image intensity value of a pixel point in a video image has obvious changes in a space domain and a time domain, namely a space-time characteristic point, defining a function: f is R2X R → R, and mixing it with Gaussian nucleusThe result of the convolution of the function is a scale space L, which is defined as: l is R2×R×R+→R
Figure FDA0002815567130000041
Delta is a correlation coefficient between pixel points, tau is a time coefficient of the same pixel point under a scene, and is specific time with the unit of second;
where g is a Gaussian kernel function:
Figure FDA0002815567130000042
in the formula, deltalIs the correlation coefficient between pixels, taulFor a time coefficient of the same pixel point appearing in a scene, t is the time of the pixel point appearing, a 2 × 2 matrix of a spatial domain is popularized to a 3 × 3 matrix of the spatial domain containing time differentiation, and the following results are obtained:
Figure FDA0002815567130000043
wherein the content of the first and second substances,
Figure FDA0002815567130000044
is shown as
Figure FDA0002815567130000045
Convolution of, LxLy、LtAre respectively as
Figure FDA0002815567130000046
The smoothing scale of the function in the space domain and the time domain is respectively:
Figure FDA0002815567130000047
and
Figure FDA0002815567130000048
and detecting these spots by Harris method;
to model a sequence of spatiotemporal images, functions f and L are used and the following functions are defined:
Figure FDA0002815567130000049
the spatio-temporal function with a gaussian kernel is:
Figure FDA0002815567130000051
Figure FDA0002815567130000052
wherein the fusion scale is:
Figure FDA0002815567130000053
and
Figure FDA0002815567130000054
finally, an extended Harris function is defined, and the maximum value of the function is the space-time characteristic point:
H=det(μ)-ktrace3(μ)
=λ1λ2λ3-k(λ123)3
in the formula, λ3Is the eigenvalue of the u matrix.
6. The human behavior recognition technology-based human interaction security monitoring system of claim 5, wherein the step S22 specifically comprises the following steps:
step S221, establishing a digital-analog definition word packet and a word book:
parameter definition: d ═ D1,d2,....,dNDenotes a set of good word packets, and W ═ W1,w2,....,wMDenotes a word book, then D and W are merged into an N M co-occurrence matrix N, which represents WjAppears at diThe number of times of (1) is as shown in the following formula
Figure RE-FDA0002923292460000055
Wherein d isiFor the ith word package, wjRepresenting the jth word, in the matrix N, the rows represent the number of times each word appears in a certain package, and the columns represent the number of times a certain word appears in each document; defining a subject variable z, p (d)i) For a word appearing at diProbability in the piece document; p (w)j|zk) To be on the subject zkIn the presence of wjThe probability of (d); p (z)k|di) As a subject zkProbability of occurrence in the ith document;
where P (Z | d) is the probability distribution of topic Z under a given topic Z, and P (w | Z) is the probability distribution of a topic w for a given topic Z;
step S222, classifying the space-time characteristic points of the robot arm and the joint points of the operator by applying a digital model:
the category d of the robot arm behavior and the operator behavior, N, can be obtained from the clustering in the video image qvA visual word is represented as NvVector of dimensions d (q):
Figure RE-FDA0002923292460000061
in the formula, n (p, v)i) The video image q representing the human body behavior d contains visual words viThe number of (2);
the space-time characteristic point feature set of the camera to the joint points of the robot arm and the operator is as follows:
X={x1,......xN},
Figure RE-FDA0002923292460000062
wherein Y is { Y ═ Y1,......yN},
Figure RE-FDA0002923292460000063
Is a reduced-dimension set of the set X, i.e. a set of noise points, and the cluster category is
Figure RE-FDA0002923292460000064
Step S223, creating a word list through a k-means clustering algorithm:
Figure RE-FDA0002923292460000065
wherein k is a cluster number, rijE {0,1} is a tag number if yiIs class j, then rij1, otherwise rij=0;
Figure RE-FDA0002923292460000066
Wherein the content of the first and second substances,
Figure RE-FDA0002923292460000067
is composed of
Figure RE-FDA0002923292460000068
The weight value after the dimension reduction is carried out,
Figure RE-FDA0002923292460000069
the probability expectation value of the x picture at the t moment and the i frame is obtained; k is the size of the word list,
Figure RE-FDA00029232924600000610
is composed of
Figure RE-FDA00029232924600000611
The weight value after dimensionality reduction; x ═ X1,......xN},
Figure RE-FDA00029232924600000612
Is divided into k classes, defining an intra-cluster variance:
Figure RE-FDA00029232924600000613
weight ε defining variance in clusterw
Figure RE-FDA00029232924600000614
Figure RE-FDA00029232924600000615
Wherein the index p represents the sensitivity degree of weight updating between classes; an empirical value of 0.7 was chosen, giving the following formula:
Figure RE-FDA00029232924600000616
firstly fixing the weight value to find a new cluster, and then calculating mk
Figure RE-FDA0002923292460000071
When w iskWhen increasing progressively, mkThe calculation formula is as follows:
Figure RE-FDA0002923292460000072
Figure RE-FDA0002923292460000073
wherein σijFor the i frame and j frame image correlation coefficient, vkIs xiiFor the difference between the ith frame image x and the ith frame probability expectation value mu,
Figure RE-FDA0002923292460000074
Figure RE-FDA0002923292460000075
the difference value of the probability expectation value mu of the jth frame image x and the ith frame image x is obtained;
the updating weight value is defined as follows in order to enhance and improve the stability of the k-means algorithm:
Figure RE-FDA0002923292460000076
wherein the content of the first and second substances,
Figure RE-FDA0002923292460000077
the difference value of the probability expectation value mu of the jth frame image x and the ith frame image x is obtained;
in the continuous iterative calculation process, the beta parameter can influence the updating of the control weight and can smooth the weight value of continuous iteration;
given weight matrix wkWe can obtain the following formula:
Figure RE-FDA0002923292460000078
wherein W is a weight matrix WkSum, X is the image matrix C for the cluster category
Figure RE-FDA0002923292460000079
T is a matrixTransposition is carried out;
the minimum value of J is the word book, i.e., the coded classification of high similarity or same behavioral actions.
7. The human behavior recognition technology-based human interaction security monitoring system of claim 6, wherein the step S23 specifically comprises the following steps:
defining a joint distribution model and a posterior probability formula:
p(w,d,z)=p(w|z,z0)p(z,z0|d)p(d)
Figure RE-FDA0002923292460000081
in order to find the maximum log-likelihood function in the EM step of the maximum expectation algorithm, the estimation value in the EM step is as follows:
Figure RE-FDA0002923292460000082
Figure RE-FDA0002923292460000083
Figure RE-FDA0002923292460000084
the parameters in the above formula are regularized by setting appropriate Lagrange multipliers alpha and beta to obtain
Figure RE-FDA0002923292460000085
Figure RE-FDA0002923292460000086
Figure RE-FDA0002923292460000087
Figure RE-FDA0002923292460000088
Bayesian estimation of p (z | w, d) and p (z) is then used0L w, d), the specific calculation steps are as follows:
Figure RE-FDA0002923292460000089
Figure RE-FDA0002923292460000091
p (z) finally obtained0And | w, d) is a probability value, and the probability value is the result of motion recognition.
8. The human behavior recognition technology-based human interaction security monitoring system of claim 1, wherein the step S3 specifically comprises the following steps:
step S31, defining an interactive behavior Case of the robot arm to be recognized and the operator, and dividing the Case into 3 parts: the previous frame has determined a man-machine interaction sub-behavior part, a current frame basic sub-behavior part and a basic correlation sub-behavior time characteristic part:
Case={{A_prior}{B_current},{B_related}}={{A1,A2,....An},{B1,B2,B3,....Bn},{t1,t2,....,tn}};
wherein { A _ prior } identifies the behavior of the previous frame; a set denoted as a1.. An, respectively;
{ B _ current } is the current behavior to be identified;
{ B _ related } is a behavior that is identified by the behavior identification algorithm above as being similar to the behavior that has been identified for A;
b is represented as a set of B1.. Bn;
t 1.. tn is the time series corresponding to it;
and step S32, judging the current man-machine interaction behavior through the previous man-machine interaction action by using a case reasoning mechanism according to the behavior category determined in the previous frame.
9. The human behavior recognition technology-based human interaction security monitoring system of claim 8, wherein the step S32 specifically comprises the following steps:
step S321, calculating the similarity attribute characteristics of the human-computer interaction behaviors of the front frame and the rear frame: defining a one-dimensional vector consisting of the interaction behaviors of the robot arm and the operator in the previous frame of image as { A _ prior }, and the current behavior part as { B _ current }, then calculating the similarity between the current behavior and the determined behavior in the previous frame:
Sim(A,B)=∑ni=1
wi×sim(Ai,Bi)
n is the characteristic number in each case base, sim (A, B) is the case of known definite behavior and ambiguous behavior to be measured, wiThe case is A, the associated weight of the case base B, the weight is the frequency of the case base A when the behavior B occurs, and the calculation formula is as follows: w is ai=Bi/A;
Step S322, calculating time attribute characteristics related to behaviors: the time sequence set of the basic sub-behaviors according to the sequence set of the front frame number and the back frame number is as follows: b _ duration, the total duration of the occurrences of each simple child behavior in vector B { T }1,T2,...,TN}; wherein, Ti is the total duration of the ith sub-behavior corresponding to the vector B _ duration;
the specific algorithm is as follows:
calculating two attribute characteristics of each simple behavior of the human body for the collected video sequence;
calculating comprehensive similarity sim (A, B);
when the maximum value of the similarity is larger than a certain threshold value T, namely sim (A, B) > T, judging that the current interactive behavior belongs to a certain interactive behavior, wherein the threshold value is determined according to a field test experience value, and adding the threshold value into a case library to update the case library in real time for classifying the subsequent human-computer interactive behavior; and sends deaf the results to the operator to effect the corresponding interaction.
CN202011396870.7A 2020-12-03 2020-12-03 Human behavior recognition technology-based human-computer interaction safety monitoring system Pending CN112454365A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202011396870.7A CN112454365A (en) 2020-12-03 2020-12-03 Human behavior recognition technology-based human-computer interaction safety monitoring system

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202011396870.7A CN112454365A (en) 2020-12-03 2020-12-03 Human behavior recognition technology-based human-computer interaction safety monitoring system

Publications (1)

Publication Number Publication Date
CN112454365A true CN112454365A (en) 2021-03-09

Family

ID=74805426

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202011396870.7A Pending CN112454365A (en) 2020-12-03 2020-12-03 Human behavior recognition technology-based human-computer interaction safety monitoring system

Country Status (1)

Country Link
CN (1) CN112454365A (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113689541A (en) * 2021-07-23 2021-11-23 电子科技大学 Two-person three-dimensional human body shape optimization reconstruction method in interactive scene
CN117576787A (en) * 2024-01-16 2024-02-20 北京大学深圳研究生院 Method, device and equipment for handing over based on active tracking and self-adaptive gesture recognition

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102930302A (en) * 2012-10-18 2013-02-13 山东大学 On-line sequential extreme learning machine-based incremental human behavior recognition method
CN106228538A (en) * 2016-07-12 2016-12-14 哈尔滨工业大学 Binocular vision indoor orientation method based on logo
CN108647582A (en) * 2018-04-19 2018-10-12 河南科技学院 Goal behavior identification and prediction technique under a kind of complex dynamic environment
CN111563446A (en) * 2020-04-30 2020-08-21 郑州轻工业大学 Human-machine interaction safety early warning and control method based on digital twin

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102930302A (en) * 2012-10-18 2013-02-13 山东大学 On-line sequential extreme learning machine-based incremental human behavior recognition method
CN106228538A (en) * 2016-07-12 2016-12-14 哈尔滨工业大学 Binocular vision indoor orientation method based on logo
CN108647582A (en) * 2018-04-19 2018-10-12 河南科技学院 Goal behavior identification and prediction technique under a kind of complex dynamic environment
CN111563446A (en) * 2020-04-30 2020-08-21 郑州轻工业大学 Human-machine interaction safety early warning and control method based on digital twin

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
涂宏斌 等: "《基于机器学习的行为识别技术研究》", 30 September 2016, 知识产权出版社 *
涂宏斌 等: "一种基于改进PLSA和案例推理的行为识别算法", 《计算机科学》 *

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113689541A (en) * 2021-07-23 2021-11-23 电子科技大学 Two-person three-dimensional human body shape optimization reconstruction method in interactive scene
CN113689541B (en) * 2021-07-23 2023-03-07 电子科技大学 Two-person three-dimensional human body shape optimization reconstruction method in interactive scene
CN117576787A (en) * 2024-01-16 2024-02-20 北京大学深圳研究生院 Method, device and equipment for handing over based on active tracking and self-adaptive gesture recognition
CN117576787B (en) * 2024-01-16 2024-04-16 北京大学深圳研究生院 Method, device and equipment for handing over based on active tracking and self-adaptive gesture recognition

Similar Documents

Publication Publication Date Title
CN108447078B (en) Interference perception tracking algorithm based on visual saliency
Smith et al. Tracking the visual focus of attention for a varying number of wandering people
Leykin et al. Thermal-visible video fusion for moving target tracking and pedestrian classification
Hermes et al. Vehicle tracking and motion prediction in complex urban scenarios
KR20190038808A (en) Object detection of video data
Salih et al. Comparison of stochastic filtering methods for 3D tracking
Leykin et al. Robust multi-pedestrian tracking in thermal-visible surveillance videos
CN112454365A (en) Human behavior recognition technology-based human-computer interaction safety monitoring system
Gallego et al. Region based foreground segmentation combining color and depth sensors via logarithmic opinion pool decision
Cilla et al. Human action recognition with sparse classification and multiple‐view learning
Chomat et al. Recognizing motion using local appearance
Qin et al. Research on intelligent image recognition technology based on equalization algorithm
Naeem et al. Multiple batches of motion history images (MB-MHIs) for multi-view human action recognition
Saptharishi et al. Agent-based moving object correspondence using differential discriminative diagnosis
Dave et al. Statistical survey on object detection and tracking methodologies
Yow et al. Scale and Orientation Invariance in Human Face Detection.
Perez-Cutino et al. Event-based human intrusion detection in UAS using deep learning
Huu et al. Action recognition application using artificial intelligence for smart social surveillance system.
Itano et al. Human actions recognition in video scenes from multiple camera viewpoints
Al-Berry et al. Human action recognition via multi-scale 3D stationary wavelet analysis
Kwak et al. Human action recognition using accumulated moving information
Halder et al. Anomalous Activity Detection from Ego View Camera of Surveillance Robots
Xiankun et al. Adaptive Tracking Algorithm for Multiperson Motion Targets Based on Local Feature Similarity
Qian et al. Recognizing human activities using appearance metric feature and kinematics feature
Abdellaoui et al. Body tracking and motion analysis using hybrid model

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
WD01 Invention patent application deemed withdrawn after publication
WD01 Invention patent application deemed withdrawn after publication

Application publication date: 20210309