CN114863352B - Personnel group behavior monitoring method based on video analysis - Google Patents

Personnel group behavior monitoring method based on video analysis Download PDF

Info

Publication number
CN114863352B
CN114863352B CN202210793997.5A CN202210793997A CN114863352B CN 114863352 B CN114863352 B CN 114863352B CN 202210793997 A CN202210793997 A CN 202210793997A CN 114863352 B CN114863352 B CN 114863352B
Authority
CN
China
Prior art keywords
key point
target person
interactive
target
person
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202210793997.5A
Other languages
Chinese (zh)
Other versions
CN114863352A (en
Inventor
刘驰
范赐恩
胡新礼
李文航
李露
李继恒
汪磊
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Optical Valley Technology Co ltd
Original Assignee
Optical Valley Technology Co ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Optical Valley Technology Co ltd filed Critical Optical Valley Technology Co ltd
Priority to CN202210793997.5A priority Critical patent/CN114863352B/en
Publication of CN114863352A publication Critical patent/CN114863352A/en
Application granted granted Critical
Publication of CN114863352B publication Critical patent/CN114863352B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/40Scenes; Scene-specific elements in video content
    • G06V20/41Higher-level, semantic clustering, classification or understanding of video scenes, e.g. detection, labelling or Markovian modelling of sport events or news items
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/50Context or environment of the image
    • G06V20/52Surveillance or monitoring of activities, e.g. for recognising suspicious objects
    • G06V20/53Recognition of crowd images, e.g. recognition of crowd congestion

Abstract

The invention relates to the technical field of data processing, in particular to a personnel group behavior monitoring method based on video analysis. The method comprises the following steps: acquiring a corresponding area image in a target place area; acquiring each key point and corresponding feature vector corresponding to each target person in the area image; obtaining a skeleton map corresponding to each target person in the region image according to each key point corresponding to each target person in the region image; acquiring an interaction domain corresponding to each target person in a target site area at the current acquisition time; splicing the skeleton map corresponding to each target person and the skeleton maps of the interactive persons corresponding to the interactive domains corresponding to the target persons to obtain the interactive maps corresponding to the target persons; and obtaining the action behaviors of the target personnel in each interactive graph according to the feature vectors corresponding to all key points in the interactive graph corresponding to each target personnel and the trained action behavior recognition network. The invention improves the accuracy of identifying the behavior of the person.

Description

Personnel group behavior monitoring method based on video analysis
Technical Field
The invention relates to the technical field of data processing, in particular to a personnel group behavior monitoring method based on video analysis.
Background
The monitoring camera is used for observing in a place with dense personnel, is a common public safety prevention and control means, and can find emergent emergencies in time, so that countermeasures for dealing with the relevant emergencies are made, and personnel are evacuated; the group behavior monitoring is a hot direction, the group behavior refers to the behavior of interaction generated by a plurality of persons, and is obviously different from the single behavior; the existing group behavior identification method is characterized in that the type of the important identification is trampling events, conflict events and other events which endanger the life safety of the masses, and the existing group behavior identification method is that monitoring professionals watch a monitoring screen in shifts to ensure real-time monitoring, and then group behaviors are analyzed and countermeasures are made manually.
In a monitoring picture, personnel are dense, the mutual shielding condition is easy to occur, and group actions have the characteristic of high complexity, so the identification difficulty is high; in addition, because the frequency of normal behaviors is high and the frequency of abnormal behaviors is low, the manual monitoring mode is easy to overlook, and the problems of strong subjectivity, easy fatigue and the like exist, so that the accuracy of identifying the behaviors of the personnel is relatively low.
Disclosure of Invention
In order to solve the problem that the accuracy of identifying the behaviors of the people based on a manual mode is relatively low in the prior art, the invention aims to provide a method for monitoring the behaviors of a group of people based on video analysis, and the adopted technical scheme is as follows:
the invention provides a personnel group behavior monitoring method based on video analysis, which comprises the following steps:
acquiring a region image corresponding to the current acquisition time in a target place region;
acquiring each key point corresponding to each target person in the area image and a feature vector corresponding to each key point; obtaining a skeleton map corresponding to each target person in the region image according to each key point corresponding to each target person in the region image; the feature vector is obtained by splicing a position vector corresponding to the key point and a body part vector;
acquiring an interaction domain corresponding to each target person in a target site area at the current acquisition time, wherein the interaction domain is an area within a preset range of the target person;
splicing the skeleton map corresponding to each target person and the skeleton maps of the interactive persons corresponding to the interactive domains corresponding to the target persons to obtain the interactive maps corresponding to the target persons;
and obtaining the action behaviors of the target personnel in each interactive graph according to the feature vectors corresponding to all key points in the interactive graph corresponding to each target personnel and the trained action behavior recognition network.
Preferably, the acquiring of each key point corresponding to each target person in the area image and the feature vector corresponding to each key point includes:
processing the area image by using a key point detection network to obtain each key point corresponding to each target person in the area image;
taking the vertex of the lower left corner of the area image as a coordinate origin, taking the horizontal direction as an x axis, and taking the vertical direction as a y axis, and obtaining position vectors corresponding to the key points corresponding to the target persons in the area image; the position vector comprises an abscissa corresponding to the key point, an ordinate corresponding to the key point and a depth value corresponding to the key point, and the depth value is obtained according to the area image;
for any target person: performing One-Hot coding on each key point corresponding to the target person to obtain body part vectors corresponding to each key point corresponding to the target person; and splicing the position vector corresponding to each key point corresponding to the target person with the body part vector to obtain the feature vector corresponding to each key point corresponding to the target person.
Preferably, the key points corresponding to the target persons in the region image are connected according to a preset connection rule to obtain a skeleton map corresponding to the target persons.
Preferably, the obtaining of the interaction domain corresponding to each target person in the target site area at the current acquisition time includes:
for any target person:
calculating to obtain a normal vector corresponding to the target person according to position vectors corresponding to the nose key point, the right eye key point and the left eye key point in the key points corresponding to the target person; the normal vector corresponding to the target person is as follows: calculating a plane normal vector constructed in the three-dimensional space by the nose key point, the right eye key point and the left eye key point by taking position vectors corresponding to the nose key point, the right eye key point and the left eye key point as coordinates in the three-dimensional space, and taking the plane normal vector as a normal vector corresponding to the target person;
taking a central point corresponding to the target person as a circular point in the target place area, and taking a first preset length as a radius to make a circle, so as to obtain a circular area corresponding to the target person; the central point corresponding to the target person is a nose key point corresponding to the target person;
taking a central point corresponding to the target person as a dot point, a normal vector corresponding to the target person as a central line, a preset angle as a fan-shaped angle and a second preset length as a radius in the target place area to obtain a fan-shaped area corresponding to the target person;
and taking the union region of the circular region and the fan-shaped region as an interaction region corresponding to the target person.
Preferably, the obtaining of the interaction map corresponding to each target person includes:
for any target person:
counting the number of other target persons not including the target person in the interaction domain corresponding to the target person, and recording as the interaction number;
if the number of the interactions corresponding to the target person is not 0, marking other target persons existing in the interaction domain corresponding to the target person as interaction persons; for any interactive person corresponding to the interactive domain corresponding to the target person: splicing the skeleton map corresponding to the target person and the skeleton map corresponding to the interactive person according to the position vector of each key point corresponding to the target person and the position vector of each key point corresponding to the interactive person to obtain an interactive map corresponding to the target person;
if the number of the interactions corresponding to the target person is 0, constructing an occupancy skeleton diagram; the space-occupying skeleton map is a skeleton map of a virtual interactive person, and position vectors of all key points in the space-occupying skeleton map are (-1, -1, -1); taking the virtual interaction personnel as interaction personnel corresponding to the interaction domain corresponding to the target personnel; connecting a left-hand key point corresponding to the target person with a left-hand key point of an interactive person corresponding to the interactive domain, connecting a right-hand key point corresponding to the target person with a right-hand key point of the interactive person corresponding to the interactive domain, and splicing the skeleton diagram corresponding to the target person with the occupancy skeleton diagram to obtain an interactive diagram corresponding to the target person; the right-hand key point and the left-hand key point are two key points in the key points.
Preferably, obtaining the interaction map corresponding to the target person includes:
connecting a key point which is closest to the right-hand key point corresponding to the target person in the key points corresponding to the interactive persons with the right-hand key point of the target person; connecting the key point which is closest to the left-hand key point corresponding to the target person in the key points corresponding to the interactive person with the left-hand key point of the target person; connecting a key point which is closest to the right-hand key point corresponding to the interactive person in the key points corresponding to the target person with the right-hand key point of the interactive person; connecting a key point which is closest to the left-hand key point corresponding to the interactive person in the key points corresponding to the target person with the left-hand key point of the interactive person; splicing the skeleton map corresponding to the target person and the skeleton map corresponding to the interactive person to obtain an interactive map corresponding to the target person;
the distance is calculated according to the position vector.
Preferably, the obtaining of the action behavior of the target person in each interactive graph includes:
for any interaction graph corresponding to any target person:
sequencing all key points in the interactive graph according to a preset sequence;
calculating the weight between each key point in the interactive map and each key point in the corresponding neighborhood set according to the feature vector corresponding to each key point in the interactive map and the normal vector corresponding to the target person; the corresponding neighborhood set comprises all key points connected with the corresponding key points in the interaction graph;
performing preset times of aggregation operation on each key point in the interactive graph according to the weight between each key point in the interactive graph and each key point in the corresponding neighborhood set to obtain a target feature vector corresponding to each key point in the interactive graph;
splicing target characteristic vectors corresponding to the key points in the interactive map according to the arrangement sequence of the key points to obtain comprehensive vectors corresponding to the interactive map;
and inputting the comprehensive vector corresponding to the interaction diagram into a multilayer perceptron to obtain the action behavior of the target person in the interaction diagram.
Preferably, the formula for calculating the weight between each key point in the interactive graph and each key point in the corresponding neighborhood set is as follows:
Figure 24901DEST_PATH_IMAGE001
wherein, the first and the second end of the pipe are connected with each other,
Figure 422516DEST_PATH_IMAGE002
for the weight of the jth keypoint to the ith keypoint in the interactive graph,
Figure 996716DEST_PATH_IMAGE003
the feature vector corresponding to the ith key point in the interactive map,
Figure 70852DEST_PATH_IMAGE004
the feature vector corresponding to the jth key point in the interaction graph,
Figure 832134DEST_PATH_IMAGE005
is the normal vector corresponding to the target person,
Figure 233772DEST_PATH_IMAGE006
the feature vector of the central point corresponding to the target person,
Figure 521534DEST_PATH_IMAGE007
in order to perform the vector splicing operation,
Figure 641937DEST_PATH_IMAGE008
in order to activate the function(s),
Figure 297040DEST_PATH_IMAGE009
is a first one of the vectors of similarity, and,
Figure 161091DEST_PATH_IMAGE010
is a second similarity degree vector, and is,
Figure 303359DEST_PATH_IMAGE011
is the transpose of the first similarity vector,
Figure 860243DEST_PATH_IMAGE012
is a transpose of the second similarity vector,
Figure 471484DEST_PATH_IMAGE013
is a first weight matrix of the weight data set,
Figure 201542DEST_PATH_IMAGE014
is a second weight matrix of the plurality of weight matrices,
Figure 339263DEST_PATH_IMAGE015
is an exponential function with e as the base,
Figure 207993DEST_PATH_IMAGE016
and (4) a neighborhood set corresponding to the ith key point in the interactive graph.
Preferably, the obtaining of the target feature vector corresponding to each key point in the interactive graph includes:
the formula for performing the first aggregation operation on each key point in the interactive graph is as follows:
Figure 165584DEST_PATH_IMAGE017
wherein the content of the first and second substances,
Figure 761651DEST_PATH_IMAGE018
for the aggregated updated feature vector corresponding to the ith key point in the interaction graph,
Figure 753878DEST_PATH_IMAGE019
for the ith key point in the interaction graphThe number of interactions of the person corresponding to the skeletal map,
Figure 462683DEST_PATH_IMAGE020
the interaction number of the person corresponding to the skeleton diagram to which the jth key point belongs in the interaction diagram is obtained; the personnel comprise the target personnel and the interactive personnel; the interaction quantity is the quantity of other target persons which do not comprise the target person and exist in the interaction domain corresponding to the target person;
and by analogy, continuously executing the aggregation operation by using the aggregated and updated feature vectors corresponding to the key points in the interactive graph, and recording the aggregated and updated feature vectors corresponding to the key points after the last aggregation as target feature vectors.
The invention has the following beneficial effects:
firstly, acquiring a region image corresponding to the current acquisition time in a region of a target place, key points corresponding to target people in the region image and feature vectors corresponding to the key points, and then acquiring a skeleton map corresponding to the target people in the region image according to the key points corresponding to the target people in the region image; then acquiring the interaction domain corresponding to each target person in the target place area at the current acquisition time, and splicing the skeleton map corresponding to each target person with the skeleton maps of the interaction persons corresponding to the interaction domains corresponding to each target person to obtain the interaction maps corresponding to each target person; and finally, according to the feature vectors corresponding to all key points in the interactive graphs corresponding to the target personnel and the trained action behavior recognition network, obtaining the action behaviors of the target personnel in the interactive graphs. The invention identifies the action behaviors of each person in monitoring in an automatic mode, overcomes the problems of strong subjectivity, easy fatigue and the like in a manual monitoring mode, and improves the accuracy of identifying the behaviors of the persons.
Drawings
In order to more clearly illustrate the embodiments of the present invention or the technical solutions and advantages of the prior art, the drawings used in the description of the embodiments or the prior art will be briefly described below, it is obvious that the drawings in the following description are only some embodiments of the present invention, and other drawings can be obtained by those skilled in the art without creative efforts.
FIG. 1 is a flow chart of a method for monitoring group behavior of people based on video analysis according to the present invention;
fig. 2 is a skeleton diagram corresponding to the target person.
Detailed Description
In order to further explain the technical means and functional effects of the present invention adopted to achieve the predetermined invention purpose, the following describes in detail a method for monitoring group behaviors of people based on video analysis according to the present invention with reference to the accompanying drawings and preferred embodiments.
Unless defined otherwise, all technical and scientific terms used herein have the same meaning as commonly understood by one of ordinary skill in the art to which this invention belongs.
The following specifically describes a specific scheme of the personnel group behavior monitoring method based on video analysis, which is provided by the invention, with reference to the accompanying drawings.
The embodiment of the personnel group behavior monitoring method based on video analysis comprises the following steps:
as shown in fig. 1, the method for monitoring the behavior of the group of people based on video analysis of the present embodiment includes the following steps:
and step S1, acquiring the area image corresponding to the current acquisition time in the target place area.
When the treading event, the collision event and other events endangering the life safety of the masses occur, corresponding countermeasures should be made in time to evacuate people so as to ensure the life safety of the masses; therefore, it is very important to monitor and identify the behaviors of each person in the group, identify abnormal behaviors in time, and make a corresponding plan. The embodiment provides a personnel group behavior monitoring method based on video analysis, which is used for monitoring people in a set place and identifying the action behaviors of the people at different moments.
In the embodiment, the public places of indoor general people when disasters occur are taken as examples (namely, places which are easy to cause people confusion), and monitoring group behaviors in the places has important significance for guaranteeing the safety of the people.
In this embodiment, a 3D simulator is established by using the universal Engine 5, and a simulation experiment is performed, so that the corresponding situation of the real environment of the final landing is as follows: the method comprises the steps of collecting the sizes of important ground objects such as building internal structures, building internal facilities, roads outside buildings, enclosing walls, green plants and the like in public places on site, and restoring a real scene by 1: 1; collecting relevant maps of real public places and baking the maps to ensure that the material maps of the environmental model are highly vivid; making sky in different time (morning, noon and evening) and different seasons (spring, summer, autumn and winter) and corresponding environment model maps to realize the switching of environment models in different time and different seasons; and (4) making environment models with three different fineness grades, namely low, medium and high, and making corresponding LOD models for optimal scheduling during virtual VR roaming and virtual training.
And modeling the 3D avatar models of 20 typical man and woman common masses and the avatar of the manager, wherein the avatar conforms to the common manager avatar and the common mass avatar. The present embodiment thus far performs scene modeling.
Setting personnel behaviors, common people can randomly execute the following behaviors, including: collision, talking, hugging, walking, jogging, sitting up, sprinting, jumping, squatting, stooping, climbing, creeping, falling down, injury, etc.
Setting an emergency disaster event, wherein an explosion accident caused by the sparks generated in the production process is assumed to be caused by flour production or other inflammable and explosive dust released into the air in a factory; and manufacturing a light and shadow effect, a dust effect and the like when an emergency disaster event occurs through a template in the non-real Engine 5, and taking the light and shadow effect, the dust effect and the like as an identification interference item of a monitoring camera, so that the robustness of a subsequent network is improved.
The modeling process can be completed by a 3D simulator designer and can be implemented by using known techniques, and the detailed description of the process is omitted.
The present embodiment corresponds the monitored condition of the real site to the position and the angle of view in the simulator to capture the picture in the simulator with the frame number of 30 FPS. The input of the monitoring data preprocessing module is the video shot by the monitoring visual angle and the environmental parameters of the 3D simulator. The present embodiment represents a simulated public place as a target place area.
In this embodiment, an image in the target location area is acquired at intervals and recorded as an area image, and then the action behaviors of each person in the target location area at the current acquisition time are identified by using the area image (that is, an area image is acquired at intervals of a preset time period, which is set as required). In this embodiment, the camera for monitoring is a depth camera, and depth information of each pixel point in the image can be acquired.
The method includes the steps that an area image corresponding to the current acquisition time in a target place area is obtained, and then the action behaviors of each person in the area image are recognized by taking the area image as an example; the present embodiment takes the general public in the area image as the target person.
Step S2, acquiring each key point corresponding to each target person in the area image and a feature vector corresponding to each key point; obtaining a skeleton map corresponding to each target person in the region image according to each key point corresponding to each target person in the region image; the feature vector is obtained by splicing a position vector corresponding to the key point and a body part vector.
In order to identify the action behavior of the target person in the image, in this embodiment, first, the key point detection network is used to process the area image to obtain each key point corresponding to each target person in the area image, and for each key point corresponding to any target person, as shown in fig. 2, where 1 is a nose key point
Figure 297784DEST_PATH_IMAGE021
2 is the key point of the neck
Figure 572907DEST_PATH_IMAGE022
And 3 is the right shoulderKey points
Figure 498269DEST_PATH_IMAGE023
And 4 is the key point of the right elbow
Figure 958069DEST_PATH_IMAGE024
5 is the key point of the right hand
Figure 890253DEST_PATH_IMAGE025
6 is the key point of the left shoulder
Figure 578855DEST_PATH_IMAGE026
7 is the key point of the left elbow
Figure 545674DEST_PATH_IMAGE027
8 is the key point of the left hand
Figure 848479DEST_PATH_IMAGE028
And 9 is the key point of right crotch
Figure 408905DEST_PATH_IMAGE029
10 is the key point of the right knee
Figure 760252DEST_PATH_IMAGE030
And 11 is the key point of the right foot
Figure 971790DEST_PATH_IMAGE031
12 is the key point of the left crotch
Figure 117601DEST_PATH_IMAGE032
13 is the key point of the left knee
Figure 896813DEST_PATH_IMAGE033
14 is the key point of the left foot
Figure 51851DEST_PATH_IMAGE034
And 15 is the key point of the right eye
Figure 55579DEST_PATH_IMAGE035
And 16 is the key point of the left eye
Figure 513236DEST_PATH_IMAGE036
17 is the key point of the right ear
Figure 907309DEST_PATH_IMAGE037
18 is the key point of the left ear
Figure 725092DEST_PATH_IMAGE038
19 is the center point of the hand held object
Figure 786589DEST_PATH_IMAGE039
19 key points (namely, one target person corresponds to 19 key points), wherein the sequence is the arrangement sequence of the key points; in this embodiment, the key point detection network is an openpos key point detection model. In this embodiment, the key point detection network and the training process thereof are prior art, and will not be described herein again.
Thus, the embodiment obtains each key point corresponding to each target person in the area image.
In the embodiment, a rectangular coordinate system is constructed by taking the vertex of the lower left corner of the regional image as the origin of coordinates, taking the horizontal direction as the x axis and taking the vertical direction as the y axis; acquiring position vectors corresponding to the key points corresponding to the target persons in the region image according to the rectangular coordinate system; the position vector corresponding to any key point corresponding to any target person is
Figure 415147DEST_PATH_IMAGE040
Wherein
Figure 30937DEST_PATH_IMAGE041
The abscissa of the kth key point corresponding to the target person,
Figure 855673DEST_PATH_IMAGE042
is the ordinate of the kth key point corresponding to the target person,
Figure 381463DEST_PATH_IMAGE043
correspond to the target personThe depth value of the kth keypoint of (1); the depth value is obtained from the area image.
For any target person: performing One-Hot coding on each key point corresponding to the target person to obtain body part vector corresponding to each key point corresponding to the target person (i.e. obtaining body part vector corresponding to each key point according to the code corresponding to each key point), such as nose
Figure 571136DEST_PATH_IMAGE021
The position code of (1,0,0, …, 0); the body part vector can reflect which body part the corresponding key point belongs to, so that the key point connection can be carried out subsequently. The One-Hot coding in this embodiment is the prior art, and is not described herein again.
For the occluded key points, the corresponding position vectors are set as
Figure 64435DEST_PATH_IMAGE044
Its corresponding body part vector is unchanged. In this embodiment, the position vector and the body part vector corresponding to each key point are spliced to obtain the feature vector corresponding to each key point. Thus, in this embodiment, according to the above process, the feature vector corresponding to each key point corresponding to each target person in the area image is obtained.
According to each key point corresponding to each target person, constructing a skeleton diagram corresponding to each target person, and for any target person: in this embodiment, the key points corresponding to the target person are connected according to a preset connection rule (i.e., the key points are connected according to positions of different parts of the human body), that is, the preset connection rule is as follows:
Figure 175086DEST_PATH_IMAGE045
Figure 680016DEST_PATH_IMAGE046
Figure 899645DEST_PATH_IMAGE047
Figure 490026DEST_PATH_IMAGE048
Figure 204036DEST_PATH_IMAGE049
Figure 829052DEST_PATH_IMAGE050
Figure 485161DEST_PATH_IMAGE051
Figure 562839DEST_PATH_IMAGE052
Figure 814960DEST_PATH_IMAGE053
Figure 622379DEST_PATH_IMAGE054
Figure 59176DEST_PATH_IMAGE055
Figure 499516DEST_PATH_IMAGE056
Figure 945541DEST_PATH_IMAGE057
Figure 873045DEST_PATH_IMAGE058
Figure 353181DEST_PATH_IMAGE059
Figure 139871DEST_PATH_IMAGE060
Figure 514221DEST_PATH_IMAGE061
Figure 968336DEST_PATH_IMAGE062
Figure 622302DEST_PATH_IMAGE063
(ii) a Connecting the key points corresponding to the target person to obtain a skeleton diagram corresponding to the target person, as shown in fig. 2; each key point in the skeleton map corresponds to a feature vector.
Thus, in this embodiment, a skeleton map corresponding to each target person is obtained according to the above process (the skeleton map includes each key point corresponding to the target person). In this embodiment, the nose key point corresponding to the target person is used as the central point of the corresponding skeleton map (i.e., the central point corresponding to the target person), so that the simulation environment and the real environment can be associated to determine the position of the target person in the target place area, thereby achieving the same recognition effect.
And step S3, obtaining the interaction domain corresponding to each target person in the target place area at the current acquisition time, wherein the interaction domain is an area within a preset range of the target person.
Considering that the distance between the target person and the person who interacts with the target person is not too far, the embodiment obtains the interaction domain corresponding to each target person in the target place area at the acquisition time, where the interaction domain is an area within a preset range of the target person (that is, the target person can interact with the person within the preset range); the process of acquiring the interaction domain corresponding to any target person in this embodiment specifically includes:
firstly, acquiring a position vector corresponding to a nose key point corresponding to the target person, a position vector corresponding to a right eye key point and a position vector corresponding to a left eye key point; in this embodiment, the position vector is used as a coordinate in a three-dimensional space, that is, the positions of the nose key point, the right eye key point and the left eye key point in the three-dimensional space are determined according to the position vector corresponding to the nose key point, the position vector corresponding to the right eye key point and the position vector corresponding to the left eye key point; because three points can uniquely determine a plane (that is, a plane can be determined by the nose key point, the right eye key point and the left eye key point in a three-dimensional space), in this embodiment, a normal vector of the plane is calculated according to a position vector corresponding to the nose key point corresponding to the target person, a position vector corresponding to the right eye key point and a position vector corresponding to the left eye key point, and a modulus of the normal vector is set to be 1 (that is, a normal vector corresponding to the target person); the present embodiment utilizes the normal vector to simulate the direction of the eye sight of the target person. The process of calculating the algorithm vector in this embodiment is the prior art, and will not be described herein again.
Then, in the embodiment, a central point (namely, a nose key point) corresponding to the target person is taken as a circular point, and a first preset length is taken as a radius to make a circle, so that a circular area corresponding to the target person is obtained; similarly, taking the central point corresponding to the target person as a dot, taking the normal vector corresponding to the target person as a central line, taking a preset angle as a fan-shaped angle, and taking a second preset length as a radius to be a fan-shaped area corresponding to the target person; taking the union region of the circular region and the fan-shaped region as an interaction region corresponding to the target person; considering that the interaction is generally performed by selecting a person closer or looking at the interaction, the visual field of the person is generally about 120 degrees, and therefore the preset angle should not be too different from 120 degrees. In this embodiment, values of the first preset length, the second preset length, and the preset angle are set according to actual needs.
So far, the embodiment obtains the interaction domain corresponding to each target person according to the above process.
And step S4, splicing the skeleton map corresponding to each target person and the skeleton maps of the interactive persons corresponding to the interactive domains corresponding to the target persons to obtain the interactive maps corresponding to the target persons.
For any interaction domain corresponding to the target person:
and counting the number of other target persons which do not include the target person and exist in the interaction domain corresponding to the target person, and recording the number as the interaction number.
If the number of interactions corresponding to the target person is not 0, namely other target persons exist in the interaction domain corresponding to the target person, marking other target persons existing in the interaction domain corresponding to the target person as interaction persons, namely interaction persons corresponding to the interaction domain corresponding to the target person; the number of the corresponding interactive personnel can be more than one and can also be 1. In this embodiment, according to the position vector of each key point corresponding to the target person and the position vector of each key point of each interactive person corresponding to the interactive domain corresponding to the target person, the skeleton map corresponding to the target person is respectively spliced with the skeleton maps corresponding to the interactive persons, so as to obtain the interactive map corresponding to the target person (if there are a plurality of interactive persons corresponding to the target person, there are a plurality of corresponding interactive maps; if there are 1 interactive person corresponding to the target person, there are only 1 corresponding interactive map); for any interactive person corresponding to the interactive domain corresponding to the target person:
firstly, the implementation connects the right-hand key point corresponding to the target person with the interactive person; respectively calculating the distance between the right-hand key point corresponding to the target person and each key point corresponding to the interactive person according to the position vector of the right-hand key point corresponding to the target person and the position vector of each key point corresponding to the interactive person, wherein the specific formula is as follows:
Figure 224185DEST_PATH_IMAGE064
wherein, the first and the second end of the pipe are connected with each other,
Figure 746433DEST_PATH_IMAGE065
the distance between the right-hand key point corresponding to the target person and the kth key point corresponding to the interactive person,
Figure 930421DEST_PATH_IMAGE066
the abscissa of the right-hand key point corresponding to the target person,
Figure 145501DEST_PATH_IMAGE067
is the ordinate of the right-hand key point corresponding to the target person,
Figure 500259DEST_PATH_IMAGE068
the depth value of the right-hand key point corresponding to the target person,
Figure 435986DEST_PATH_IMAGE069
the abscissa of the kth key point corresponding to the interactive person,
Figure 130272DEST_PATH_IMAGE070
the ordinate of the kth key point corresponding to the interactive person,
Figure 375309DEST_PATH_IMAGE071
the depth value of the k-th key point corresponding to the interactive person.
Therefore, the distance between the right-hand key point corresponding to the target person and each key point corresponding to the interactive person is obtained in the embodiment; and selecting the key point which is closest to the right-hand key point corresponding to the target person from the key points corresponding to the interactive person, and connecting the key point with the right-hand key point corresponding to the target person.
Similarly, connecting the key point which is closest to the left-hand key point corresponding to the target person in the key points corresponding to the interactive person with the left-hand key point of the target person; connecting the key point which is closest to the right-hand key point corresponding to the interactive person in the key points corresponding to the target person with the right-hand key point of the interactive person; connecting the key point which is closest to the left-hand key point corresponding to the interactive person in the key points corresponding to the target person with the left-hand key point of the interactive person; and further splicing the skeleton graph corresponding to the target person and the skeleton graph of the interactive person to obtain an interactive graph corresponding to the target person. For an interactive graph, the interactive graph comprises key points corresponding to two persons and feature vectors corresponding to the key points.
If the number of interactions corresponding to the target person is 0, that is, no interactive person exists in the interaction domain corresponding to the target person (that is, the target person does not have a corresponding interactive person), the embodiment constructs an occupancy skeleton map (that is, constructs a virtual interactive person), and takes the occupancy skeleton map as a skeleton map of the virtual interactive person; taking the virtual interactive personnel as the interactive personnel corresponding to the interactive domain corresponding to the target personnel; the position vectors of all key points in the space occupying skeleton map are (-1, -1, -1), and the body part vectors are unchanged. In this embodiment, the left-hand key point corresponding to the target person is connected with the left-hand key point of the interactive person corresponding to the interactive domain, and the right-hand key point corresponding to the target person is connected with the right-hand key point of the interactive person corresponding to the interactive domain, so as to obtain the interactive map corresponding to the target person; at this time, there is only one interactive map, and the number of interactions corresponding to the corresponding interactive person (i.e., the virtual interactive person) is 0.
And step S5, obtaining the action behaviors of the target person in each interactive graph according to the feature vectors corresponding to all the key points in the interactive graph corresponding to each target person and the trained action behavior recognition network.
In this embodiment, an interaction graph corresponding to each target person is obtained according to step S4 (the interaction graph corresponding to the target person may be one or multiple, and specifically relates to the number of the interactive persons). All key points in the interaction graph are sequenced, specifically: the interaction graph comprises key points of a target person and key points of an interaction person, and the key points corresponding to one person are ordered, in this embodiment, the key points of the interaction person are ordered next to the key points of the target person, that is, the key points corresponding to the target person are arranged to 19, and the key points of the interaction person are ordered from 20.
In order to identify the action behaviors of each target person in the area image, an action behavior identification network is constructed in the embodiment; in this embodiment, the interaction graph corresponding to each target person is input into the trained action behavior recognition network, and is output as the action behavior of each target person in the corresponding interaction graph. The action behavior recognition network comprises an attention mechanism and a multi-layer perceptron.
The action behavior recognition network firstly utilizes an attention mechanism to carry out aggregation processing on the feature vectors of the key points corresponding to the input interactive graphs, and after aggregation is completed, the target feature vectors of the key points corresponding to the input interactive graphs are obtained.
For any interaction graph corresponding to any input target person:
in this embodiment, according to the feature vector corresponding to each key point in the interactive map and the normal vector corresponding to the target person, the weight between each key point in the interactive map and each key point in the corresponding neighborhood set is calculated; the neighborhood set corresponding to the key point is a set formed by all key points connected with the corresponding key point in the interactive graph, namely:
Figure 623887DEST_PATH_IMAGE072
wherein the content of the first and second substances,
Figure 563637DEST_PATH_IMAGE002
for the weight between the ith and jth keypoints in the interaction graph (i.e. the weight of the jth keypoint to the ith keypoint),
Figure 971485DEST_PATH_IMAGE003
the feature vector corresponding to the ith key point in the interactive map,
Figure 528368DEST_PATH_IMAGE004
the feature vector corresponding to the jth key point in the interactive graph,
Figure 139609DEST_PATH_IMAGE005
is the normal vector corresponding to the target person,
Figure 807351DEST_PATH_IMAGE006
the feature vector of the central point corresponding to the target person (i.e. the feature vector corresponding to the nose key point corresponding to the target person),
Figure 272967DEST_PATH_IMAGE007
in order to perform the vector splicing operation,
Figure 876118DEST_PATH_IMAGE008
in order to activate the function(s),
Figure 99289DEST_PATH_IMAGE009
is the first similarity vector and the second similarity vector,
Figure 695355DEST_PATH_IMAGE010
is a second similarity degree vector, and is,
Figure 687582DEST_PATH_IMAGE011
is the transpose of the first similarity vector,
Figure 196055DEST_PATH_IMAGE012
is a transpose of the second similarity vector,
Figure 906522DEST_PATH_IMAGE013
is a first weight matrix of the weight data set,
Figure 243962DEST_PATH_IMAGE014
is a second weight matrix of the plurality of weight matrices,
Figure 252149DEST_PATH_IMAGE015
is an exponential function with e as the base,
Figure 321736DEST_PATH_IMAGE016
and a neighborhood set corresponding to the ith key point in the interaction graph.
According to the above-mentioned formula,
Figure 378554DEST_PATH_IMAGE073
is to measure two vectors
Figure 191789DEST_PATH_IMAGE074
And
Figure 768395DEST_PATH_IMAGE011
of (2), wherein
Figure 336779DEST_PATH_IMAGE009
Of dimension and
Figure 756260DEST_PATH_IMAGE074
are the same in dimension; in the same way
Figure 248552DEST_PATH_IMAGE075
Is to measure two vectors
Figure 397773DEST_PATH_IMAGE076
And
Figure 74742DEST_PATH_IMAGE012
of (2), wherein
Figure 856885DEST_PATH_IMAGE012
Dimension of and
Figure 339819DEST_PATH_IMAGE076
are the same. The first weight matrix and the second weight matrix in this embodiment are used for performing linear transformation to implement dimension reduction, that is, vectors are subjected to different linear transformations to implement dimension reduction processing; the values of the first similarity vector, the second similarity vector, the first weight matrix and the second weight matrix can be obtained by training of a neural network, and are not repeated herein.
According to the above process, the weight between each key point in the interactive graph and each key point in the corresponding neighborhood set can be obtained.
Further, performing aggregation operation on each key point in the interactive graph according to the weight between each key point in the interactive graph and each key point in the corresponding neighborhood set to obtain an aggregated and updated feature vector corresponding to each key point corresponding to the interactive graph; for the first aggregation operation on each key point in the interactive graph, the formula for obtaining the aggregated updated feature vector corresponding to each key point is as follows:
Figure 15651DEST_PATH_IMAGE017
wherein, the first and the second end of the pipe are connected with each other,
Figure 470378DEST_PATH_IMAGE018
for the aggregated updated feature vector corresponding to the ith key point in the interactive graph,
Figure 864451DEST_PATH_IMAGE019
the number of interaction of the person corresponding to the skeleton diagram to which the ith key point belongs in the interaction diagram,
Figure 947813DEST_PATH_IMAGE020
the number of interactions for the person corresponding to the skeleton diagram to which the jth key point belongs in the interaction diagram (that is, the ith key point and the jth key point may belong to the skeleton diagram corresponding to the target person or may belong to the skeleton diagram corresponding to the interaction person, and the persons include the target person and the interaction person).
Figure 822359DEST_PATH_IMAGE077
The function of (1) is to measure the degree of visual attention when
Figure 309972DEST_PATH_IMAGE077
The larger the number of people around the screen, the larger the number of people who touch the screen, the more likely the action is due to congestion and the more unintentional the action, and therefore the size of the weight is appropriately reduced.
After each key point in the interactive graph is updated once, repeating the steps, and performing second aggregation operation by using the aggregated updated feature vector corresponding to each key point so as to ensure that each key point fully senses neighborhood information and improve the identification accuracy, wherein the specific aggregation times can be set according to actual needs; and when the feature vectors corresponding to the key points in the interactive graph are updated through all aggregation, obtaining target feature vectors corresponding to the key points in the interactive graph (namely, the feature vectors updated through aggregation corresponding to the key points after the last aggregation are recorded as the target feature vectors).
In this embodiment, target feature vectors corresponding to the key points in the interactive map are spliced according to the arrangement sequence of the key points, so as to obtain a comprehensive vector corresponding to the interactive map; then, the comprehensive vector corresponding to the interaction diagram is input into a multi-layer perceptron MLP, and a recognition result of the action behavior of the target person in the interaction diagram is obtained (the action behavior of the target person in the interaction diagram is relative to the interaction person). The behavior actions comprise daily behaviors: walking, sitting up, talking, queuing, jogging and stooping; the non-daily behaviors are as follows: conflict, hug, crowd, climb, crawl, fall, injury, jump, squat, sprint, others; the specific action behaviors can be set according to actual needs.
In the embodiment, the loss function of the training action and behavior recognition network is a cross entropy loss function, and the RMSProp is used for an optimization algorithm; the specific training method is the prior art, and is not described herein again.
Thus, the embodiment obtains the action behaviors of each target person in the corresponding interaction graph in the area image so as to reflect the group behaviors in the target place area at the current acquisition time; if the emergency event is judged to occur through the action behavior of each target person, the management person needs to go to a specific position to control through a disordered scene; in order to enable the manager to reach the designated location more quickly, the route is planned in the embodiment, specifically:
according to the action behavior recognition result, marking the position where the emergency event occurs as a blocked area, and further obtaining an emergency event distribution map of the target place area; in the embodiment, the blocked area in the emergency distribution map is regarded as an impassable area and needs to be bypassed, so that the ant colony algorithm is used for processing the emergency distribution map and planning an optimal path to the blocked area; the ant colony algorithm has the advantages of a positive feedback mechanism, distributed computation, strong robustness and the like, the candidate solution construction process is similar to the path planning process, and the shortest foraging path can be found without prior knowledge; the ant colony algorithm is the prior art, and is not described herein again.
The method includes the steps that firstly, an area image corresponding to the current acquisition time in an area of a target place, key points corresponding to target people in the area image and feature vectors corresponding to the key points are obtained, and then a skeleton map corresponding to the target people in the area image is obtained according to the key points corresponding to the target people in the area image; then acquiring the interaction domain corresponding to each target person in the target place area at the current acquisition time, and splicing the skeleton map corresponding to each target person with the skeleton maps of the interaction persons corresponding to the interaction domains corresponding to each target person to obtain the interaction maps corresponding to each target person; and finally, according to the feature vectors corresponding to all key points in the interactive graphs corresponding to the target personnel and the trained action behavior recognition network, obtaining the action behaviors of the target personnel in the interactive graphs. The embodiment identifies the action behaviors of each person in the monitoring in an automatic mode, overcomes the problems of strong subjectivity, easy fatigue and the like in a manual monitoring mode, and improves the accuracy of identifying the behaviors of the persons.
It should be noted that: the above description is only for the purpose of illustrating the preferred embodiments of the present invention and is not to be construed as limiting the invention, and any modifications, equivalents, improvements and the like that fall within the spirit and principle of the present invention are intended to be included therein.

Claims (7)

1. A personnel group behavior monitoring method based on video analysis is characterized by comprising the following steps:
acquiring a region image corresponding to the current acquisition time in a target place region;
acquiring each key point corresponding to each target person in the area image and a feature vector corresponding to each key point; obtaining a skeleton map corresponding to each target person in the region image according to each key point corresponding to each target person in the region image; the characteristic vector is obtained by splicing a position vector corresponding to the key point and a body part vector;
acquiring an interaction domain corresponding to each target person in a target site area at the current acquisition time, wherein the interaction domain is an area within a preset range of the target person;
splicing the skeleton map corresponding to each target person and the skeleton maps of the interactive persons corresponding to the interactive domains corresponding to the target persons to obtain interactive maps corresponding to the target persons;
according to the feature vectors corresponding to all key points in the interactive graphs corresponding to the target personnel and the trained action behavior recognition network, obtaining the action behaviors of the target personnel in the interactive graphs;
the obtaining of the action behaviors of the target person in each interactive graph includes:
for any interaction graph corresponding to any target person:
sequencing all key points in the interactive graph according to a preset sequence;
calculating the weight between each key point in the interactive map and each key point in the corresponding neighborhood set according to the feature vector corresponding to each key point in the interactive map and the normal vector corresponding to the target person; the corresponding neighborhood set comprises all key points connected with the corresponding key points in the interactive graph;
performing preset times of aggregation operation on each key point in the interactive graph according to the weight between each key point in the interactive graph and each key point in the corresponding neighborhood set to obtain a target feature vector corresponding to each key point in the interactive graph;
splicing target characteristic vectors corresponding to the key points in the interactive map according to the arrangement sequence of the key points to obtain comprehensive vectors corresponding to the interactive map;
inputting the comprehensive vector corresponding to the interactive map into a multilayer perceptron to obtain the action behavior of the target person in the interactive map;
the formula for calculating the weight between each key point in the interactive graph and each key point in the corresponding neighborhood set is as follows:
Figure 488008DEST_PATH_IMAGE001
wherein, the first and the second end of the pipe are connected with each other,
Figure 292016DEST_PATH_IMAGE002
for the weight of the jth key point to the ith key point in the interactive graph,
Figure 224200DEST_PATH_IMAGE003
The feature vector corresponding to the ith key point in the interaction graph,
Figure 506276DEST_PATH_IMAGE004
the feature vector corresponding to the jth key point in the interaction graph,
Figure 207516DEST_PATH_IMAGE005
is the normal vector corresponding to the target person,
Figure 916846DEST_PATH_IMAGE006
the feature vector of the central point corresponding to the target person,
Figure 805168DEST_PATH_IMAGE007
in order to perform the vector splicing operation,
Figure 156515DEST_PATH_IMAGE008
in order to activate the function(s),
Figure 977840DEST_PATH_IMAGE009
is the first similarity vector and the second similarity vector,
Figure 123651DEST_PATH_IMAGE010
is the second similarity vector, and is the second similarity vector,
Figure 764848DEST_PATH_IMAGE011
is a transpose of the first similarity vector,
Figure 123148DEST_PATH_IMAGE012
is a transpose of the second similarity vector,
Figure 64559DEST_PATH_IMAGE013
is a first weight matrix of the weight data set,
Figure 375411DEST_PATH_IMAGE014
is a second weight matrix that is a function of,
Figure 503904DEST_PATH_IMAGE015
is an exponential function with e as the base,
Figure 931475DEST_PATH_IMAGE016
and (4) a neighborhood set corresponding to the ith key point in the interactive graph.
2. The method for monitoring the group behaviors of people based on video analysis according to claim 1, wherein the obtaining of the key points corresponding to the target people in the area image and the feature vectors corresponding to the key points comprises:
processing the area image by using a key point detection network to obtain each key point corresponding to each target person in the area image;
taking the vertex of the lower left corner of the area image as a coordinate origin, taking the horizontal direction as an x axis and taking the vertical direction as a y axis, and obtaining a position vector corresponding to each key point corresponding to each target person in the area image; the position vector comprises an abscissa corresponding to the key point, an ordinate corresponding to the key point and a depth value corresponding to the key point, and the depth value is obtained according to the area image;
for any target person: performing One-Hot coding on each key point corresponding to the target person to obtain body part vectors corresponding to the key points corresponding to the target person; and splicing the position vector and the body part vector corresponding to each key point corresponding to the target person to obtain the feature vector corresponding to each key point corresponding to the target person.
3. The method for monitoring the group behaviors of people based on video analysis according to claim 1, wherein the key points corresponding to the target people in the region image are connected according to a preset connection rule to obtain a skeleton map corresponding to the target people.
4. The method for monitoring the group behaviors of people based on video analysis according to claim 1, wherein the step of obtaining the interaction domain corresponding to each target person in the target site area at the current acquisition time comprises the following steps:
for any target person:
calculating to obtain a normal vector corresponding to the target person according to position vectors corresponding to the nose key point, the right-eye key point and the left-eye key point in the key points corresponding to the target person; the normal vector corresponding to the target person is as follows: calculating a plane normal vector constructed in the three-dimensional space by the nose key point, the right eye key point and the left eye key point by taking position vectors corresponding to the nose key point, the right eye key point and the left eye key point as coordinates in the three-dimensional space, and taking the plane normal vector as a normal vector corresponding to the target person;
taking a central point corresponding to the target person as a circular point in the target place area, and taking a first preset length as a radius to make a circle, so as to obtain a circular area corresponding to the target person; the central point corresponding to the target person is a nose key point corresponding to the target person;
taking a central point corresponding to the target person as a dot point, a normal vector corresponding to the target person as a central line, a preset angle as a fan-shaped angle and a second preset length as a radius in the target place area to obtain a fan-shaped area corresponding to the target person;
and taking the union region of the circular region and the fan-shaped region as an interaction region corresponding to the target person.
5. The method for monitoring the group behaviors of people based on video analysis according to claim 1, wherein the obtaining of the interaction graph corresponding to each target person comprises:
for any target person:
counting the number of other target persons not including the target person in the interaction domain corresponding to the target person, and recording as the interaction number;
if the number of the interactions corresponding to the target person is not 0, marking other target persons existing in the interaction domain corresponding to the target person as interaction persons; for any interactive person corresponding to the interactive domain corresponding to the target person: splicing the skeleton map corresponding to the target person and the skeleton map corresponding to the interactive person according to the position vector of each key point corresponding to the target person and the position vector of each key point corresponding to the interactive person to obtain an interactive map corresponding to the target person;
if the number of the interactions corresponding to the target person is 0, constructing an occupancy skeleton diagram; the space-occupying skeleton map is a skeleton map of a virtual interactive person, and position vectors of all key points in the space-occupying skeleton map are (-1, -1, -1); taking the virtual interaction personnel as interaction personnel corresponding to the interaction domain corresponding to the target personnel; connecting a left-hand key point corresponding to the target person with a left-hand key point of an interactive person corresponding to the interactive domain, connecting a right-hand key point corresponding to the target person with a right-hand key point of the interactive person corresponding to the interactive domain, and splicing the skeleton diagram corresponding to the target person with the occupancy skeleton diagram to obtain an interactive diagram corresponding to the target person; the right-hand key point and the left-hand key point are two key points in the key points.
6. The method for monitoring the group behaviors of people based on video analysis according to claim 5, wherein obtaining the interaction graph corresponding to the target person comprises:
connecting a key point which is closest to the right-hand key point corresponding to the target person in the key points corresponding to the interactive persons with the right-hand key point of the target person; connecting a key point which is closest to the left-hand key point corresponding to the target person in the key points corresponding to the interactive persons with the left-hand key point of the target person; connecting a key point which is closest to the right-hand key point corresponding to the interactive person in the key points corresponding to the target person with the right-hand key point of the interactive person; connecting a key point which is closest to the left-hand key point corresponding to the interactive person in the key points corresponding to the target person with the left-hand key point of the interactive person; splicing the skeleton graph corresponding to the target person and the skeleton graph corresponding to the interactive person to obtain an interactive graph corresponding to the target person;
the distance is calculated according to the position vector.
7. The method for monitoring the behaviors of the people group based on the video analysis according to claim 1, wherein the obtaining of the target feature vector corresponding to each key point in the interactive map comprises:
the formula for performing the first aggregation operation on each key point in the interactive graph is as follows:
Figure 196234DEST_PATH_IMAGE017
wherein the content of the first and second substances,
Figure 418268DEST_PATH_IMAGE018
for the aggregated updated feature vector corresponding to the ith key point in the interactive graph,
Figure 34057DEST_PATH_IMAGE019
the number of interaction of the person corresponding to the skeleton diagram to which the ith key point belongs in the interaction diagram,
Figure 265318DEST_PATH_IMAGE020
the interaction number of the person corresponding to the skeleton diagram to which the jth key point belongs in the interaction diagram is obtained; the personnel comprise the target personnel and the interactive personnel; the interaction quantity is the quantity of other target persons which do not comprise the target person and exist in the interaction domain corresponding to the target person;
and by analogy, continuously executing the aggregation operation by using the aggregated and updated feature vectors corresponding to the key points in the interactive graph, and recording the aggregated and updated feature vectors corresponding to the key points after the last aggregation as target feature vectors.
CN202210793997.5A 2022-07-07 2022-07-07 Personnel group behavior monitoring method based on video analysis Active CN114863352B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202210793997.5A CN114863352B (en) 2022-07-07 2022-07-07 Personnel group behavior monitoring method based on video analysis

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202210793997.5A CN114863352B (en) 2022-07-07 2022-07-07 Personnel group behavior monitoring method based on video analysis

Publications (2)

Publication Number Publication Date
CN114863352A CN114863352A (en) 2022-08-05
CN114863352B true CN114863352B (en) 2022-09-30

Family

ID=82626219

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202210793997.5A Active CN114863352B (en) 2022-07-07 2022-07-07 Personnel group behavior monitoring method based on video analysis

Country Status (1)

Country Link
CN (1) CN114863352B (en)

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115150818B (en) * 2022-09-05 2022-11-04 光谷技术有限公司 Communication transmission encryption method based on artificial intelligence

Family Cites Families (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111079578A (en) * 2019-12-02 2020-04-28 海信集团有限公司 Behavior detection method and device
CN112633224B (en) * 2020-12-30 2024-03-26 深圳云天励飞技术股份有限公司 Social relation recognition method and device, electronic equipment and storage medium
CN113128383A (en) * 2021-04-07 2021-07-16 杭州海宴科技有限公司 Recognition method for campus student cheating behavior
CN114155595A (en) * 2021-09-30 2022-03-08 深圳市爱深盈通信息技术有限公司 Behavior detection monitoring method, intelligent camera and intelligent monitoring system
CN113673489B (en) * 2021-10-21 2022-04-08 之江实验室 Video group behavior identification method based on cascade Transformer
CN114241372A (en) * 2021-12-09 2022-03-25 江苏和正特种装备有限公司 Target identification method applied to sector-scan splicing

Also Published As

Publication number Publication date
CN114863352A (en) 2022-08-05

Similar Documents

Publication Publication Date Title
CN111860205B (en) Forest fire evaluation method based on multisource remote sensing images and grids and storage medium
CN111639825B (en) Forest fire indication escape path method and system based on A-Star algorithm
CN107292989A (en) High voltage power transmission cruising inspection system based on 3DGIS technologies
CN113240249B (en) Urban engineering quality intelligent evaluation method and system based on unmanned aerial vehicle augmented reality
Dhall et al. A survey on systematic approaches in managing forest fires
CN114863352B (en) Personnel group behavior monitoring method based on video analysis
CN110414400A (en) A kind of construction site safety cap wearing automatic testing method and system
CN114582030B (en) Behavior recognition method based on service robot
CN114155595A (en) Behavior detection monitoring method, intelligent camera and intelligent monitoring system
Sanchez-Fernandez et al. VPP: visibility-based path planning heuristic for monitoring large regions of complex terrain using a UAV onboard camera
CN115797864A (en) Safety management system applied to smart community
CN115567690A (en) Intelligent monitoring system capable of automatically identifying dangerous points of field operation
CN110276379A (en) A kind of the condition of a disaster information rapid extracting method based on video image analysis
Usman et al. A social distancing index: Evaluating navigational policies on human proximity using crowd simulations
CN112991544A (en) Group evacuation behavior simulation method based on panoramic image modeling
CN112597802A (en) Pedestrian motion simulation method based on visual perception network deep learning
CN113361968B (en) Power grid infrastructure worker safety risk assessment method based on artificial intelligence and big data
CN108304809A (en) The damaged appraisal procedure of near real-time based on aerial images after shake
Koutamanis Multilevel analysis of fire escape routes in a virtual environment
Tadokoro et al. The robocup rescue project: a multiagent approach to the disaster mitigation problem
CN113743015A (en) Fire scene data acquisition method, medium and electronic device
CN113096479A (en) Fire drill virtual training method
CN116913147A (en) Fire-fighting simulation training scheme design method and system
CN111723741A (en) Temporary fence movement detection alarm system based on visual analysis
CN111191511A (en) Method and system for identifying dynamic real-time behaviors of prisons

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant