WO2022041953A1 - Behavior recognition method and apparatus, and storage medium - Google Patents

Behavior recognition method and apparatus, and storage medium Download PDF

Info

Publication number
WO2022041953A1
WO2022041953A1 PCT/CN2021/100379 CN2021100379W WO2022041953A1 WO 2022041953 A1 WO2022041953 A1 WO 2022041953A1 CN 2021100379 W CN2021100379 W CN 2021100379W WO 2022041953 A1 WO2022041953 A1 WO 2022041953A1
Authority
WO
WIPO (PCT)
Prior art keywords
height
user
identified
posture
joint point
Prior art date
Application number
PCT/CN2021/100379
Other languages
French (fr)
Chinese (zh)
Inventor
慕晨
黄伟
郭红星
王春利
梁敬柏
Original Assignee
中兴通讯股份有限公司
长安大学
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 中兴通讯股份有限公司, 长安大学 filed Critical 中兴通讯股份有限公司
Publication of WO2022041953A1 publication Critical patent/WO2022041953A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T5/00Image enhancement or restoration
    • G06T5/70Denoising; Smoothing
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/11Region-based segmentation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/13Edge detection
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/194Segmentation; Edge detection involving foreground-background segmentation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/70Determining position or orientation of objects or cameras
    • G06T7/73Determining position or orientation of objects or cameras using feature-based methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/20Movements or behaviour, e.g. gesture recognition
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10016Video; Image sequence
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20024Filtering details
    • G06T2207/20032Median filtering
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30196Human being; Person

Definitions

  • the embodiments of the present application relate to the technical field of image processing, and in particular, to a behavior recognition method, device, and storage medium.
  • the commonly used human behavior recognition method is the behavior recognition method based on wearable devices.
  • the wearable device-based behavior recognition method collects the motion data of the human body through the motion sensor worn on the person.
  • the disadvantage is that the user needs to wear the sensor, which is not convenient and the recognition accuracy is low.
  • An embodiment of the present application provides a method for behavior recognition, including: acquiring a video image frame corresponding to a user to be recognized, wherein the video image frame includes a depth image and a skeleton image; The gesture characteristic parameter corresponding to the user to be recognized is determined; the current behavior state of the user to be recognized is determined according to the gesture characteristic parameter corresponding to the user to be recognized.
  • An embodiment of the present application further provides a behavior recognition device, including: a processor and a memory; the memory is used to store a program; the processor is used to execute the program and implement the above behavior recognition when the program is executed method.
  • Embodiments of the present application further provide a storage medium for readable storage, where the storage medium stores one or more programs, and the one or more programs can be executed by one or more processors to implement the following: The above-mentioned behavior recognition method.
  • FIG. 1 is a schematic structural diagram of a behavior recognition system provided by an embodiment of the present application.
  • FIG. 2 is a schematic diagram of performing part segmentation on a human body region provided by an embodiment of the present application
  • FIG. 3 is a schematic structural diagram of a behavior recognition device provided by an embodiment of the present application.
  • FIG. 4 is a schematic flowchart of a behavior recognition method provided by an embodiment of the present application.
  • FIG. 5 is a schematic diagram of a video image frame provided by an embodiment of the present application.
  • FIG. 6 is a schematic diagram of a human body region and a human body center of gravity provided by an embodiment of the present application;
  • FIG. 7 is a schematic diagram of determining a head height corresponding to a user to be identified according to an embodiment of the present application.
  • FIG. 8 is a schematic diagram of determining the posture height of a user to be recognized provided by an embodiment of the present application.
  • Fig. 9 is a schematic block diagram of the sub-steps of determining the current behavioral state of the user to be identified in Fig. 4;
  • FIG. 10 is a schematic flowchart of the sub-steps of judging whether the user to be identified has fallen in FIG. 9 .
  • Embodiments of the present application provide a behavior recognition method, device, and storage medium, so as to at least solve the problems of low accuracy and poor convenience in recognizing human behavior in the related art.
  • An embodiment of the present application provides a behavior recognition method, apparatus, system, and storage medium.
  • the behavior recognition method can be applied to a behavior recognition device to realize the determination of the posture feature parameters according to the depth image and the skeleton image, and to determine the current behavior parameters of the user to be recognized through the posture feature parameters, which can more conveniently and accurately identify the user to be recognized. The corresponding current behavior state.
  • the behavior recognition device may include a server or a terminal.
  • the server may be an independent server or a server cluster;
  • the terminal may be an electronic device such as a smart phone, a tablet computer, a notebook computer, and a desktop computer.
  • FIG. 1 is a schematic structural diagram of a behavior recognition system provided by an embodiment of the present application.
  • the behavior recognition system includes a behavior recognition device 10 and a photographing device 20 .
  • the behavior recognition device 10 may be connected with the photographing device 20 in wired or wireless communication.
  • the photographing device 20 is used to collect video image frames including the user to be recognized, and the behavior recognition device 10 is used to perform image processing on the video image frames collected by the photographing device 20 to determine the current behavior state corresponding to the user to be recognized.
  • the photographing device 20 may collect video image frames including the corresponding user to be identified, and perform human body recognition, human body part recognition, and skeletal joint point positioning on the video image frames, and the processed video image frames include depth images. and skeleton images.
  • the behavior recognition device 10 can acquire the video image frames processed by the shooting device 20, determine the body center of gravity information corresponding to the user to be recognized according to the depth image, and determine the posture height and head height corresponding to the user to be recognized according to the skeleton image; then, according to the body center of gravity information, posture height, and head height to determine the current behavioral state of the user to be identified.
  • the photographing apparatus 20 may include an electronic device such as a 3D camera, such as a somatosensory that may capture video image frames.
  • the somatosensory may be used to collect video image frames including the user to be identified.
  • the somatosensory may include a depth camera, a color camera and a light source emitter, and the somatosensory may acquire depth images, color images and three-dimensional data information of the background space.
  • the somatosensory can acquire depth information.
  • the working principle of obtaining depth information includes: projecting the light emitted by the light source emitter into the real scene. Since the emitted light will change due to the different surface shapes of the object, this light can be collected and encoded. The distance difference between each pixel in the scene and the depth camera can be obtained, and then the position and depth information of the object can be obtained.
  • the body sensor performs human body recognition on the video image frame; for example, the background and the person in the video image frame are segmented according to a preset segmentation strategy, so as to determine the human body region or the human body contour corresponding to the user to be identified. information, and the obtained depth image includes the body region or body contour information corresponding to the user to be identified.
  • the body sensor performs human body part recognition on the depth image obtained by human body recognition.
  • FIG. 2 is a schematic diagram of part segmentation of the human body area; part segmentation is performed on the human body area in the depth image to obtain multiple part images, such as head, arms, legs, limbs and torso, etc. Part images; perform feature value classification and matching on multiple part images to determine the body part corresponding to each part image.
  • the somatosensory performs skeletal joint point positioning on the video image frame to obtain a skeletal image.
  • the identified body parts are added to the virtual skeleton model, and adjusted according to the position information of the body parts to obtain a skeleton image including a plurality of joint points.
  • the skeleton image may include, but is not limited to, joint points such as head joint points, neck joint points, knee joint points, elbow joint points, hip joint points, or ankle joint points.
  • the behavior recognition device 10 may include a processor 11 and a memory 12, wherein the processor 11 and the memory 12 may be connected through a bus, such as an integrated circuit bus (Inter-integrated Circuit, referred to as: IIC) and other suitable the bus.
  • IIC Inter-integrated Circuit
  • the memory 12 may include a non-volatile storage medium and an internal memory.
  • the nonvolatile storage medium can store operating systems and computer programs.
  • the computer program includes program instructions that, when executed, can cause the processor to perform any behavior recognition method.
  • the processor 11 is used to provide computing and control capabilities to support the operation of the entire behavior recognition device 10 .
  • the processor 11 may be a central processing unit (Central Processing Unit, CPU), and the processor may also be other general-purpose processors, digital signal processors (Digital Signal Processors, DSP), application specific integrated circuits (application specific integrated circuits) circuit, ASIC), Field-Programmable Gate Array (FPGA) or other programmable logic devices, discrete gate or transistor logic devices, discrete hardware components, etc.
  • a general purpose processor may be a microprocessor or the processor may be any conventional processor or the like.
  • FIG. 4 is a schematic flowchart of a behavior recognition method provided by an embodiment of the present application.
  • the behavior recognition method can be applied to a behavior recognition device to realize the determination of the posture feature parameters according to the depth image and the skeleton image, and to determine the current behavior parameters of the user to be recognized through the posture feature parameters, which can more conveniently and accurately identify the corresponding user to be recognized.
  • the behavior recognition method includes steps S10 to S30.
  • Step S10 Obtain a video image frame corresponding to the user to be identified, wherein the video image frame includes a depth image and a skeleton image.
  • the video image frame corresponding to the user to be identified may be collected by the somatosensory sensor.
  • the somatosensory sensor can be installed indoors to monitor the indoor environment and people in real time.
  • the somatosensory can collect video image frames corresponding to the user to be recognized through a built-in software development kit, and perform human body recognition, human body part recognition, and skeletal joint point positioning on the video image frames.
  • Image frames include depth images and skeletal images.
  • a processed video image frame sent by a body sensor may be received, wherein the video image frame includes a depth image and a skeleton image.
  • the depth image and the skeleton image overlap each other in the same image, exemplarily, as shown in FIG. 5 , which is a video image frame.
  • the depth image includes the human body region corresponding to the user to be identified, and the skeleton image includes bone information corresponding to the user to be identified.
  • the skeleton information may include different joint points and connection relationships between the joint points, and the like.
  • obtaining the depth image and the skeleton image corresponding to the user to be recognized can improve the accuracy of determining the posture feature parameters corresponding to the user to be recognized according to the depth image and the skeleton image.
  • Step S20 Determine, according to the depth image and the skeleton image, a posture feature parameter corresponding to the user to be recognized.
  • the posture feature parameters may include information on the body's center of gravity, posture height, and head height.
  • the posture height refers to the height of the body of the user to be identified in different postures. Exemplarily, a person's body height when standing, body height when sitting cross-legged, body height when sitting on a chair, and the like.
  • the information on the center of gravity of the human body may include, but is not limited to, the center of gravity point, the coordinates of the center of gravity, and the speed of falling of the center of gravity, and the like.
  • the head height refers to the height of the entire head of the user to be identified.
  • the center of gravity information of the human body corresponding to the user to be identified may be determined according to the depth image, and the posture height and the height of the head corresponding to the user to be identified may be determined according to the skeleton image.
  • the method may further include: performing format conversion on the depth data in the initial depth image according to a preset data format, to obtain the depth after the format conversion image.
  • the preset data format may be Mat format.
  • the initial depth image refers to the depth image in the video image frame obtained from the somatosensory sensor.
  • the recognition process of determining the current behavior state of the user to be recognized according to the depth image and the skeleton image is completed on the basis of the computer vision platform.
  • Format conversion of the depth data in the original depth image is required.
  • the depth data in the original depth image is format-converted, and the depth data in the format-converted depth image is in the Mat format.
  • the method may further include: smoothing the initial skeleton image according to a preset smoothing strategy to obtain a smoothed skeleton image.
  • the initial skeleton image refers to the skeleton image in the video image frame acquired from the somatosensory sensor.
  • the posture height and head height corresponding to the user to be recognized need to be determined according to the skeleton image; during the recognition process, the height, movement and real-time performance of the user to be recognized are highly recognized. It is required that if the skeleton image is not smoothed, it may cause the computer vision platform to shake or even crash.
  • smoothing is also called filtering. By smoothing the skeleton image, noise or distortion in the skeleton image can be reduced.
  • the preset smoothing processing strategy may include, but is not limited to, a mean filtering algorithm, a median filtering algorithm, a Gaussian filtering algorithm, a bilateral filtering algorithm, and the like.
  • the initial skeleton image is smoothed to obtain the smoothed skeleton image.
  • the initial skeleton image is smoothed to obtain the smoothed skeleton image.
  • the depth image corresponding to the user to be recognized may be determined according to the format-converted depth image. Human body center of gravity information.
  • the information on the center of gravity of the human body may include a center of gravity point of the human body; as shown in FIG. 6 , FIG. 6 is a schematic diagram of the area of the human body and the center of gravity point of the human body.
  • determining the center of gravity information of the human body corresponding to the user to be identified according to the depth image may include: acquiring the total number of pixels in the human body region in the format-converted depth image, and acquiring the corresponding pixel points in the human body region.
  • the abscissa and the ordinate determine the sum of the abscissas and ordinates corresponding to all the pixels in the human body area; according to the total number of pixels and the sum of the abscissas, determine the mean of the abscissas corresponding to all the pixels, and according to the pixel
  • the sum of the total number of points and the ordinate is to determine the mean of the ordinates corresponding to all the pixels; the mean of the abscissas corresponding to all the pixels is taken as the abscissa of the center of gravity of the human body, and the mean of the ordinates corresponding to all the pixels is taken as the human body
  • the ordinate of the center of gravity to get the center of gravity of the human body in the depth image.
  • a rectangular coordinate system can be established in the depth image, the total number of pixels in the human body region in the depth image is obtained, and the abscissa and ordinate corresponding to all the pixels in the human body region are determined.
  • the center of gravity coordinates corresponding to the center of gravity of the human body may be calculated according to the center of gravity calculation formula.
  • the calculation formula of the center of gravity is as follows:
  • (X 0 , Y 0 ) represents the barycentric coordinates; s represents the total number of pixels in the human body area; i represents the ith pixel in the human body area; x i represents the abscissa corresponding to the ith pixel; y i Indicates the ordinate corresponding to the i-th pixel.
  • the mean value X 0 of the abscissa and the mean value Y 0 of the ordinate corresponding to all the pixel points in the human body area can be determined. Then, the mean value X 0 is taken as the abscissa of the human body's center of gravity, and the mean value Y 0 is taken as the ordinate of the human body's center of gravity, to obtain the barycentric coordinates (X 0 , Y 0 ) corresponding to the human body's center of gravity in the depth image.
  • the center of gravity of the user to be identified can be more accurately determined.
  • the posture height corresponding to the user to be recognized and head height after smoothing an initial skeleton image according to a preset smoothing strategy to obtain a smoothed skeleton image, the posture height corresponding to the user to be recognized and head height.
  • the joint point information in the smoothed skeleton image is extracted, and the posture height and the head height corresponding to the user to be recognized are determined according to the joint point information.
  • the joint point information includes joint point coordinates. For example, the joint point coordinates corresponding to the head joint point, the joint point coordinates corresponding to the neck joint point, and so on.
  • extracting joint point information in the smoothed bone image may include: establishing a two-dimensional coordinate system in the bone image and determining a coordinate origin; according to the two-dimensional coordinate system of each joint point in the bone image In the position, determine the joint point coordinates of each joint point. For example, the joint point coordinates corresponding to the head joint point are (X 1 , Y 1 ), and the joint point coordinates corresponding to the neck joint point are (X 2 , Y 2 ).
  • the method may further include: acquiring the head joint points in the skeleton image; when the head joint points are located in the human body region in the depth image Inside, the posture height and head height corresponding to the user to be recognized are determined according to the joint point information.
  • the head joint points can be positioned according to the connection relationship between the joint points in the skeleton image to determine the specific position of the head joint point in the skeleton image and whether the head joint point is located in the depth image. in the human body area.
  • the head of the user to be identified is not at the highest position or the head is blocked by other parts of the body, the head height cannot be accurately determined through the joint point information in the skeleton image.
  • the head joint point in the skeleton image when the head joint point in the skeleton image is located in the human body area in the depth image, it means that the head of the user to be identified is at the highest position or the head is not blocked, at this time, it can be accurately determined according to the joint point information.
  • the head joint points in the skeleton image are not within the human body area in the depth image, it means that the head of the user to be identified is not in the highest position or the head is occluded. At this time, the joint point information cannot be accurately determined. Identify the user's corresponding posture height and head height.
  • the accuracy of determining the posture height and the head height corresponding to the user to be recognized can be improved.
  • the posture height and the head height corresponding to the user to be identified may be determined according to the joint point information.
  • determining the height of the head corresponding to the user to be identified according to the joint point information may include: determining the head joint point and the neck joint point according to the joint point coordinates corresponding to the head joint point and the joint point coordinates corresponding to the neck joint point. The first height difference between the neck joint points; according to the product of the preset height ratio and the first height difference, the height of the head corresponding to the user to be identified is obtained.
  • FIG. 7 is a schematic diagram of determining the height of the head corresponding to the user to be identified.
  • the head The first height difference h 1 between the neck joint point and the neck joint point is
  • the head height can be obtained more accurately, which improves the accuracy of subsequent judgment of the current behavior state of the user to be identified.
  • determining the posture height corresponding to the user to be recognized according to the joint point information may include: determining the highest joint point and the lowest joint point in the skeleton image; according to the joint point coordinates corresponding to the highest joint point and the lowest joint point The joint point coordinates are used to determine the second height difference between the highest joint point and the lowest joint point, and the second height difference is used as the posture height corresponding to the user to be identified.
  • the highest joint point in the skeleton image is the head joint point, and the lowest joint point is the ankle joint point.
  • the highest joint point in the skeleton image is the head joint point, and the lowest joint point is the hip joint point.
  • the highest joint point in the skeleton image is the head joint point, and the lowest joint point is the ankle joint point; the joint point coordinates corresponding to the head joint point and the foot joint point can be used according to The joint point coordinates corresponding to the wrist joint point determine the second height difference between the head joint point and the ankle joint point, that is, the posture height of the user to be identified is obtained.
  • the posture height is the height of the body when the user to be recognized is standing.
  • FIG. 8 is a schematic diagram of determining the gesture height of the user to be recognized.
  • the highest joint point in the skeletal image is the head joint point, and the lowest joint point is the hip joint point; the joint point coordinates corresponding to the head joint point and the hip joint The joint point coordinates corresponding to the points determine the second height difference between the head joint point and the hip joint point, that is, the posture height of the user to be recognized is obtained.
  • the posture height is the height of the body when the user to be identified sits cross-legged.
  • the head joint point is The second height difference between it and the hip joint point is
  • the posture height corresponding to the user to be recognized can be more truly and accurately reflected.
  • Step S30 Determine the current behavior state of the to-be-identified user according to the gesture characteristic parameter corresponding to the to-be-identified user.
  • determining the current behavioral state of the to-be-recognized user according to the posture feature parameters corresponding to the to-be-recognized user may include: determining the current behavioral state of the to-be-recognized user according to the body's center of gravity information, posture height, and head height.
  • the current behavioral state of the user to be identified is comprehensively determined according to parameters such as the user's body center of gravity information, posture height, and head height, and no additional wearable equipment is required, and the to-be-identified user can be more conveniently and accurately determined.
  • the current behavioral state of the user is comprehensively determined according to parameters such as the user's body center of gravity information, posture height, and head height, and no additional wearable equipment is required, and the to-be-identified user can be more conveniently and accurately determined.
  • determining the current behavioral state of the user to be identified according to the body's center of gravity information, posture height, and head height may include the following step S31 or step S32.
  • Step S31 If the ratio of the posture height to the head height is within a preset ratio range, determine the posture corresponding to the to-be-identified user according to the preset correspondence between the ratio range and the posture type Types of.
  • the preset ratio range may include a first ratio range, a second ratio range and a third ratio range.
  • posture types may include, but are not limited to, standing, cross-sitting, sitting, kneeling, and the like.
  • the first ratio range is the ratio of the standing height of the human body to the height of the head
  • the second ratio range is the ratio of the sitting height of the human body to the height of the head
  • the third ratio range is the sitting height of the human body and the head height The ratio of the height of the waist or the height of the kneeling position to the height of the head.
  • the first ratio range, the second ratio range, and the third ratio range may be determined from the ratio measurement data between the body height and the head height.
  • the measurement data in Table 1 are obtained by measuring a preset number of testers.
  • the measurement data corresponding to the 1% column represents the measurement data corresponding to 1% of the testers;
  • the measurement data corresponding to the 99% column represents the measurement data corresponding to 99% of the testers.
  • A is the average of the ratio of height to head height for all people in each percentile; H is the ratio of the posture height of 99% of testers to the head height of 1% of testers; L is the ratio of 1% of testers The ratio of the posture height to the head height of 99% of the subjects; P represents the average of all ratios in each row; H/H represents the ratio of standing height to head height; S/H represents the ratio of sitting height to head height ; C/H represents the ratio of the height of the cross sitting posture to the height of the head.
  • the average ratio of the standing height to the head height is 7.551 for the males and 7.322 for the females; the average of the ratio of the sitting height to the head height, Males were 5.953 and females were 5.726; the average ratio of the height of the cross-sitting posture to the head height was 4.084 for males and 3.981 for females.
  • the Chinese national standard GB 10000-88 adult body shape and related standardization literature show that when the ratio is greater than or equal to 6.5, it can be determined that the human body is in a standing state; when the ratio is [3.3, 4.5], it can be determined that the human body is in a cross-sitting state; When the ratio is [5,6], it can be determined that the human body is in a sitting or kneeling state. Therefore, in the embodiment of the present invention, the first ratio range is set to [6.5, + ⁇ ], the second ratio range is [3.3, 4.5], and the third ratio range is [5, 6].
  • the preset correspondence between the ratio range and the gesture type may be as shown in Table 2.
  • the gesture type corresponding to the user to be recognized may be determined according to a preset correspondence between the ratio range and the gesture type.
  • the posture type corresponding to the user to be recognized is a standing posture.
  • the posture type corresponding to the user to be identified is the cross-sitting posture.
  • the ratio of the posture height to the head height is in the third ratio range, then according to the positional relationship between the lowest joint point corresponding to the to-be-identified user and the head joint point, it is determined that the posture type corresponding to the to-be-identified user is: Sitting or kneeling. It can be understood that, when the ratio of the posture height to the head height corresponding to the user to be identified is in the third ratio range, since the posture type corresponding to the third ratio range is a sitting posture or a kneeling posture, it is necessary to further determine the posture of the user to be identified. Posture type.
  • the lowest joint point in the skeleton image is the ankle joint point, and the ankle joint point and the head joint point are not on the same vertical line; if If the user to be identified is in a kneeling posture, the lowest joint point in the skeleton image is the knee joint point, and at this time the knee joint point and the head joint point are on the same vertical line. Therefore, according to the positional relationship between the lowest joint point in the skeleton image and the head joint point, it can be determined that the posture type corresponding to the user to be identified is a sitting posture or a kneeling posture.
  • the posture type corresponding to the user to be identified is a kneeling posture.
  • the posture type corresponding to the user to be identified is a sitting posture.
  • the gesture type corresponding to the user to be recognized can be more accurately determined.
  • the posture type of the user to be recognized is a sitting posture or a kneeling posture.
  • Step S32 if the ratio of the posture height to the head height is not within the preset ratio range, determine whether the user to be identified falls according to the body center of gravity information according to a preset detection strategy.
  • the information about the center of gravity of the human body may also include the rate of descent of the center of gravity.
  • the detection strategy may include: determining whether the center of gravity descending rate of the user to be identified is greater than a preset descending rate; when the descending rate of the center of gravity is greater than the preset descending rate, determining a hip height value corresponding to the user to be identified; when the hip height value is less than the preset descending rate When the height value is , it is determined that the user to be identified is in a falling state, wherein the height value of the buttocks is the distance between the buttocks of the user to be identified and the ground.
  • FIG. 10 is a schematic flowchart of determining whether the user to be identified falls according to the information of the center of gravity of the human body according to the preset detection strategy in step S32 , which may specifically include the following steps S321 to S323 .
  • Step S321 Detect the drop rate of the center of gravity of the user to be identified.
  • the method before detecting the drop rate of the center of gravity of the user to be identified, the method further includes: acquiring the center of gravity of the human body in the depth image, and marking the center of gravity of the human body in the skeleton image.
  • the human body center of gravity point corresponding to the user to be identified has been determined according to the depth image, wherein the human body center point of gravity in the depth image can be directly obtained and marked in the skeleton image.
  • a first skeletal image and a second skeletal image at adjacent preset intervals in the video image frame are acquired, the first skeletal image includes the center of gravity of the first human body, and the second skeletal image includes the center of gravity of the second human body point; according to the coordinates corresponding to the center of gravity of the first person, the coordinates corresponding to the center of gravity of the second person, and the interval time, determine the drop rate of the center of gravity of the user to be identified, wherein the preset interval time can be set according to the actual situation, and the specific value It is not limited here.
  • the skeleton image in the video image frame of the first frame is used as the first skeleton image
  • the skeleton image in the video image frame of the 10th frame is used as the interval time t.
  • Second bone image Since the center of gravity of the human body has been marked in the skeleton image, the first skeleton image includes the center of gravity of the first human body, and the second skeleton image includes the center of gravity of the second human body.
  • the coordinates corresponding to the center of gravity of the first person are (X 01 , Y 01 ) and the coordinates corresponding to the center of gravity of the second person (X 10 , Y 10 ), then according to the coordinates corresponding to the center of gravity of the first person (X 01 , Y 01 ), the coordinates (X 10 , Y 10 ) corresponding to the center of gravity of the second person, and the interval time t, can determine the drop rate of the center of gravity of the user to be identified.
  • the descent rate of the center of gravity is represented by v, and the descent rate of the center of gravity v can be calculated by the following formula:
  • Step S322 When the lowering rate of the center of gravity is greater than a preset lowering rate, detect the hip height value of the user to be identified.
  • the preset falling rate may be represented by V, where the falling rate V may be set according to the actual situation, and the specific value is not limited herein.
  • the hip height value of the user to be identified is detected.
  • acquiring the skeleton image in the video image frame for example, using the skeleton image in the video image frame of the 11th frame as the first skeleton image, and using the skeleton image in the video image frame of the 20th frame as the second skeleton image; continue According to the coordinates and the interval time corresponding to the center of gravity point in the first skeleton image and the second skeleton image, the drop rate of the center of gravity of the user to be identified is determined.
  • detecting the height value of the buttocks of the user to be identified may include: acquiring the coordinates corresponding to the hip joint points in the skeleton image; determining the vertical distance between the coordinates corresponding to the hip joint points and the ground, and using the vertical distance as the to-be-identified distance The user's hip height value.
  • the coordinates corresponding to the hip joint points in the skeleton image can be directly obtained; exemplarily, in the skeleton image, the coordinates corresponding to the hip joint points are (X 4 , Y4 ) .
  • a three-dimensional coordinate system can be established with the somatosensory as the coordinate origin, wherein the distance between the hip joint point and the somatosensory can be determined by the depth information in the depth image; for example, if the distance between the hip joint point and the somatosensory If the distance is Z 4 , in the three-dimensional coordinate system, the coordinates corresponding to the hip joint point are (X 4 , Y 4 , Z 4 ).
  • the constant D represents the distance between the somatosensory sensor and the ground.
  • the coordinates corresponding to the hip joint point are determined to be the vertical distance between (X 4 , Y 4 , Z 4 ) and the ground, which can be determined by the point-to-surface distance formula, as shown below:
  • H represents the vertical distance between the hip joint point and the ground, that is, the hip height value of the user to be identified is H.
  • Step S323 When the hip height value is less than a preset height value, determine that the user to be identified is in a falling state.
  • the preset height value can be set according to the average width of the waist and buttocks of an adult male and the average width of the waist and buttocks of an adult female, and the specific value is not limited herein.
  • hip height value H is smaller than the preset height value, it means that the buttocks of the user to be identified are relatively close to the ground, and at this time, it can be determined that the user to be identified is in a falling state.
  • the method further includes: sending an emergency notification to the family or hospital corresponding to the to-be-identified user, so that the family or hospital finds that the to-be-identified user has fallen and handles it in time according to the emergency notification.
  • the manner of sending the emergency notification may include, but is not limited to, text messages, phone calls, emails, and the like.
  • the emergency notification may include location information of the user to be identified, and may also include depth images and skeleton images corresponding to the user to be identified.
  • the fall of the user to be identified can be detected in time and time can be saved for rescue.
  • the behavior recognition method, device, system and storage medium provided by the above embodiments can more conveniently determine the body center of gravity information according to the depth image and determine the posture height and head according to the skeleton image by acquiring the depth image and skeleton image corresponding to the user to be recognized.
  • the center of gravity of the user to be identified can be more accurately determined according to the total number of pixels in the human body area in the depth image and the coordinates of each pixel;
  • the coordinates of the joint points corresponding to the points can obtain the height of the head more accurately, which improves the accuracy of the subsequent judgment of the current behavior state of the user to be identified; difference, determine the posture height corresponding to the user to be recognized, which can more truly and accurately reflect the posture height corresponding to the user to be recognized in the current state; by judging whether the ratio of the posture height corresponding to the user to be recognized to the head height is within the first ratio range , the second ratio range or the third ratio range, the gesture type corresponding to the user to be recognized can be more accurately determined.
  • the posture type of the user to be identified is sitting or kneeling; by first judging whether the downward rate of the center of gravity of the user to be identified is greater than the preset drop Then, the hip height value of the user to be identified can be detected, so that it can be determined whether the user to be identified is in a fall state by combining the lowering rate of the center of gravity and the hip height value, which greatly improves the identification accuracy.
  • the embodiments of the present application further provide a storage medium for readable storage, the storage medium stores a program, the program includes program instructions, and the processor executes the program instructions to implement the embodiments of the present application Any of the behavioral identification methods provided.
  • the program is loaded by the processor and can perform the following steps:
  • the storage medium may be an internal storage unit of the behavior recognition apparatus described in the foregoing embodiments, such as a hard disk or a memory of the behavior recognition apparatus.
  • the storage medium may also be an external storage device of the behavior recognition device, such as a plug-in hard disk equipped on the behavior recognition device, a smart memory card (Smart Media Card, SMC), a Secure Digital Card (Secure Digital Card, SD Card), flash memory card (Flash Card), etc.
  • the division between functional modules/units mentioned in the above description does not necessarily correspond to the division of physical components; for example, one physical component may have multiple functions, or one function or step may be composed of several physical components Components execute cooperatively.
  • Some or all physical components may be implemented as software executed by a processor, such as a central processing unit, digital signal processor or microprocessor, or as hardware, or as an integrated circuit, such as an application specific integrated circuit .
  • a processor such as a central processing unit, digital signal processor or microprocessor
  • Such software may be distributed on storable media, which may include computer storage media (or non-transitory media) and communication media (or transitory media).
  • storage medium includes both volatile and nonvolatile implemented in any method or technology for storage of information, such as computer readable instructions, data structures, program modules or other data , removable and non-removable media.
  • Computer storage media include, but are not limited to, RAM, ROM, EEPROM, flash memory or other memory technology, CD-ROM, digital versatile disk (DVD) or other optical disk storage, magnetic cartridges, magnetic tape, magnetic disk storage or other magnetic storage devices, or may Any other medium used to store desired information and which can be accessed by a computer.
  • communication media typically embodies computer readable instructions, data structures, program modules, or other data in a modulated data signal such as a carrier wave or other transport mechanism, and can include any information delivery media, as is well known to those of ordinary skill in the art .

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Human Computer Interaction (AREA)
  • Multimedia (AREA)
  • Health & Medical Sciences (AREA)
  • General Health & Medical Sciences (AREA)
  • Psychiatry (AREA)
  • Social Psychology (AREA)
  • Image Analysis (AREA)

Abstract

A behavior recognition method and apparatus, and a storage medium. The behavior recognition method comprises: acquiring a video image frame corresponding to a user to be subjected to recognition, wherein the video image frame comprises a depth image and a skeleton image (S10); determining a posture feature parameter corresponding to said user according to the depth image and the skeleton image (S20); and determining the current behavioral state of said user according to the posture feature parameter corresponding to said user (S30).

Description

行为识别方法、装置和存储介质Behavior recognition method, device and storage medium
交叉引用cross reference
本申请基于申请号为“202010901631.6”、申请日为2020年08月31日的中国专利申请提出,并要求该中国专利申请的优先权,该中国专利申请的全部内容在此以引入方式并入本申请。This application is based on the Chinese patent application with the application number "202010901631.6" and the application date is August 31, 2020, and claims the priority of the Chinese patent application. The entire content of the Chinese patent application is hereby incorporated by reference. Apply.
技术领域technical field
本申请实施例涉及图像处理技术领域,尤其涉及一种行为识别方法、装置和存储介质。The embodiments of the present application relate to the technical field of image processing, and in particular, to a behavior recognition method, device, and storage medium.
背景技术Background technique
随着人工智能技术的不断发展,人体行为的识别成为一个新兴研究领域。常用的人体行为识别方法是基于可穿戴设备的行为识别方法。基于可穿戴设备的行为识别方法通过佩戴在人身上的运动传感器对人体的运动数据进行采集,缺点是需要用户佩戴传感器,不够便捷,而且识别的准确度较低。With the continuous development of artificial intelligence technology, human behavior recognition has become an emerging research field. The commonly used human behavior recognition method is the behavior recognition method based on wearable devices. The wearable device-based behavior recognition method collects the motion data of the human body through the motion sensor worn on the person. The disadvantage is that the user needs to wear the sensor, which is not convenient and the recognition accuracy is low.
因此,如何提高识别人体行为的准确度和便捷性成为亟需解决的问题。Therefore, how to improve the accuracy and convenience of recognizing human behavior has become an urgent problem to be solved.
发明内容SUMMARY OF THE INVENTION
本申请实施例提供了一种行为识别方法,包括:获取待识别用户对应的视频图像帧,其中,所述视频图像帧包括深度图像和骨骼图像;根据所述深度图像和所述骨骼图像确定所述待识别用户对应的姿势特征参数;根据所述待识别用户对应的姿势特征参数,确定所述待识别用户的当前行为状态。An embodiment of the present application provides a method for behavior recognition, including: acquiring a video image frame corresponding to a user to be recognized, wherein the video image frame includes a depth image and a skeleton image; The gesture characteristic parameter corresponding to the user to be recognized is determined; the current behavior state of the user to be recognized is determined according to the gesture characteristic parameter corresponding to the user to be recognized.
本申请实施例还提供了一种行为识别装置,包括:处理器和存储器;所述存储器用于存储程序;所述处理器用于执行所述程序并在执行所述程序时实现如上述的行为识别方法。An embodiment of the present application further provides a behavior recognition device, including: a processor and a memory; the memory is used to store a program; the processor is used to execute the program and implement the above behavior recognition when the program is executed method.
本申请实施例还提供了一种存储介质,用于可读存储,所述存储介质存储有一个或者多个程序,所述一个或者多个程序可被一个或者多个处理器执行,以实现如上述的行为识别方法。Embodiments of the present application further provide a storage medium for readable storage, where the storage medium stores one or more programs, and the one or more programs can be executed by one or more processors to implement the following: The above-mentioned behavior recognition method.
附图说明Description of drawings
为了更清楚地说明本申请实施例的技术方案,下面将对实施例描述中所需要使用的附图作简单地介绍,显而易见地,下面描述中的附图是本申请的一些实施例,对于本领域普通技术人员来讲,在不付出创造性劳动的前提下,还可以根据这些附图获得其他的附图。In order to explain the technical solutions of the embodiments of the present application more clearly, the following briefly introduces the accompanying drawings used in the description of the embodiments. Obviously, the drawings in the following description are some embodiments of the present application. For those of ordinary skill in the art, other drawings can also be obtained from these drawings without any creative effort.
图1是本申请的实施例提供的一种行为识别系统的结构示意图;1 is a schematic structural diagram of a behavior recognition system provided by an embodiment of the present application;
图2是本申请的实施例提供的对人体区域进行部位分割的示意图;2 is a schematic diagram of performing part segmentation on a human body region provided by an embodiment of the present application;
图3是本申请的实施例提供的一种行为识别装置的结构示意图;3 is a schematic structural diagram of a behavior recognition device provided by an embodiment of the present application;
图4是本申请的实施例提供的一种行为识别方法的示意流程图;FIG. 4 is a schematic flowchart of a behavior recognition method provided by an embodiment of the present application;
图5是本申请的实施例提供的视频图像帧的示意图;5 is a schematic diagram of a video image frame provided by an embodiment of the present application;
图6是本申请的实施例提供的人体区域与人体重心点的示意图;6 is a schematic diagram of a human body region and a human body center of gravity provided by an embodiment of the present application;
图7是本申请的实施例提供的确定待识别用户对应的头部高度的示意图;7 is a schematic diagram of determining a head height corresponding to a user to be identified according to an embodiment of the present application;
图8是本申请的实施例提供的确定待识别用户的姿势高度的示意图;8 is a schematic diagram of determining the posture height of a user to be recognized provided by an embodiment of the present application;
图9是图4中确定待识别用户的当前行为状态的子步骤的示意性框图;Fig. 9 is a schematic block diagram of the sub-steps of determining the current behavioral state of the user to be identified in Fig. 4;
图10是图9中判断待识别用户是否摔倒的子步骤的示意性流程图。FIG. 10 is a schematic flowchart of the sub-steps of judging whether the user to be identified has fallen in FIG. 9 .
具体实施方式detailed description
本申请实施例提供了一种行为识别方法、装置和存储介质,以至少解决相关技术中识别人体行为的准确度低、便捷性差的问题。Embodiments of the present application provide a behavior recognition method, device, and storage medium, so as to at least solve the problems of low accuracy and poor convenience in recognizing human behavior in the related art.
下面将结合本申请实施例中的附图,对本申请实施例中的技术方案进行清楚、完整地描述,显然,所描述的实施例是本申请一部分实施例,而不是全部的实施例。基于本申请中的实施例,本领域普通技术人员在没有做出创造性劳动前提下所获得的所有其他实施例,都属于本申请保护的范围。The technical solutions in the embodiments of the present application will be clearly and completely described below with reference to the accompanying drawings in the embodiments of the present application. Obviously, the described embodiments are part of the embodiments of the present application, not all of the embodiments. Based on the embodiments in the present application, all other embodiments obtained by those of ordinary skill in the art without creative efforts shall fall within the protection scope of the present application.
附图中所示的流程图仅是示例说明,不是必须包括所有的内容和操作/步骤,也不是必须按所描述的顺序执行。例如,有的操作/步骤还可以分解、组合或部分合并,因此实际执行的顺序有可能根据实际情况改变。The flowcharts shown in the figures are for illustration only, and do not necessarily include all contents and operations/steps, nor do they have to be performed in the order described. For example, some operations/steps can also be decomposed, combined or partially combined, so the actual execution order may be changed according to the actual situation.
应当理解,在此本申请说明书中所使用的术语仅仅是出于描述特定实施例的目的而并不意在限制本申请。如在本申请说明书和所附权利要求书中所使用的那样,除非上下文清楚地指明其它情况,否则单数形式的“一”、“一个”及“该”意在包括复数形式。It should be understood that the terms used in the specification of the present application herein are for the purpose of describing particular embodiments only and are not intended to limit the present application. As used in this specification and the appended claims, the singular forms "a," "an," and "the" are intended to include the plural unless the context clearly dictates otherwise.
还应当理解,在本申请说明书和所附权利要求书中使用的术语“和/或”是指相关联列出的项中的一个或多个的任何组合以及所有可能组合,并且包括这些组合。It will also be understood that, as used in this specification and the appended claims, the term "and/or" refers to and including any and all possible combinations of one or more of the associated listed items.
应当理解,此处所描述的具体实施例仅仅用以解释本申请,并不用于限定本申请。It should be understood that the specific embodiments described herein are only used to explain the present application, but not to limit the present application.
在后续的描述中,使用用于表示元件的诸如“模块”、“部件”或“单元”的后缀仅为了有利于本申请的说明,其本身没有特有的意义。因此,“模块”、“部件”或“单元”可以混合地使用。In the following description, suffixes such as 'module', 'component' or 'unit' used to represent elements are used only to facilitate the description of the present application, and have no specific meaning per se. Thus, "module", "component" or "unit" may be used interchangeably.
本申请的一个实施例提供了一种行为识别方法、装置、系统和存储介质。其中,该行为识别方法可以应用于行为识别装置中,实现根据深度图像和骨骼图像确定姿势特征参数,并通过姿势特征参数确定待识别用户的当前行为参数,可以更加便捷和准确地识别待识别用户对应的当前行为状态。An embodiment of the present application provides a behavior recognition method, apparatus, system, and storage medium. Among them, the behavior recognition method can be applied to a behavior recognition device to realize the determination of the posture feature parameters according to the depth image and the skeleton image, and to determine the current behavior parameters of the user to be recognized through the posture feature parameters, which can more conveniently and accurately identify the user to be recognized. The corresponding current behavior state.
示例性的,行为识别装置可以包括服务器或终端。其中,服务器可以为独立的服务器,也可以为服务器集群;终端可以是智能手机、平板电脑、笔记本电脑和台式电脑等电子设备。Exemplarily, the behavior recognition device may include a server or a terminal. The server may be an independent server or a server cluster; the terminal may be an electronic device such as a smart phone, a tablet computer, a notebook computer, and a desktop computer.
请参阅图1,图1是本申请的实施例提供的一种行为识别系统的结构示意图。该行为识别系统包括行为识别装置10与拍摄装置20。其中,行为识别装置10可以与拍摄装置20有线或无线通信连接。拍摄装置20用于采集包括待识别用户的视频图像帧,行为识别装置10用于对拍摄装置20采集的视频图像帧进行图像处理,以确定待识别用户对应的当前行为状态。Please refer to FIG. 1. FIG. 1 is a schematic structural diagram of a behavior recognition system provided by an embodiment of the present application. The behavior recognition system includes a behavior recognition device 10 and a photographing device 20 . Wherein, the behavior recognition device 10 may be connected with the photographing device 20 in wired or wireless communication. The photographing device 20 is used to collect video image frames including the user to be recognized, and the behavior recognition device 10 is used to perform image processing on the video image frames collected by the photographing device 20 to determine the current behavior state corresponding to the user to be recognized.
在一些实施例中,拍摄装置20可以采集包括待识别用户对应的视频图像帧,并对视频图像帧进行人体识别、人体部位识别以及骨骼关节点定位等处理,处理后的视频图像帧包括深度图像和骨骼图像。行为识别装置10可以获取拍摄装置20处理后的视频图像帧,根据深度图像确定待识别用户对应的人体重心信息以及根据骨骼图像确定待识别用户对应的姿势高度和头部高度;然后,根据人体重心信息、姿势高度以及头部高度,确定待识别用户的当前行为状态。In some embodiments, the photographing device 20 may collect video image frames including the corresponding user to be identified, and perform human body recognition, human body part recognition, and skeletal joint point positioning on the video image frames, and the processed video image frames include depth images. and skeleton images. The behavior recognition device 10 can acquire the video image frames processed by the shooting device 20, determine the body center of gravity information corresponding to the user to be recognized according to the depth image, and determine the posture height and head height corresponding to the user to be recognized according to the skeleton image; then, according to the body center of gravity information, posture height, and head height to determine the current behavioral state of the user to be identified.
示例性的,拍摄装置20可以包括3D摄像头的电子设备,例如可以采集视频图像帧的体感器。在本申请的实施例中,可以使用所述体感器采集包括待识别用户的视频图像帧。Exemplarily, the photographing apparatus 20 may include an electronic device such as a 3D camera, such as a somatosensory that may capture video image frames. In the embodiment of the present application, the somatosensory may be used to collect video image frames including the user to be identified.
需要说明的是,所述体感器可以包括深度摄像头、彩色摄像头以及光源发射器,所述体感器可以获取背景空间的深度图像、彩色图像以及三维数据信息。It should be noted that the somatosensory may include a depth camera, a color camera and a light source emitter, and the somatosensory may acquire depth images, color images and three-dimensional data information of the background space.
示例性的,所述体感器可以获取深度信息。获取深度信息的工作原理包括:将光源发射器发出的光投射到现实的场景中,由于发射出去的光会因为物体表面形状的不同会产生改变,由此可以将这种光进行收集并进行编码就可以得到场景中各个像素点与深度摄像头的之间的距离差值,进而得到物体的位置和深度信息。Exemplarily, the somatosensory can acquire depth information. The working principle of obtaining depth information includes: projecting the light emitted by the light source emitter into the real scene. Since the emitted light will change due to the different surface shapes of the object, this light can be collected and encoded. The distance difference between each pixel in the scene and the depth camera can be obtained, and then the position and depth information of the object can be obtained.
在一些实施例中,所述体感器对视频图像帧进行人体识别;例如,根据预设的分割策略对视频图像帧中的背景和人物进行分割,以确定待识别用户对应的人体区域或人体轮廓信息,得到的深度图像包括待识别用户对应的人体区域或人体轮廓信息。In some embodiments, the body sensor performs human body recognition on the video image frame; for example, the background and the person in the video image frame are segmented according to a preset segmentation strategy, so as to determine the human body region or the human body contour corresponding to the user to be identified. information, and the obtained depth image includes the body region or body contour information corresponding to the user to be identified.
在一些实施例中,所述体感器对人体识别得到的深度图像进行人体部位识别。示例性的,如图2所示,图2是对人体区域进行部位分割的示意图;对深度图像中的人体区域进行部位分割,得到多个部位图像,例如头、手臂、腿、四肢以及躯干等部位图像;对多个部位图像进行特征值分类匹配,以确定各部位图像对应的人体部位。In some embodiments, the body sensor performs human body part recognition on the depth image obtained by human body recognition. Exemplarily, as shown in FIG. 2, FIG. 2 is a schematic diagram of part segmentation of the human body area; part segmentation is performed on the human body area in the depth image to obtain multiple part images, such as head, arms, legs, limbs and torso, etc. Part images; perform feature value classification and matching on multiple part images to determine the body part corresponding to each part image.
在一些实施例中,所述体感器对视频图像帧进行骨骼关节点定位,得到骨骼图像。具体地,将识别出的人体部位添加至虚拟的骨骼模型中,并根据人体部位的位置信息进行调整,得到包括多个关节点的骨骼图像。In some embodiments, the somatosensory performs skeletal joint point positioning on the video image frame to obtain a skeletal image. Specifically, the identified body parts are added to the virtual skeleton model, and adjusted according to the position information of the body parts to obtain a skeleton image including a plurality of joint points.
示例性的,骨骼图像可以包括但不限于:头部关节点、颈部关节点、膝部关节点、肘部关节点、髋部关节点或脚腕关节点等关节点。Exemplarily, the skeleton image may include, but is not limited to, joint points such as head joint points, neck joint points, knee joint points, elbow joint points, hip joint points, or ankle joint points.
请参阅图3,图3是本申请的实施例提供的一种行为识别装置10的结构示意图。行为识别装置10可以包括处理器11和存储器12,其中,所述处理器11和所述存储器12可以通过总线连接,该总线比如为集成电路总线(Inter-integrated Circuit,简称:IIC)等任意适用的总线。Please refer to FIG. 3 , which is a schematic structural diagram of a behavior recognition apparatus 10 provided by an embodiment of the present application. The behavior recognition device 10 may include a processor 11 and a memory 12, wherein the processor 11 and the memory 12 may be connected through a bus, such as an integrated circuit bus (Inter-integrated Circuit, referred to as: IIC) and other suitable the bus.
其中,所述存储器12可以包括非易失性存储介质和内存储器。非易失性存储介质可存储操作系统和计算机程序。该计算机程序包括程序指令,该程序指令被执行时,可使得处理器执行任意一种行为识别方法。Wherein, the memory 12 may include a non-volatile storage medium and an internal memory. The nonvolatile storage medium can store operating systems and computer programs. The computer program includes program instructions that, when executed, can cause the processor to perform any behavior recognition method.
其中,所述处理器11用于提供计算和控制能力,支撑整个行为识别装置10的运行。Wherein, the processor 11 is used to provide computing and control capabilities to support the operation of the entire behavior recognition device 10 .
其中,所述处理器11可以是中央处理单元(Central Processing Unit,CPU),该处理器还可以是其他通用处理器、数字信号处理器(Digital Signal Processor,DSP)、专用集成电路(application specific integrated circuit,ASIC)、现场可编程门阵列(Field-Programmable Gate Array,FPGA)或者其他可编程逻辑器件、分立门或者晶体管逻辑器件、分立硬件组件等。通用处理器可以是微处理器或者该处理器也可以是任何常规的处理器等。The processor 11 may be a central processing unit (Central Processing Unit, CPU), and the processor may also be other general-purpose processors, digital signal processors (Digital Signal Processors, DSP), application specific integrated circuits (application specific integrated circuits) circuit, ASIC), Field-Programmable Gate Array (FPGA) or other programmable logic devices, discrete gate or transistor logic devices, discrete hardware components, etc. A general purpose processor may be a microprocessor or the processor may be any conventional processor or the like.
下面结合附图,对本申请的一些实施方式作详细说明。在不冲突的情况下,下述的实施例及实施例中的特征可以相互组合。Some embodiments of the present application will be described in detail below with reference to the accompanying drawings. The embodiments described below and features in the embodiments may be combined with each other without conflict.
如图4所示,图4是本申请的实施例提供的一种行为识别方法的示意性流程图。该行为识别方法可以应用于行为识别装置中,实现根据深度图像和骨骼图像确定姿势特征参数,并通过姿势特征参数确定待识别用户的当前行为参数,可以更加便捷和准确地识别待识别用户对应的当前行为状态。该行为识别方法包括步骤S10至步骤S30。As shown in FIG. 4 , FIG. 4 is a schematic flowchart of a behavior recognition method provided by an embodiment of the present application. The behavior recognition method can be applied to a behavior recognition device to realize the determination of the posture feature parameters according to the depth image and the skeleton image, and to determine the current behavior parameters of the user to be recognized through the posture feature parameters, which can more conveniently and accurately identify the corresponding user to be recognized. Current behavior state. The behavior recognition method includes steps S10 to S30.
步骤S10:获取待识别用户对应的视频图像帧,其中,所述视频图像帧包括深度图像和 骨骼图像。Step S10: Obtain a video image frame corresponding to the user to be identified, wherein the video image frame includes a depth image and a skeleton image.
需要说明的是,在本申请的实施例中,可以通过体感器采集待识别用户对应的视频图像帧。示例性的,体感器可以安装在室内,对室内的环境和人物进行实时监控。It should be noted that, in the embodiment of the present application, the video image frame corresponding to the user to be identified may be collected by the somatosensory sensor. Exemplarily, the somatosensory sensor can be installed indoors to monitor the indoor environment and people in real time.
在一些实施例中,体感器可以通过内置的软件开发工具包采集待识别用户对应的视频图像帧,并对视频图像帧进行人体识别、人体部位识别以及骨骼关节点定位等处理,处理后的视频图像帧包括深度图像和骨骼图像。In some embodiments, the somatosensory can collect video image frames corresponding to the user to be recognized through a built-in software development kit, and perform human body recognition, human body part recognition, and skeletal joint point positioning on the video image frames. Image frames include depth images and skeletal images.
示例性的,可以接收体感器发送的经过处理后的视频图像帧,其中,视频图像帧包括深度图像和骨骼图像。深度图像和骨骼图像相互重叠在同一图像中,示例性的,如图5所示,图5是视频图像帧。深度图像包括待识别用户对应的人体区域,骨骼图像包括待识别用户对应的骨骼信息。其中,骨骼信息可以包括不同的关节点和各关节点之间的连接关系等。Exemplarily, a processed video image frame sent by a body sensor may be received, wherein the video image frame includes a depth image and a skeleton image. The depth image and the skeleton image overlap each other in the same image, exemplarily, as shown in FIG. 5 , which is a video image frame. The depth image includes the human body region corresponding to the user to be identified, and the skeleton image includes bone information corresponding to the user to be identified. The skeleton information may include different joint points and connection relationships between the joint points, and the like.
由于深度图像和骨骼图像不易受外界环境的干扰,获取待识别用户对应的深度图像和骨骼图像,可以提高根据深度图像和骨骼图像确定待识别用户对应的姿势特征参数的准确度。Since the depth image and the skeleton image are not easily disturbed by the external environment, obtaining the depth image and the skeleton image corresponding to the user to be recognized can improve the accuracy of determining the posture feature parameters corresponding to the user to be recognized according to the depth image and the skeleton image.
步骤S20:根据所述深度图像和所述骨骼图像确定所述待识别用户对应的姿势特征参数。Step S20: Determine, according to the depth image and the skeleton image, a posture feature parameter corresponding to the user to be recognized.
示例性的,姿势特征参数可以包括人体重心信息、姿势高度和头部高度。Exemplarily, the posture feature parameters may include information on the body's center of gravity, posture height, and head height.
可以理解的是,姿势高度是指待识别用户在不同姿势时的身体高度。示例性的,人在站立时的身体高度、在盘坐时的身体高度以及坐在椅子上时的身体高度等等。It can be understood that the posture height refers to the height of the body of the user to be identified in different postures. Exemplarily, a person's body height when standing, body height when sitting cross-legged, body height when sitting on a chair, and the like.
示例性的,人体重心信息可以包括但不限于重心点、重心坐标以及重心下降速度等等。头部高度是指待识别用户的整个头部的高度。Exemplarily, the information on the center of gravity of the human body may include, but is not limited to, the center of gravity point, the coordinates of the center of gravity, and the speed of falling of the center of gravity, and the like. The head height refers to the height of the entire head of the user to be identified.
在一些实施例中,可以根据深度图像确定待识别用户对应的人体重心信息,以及根据骨骼图像确定待识别用户对应的姿势高度和头部高度。In some embodiments, the center of gravity information of the human body corresponding to the user to be identified may be determined according to the depth image, and the posture height and the height of the head corresponding to the user to be identified may be determined according to the skeleton image.
在一些实施例中,根据深度图像确定待识别用户对应的人体重心信息之前,还可以包括:根据预设的数据格式,对初始的深度图像中的深度数据进行格式转换,得到格式转换后的深度图像。其中,预设的数据格式可以是Mat格式。初始的深度图像是指从体感器获取的视频图像帧中的深度图像。In some embodiments, before determining the body center of gravity information corresponding to the user to be identified according to the depth image, the method may further include: performing format conversion on the depth data in the initial depth image according to a preset data format, to obtain the depth after the format conversion image. The preset data format may be Mat format. The initial depth image refers to the depth image in the video image frame obtained from the somatosensory sensor.
需要说明的是,在本申请的实施例中,基于计算机视觉平台的基础上完成根据深度图像和骨骼图像确定待识别用户的当前行为状态的识别过程,为了使深度图像有更好的显示效果,需要对初始的深度图像中的深度数据进行格式转换。It should be noted that, in the embodiment of the present application, the recognition process of determining the current behavior state of the user to be recognized according to the depth image and the skeleton image is completed on the basis of the computer vision platform. In order to make the depth image have a better display effect, Format conversion of the depth data in the original depth image is required.
示例性的,根据Mat格式,对初始的深度图像中的深度数据进行格式转换,格式转换后的深度图像中的深度数据为Mat格式。Exemplarily, according to the Mat format, the depth data in the original depth image is format-converted, and the depth data in the format-converted depth image is in the Mat format.
在一些实施例中,根据骨骼图像确定待识别用户对应的姿势高度以及头部高度之前,还可以包括:根据预设的平滑处理策略,对初始的骨骼图像进行平滑处理,得到平滑处理后的骨骼图像。其中,初始的骨骼图像是指从体感器获取的视频图像帧中的骨骼图像。In some embodiments, before determining the posture height and the head height corresponding to the user to be recognized according to the skeleton image, the method may further include: smoothing the initial skeleton image according to a preset smoothing strategy to obtain a smoothed skeleton image. Wherein, the initial skeleton image refers to the skeleton image in the video image frame acquired from the somatosensory sensor.
需要说明的是,在本申请的实施例中,需要根据骨骼图像确定待识别用户对应的姿势高度和头部高度;在识别过程中,对待识别用户的身高、动作和实时性有较高的识别要求,若不对骨骼图像进行平滑处理,可能导致计算机视觉平台的抖动甚至崩溃。It should be noted that, in the embodiments of the present application, the posture height and head height corresponding to the user to be recognized need to be determined according to the skeleton image; during the recognition process, the height, movement and real-time performance of the user to be recognized are highly recognized. It is required that if the skeleton image is not smoothed, it may cause the computer vision platform to shake or even crash.
可以理解的是,平滑处理也叫滤波处理。通过对骨骼图像进行平滑处理,可以减少骨骼图像中的噪声或者失真。It can be understood that smoothing is also called filtering. By smoothing the skeleton image, noise or distortion in the skeleton image can be reduced.
示例性的,预设的平滑处理策略可以包括但不限于均值滤波算法、中值滤波算法、高斯滤波算法以及双边滤波算法等等。Exemplarily, the preset smoothing processing strategy may include, but is not limited to, a mean filtering algorithm, a median filtering algorithm, a Gaussian filtering algorithm, a bilateral filtering algorithm, and the like.
例如,根据均值滤波算法,对初始的骨骼图像进行平滑处理,得到平滑处理后的骨骼图像。For example, according to the mean filtering algorithm, the initial skeleton image is smoothed to obtain the smoothed skeleton image.
例如,根据高斯滤波算法,对初始的骨骼图像进行平滑处理,得到平滑处理后的骨骼图像。For example, according to the Gaussian filtering algorithm, the initial skeleton image is smoothed to obtain the smoothed skeleton image.
在一些实施例中,在根据预设的数据格式,对初始的深度图像中的深度数据进行格式转换,得到格式转换后的深度图像之后,可以根据格式转换后的深度图像确定待识别用户对应的人体重心信息。In some embodiments, after format conversion is performed on the depth data in the initial depth image according to a preset data format to obtain a format-converted depth image, the depth image corresponding to the user to be recognized may be determined according to the format-converted depth image. Human body center of gravity information.
示例性的,人体重心信息可以包括人体重心点;如图6所示,图6是人体区域与人体重心点的示意图。Exemplarily, the information on the center of gravity of the human body may include a center of gravity point of the human body; as shown in FIG. 6 , FIG. 6 is a schematic diagram of the area of the human body and the center of gravity point of the human body.
在一些实施例中,根据深度图像确定待识别用户对应的人体重心信息,可以包括:获取格式转换后的深度图像中的人体区域内的像素点总数,以及获取人体区域中的全部像素点对应的横坐标和纵坐标;确定人体区域中的全部像素点对应的横坐标之和与纵坐标之和;根据像素点总数和横坐标之和,确定全部像素点对应的横坐标的均值,以及根据像素点总数和纵坐标之和,确定全部像素点对应的纵坐标的均值;将全部像素点对应的横坐标的均值作为人体重心点的横坐标,以及将全部像素点对应的纵坐标的均值作为人体重心点的纵坐标,得到深度图像中的人体重心点。In some embodiments, determining the center of gravity information of the human body corresponding to the user to be identified according to the depth image may include: acquiring the total number of pixels in the human body region in the format-converted depth image, and acquiring the corresponding pixel points in the human body region. The abscissa and the ordinate; determine the sum of the abscissas and ordinates corresponding to all the pixels in the human body area; according to the total number of pixels and the sum of the abscissas, determine the mean of the abscissas corresponding to all the pixels, and according to the pixel The sum of the total number of points and the ordinate is to determine the mean of the ordinates corresponding to all the pixels; the mean of the abscissas corresponding to all the pixels is taken as the abscissa of the center of gravity of the human body, and the mean of the ordinates corresponding to all the pixels is taken as the human body The ordinate of the center of gravity, to get the center of gravity of the human body in the depth image.
示例性的,可以在深度图像中建立直角坐标系,获取深度图像中的人体区域中的像素点总数,并确定人体区域中的全部像素点对应的横坐标和纵坐标。Exemplarily, a rectangular coordinate system can be established in the depth image, the total number of pixels in the human body region in the depth image is obtained, and the abscissa and ordinate corresponding to all the pixels in the human body region are determined.
在一些实施例中,可以根据重心计算公式,计算出人体重心点对应的重心坐标。示例性的,重心计算公式如下:In some embodiments, the center of gravity coordinates corresponding to the center of gravity of the human body may be calculated according to the center of gravity calculation formula. Exemplarily, the calculation formula of the center of gravity is as follows:
Figure PCTCN2021100379-appb-000001
Figure PCTCN2021100379-appb-000001
式中,(X 0,Y 0)表示重心坐标;s表示人体区域内的像素点总数;i表示人体区域中第i个像素点;x i表示第i个像素点对应的横坐标;y i表示第i个像素点对应的纵坐标。 In the formula, (X 0 , Y 0 ) represents the barycentric coordinates; s represents the total number of pixels in the human body area; i represents the ith pixel in the human body area; x i represents the abscissa corresponding to the ith pixel; y i Indicates the ordinate corresponding to the i-th pixel.
根据重心计算公式,可以确定人体区域内的全部像素点对应的横坐标的均值X 0和纵坐标的均值Y 0。然后将均值X 0作为人体重心点的横坐标,将均值Y 0作为人体重心点的纵坐标,得到深度图像中的人体重心点对应的重心坐标(X 0,Y 0)。 According to the calculation formula of the center of gravity, the mean value X 0 of the abscissa and the mean value Y 0 of the ordinate corresponding to all the pixel points in the human body area can be determined. Then, the mean value X 0 is taken as the abscissa of the human body's center of gravity, and the mean value Y 0 is taken as the ordinate of the human body's center of gravity, to obtain the barycentric coordinates (X 0 , Y 0 ) corresponding to the human body's center of gravity in the depth image.
通过根据深度图像中的人体区域内的像素点总数和各像素点的坐标,可以更加准确地确定待识别用户的人体重心点。By using the total number of pixels in the human body region in the depth image and the coordinates of each pixel, the center of gravity of the user to be identified can be more accurately determined.
在一些实施例中,在根据预设的平滑处理策略,对初始的骨骼图像进行平滑处理,得到平滑处理后的骨骼图像之后,可以根据平滑处理后的骨骼图像确定待识别用户对应的姿势高度以及头部高度。In some embodiments, after smoothing an initial skeleton image according to a preset smoothing strategy to obtain a smoothed skeleton image, the posture height corresponding to the user to be recognized and head height.
示例性的,提取平滑处理后的骨骼图像中的关节点信息,根据关节点信息确定待识别用户对应的姿势高度以及头部高度。其中,关节点信息包括关节点坐标。例如头部关节点对应的关节点坐标、颈部关节点对应的关节点坐标等等。Exemplarily, the joint point information in the smoothed skeleton image is extracted, and the posture height and the head height corresponding to the user to be recognized are determined according to the joint point information. Wherein, the joint point information includes joint point coordinates. For example, the joint point coordinates corresponding to the head joint point, the joint point coordinates corresponding to the neck joint point, and so on.
在一些实施例中,提取平滑处理后的骨骼图像中的关节点信息,可以包括:在骨骼图像中建立二维坐标系并确定坐标原点;根据骨骼图像中的各关节点的在二维坐标系中的位置,确定各关节点的关节点坐标。例如,头部关节点对应的关节点坐标为(X 1,Y 1),颈部关节点对 应的关节点坐标为(X 2,Y 2)。 In some embodiments, extracting joint point information in the smoothed bone image may include: establishing a two-dimensional coordinate system in the bone image and determining a coordinate origin; according to the two-dimensional coordinate system of each joint point in the bone image In the position, determine the joint point coordinates of each joint point. For example, the joint point coordinates corresponding to the head joint point are (X 1 , Y 1 ), and the joint point coordinates corresponding to the neck joint point are (X 2 , Y 2 ).
在一些实施例中,根据关节点信息确定待识别用户对应的姿势高度以及头部高度之前,还可以包括:获取骨骼图像中的头部关节点;当头部关节点位于深度图像中的人体区域内,根据关节点信息确定待识别用户对应的姿势高度以及头部高度。In some embodiments, before determining the posture height and the head height corresponding to the user to be recognized according to the joint point information, the method may further include: acquiring the head joint points in the skeleton image; when the head joint points are located in the human body region in the depth image Inside, the posture height and head height corresponding to the user to be recognized are determined according to the joint point information.
示例性的,可以根据骨骼图像中各关节点之间的连接关系,对头部关节点进行定位,以确定头部关节点在骨骼图像中的具体位置以及头部关节点是否位于深度图像中的人体区域内。Exemplarily, the head joint points can be positioned according to the connection relationship between the joint points in the skeleton image to determine the specific position of the head joint point in the skeleton image and whether the head joint point is located in the depth image. in the human body area.
需要说明的是,当待识别用户的头部不是最高位置或头部被身体的其它部位遮挡时,通过骨骼图像中的关节点信息无法准确地确定头部高度。It should be noted that, when the head of the user to be identified is not at the highest position or the head is blocked by other parts of the body, the head height cannot be accurately determined through the joint point information in the skeleton image.
示例性的,当骨骼图像中的头部关节点位于深度图像中的人体区域内时,说明待识别用户的头部是最高位置或头部没有被遮挡,此时可以准确地根据关节点信息确定待识别用户对应的姿势高度以及头部高度。Exemplarily, when the head joint point in the skeleton image is located in the human body area in the depth image, it means that the head of the user to be identified is at the highest position or the head is not blocked, at this time, it can be accurately determined according to the joint point information. The posture height and head height corresponding to the user to be recognized.
示例性的,当骨骼图像中的头部关节点不在深度图像中的人体区域内时,说明待识别用户的头部不是最高位置或头部被遮挡,此时无法准确地根据关节点信息确定待识别用户对应的姿势高度以及头部高度。Exemplarily, when the head joint points in the skeleton image are not within the human body area in the depth image, it means that the head of the user to be identified is not in the highest position or the head is occluded. At this time, the joint point information cannot be accurately determined. Identify the user's corresponding posture height and head height.
通过判断头部关节点是否位于深度图像中的人体区域内,可以提高确定待识别用户对应的姿势高度以及头部高度的准确度。By judging whether the head joint point is located in the human body area in the depth image, the accuracy of determining the posture height and the head height corresponding to the user to be recognized can be improved.
在一些实施例中,在确定头部关节点位于深度图像中的人体区域内之后,可以根据关节点信息确定待识别用户对应的姿势高度以及头部高度。In some embodiments, after it is determined that the head joint points are located in the human body region in the depth image, the posture height and the head height corresponding to the user to be identified may be determined according to the joint point information.
在一些实施例中,根据关节点信息确定待识别用户对应的头部高度,可以包括:根据头部关节点对应的关节点坐标和颈部关节点对应的关节点坐标,确定头部关节点与颈部关节点之间的第一高度差;根据预设的高度比值与第一高度差的积,得到待识别用户对应的头部高度。In some embodiments, determining the height of the head corresponding to the user to be identified according to the joint point information may include: determining the head joint point and the neck joint point according to the joint point coordinates corresponding to the head joint point and the joint point coordinates corresponding to the neck joint point. The first height difference between the neck joint points; according to the product of the preset height ratio and the first height difference, the height of the head corresponding to the user to be identified is obtained.
示例性的,如图7所示,图7是确定待识别用户对应的头部高度的示意图。在骨骼图像建立的二维坐标系中,若头部关节点对应的关节点坐标为(X 1,Y 1),颈部关节点对应的关节点坐标为(X 2,Y 2),则头部关节点与颈部关节点之间的第一高度差h 1为|Y 1-Y 2|。 Exemplarily, as shown in FIG. 7 , FIG. 7 is a schematic diagram of determining the height of the head corresponding to the user to be identified. In the two-dimensional coordinate system established by the skeleton image, if the joint point coordinates corresponding to the head joint points are (X 1 , Y 1 ), and the joint point coordinates corresponding to the neck joint points are (X 2 , Y 2 ), then the head The first height difference h 1 between the neck joint point and the neck joint point is |Y 1 -Y 2 |.
在本申请的实施例中,若预设的高度比值为1.8726,则根据预设的高度比值与第一高度差h 1的积,可以得到待识别用户对应的头部高度H 1;其中,H 1=1.8726h 1In the embodiment of the present application, if the preset height ratio is 1.8726, then according to the product of the preset height ratio and the first height difference h1 , the head height H1 corresponding to the user to be identified can be obtained; wherein, H1 1 = 1.8726h 1 .
通过根据头部关节点对应的关节点坐标和颈部关节点对应的关节点坐标,可以更加准确地得到头部高度,提高了后续判断待识别用户的当前行为状态的准确性。By using the joint point coordinates corresponding to the head joint points and the joint point coordinates corresponding to the neck joint points, the head height can be obtained more accurately, which improves the accuracy of subsequent judgment of the current behavior state of the user to be identified.
在一些实施例中,根据关节点信息确定待识别用户对应的姿势高度,可以包括:确定骨骼图像中的最高关节点和最低关节点;根据最高关节点对应的关节点坐标和最低关节点对应的关节点坐标,确定最高关节点和最低关节点之间的第二高度差,将第二高度差作为待识别用户对应的姿势高度。In some embodiments, determining the posture height corresponding to the user to be recognized according to the joint point information may include: determining the highest joint point and the lowest joint point in the skeleton image; according to the joint point coordinates corresponding to the highest joint point and the lowest joint point The joint point coordinates are used to determine the second height difference between the highest joint point and the lowest joint point, and the second height difference is used as the posture height corresponding to the user to be identified.
可以理解的是,当待识别用户为站姿时,骨骼图像中的最高关节点为头部关节点,最低关节点为脚腕关节点。当待识别用户为盘坐姿时,骨骼图像中的最高关节点为头部关节点,最低关节点为臀部关节点。It can be understood that when the user to be recognized is standing, the highest joint point in the skeleton image is the head joint point, and the lowest joint point is the ankle joint point. When the user to be identified is sitting cross-legged, the highest joint point in the skeleton image is the head joint point, and the lowest joint point is the hip joint point.
在一些实施例中,当待识别用户为站姿时,骨骼图像中的最高关节点为头部关节点,最低关节点为脚腕关节点;可以根据头部关节点对应的关节点坐标以及脚腕关节点对应的关节点坐标确定头部关节点和脚腕关节点之间的第二高度差,即得到待识别用户的姿势高度。此 时姿势高度为待识别用户站立时的身体高度。In some embodiments, when the user to be identified is standing, the highest joint point in the skeleton image is the head joint point, and the lowest joint point is the ankle joint point; the joint point coordinates corresponding to the head joint point and the foot joint point can be used according to The joint point coordinates corresponding to the wrist joint point determine the second height difference between the head joint point and the ankle joint point, that is, the posture height of the user to be identified is obtained. At this time, the posture height is the height of the body when the user to be recognized is standing.
示例性的,若头部关节点对应的关节点坐标为(X 1,Y 1),脚腕关节点对应的关节点坐标为(X 3,Y 3),则在骨骼图像中,头部关节点和脚腕关节点之间的第二高度差为|Y 1-Y 3|,即待识别用户的姿势高度为|Y 1-Y 3|。如图8所示,图8是确定待识别用户的姿势高度的示意图。 Exemplarily, if the joint point coordinates corresponding to the head joint point are (X 1 , Y 1 ), and the joint point coordinates corresponding to the ankle joint point are (X 3 , Y 3 ), then in the skeleton image, the head joint The second height difference between the point and the ankle joint point is |Y 1 -Y 3 |, that is, the height of the gesture of the user to be recognized is |Y 1 -Y 3 |. As shown in FIG. 8 , FIG. 8 is a schematic diagram of determining the gesture height of the user to be recognized.
在一些实施例中,当待识别用户为盘坐姿时,骨骼图像中的最高关节点为头部关节点,最低关节点为臀部关节点;可以根据头部关节点对应的关节点坐标以及臀部关节点对应的关节点坐标确定头部关节点和臀部关节点之间的第二高度差,即得到待识别用户的姿势高度。此时姿势高度为待识别用户盘坐时的身体高度。In some embodiments, when the user to be identified is sitting cross-legged, the highest joint point in the skeletal image is the head joint point, and the lowest joint point is the hip joint point; the joint point coordinates corresponding to the head joint point and the hip joint The joint point coordinates corresponding to the points determine the second height difference between the head joint point and the hip joint point, that is, the posture height of the user to be recognized is obtained. At this time, the posture height is the height of the body when the user to be identified sits cross-legged.
示例性的,若头部关节点对应的关节点坐标为(X 1,Y 1),臀部关节点对应的关节点坐标为(X 4,Y 4),则在骨骼图像中,头部关节点和臀部关节点之间的第二高度差为|Y 1-Y 4|,即待识别用户的姿势高度为|Y 1-Y 4|。 Exemplarily, if the joint point coordinates corresponding to the head joint point are (X 1 , Y 1 ), and the joint point coordinates corresponding to the hip joint point are (X 4 , Y 4 ), then in the skeleton image, the head joint point is The second height difference between it and the hip joint point is |Y 1 -Y 4 |, that is, the height of the posture of the user to be recognized is |Y 1 -Y 4 |.
通过根据骨骼图像中的最高关节点与最低关节点之间的第二高度差,确定待识别用户对应的姿势高度,可以更加真实和准确地反映待识别用户在当前状态对应的姿势高度。By determining the posture height corresponding to the user to be recognized according to the second height difference between the highest joint point and the lowest joint point in the skeleton image, the posture height corresponding to the current state of the user to be recognized can be more truly and accurately reflected.
步骤S30:根据所述待识别用户对应的姿势特征参数,确定所述待识别用户的当前行为状态。Step S30: Determine the current behavior state of the to-be-identified user according to the gesture characteristic parameter corresponding to the to-be-identified user.
在一些实施例中,根据待识别用户对应的姿势特征参数,确定待识别用户的当前行为状态,可以包括:根据人体重心信息、姿势高度以及头部高度,确定待识别用户的当前行为状态。In some embodiments, determining the current behavioral state of the to-be-recognized user according to the posture feature parameters corresponding to the to-be-recognized user may include: determining the current behavioral state of the to-be-recognized user according to the body's center of gravity information, posture height, and head height.
在本申请的实施例中,根据待识别用户的人体重心信息、姿势高度以及头部高度等参数综合确定待识别用户的当前行为状态,无须额外的穿戴设备,可以更加便捷和准确地确定待识别用户的当前行为状态。In the embodiments of the present application, the current behavioral state of the user to be identified is comprehensively determined according to parameters such as the user's body center of gravity information, posture height, and head height, and no additional wearable equipment is required, and the to-be-identified user can be more conveniently and accurately determined. The current behavioral state of the user.
请参阅图9,根据人体重心信息、姿势高度以及头部高度,确定待识别用户的当前行为状态,可以包括以下步骤S31或步骤S32。Referring to FIG. 9 , determining the current behavioral state of the user to be identified according to the body's center of gravity information, posture height, and head height may include the following step S31 or step S32.
步骤S31:若所述姿势高度与所述头部高度的比率处于预设的比率范围中,则根据所述比率范围与姿势类型之间预设的对应关系,确定所述待识别用户对应的姿势类型。Step S31: If the ratio of the posture height to the head height is within a preset ratio range, determine the posture corresponding to the to-be-identified user according to the preset correspondence between the ratio range and the posture type Types of.
示例性的,预设的比率范围可以包括第一比率范围、第二比率范围以及第三比率范围。Exemplarily, the preset ratio range may include a first ratio range, a second ratio range and a third ratio range.
示例性的,姿势类型可以包括但不限于站姿、盘坐姿、坐姿以及跪姿等等。Exemplarily, posture types may include, but are not limited to, standing, cross-sitting, sitting, kneeling, and the like.
其中,第一比率范围为人体的站立高度与头部高度之比;第二比率范围为第二比率为人体的盘坐高度与头部高度之比;第三比率范围为人体的坐姿高度与头部高度之比或跪姿高度与头部高度之比。Among them, the first ratio range is the ratio of the standing height of the human body to the height of the head; the second ratio range is the ratio of the sitting height of the human body to the height of the head; the third ratio range is the sitting height of the human body and the head height The ratio of the height of the waist or the height of the kneeling position to the height of the head.
在一些实施例中,可以根据人体高度与头部高度之间的比率测量数据,确定第一比率范围、第二比率范围以及第三比率范围。In some embodiments, the first ratio range, the second ratio range, and the third ratio range may be determined from the ratio measurement data between the body height and the head height.
示例性的,人体高度与头部高度之间的比率测量数据,如表1所示。Exemplarily, the ratio measurement data between body height and head height are shown in Table 1.
表1:测量数据Table 1: Measurement data
Figure PCTCN2021100379-appb-000002
Figure PCTCN2021100379-appb-000002
Figure PCTCN2021100379-appb-000003
Figure PCTCN2021100379-appb-000003
需要说明的是,表1中的测量数据是根据预设数量的测试者进行测量得到的。表1中的1%所在列对应的测量数据表示1%的测试者对应的测量数据;99%所在列对应的测量数据表示99%的测试者对应的测量数据。A表示每个百分位中所有人的身高与头部身高之比的平均值;H表示99%的测试者的姿势高度与1%测试者的头部高度之比;L表示1%测试者的姿势高度与99%的测试者头部高度之比;P表示各行中所有比值的平均值;H/H表示站立高度与头部高度之比;S/H代表坐姿高度与头部高度之比;C/H代表盘坐姿高度与头部高度之比。It should be noted that, the measurement data in Table 1 are obtained by measuring a preset number of testers. In Table 1, the measurement data corresponding to the 1% column represents the measurement data corresponding to 1% of the testers; the measurement data corresponding to the 99% column represents the measurement data corresponding to 99% of the testers. A is the average of the ratio of height to head height for all people in each percentile; H is the ratio of the posture height of 99% of testers to the head height of 1% of testers; L is the ratio of 1% of testers The ratio of the posture height to the head height of 99% of the subjects; P represents the average of all ratios in each row; H/H represents the ratio of standing height to head height; S/H represents the ratio of sitting height to head height ; C/H represents the ratio of the height of the cross sitting posture to the height of the head.
示例性的,根据表1中的测量数据可以确定:站立高度与头部高度之比的平均值,测试者中的男性为7.551,女性为7.322;坐姿高度与头部高度的比率的平均值,男性为5.953,女性为5.726;盘坐姿高度与头部高度之比的平均值,男性为4.084,女性为3.981。Exemplarily, according to the measurement data in Table 1, it can be determined: the average ratio of the standing height to the head height is 7.551 for the males and 7.322 for the females; the average of the ratio of the sitting height to the head height, Males were 5.953 and females were 5.726; the average ratio of the height of the cross-sitting posture to the head height was 4.084 for males and 3.981 for females.
中国国家标准GB 10000-88的成人体型及有关标准化的文献表明,当比率≥6.5时,可以确定人体处于站立状态;当比率为[3.3,4.5]时,可以确定人体处于盘坐的状态;当比率为[5,6]时,可以确定人体处于坐姿或跪姿状态。因此在本发明实施例中,第一比率范围为设为[6.5,+∞],第二比率范围为[3.3,4.5],第三比率范围为[5,6]。The Chinese national standard GB 10000-88 adult body shape and related standardization literature show that when the ratio is greater than or equal to 6.5, it can be determined that the human body is in a standing state; when the ratio is [3.3, 4.5], it can be determined that the human body is in a cross-sitting state; When the ratio is [5,6], it can be determined that the human body is in a sitting or kneeling state. Therefore, in the embodiment of the present invention, the first ratio range is set to [6.5, +∞], the second ratio range is [3.3, 4.5], and the third ratio range is [5, 6].
示例性的,比率范围与姿势类型之间预设的对应关系,可以如表2所示。Exemplarily, the preset correspondence between the ratio range and the gesture type may be as shown in Table 2.
表2:比率范围与姿势类型对照表Table 2: Comparison of Ratio Ranges and Posture Types
比率范围Ratio range 姿势类型Posture Type
第一比率范围first ratio range 站姿standing
第二比率范围Second Ratio Range 盘坐姿sitting cross-legged
第三比率范围third ratio range 坐姿或跪姿Sitting or kneeling
在一些实施例中,可以根据比率范围与姿势类型之间预设的对应关系,确定待识别用户对应的姿势类型。In some embodiments, the gesture type corresponding to the user to be recognized may be determined according to a preset correspondence between the ratio range and the gesture type.
示例性的,若姿势高度与头部高度的比率处于第一比率范围中,则确定待识别用户对应的姿势类型为站姿。Exemplarily, if the ratio of the posture height to the head height is in the first ratio range, it is determined that the posture type corresponding to the user to be recognized is a standing posture.
示例性的,若姿势高度与头部高度的比率处于第二比率范围中,则确定待识别用户对应的姿势类型为盘坐姿。Exemplarily, if the ratio of the posture height to the head height is within the second ratio range, it is determined that the posture type corresponding to the user to be identified is the cross-sitting posture.
示例性的,若姿势高度与头部高度的比率处于第三比率范围中,则根据待识别用户对应的最低关节点与头部关节点之间的位置关系,判定待识别用户对应的姿势类型为坐姿或跪姿。可以理解的是,当待识别用户对应的姿势高度与头部高度的比率处于第三比率范围中时,由于第三比率范围对应的姿势类型为坐姿或跪姿,因此需要进一步确定待识别用户的姿势类型。Exemplarily, if the ratio of the posture height to the head height is in the third ratio range, then according to the positional relationship between the lowest joint point corresponding to the to-be-identified user and the head joint point, it is determined that the posture type corresponding to the to-be-identified user is: Sitting or kneeling. It can be understood that, when the ratio of the posture height to the head height corresponding to the user to be identified is in the third ratio range, since the posture type corresponding to the third ratio range is a sitting posture or a kneeling posture, it is necessary to further determine the posture of the user to be identified. Posture type.
需要说明的是,在骨骼图像中,若待识别用户为坐姿,则骨骼图像中的最低关节点为脚腕关节点,此时脚腕关节点与头部关节点不处于同一垂直线上;若待识别用户为跪姿,则骨骼图像中的最低关节点为膝部关节点,此时膝部关节点与头部关节点处于同一垂直线上。因此,可以根据骨骼图像中的最低关节点与头部关节点之间的位置关系,判定待识别用户对应的姿势类型为坐姿或跪姿。It should be noted that, in the skeleton image, if the user to be identified is in a sitting posture, the lowest joint point in the skeleton image is the ankle joint point, and the ankle joint point and the head joint point are not on the same vertical line; if If the user to be identified is in a kneeling posture, the lowest joint point in the skeleton image is the knee joint point, and at this time the knee joint point and the head joint point are on the same vertical line. Therefore, according to the positional relationship between the lowest joint point in the skeleton image and the head joint point, it can be determined that the posture type corresponding to the user to be identified is a sitting posture or a kneeling posture.
在一些实施例中,若最低关节点与头部关节点处于同一垂直线上,则确定待识别用户对应的姿势类型为跪姿。In some embodiments, if the lowest joint point and the head joint point are on the same vertical line, it is determined that the posture type corresponding to the user to be identified is a kneeling posture.
在另一些实施例中,若最低关节点与头部关节点不处于同一垂直线上,则确定待识别用户对应的姿势类型为坐姿。In other embodiments, if the lowest joint point and the head joint point are not on the same vertical line, it is determined that the posture type corresponding to the user to be identified is a sitting posture.
通过判断待识别用户对应的姿势高度与头部高度的比率是否处于第一比率范围、第二比率范围或第三比率范围,可以更加准确地确定待识别用户对应的姿势类型。通过判断最低关节点与头部关节点是否处于同一垂直线上,可以更加准确地确定待识别用户的姿势类型为坐姿或跪姿。By judging whether the ratio of the height of the gesture corresponding to the user to be recognized to the height of the head is within the first ratio range, the second ratio range or the third ratio range, the gesture type corresponding to the user to be recognized can be more accurately determined. By judging whether the lowest joint point and the head joint point are on the same vertical line, it can be more accurately determined that the posture type of the user to be recognized is a sitting posture or a kneeling posture.
步骤S32:若所述姿势高度与所述头部高度的比率不处于预设的比率范围中,则根据所述人体重心信息按照预设的检测策略判断所述待识别用户是否摔倒。Step S32 : if the ratio of the posture height to the head height is not within the preset ratio range, determine whether the user to be identified falls according to the body center of gravity information according to a preset detection strategy.
示例性的,人体重心信息除了包括重心点,还可以包括重心下降速率。检测策略可以包括:确定待识别用户的重心下降速率是否大于预设的下降速率;当重心下降速率大于预设的下降速率时,确定待识别用户对应的臀部高度值;当臀部高度值小于预设的高度值时,则判定待识别用户为摔倒状态,其中,臀部高度值为待识别用户的臀部与地面之间的距离。Exemplarily, in addition to the center of gravity point, the information about the center of gravity of the human body may also include the rate of descent of the center of gravity. The detection strategy may include: determining whether the center of gravity descending rate of the user to be identified is greater than a preset descending rate; when the descending rate of the center of gravity is greater than the preset descending rate, determining a hip height value corresponding to the user to be identified; when the hip height value is less than the preset descending rate When the height value is , it is determined that the user to be identified is in a falling state, wherein the height value of the buttocks is the distance between the buttocks of the user to be identified and the ground.
请参阅图10,图10是步骤S32中根据人体重心信息按照预设的检测策略判断待识别用户是否摔倒的示意性流程图,具体可以包括以下步骤S321至步骤S323。Please refer to FIG. 10 . FIG. 10 is a schematic flowchart of determining whether the user to be identified falls according to the information of the center of gravity of the human body according to the preset detection strategy in step S32 , which may specifically include the following steps S321 to S323 .
步骤S321:检测所述待识别用户的重心下降速率。Step S321: Detect the drop rate of the center of gravity of the user to be identified.
在一些实施例中,检测待识别用户的重心下降速率之前,还包括:获取深度图像中的人体重心点,将人体重心点标注在骨骼图像中。In some embodiments, before detecting the drop rate of the center of gravity of the user to be identified, the method further includes: acquiring the center of gravity of the human body in the depth image, and marking the center of gravity of the human body in the skeleton image.
需要说明的是,由于在上述实施例中已经根据深度图像确定待识别用户对应的人体重心点,其中,因此可以直接获取深度图像中的人体重心点,并将人体重心点标注在骨骼图像中。It should be noted that, in the above-mentioned embodiments, the human body center of gravity point corresponding to the user to be identified has been determined according to the depth image, wherein the human body center point of gravity in the depth image can be directly obtained and marked in the skeleton image.
通过将人体重心点标注在骨骼图像中,可以通过骨骼图像中的人体重心点在预设的间隔时间内的位置变化,确定待识别用户对应的重心下降速率。By marking the center of gravity of the human body in the skeleton image, it is possible to determine the drop rate of the center of gravity corresponding to the user to be identified by changing the position of the center of gravity of the human body in the skeleton image within a preset interval.
在一些实施例中,获取视频图像帧中相邻预设的间隔时间的第一骨骼图像和第二骨骼图像,第一骨骼图像包括第一人体重心点,第二骨骼图像包括第二人体重心点;根据第一人体重心点对应的坐标、第二人体重心点对应的坐标以及间隔时间,确定待识别用户的重心下降速率,其中,预设的间隔时间可以根据实际情况设定,具体数值在此不作限定。In some embodiments, a first skeletal image and a second skeletal image at adjacent preset intervals in the video image frame are acquired, the first skeletal image includes the center of gravity of the first human body, and the second skeletal image includes the center of gravity of the second human body point; according to the coordinates corresponding to the center of gravity of the first person, the coordinates corresponding to the center of gravity of the second person, and the interval time, determine the drop rate of the center of gravity of the user to be identified, wherein the preset interval time can be set according to the actual situation, and the specific value It is not limited here.
示例性的,以获取10帧视频图像帧的时间为间隔时间t,例如将第1帧的视频图像帧中的骨骼图像作为第一骨骼图像,将第10帧的视频图像帧中的骨骼图像作为第二骨骼图像。由于已经在骨骼图像中标注了人体重心点,因此第一骨骼图像包括第一人体重心点,第二骨骼图像包括第二人体重心点。Exemplarily, take the time of acquiring 10 video image frames as the interval time t, for example, the skeleton image in the video image frame of the first frame is used as the first skeleton image, and the skeleton image in the video image frame of the 10th frame is used as the interval time t. Second bone image. Since the center of gravity of the human body has been marked in the skeleton image, the first skeleton image includes the center of gravity of the first human body, and the second skeleton image includes the center of gravity of the second human body.
示例性的,若第一人体重心点对应的坐标为(X 01,Y 01)、第二人体重心点对应的坐标(X 10,Y 10),则根据第一人体重心点对应的坐标(X 01,Y 01)、第二人体重心点对应的坐标(X 10,Y 10)以及间隔时间t,可以确定待识别用户的重心下降速率。重心下降速率用v表示,重心下降速率v可以通过下式计算得到: Exemplarily, if the coordinates corresponding to the center of gravity of the first person are (X 01 , Y 01 ) and the coordinates corresponding to the center of gravity of the second person (X 10 , Y 10 ), then according to the coordinates corresponding to the center of gravity of the first person (X 01 , Y 01 ), the coordinates (X 10 , Y 10 ) corresponding to the center of gravity of the second person, and the interval time t, can determine the drop rate of the center of gravity of the user to be identified. The descent rate of the center of gravity is represented by v, and the descent rate of the center of gravity v can be calculated by the following formula:
Figure PCTCN2021100379-appb-000004
Figure PCTCN2021100379-appb-000004
步骤S322:当所述重心下降速率大于预设的下降速率时,检测所述待识别用户的臀部高度值。Step S322: When the lowering rate of the center of gravity is greater than a preset lowering rate, detect the hip height value of the user to be identified.
示例性的,预设的下降速率可以用V表示,其中,下降速率V可以根据实际情况设定,具体数值在此不作限定。Exemplarily, the preset falling rate may be represented by V, where the falling rate V may be set according to the actual situation, and the specific value is not limited herein.
在一些实施例中,当重心下降速率v大于预设的下降速率V时,检测待识别用户的臀部高度值。In some embodiments, when the gravity center descending rate v is greater than the preset descending rate V, the hip height value of the user to be identified is detected.
在另一些实施例中,当重心下降速率v不大于预设的下降速率V时,继续检测待识别用户的重心下降速率v,直至检测到大于预设的下降速率V的重心下降速率v。In other embodiments, when the gravity center descending rate v is not greater than the preset descending rate V, continue to detect the gravity center descending rate v of the user to be identified until a gravity center descending rate v greater than the preset descending rate V is detected.
示例性的,获取视频图像帧中骨骼图像,例如将第11帧的视频图像帧中的骨骼图像作为第一骨骼图像,将第20帧的视频图像帧中的骨骼图像作为第二骨骼图像;继续根据第一骨骼图像与第二骨骼图像中的重心点对应的坐标以及间隔时间,确定待识别用户的重心下降速率。Exemplarily, acquiring the skeleton image in the video image frame, for example, using the skeleton image in the video image frame of the 11th frame as the first skeleton image, and using the skeleton image in the video image frame of the 20th frame as the second skeleton image; continue According to the coordinates and the interval time corresponding to the center of gravity point in the first skeleton image and the second skeleton image, the drop rate of the center of gravity of the user to be identified is determined.
在一些实施例中,检测待识别用户的臀部高度值可以包括:获取骨骼图像中的臀部关节点对应的坐标;确定臀部关节点对应的坐标与地面之间的垂直距离,将垂直距离作为待识别用户的臀部高度值。In some embodiments, detecting the height value of the buttocks of the user to be identified may include: acquiring the coordinates corresponding to the hip joint points in the skeleton image; determining the vertical distance between the coordinates corresponding to the hip joint points and the ground, and using the vertical distance as the to-be-identified distance The user's hip height value.
由于上述实施例中已经获取骨骼图像中的关节点坐标,因此可以直接得到骨骼图像中的臀部关节点对应的坐标;示例性的,在骨骼图像中,臀部关节点对应的坐标为(X 4,Y 4)。 Since the coordinates of the joint points in the skeleton image have been obtained in the above embodiment, the coordinates corresponding to the hip joint points in the skeleton image can be directly obtained; exemplarily, in the skeleton image, the coordinates corresponding to the hip joint points are (X 4 , Y4 ) .
示例性的,可以以体感器为坐标原点建立三维坐标系,其中,臀部关节点与体感器之间的距离可以通过深度图像中的深度信息确定;例如,若臀部关节点与体感器之间的距离为Z 4,则在三维坐标系中,臀部关节点对应的坐标为(X 4,Y 4,Z 4)。 Exemplarily, a three-dimensional coordinate system can be established with the somatosensory as the coordinate origin, wherein the distance between the hip joint point and the somatosensory can be determined by the depth information in the depth image; for example, if the distance between the hip joint point and the somatosensory If the distance is Z 4 , in the three-dimensional coordinate system, the coordinates corresponding to the hip joint point are (X 4 , Y 4 , Z 4 ).
示例性的,地面可以表示为AX+BY+CZ+D=0;式中,A、B、C为系数,D为常数。Exemplarily, the ground can be expressed as AX+BY+CZ+D=0; in the formula, A, B, C are coefficients, and D is a constant.
在本申请的实施例中,常数D表示体感器与地面之间的距离。在一些实施例中,可以通过kinect SDK中的get_FloorClipPlane函数获取三个不在同一直线上的点坐标,将三个点坐标代入AX+BY+CZ+D=0中,可以确定系数A、B、C。In the embodiments of the present application, the constant D represents the distance between the somatosensory sensor and the ground. In some embodiments, the get_FloorClipPlane function in the kinect SDK can be used to obtain three point coordinates that are not on the same straight line, and the three point coordinates can be substituted into AX+BY+CZ+D=0 to determine the coefficients A, B, C .
示例性的,确定臀部关节点对应的坐标为(X 4,Y 4,Z 4)与地面之间的垂直距离,可以通过点到面的距离公式进行确定,如下所示: Exemplarily, the coordinates corresponding to the hip joint point are determined to be the vertical distance between (X 4 , Y 4 , Z 4 ) and the ground, which can be determined by the point-to-surface distance formula, as shown below:
Figure PCTCN2021100379-appb-000005
Figure PCTCN2021100379-appb-000005
式中,H表示臀部关节点与地面之间垂直距离,即待识别用户的臀部高度值为H。In the formula, H represents the vertical distance between the hip joint point and the ground, that is, the hip height value of the user to be identified is H.
步骤S323:当所述臀部高度值小于预设的高度值时,确定所述待识别用户为摔倒状态。其中,预设的高度值可以根据成年男性的腰部、臀部的平均宽度以及成年女性的腰部、臀部平均宽度进行设定,具体数值在此不作限定。Step S323: When the hip height value is less than a preset height value, determine that the user to be identified is in a falling state. Wherein, the preset height value can be set according to the average width of the waist and buttocks of an adult male and the average width of the waist and buttocks of an adult female, and the specific value is not limited herein.
示例性的,当臀部高度值H小于预设的高度值时,说明待识别用户的臀部距离地面较近,此时可以确定待识别用户为摔倒状态。Exemplarily, when the hip height value H is smaller than the preset height value, it means that the buttocks of the user to be identified are relatively close to the ground, and at this time, it can be determined that the user to be identified is in a falling state.
通过先判断待识别用户的重心下降速率是否大于预设的下降速率,然后再检测待识别用户的臀部高度值,从而可以结合重心下降速率和臀部高度值判断待识别用户是否为摔倒状态,极大地提高了准确度。By first judging whether the lowering rate of the center of gravity of the user to be recognized is greater than the preset lowering rate, and then detecting the height of the hips of the user to be recognized, it is possible to combine the lowering rate of the center of gravity and the height of the hips to determine whether the user to be recognized is in a fall state, and it is extremely Greatly improved accuracy.
在一些实施例中,确定待识别用户为摔倒状态之后,还包括:向待识别用户对应的家属或医院发送紧急通知,以使家属或医院根据紧急通知发现待识别用户摔倒并及时处理。In some embodiments, after determining that the to-be-identified user is in a fall state, the method further includes: sending an emergency notification to the family or hospital corresponding to the to-be-identified user, so that the family or hospital finds that the to-be-identified user has fallen and handles it in time according to the emergency notification.
示例性的,发送紧急通知的方式可以包括但不限于短信、电话以及邮件等等。Exemplarily, the manner of sending the emergency notification may include, but is not limited to, text messages, phone calls, emails, and the like.
示例性的,紧急通知可以包括待识别用户的位置信息,还可以包括待识别用户对应的深度图像以及骨骼图像等等。Exemplarily, the emergency notification may include location information of the user to be identified, and may also include depth images and skeleton images corresponding to the user to be identified.
通过在确定待识别用户为摔倒状态时,向待识别用户对应的家属或医院发送紧急通知,可以及时发现待识别用户的摔倒并为抢救争取时间。By sending an emergency notification to a family member or hospital corresponding to the user to be identified when it is determined that the user to be identified is in a fall state, the fall of the user to be identified can be detected in time and time can be saved for rescue.
上述实施例提供的行为识别方法、装置、系统和存储介质,通过获取待识别用户对应的深度图像和骨骼图像,可以更加便捷地根据深度图像确定人体重心信息以及根据骨骼图像确定姿势高度和头部高度;通过根据深度图像中的人体区域内的像素点总数和各像素点的坐标,可以更加准确地确定待识别用户的人体重心点;通过根据头部关节点对应的关节点坐标和颈部关节点对应的关节点坐标,可以更加准确地得到头部高度,提高了后续判断待识别用户的当前行为状态的准确性;通过根据骨骼图像中的最高关节点与最低关节点之间的第二高度差,确定待识别用户对应的姿势高度,可以更加真实和准确地反映待识别用户在当前状态对应的姿势高度;通过判断待识别用户对应的姿势高度与头部高度的比率是否处于第一比率范围、第二比率范围或第三比率范围,可以更加准确地确定待识别用户对应的姿势类型。通过判断最低关节点与头部关节点是否处于同一垂直线上,可以更加准确地确定待识别用户的姿势类型为坐姿或跪姿;通过先判断待识别用户的重心下降速率是否大于预设的下降速率,然后再检测待识别用户的臀部高度值,从而可以结合重心下降速率和臀部高度值判断待识别用户是否为摔倒状态,极大地提高了识别准确度。The behavior recognition method, device, system and storage medium provided by the above embodiments can more conveniently determine the body center of gravity information according to the depth image and determine the posture height and head according to the skeleton image by acquiring the depth image and skeleton image corresponding to the user to be recognized. Height; the center of gravity of the user to be identified can be more accurately determined according to the total number of pixels in the human body area in the depth image and the coordinates of each pixel; The coordinates of the joint points corresponding to the points can obtain the height of the head more accurately, which improves the accuracy of the subsequent judgment of the current behavior state of the user to be identified; difference, determine the posture height corresponding to the user to be recognized, which can more truly and accurately reflect the posture height corresponding to the user to be recognized in the current state; by judging whether the ratio of the posture height corresponding to the user to be recognized to the head height is within the first ratio range , the second ratio range or the third ratio range, the gesture type corresponding to the user to be recognized can be more accurately determined. By judging whether the lowest joint point and the head joint point are on the same vertical line, it can be more accurately determined that the posture type of the user to be identified is sitting or kneeling; by first judging whether the downward rate of the center of gravity of the user to be identified is greater than the preset drop Then, the hip height value of the user to be identified can be detected, so that it can be determined whether the user to be identified is in a fall state by combining the lowering rate of the center of gravity and the hip height value, which greatly improves the identification accuracy.
本申请的实施例中还提供一种存储介质,用于可读存储,所述存储介质存储有程序,所述程序中包括程序指令,所述处理器执行所述程序指令,实现本申请实施例提供的任一项行为识别方法。The embodiments of the present application further provide a storage medium for readable storage, the storage medium stores a program, the program includes program instructions, and the processor executes the program instructions to implement the embodiments of the present application Any of the behavioral identification methods provided.
例如,该程序被处理器加载,可以执行如下步骤:For example, the program is loaded by the processor and can perform the following steps:
获取待识别用户对应的视频图像帧,其中,所述视频图像帧包括深度图像和骨骼图像;根据所述深度图像和所述骨骼图像确定所述待识别用户对应的姿势特征参数;根据所述待识别用户对应的姿势特征参数,确定所述待识别用户的当前行为状态。Obtain a video image frame corresponding to the user to be identified, wherein the video image frame includes a depth image and a skeleton image; determine the posture feature parameter corresponding to the user to be identified according to the depth image and the skeleton image; The gesture characteristic parameters corresponding to the user are identified, and the current behavior state of the to-be-identified user is determined.
其中,所述存储介质可以是前述实施例所述的行为识别装置的内部存储单元,例如所述行为识别装置的硬盘或内存。所述存储介质也可以是所述行为识别装置的外部存储设备,例如所述行为识别装置上配备的插接式硬盘,智能存储卡(Smart Media Card,SMC),安全数字卡(Secure Digital Card,SD Card),闪存卡(Flash Card)等。Wherein, the storage medium may be an internal storage unit of the behavior recognition apparatus described in the foregoing embodiments, such as a hard disk or a memory of the behavior recognition apparatus. The storage medium may also be an external storage device of the behavior recognition device, such as a plug-in hard disk equipped on the behavior recognition device, a smart memory card (Smart Media Card, SMC), a Secure Digital Card (Secure Digital Card, SD Card), flash memory card (Flash Card), etc.
本领域普通技术人员可以理解,上文中所公开方法中的全部或某些步骤、系统、设备中的功能模块/单元可以被实施为软件、固件、硬件及其适当的组合。Those of ordinary skill in the art can understand that all or some of the steps in the methods disclosed above, functional modules/units in the systems, and devices can be implemented as software, firmware, hardware, and appropriate combinations thereof.
在硬件实施例中,在以上描述中提及的功能模块/单元之间的划分不一定对应于物理组件的划分;例如,一个物理组件可以具有多个功能,或者一个功能或步骤可以由若干物理组件合作执行。某些物理组件或所有物理组件可以被实施为由处理器,如中央处理器、数字信号处理器或微处理器执行的软件,或者被实施为硬件,或者被实施为集成电路,如专用集成电路。这样的软件可以分布在可存储介质上,存储介质可以包括计算机存储介质(或非暂时性介质)和通信介质(或暂时性介质)。如本领域普通技术人员公知的,术语存储介质包括在用于存储信息(诸如计算机可读指令、数据结构、程序模块或其他数据)的任何方法或技术中实施的易失性和非易失性、可移除和不可移除介质。计算机存储介质包括但不限于RAM、ROM、EEPROM、闪存或其他存储器技术、CD-ROM、数字多功能盘(DVD)或其他光盘存储、磁盒、磁带、磁盘存储或其他磁存储装置、或者可以用于存储期望的信息并且可以被计算机访问的任何其他的介质。此外,本领域普通技术人员公知的是,通信介质通常包含计算机可读指令、 数据结构、程序模块或者诸如载波或其他传输机制之类的调制数据信号中的其他数据,并且可包括任何信息递送介质。In hardware embodiments, the division between functional modules/units mentioned in the above description does not necessarily correspond to the division of physical components; for example, one physical component may have multiple functions, or one function or step may be composed of several physical components Components execute cooperatively. Some or all physical components may be implemented as software executed by a processor, such as a central processing unit, digital signal processor or microprocessor, or as hardware, or as an integrated circuit, such as an application specific integrated circuit . Such software may be distributed on storable media, which may include computer storage media (or non-transitory media) and communication media (or transitory media). As is known to those of ordinary skill in the art, the term storage medium includes both volatile and nonvolatile implemented in any method or technology for storage of information, such as computer readable instructions, data structures, program modules or other data , removable and non-removable media. Computer storage media include, but are not limited to, RAM, ROM, EEPROM, flash memory or other memory technology, CD-ROM, digital versatile disk (DVD) or other optical disk storage, magnetic cartridges, magnetic tape, magnetic disk storage or other magnetic storage devices, or may Any other medium used to store desired information and which can be accessed by a computer. In addition, communication media typically embodies computer readable instructions, data structures, program modules, or other data in a modulated data signal such as a carrier wave or other transport mechanism, and can include any information delivery media, as is well known to those of ordinary skill in the art .
以上参照附图说明了本发明的优选实施例,并非因此局限本发明的权利范围。本领域技术人员不脱离本发明的范围和实质内所作的任何修改、等同替换和改进,均应在本发明的权利范围之内。The preferred embodiments of the present invention have been described above with reference to the accompanying drawings, but are not intended to limit the scope of the rights of the present invention. Any modifications, equivalent substitutions and improvements made by those skilled in the art without departing from the scope and spirit of the present invention shall fall within the right scope of the present invention.

Claims (15)

  1. 一种行为识别方法,包括:A behavioral identification method comprising:
    获取待识别用户对应的视频图像帧,其中,所述视频图像帧包括深度图像和骨骼图像;acquiring a video image frame corresponding to the user to be identified, wherein the video image frame includes a depth image and a skeleton image;
    根据所述深度图像和所述骨骼图像确定所述待识别用户对应的姿势特征参数;Determine the posture feature parameter corresponding to the user to be recognized according to the depth image and the skeleton image;
    根据所述待识别用户对应的姿势特征参数,确定所述待识别用户的当前行为状态。The current behavior state of the to-be-identified user is determined according to the gesture characteristic parameter corresponding to the to-be-identified user.
  2. 根据权利要求1所述的行为识别方法,其中,所述姿势特征参数包括人体重心信息、姿势高度和头部高度;所述根据所述深度图像和所述骨骼图像确定所述待识别用户对应的姿势特征参数,包括:The behavior recognition method according to claim 1, wherein the posture feature parameters include body center of gravity information, posture height and head height; Pose feature parameters, including:
    根据所述深度图像确定所述待识别用户对应的人体重心信息,以及根据所述骨骼图像确定所述待识别用户对应的姿势高度和头部高度;Determine the body center of gravity information corresponding to the to-be-identified user according to the depth image, and determine the posture height and head height corresponding to the to-be-identified user according to the skeleton image;
    所述根据所述待识别用户对应的姿势特征参数,确定所述待识别用户的当前行为状态,包括:The determining the current behavioral state of the user to be identified according to the gesture characteristic parameter corresponding to the user to be identified includes:
    根据所述人体重心信息、所述姿势高度以及所述头部高度,确定所述待识别用户的当前行为状态。The current behavior state of the user to be identified is determined according to the information on the center of gravity of the human body, the posture height, and the head height.
  3. 根据权利要求2所述的行为识别方法,其中,在所述根据所述深度图像确定所述待识别用户对应的人体重心信息之前,还包括:The behavior recognition method according to claim 2, wherein before the determining according to the depth image the information on the body center of gravity corresponding to the user to be recognized, the method further comprises:
    根据预设的数据格式,对初始的深度图像中的深度数据进行格式转换,得到格式转换后的所述深度图像;According to a preset data format, format conversion is performed on the depth data in the initial depth image to obtain the format-converted depth image;
    在所述根据所述骨骼图像确定所述待识别用户对应的姿势高度以及头部高度之前,还包括:Before determining the posture height and the head height corresponding to the user to be identified according to the skeleton image, the method further includes:
    根据预设的平滑处理策略,对初始的骨骼图像进行平滑处理,得到平滑处理后的所述骨骼图像。According to a preset smoothing processing strategy, the initial skeleton image is smoothed to obtain the smoothed skeleton image.
  4. 根据权利要求3所述的行为识别方法,其中,所述人体重心信息包括人体重心点;所述根据所述深度图像确定所述待识别用户对应的人体重心信息,包括:The behavior recognition method according to claim 3, wherein the information on the center of gravity of the human body includes a center of gravity point of the human body; and the determining the information on the center of gravity of the human body corresponding to the user to be identified according to the depth image comprises:
    获取格式转换后的所述深度图像中的人体区域内的像素点总数,以及获取所述人体区域中的全部像素点对应的横坐标和纵坐标;Obtain the total number of pixels in the human body region in the depth image after format conversion, and obtain the abscissa and vertical coordinates corresponding to all the pixels in the human body region;
    确定所述人体区域中的全部像素点对应的横坐标之和与纵坐标之和;Determine the sum of the abscissas and ordinates corresponding to all the pixels in the human body area;
    根据所述像素点总数和所述横坐标之和,确定所述全部像素点对应的横坐标的均值,以及根据所述像素点总数和所述纵坐标之和,确定所述全部像素点对应的纵坐标的均值;According to the sum of the total number of pixels and the abscissa, the mean value of the abscissa corresponding to all the pixels is determined, and according to the sum of the total number of pixels and the ordinate, the average value of the corresponding the mean of the ordinate;
    将所述全部像素点对应的横坐标的均值作为所述人体重心点的横坐标,以及将所述全部像素点对应的纵坐标的均值作为所述人体重心点的纵坐标,得到所述深度图像中的人体重心点。Taking the mean value of the abscissas corresponding to all the pixels as the abscissa of the body's center of gravity, and using the mean of the ordinates corresponding to all the pixels as the ordinate of the body's center of gravity to obtain the depth image The center of gravity of the human body.
  5. 根据权利要求3或4所述的行为识别方法,其中,所述根据所述骨骼图像确定所述待识别用户对应的姿势高度以及头部高度,包括:The behavior recognition method according to claim 3 or 4, wherein the determining the posture height and the head height corresponding to the to-be-recognized user according to the skeleton image, comprises:
    提取平滑处理后的所述骨骼图像中的关节点信息,根据所述关节点信息确定所述待识别用户对应的姿势高度以及头部高度。Extracting joint point information in the smoothed skeleton image, and determining a posture height and a head height corresponding to the user to be identified according to the joint point information.
  6. 根据权利要求5所述的行为识别方法,其中,所述骨骼图像包括头部关节点;在所述根据所述关节点信息确定所述待识别用户对应的姿势高度以及头部高度之前,还包括:The behavior recognition method according to claim 5, wherein the skeleton image includes head joint points; before determining the posture height and the head height corresponding to the user to be recognized according to the joint point information, further comprising: :
    获取所述骨骼图像中的头部关节点;obtaining the head joint points in the skeleton image;
    当所述头部关节点位于所述深度图像中的人体区域内,根据所述关节点信息确定所述待识别用户对应的姿势高度以及头部高度。When the head joint point is located in the human body area in the depth image, the posture height and the head height corresponding to the user to be identified are determined according to the joint point information.
  7. 根据权利要求6所述的行为识别方法,其中,所述骨骼图像还包括颈部关节点;所述根据所述关节点信息确定所述待识别用户对应的头部高度,包括:The behavior recognition method according to claim 6, wherein the skeleton image further includes neck joint points; and the determining the head height corresponding to the user to be identified according to the joint point information comprises:
    根据所述头部关节点对应的关节点坐标和所述颈部关节点对应的关节点坐标,确定所述头部关节点与所述颈部关节点之间的第一高度差;determining a first height difference between the head joint point and the neck joint point according to the joint point coordinates corresponding to the head joint point and the joint point coordinates corresponding to the neck joint point;
    根据预设的高度比值与所述第一高度差的积,得到所述待识别用户对应的头部高度;obtaining the head height corresponding to the user to be identified according to the product of the preset height ratio and the first height difference;
    所述根据所述关节点信息确定所述待识别用户对应的姿势高度,包括:The determining the posture height corresponding to the user to be identified according to the joint point information includes:
    确定所述骨骼图像中的最高关节点和最低关节点;determining the highest joint point and the lowest joint point in the skeleton image;
    根据所述最高关节点对应的关节点坐标和所述最低关节点对应的关节点坐标,确定所述最高关节点和所述最低关节点之间的第二高度差,将所述第二高度差作为所述待识别用户对应的姿势高度。According to the joint point coordinates corresponding to the highest joint point and the joint point coordinates corresponding to the lowest joint point, a second height difference between the highest joint point and the lowest joint point is determined, and the second height difference as the gesture height corresponding to the to-be-identified user.
  8. 根据权利要求2-7中任一项所述的行为识别方法,其中,所述当前行为状态包括姿势类型和是否摔倒;所述根据所述人体重心信息、所述姿势高度以及所述头部高度,确定所述 待识别用户的当前行为状态,包括:The behavior recognition method according to any one of claims 2-7, wherein the current behavior state includes posture type and whether to fall; height, determine the current behavior status of the user to be identified, including:
    若所述姿势高度与所述头部高度的比率处于预设的比率范围中,则根据所述比率范围与姿势类型之间预设的对应关系,确定所述待识别用户对应的姿势类型;或If the ratio of the gesture height to the head height is within a preset ratio range, determine the gesture type corresponding to the to-be-identified user according to the preset correspondence between the ratio range and the gesture type; or
    若所述姿势高度与所述头部高度的比率不处于预设的比率范围中,则根据所述人体重心信息按照预设的检测策略判断所述待识别用户是否摔倒。If the ratio of the posture height to the head height is not within a preset ratio range, it is determined whether the user to be identified falls down according to the information on the center of gravity of the human body according to a preset detection strategy.
  9. 根据权利要求8所述的行为识别方法,其中,所述预设的比率范围包括第一比率范围、第二比率范围以及第三比率范围,所述姿势类型包括站姿、盘坐姿、坐姿以及跪姿;The behavior recognition method according to claim 8, wherein the preset ratio range includes a first ratio range, a second ratio range and a third ratio range, and the posture types include standing posture, cross-sitting posture, sitting posture and kneeling posture posture;
    其中,所述第一比率范围为人体的站立高度与头部高度之比,所述第二比率范围为第二比率为人体的盘坐高度与头部高度之比,所述第三比率范围为人体的坐姿高度与头部高度之比或跪姿高度与头部高度之比。Wherein, the first ratio range is the ratio of the standing height of the human body to the height of the head, the second ratio range is that the second ratio is the ratio of the sitting height of the human body to the head height, and the third ratio range is The ratio of the sitting height to the head height of the human body or the ratio of the kneeling height to the head height.
  10. 根据权利要求9所述的行为识别方法,其中,所述若所述姿势高度与所述头部高度的比率处于预设的比率范围中,则根据所述比率范围与姿势类型之间预设的对应关系,确定所述待识别用户对应的姿势类型,包括:The behavior recognition method according to claim 9, wherein, if the ratio of the posture height to the head height is in a preset ratio range, then the ratio between the ratio range and the posture type is determined according to a preset ratio. Correspondence, determine the gesture type corresponding to the user to be identified, including:
    若所述姿势高度与所述头部高度的比率处于所述第一比率范围中,则确定所述待识别用户对应的姿势类型为站姿;或If the ratio of the posture height to the head height is within the first ratio range, determining that the posture type corresponding to the user to be identified is a standing posture; or
    若所述姿势高度与所述头部高度的比率处于所述第二比率范围中,则确定所述待识别用户对应的姿势类型为盘坐姿;或If the ratio of the posture height to the head height is within the second ratio range, determining that the posture type corresponding to the user to be identified is a cross-sitting posture; or
    若所述姿势高度与所述头部高度的比率处于所述第三比率范围中,则根据所述待识别用户对应的最低关节点与头部关节点之间的位置关系,判定所述待识别用户对应的姿势类型为坐姿或跪姿。If the ratio of the posture height to the head height is within the third ratio range, determine the to-be-identified user according to the positional relationship between the lowest joint point corresponding to the to-be-identified user and the head joint point The user's corresponding posture type is sitting or kneeling.
  11. 根据权利要求10所述的行为识别方法,其中,所述根据所述待识别用户对应的最低关节点与头部关节点之间的位置关系,判定所述待识别用户对应的姿势类型为坐姿或跪姿,包括:The behavior recognition method according to claim 10, wherein, according to the positional relationship between the lowest joint point corresponding to the to-be-identified user and the head joint point, it is determined that the posture type corresponding to the to-be-identified user is a sitting posture or a Kneeling positions, including:
    若所述最低关节点与所述头部关节点处于同一垂直线上,则确定所述待识别用户对应的姿势类型为跪姿;或If the lowest joint point and the head joint point are on the same vertical line, it is determined that the posture type corresponding to the to-be-identified user is kneeling posture; or
    若所述最低关节点与所述头部关节点不处于同一垂直线上,则确定所述待识别用户对应的姿势类型为坐姿。If the lowest joint point and the head joint point are not on the same vertical line, it is determined that the posture type corresponding to the to-be-identified user is a sitting posture.
  12. 根据权利要求8-11中任一项所述的行为识别方法,其中,所述人体重心信息还包括重心下降速率;所述若所述姿势高度与所述头部高度的比率不处于预设的比率范围中,则根据所述人体重心信息按照预设的检测策略判断所述待识别用户是否摔倒,包括:The behavior recognition method according to any one of claims 8-11, wherein the information on the center of gravity of the human body further includes a rate of descent of the center of gravity; if the ratio of the posture height to the head height is not within a preset In the ratio range, according to the information of the center of gravity of the human body, it is determined whether the user to be identified falls according to the preset detection strategy, including:
    检测所述待识别用户的重心下降速率;Detecting the drop rate of the center of gravity of the user to be identified;
    当所述重心下降速率大于预设的下降速率时,检测所述待识别用户的臀部高度值;When the lowering rate of the center of gravity is greater than a preset lowering rate, detecting the height value of the buttocks of the user to be identified;
    当所述臀部高度值小于预设的高度值时,确定所述待识别用户为摔倒状态。When the hip height value is smaller than a preset height value, it is determined that the user to be identified is in a falling state.
  13. 根据权利要求12所述的行为识别方法,其中,所述检测所述待识别用户的重心下降速率,包括:The behavior identification method according to claim 12, wherein the detecting the gravity center drop rate of the to-be-identified user comprises:
    获取所述视频图像帧中相邻预设的间隔时间的第一骨骼图像和第二骨骼图像,其中,所述第一骨骼图像包括第一人体重心点,所述第二骨骼图像包括第二人体重心点;Acquire a first skeleton image and a second skeleton image at adjacent preset intervals in the video image frame, wherein the first skeleton image includes a first human body center of gravity, and the second skeleton image includes a second skeleton image. center of gravity
    根据所述第一人体重心点对应的坐标、所述第二人体重心点对应的坐标以及所述间隔时间,确定所述待识别用户的重心下降速率;According to the coordinates corresponding to the center of gravity of the first person, the coordinates corresponding to the center of gravity of the second person, and the interval time, determine the rate of descent of the center of gravity of the user to be identified;
    所述检测所述待识别用户的臀部高度值,包括:The detecting the hip height value of the user to be identified includes:
    获取所述骨骼图像中的臀部关节点对应的坐标;obtaining the coordinates corresponding to the hip joint points in the skeleton image;
    确定所述臀部关节点对应的坐标与地面之间的垂直距离,将所述垂直距离作为所述待识别用户的臀部高度值。Determine the vertical distance between the coordinates corresponding to the hip joint point and the ground, and use the vertical distance as the hip height value of the user to be identified.
  14. 一种行为识别装置,包括:A behavior recognition device, comprising:
    存储器,用于存储程序;memory for storing programs;
    处理器,用于执行所述程序并在执行所述程序时实现如权利要求1至13任一项所述的行为识别方法。The processor is configured to execute the program and implement the behavior recognition method according to any one of claims 1 to 13 when the program is executed.
  15. 一种存储介质,用于可读存储,所述存储介质存储有一个或者多个程序,所述一个或者多个程序可被一个或者多个处理器执行,以实现如权利要求1至13任一项所述的行为识别方法。A storage medium for readable storage, the storage medium stores one or more programs, and the one or more programs can be executed by one or more processors to realize any one of claims 1 to 13 The behavior recognition method described in item.
PCT/CN2021/100379 2020-08-31 2021-06-16 Behavior recognition method and apparatus, and storage medium WO2022041953A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN202010901631.6 2020-08-31
CN202010901631.6A CN114202797A (en) 2020-08-31 2020-08-31 Behavior recognition method, behavior recognition device and storage medium

Publications (1)

Publication Number Publication Date
WO2022041953A1 true WO2022041953A1 (en) 2022-03-03

Family

ID=80354540

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2021/100379 WO2022041953A1 (en) 2020-08-31 2021-06-16 Behavior recognition method and apparatus, and storage medium

Country Status (2)

Country Link
CN (1) CN114202797A (en)
WO (1) WO2022041953A1 (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114919472A (en) * 2022-06-30 2022-08-19 苏州浪潮智能科技有限公司 Method, device, equipment and medium for acquiring sitting posture height of vehicle driving user

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114783059B (en) * 2022-04-20 2022-10-25 浙江东昊信息工程有限公司 Temple incense and worship participation management method and system based on depth camera

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20180035082A1 (en) * 2016-07-28 2018-02-01 Chigru Innovations (OPC) Private Limited Infant monitoring system
CN110245623A (en) * 2019-06-18 2019-09-17 重庆大学 A kind of real time human movement posture correcting method and system
CN111008617A (en) * 2020-01-01 2020-04-14 王淞 Desk lamp with intelligent detection function

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20180035082A1 (en) * 2016-07-28 2018-02-01 Chigru Innovations (OPC) Private Limited Infant monitoring system
CN110245623A (en) * 2019-06-18 2019-09-17 重庆大学 A kind of real time human movement posture correcting method and system
CN111008617A (en) * 2020-01-01 2020-04-14 王淞 Desk lamp with intelligent detection function

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114919472A (en) * 2022-06-30 2022-08-19 苏州浪潮智能科技有限公司 Method, device, equipment and medium for acquiring sitting posture height of vehicle driving user

Also Published As

Publication number Publication date
CN114202797A (en) 2022-03-18

Similar Documents

Publication Publication Date Title
US10417775B2 (en) Method for implementing human skeleton tracking system based on depth data
KR101894686B1 (en) Imaging a body
JP6125188B2 (en) Video processing method and apparatus
CN104380338B (en) Information processor and information processing method
WO2022041953A1 (en) Behavior recognition method and apparatus, and storage medium
US9443325B2 (en) Image processing apparatus, image processing method, and computer program
JP2019121045A (en) Posture estimation system, behavior estimation system, and posture estimation program
CN109271918A (en) The method for distinguishing balanced capacity obstacle crowd based on centre-of gravity shift model
US11182636B2 (en) Method and computing device for adjusting region of interest
JP2018088049A (en) Device, method and program for image processing
JP7173341B2 (en) Human state detection device, human state detection method and program
WO2022143264A1 (en) Face orientation estimation method and apparatus, electronic device, and storage medium
US11995840B2 (en) Anthropometric data portable acquisition device and method of collecting anthropometric data
US10229489B1 (en) Medical environment monitoring system
JP7197011B2 (en) Height estimation device, height estimation method and program
EP3699865B1 (en) Three-dimensional face shape derivation device, three-dimensional face shape deriving method, and non-transitory computer readable medium
JP7136344B2 (en) Camera calibration method, camera and program
CN114758354A (en) Sitting posture detection method and device, electronic equipment, storage medium and program product
CN114463663A (en) Method and device for calculating height of person, electronic equipment and storage medium
US20240087353A1 (en) Image processing apparatus, image processing method, and non-transitory computer readable medium storing image processing program
KR20130081126A (en) Method for hand-gesture recognition and apparatus thereof
CN110415486B (en) Attitude determination method, electronic system and non-transitory computer readable recording medium
CN112767415A (en) Chest scanning area automatic determination method, device, equipment and storage medium
JP2022123391A (en) Information processing apparatus, information processing method, and program
WO2023152841A1 (en) Image processing system, image processing method, and non-transitory computer-readable medium

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 21859797

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

32PN Ep: public notification in the ep bulletin as address of the adressee cannot be established

Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205A DATED 11.07.2023)

122 Ep: pct application non-entry in european phase

Ref document number: 21859797

Country of ref document: EP

Kind code of ref document: A1