WO2020107847A1 - Bone point-based fall detection method and fall detection device therefor - Google Patents

Bone point-based fall detection method and fall detection device therefor Download PDF

Info

Publication number
WO2020107847A1
WO2020107847A1 PCT/CN2019/089500 CN2019089500W WO2020107847A1 WO 2020107847 A1 WO2020107847 A1 WO 2020107847A1 CN 2019089500 W CN2019089500 W CN 2019089500W WO 2020107847 A1 WO2020107847 A1 WO 2020107847A1
Authority
WO
WIPO (PCT)
Prior art keywords
feature
neural network
points
layer
behavior
Prior art date
Application number
PCT/CN2019/089500
Other languages
French (fr)
Chinese (zh)
Inventor
周涛涛
周宝
陈远旭
肖京
Original Assignee
平安科技(深圳)有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 平安科技(深圳)有限公司 filed Critical 平安科技(深圳)有限公司
Publication of WO2020107847A1 publication Critical patent/WO2020107847A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/20Movements or behaviour, e.g. gesture recognition
    • G06V40/23Recognition of whole body movements, e.g. for sport training
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/103Detecting, measuring or recording devices for testing the shape, pattern, colour, size or movement of the body or parts thereof, for diagnostic purposes
    • A61B5/11Measuring movement of the entire body or parts thereof, e.g. head or hand tremor, mobility of a limb
    • A61B5/1116Determining posture transitions
    • A61B5/1117Fall detection
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/103Detecting, measuring or recording devices for testing the shape, pattern, colour, size or movement of the body or parts thereof, for diagnostic purposes
    • A61B5/11Measuring movement of the entire body or parts thereof, e.g. head or hand tremor, mobility of a limb
    • A61B5/1126Measuring movement of the entire body or parts thereof, e.g. head or hand tremor, mobility of a limb using a particular sensing technique
    • A61B5/1128Measuring movement of the entire body or parts thereof, e.g. head or hand tremor, mobility of a limb using a particular sensing technique using image analysis
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques

Definitions

  • the present application relates to the field of machine vision deep learning technology, and in particular to a fall detection method, device, computer equipment, and storage medium based on bone points.
  • fall detection based on wearable devices fall detection based on depth cameras and fall detection based on ordinary cameras.
  • the method based on wearable devices must be carried at all times, which causes great inconvenience to users and has little practical application value; the method based on depth cameras is expensive and difficult to promote in practice; and the method based on ordinary cameras is cheap and easy to use Convenient, but requires higher algorithm.
  • the purpose of this application is to provide a fall detection method, device, computer equipment and storage medium based on bone points, which are used to solve the problems in the prior art.
  • the present application provides a bone point-based fall detection method, including the following steps:
  • a first feature extraction neural network through a first picture sample, the first feature extraction neural network is used to extract a plurality of first feature points in the first picture sample, the first feature points represent key points on the human body Bone point
  • Input a second video sample into the trained first feature extraction neural network to obtain a plurality of second feature points characterizing key bone points of the human body in the second video sample;
  • the video data of the monitored object is sequentially input into the trained first feature extraction neural network and the second behavior classification neural network to output the behavior category of the monitored object.
  • the present application also proposes a fall detection device based on bone points, including:
  • the first neural network training module is adapted to train a first feature extraction neural network through a first picture sample, the first feature extraction neural network is used to extract multiple first feature points in the first picture sample, the The first feature point represents the key bone point on the human body;
  • the feature point extraction module is adapted to input the second video sample into the trained first feature extraction neural network to obtain a plurality of second feature points characterizing key bone points of the human body in the second video sample;
  • a feature map generation module adapted to encode the multiple second feature points to generate a predicted feature map characterizing the distribution of the multiple second feature points
  • a second neural network training module adapted to train a second behavior classification neural network through the predicted feature map, and the second behavior classification neural network is used to classify the behavior represented in the predicted feature map;
  • the classification module is adapted to sequentially input the video data of the monitored object into the trained first feature extraction neural network and the second behavior classification neural network to output the behavior category of the monitored object.
  • the present application also provides a computer device, including a memory, a processor, and a computer program stored on the memory and executable on the processor.
  • a computer program stored on the memory and executable on the processor.
  • a first feature extraction neural network through a first picture sample, the first feature extraction neural network is used to extract a plurality of first feature points in the first picture sample, the first feature points represent key points on the human body Bone point
  • Input a second video sample into the trained first feature extraction neural network to obtain a plurality of second feature points characterizing key bone points of the human body in the second video sample;
  • the video data of the monitored object is sequentially input into the trained first feature extraction neural network and the second behavior classification neural network to output the behavior category of the monitored object.
  • the present application also provides a computer-readable storage medium on which a computer program is stored, and when the computer program is executed by the processor, the following steps are realized:
  • a first feature extraction neural network through a first picture sample, the first feature extraction neural network is used to extract a plurality of first feature points in the first picture sample, the first feature points represent key points on the human body Bone point
  • Input a second video sample into the trained first feature extraction neural network to obtain a plurality of second feature points characterizing key bone points of the human body in the second video sample;
  • the video data of the monitored object is sequentially input into the trained first feature extraction neural network and the second behavior classification neural network to output the behavior category of the monitored object.
  • This application addresses the problem of insufficient fall detection data in the prior art, and uses other data to train human bone point feature extraction neural networks; for the problem of using frame information to detect fall behavior, use bone point information to classify fall behavior.
  • This application trains the first feature extraction neural network through the image sample library to extract the key bone point information in the human body; trains the second behavior classification neural network through the video sample library, and judges the video based on the extracted key bone point information Whether the human movement in is a fall behavior.
  • the bone point information of the monitored object can be accurately extracted, and according to the bone point information, it can be judged in time whether the monitored object has fallen down.
  • the provision of timely and effective care for the handicapped elderly and disabled persons is conducive to improving people's quality of life.
  • FIG. 1 is a flowchart of Embodiment 1 of a fall detection method based on bone points of the present application
  • FIG. 2 is a schematic structural diagram of a first feature extraction neural network in Embodiment 1 of the present application.
  • FIG. 3 is a schematic structural diagram of a second behavior classification neural network in Embodiment 1 of the present application.
  • FIG. 4 is a schematic diagram of a program module of a first embodiment of a fall detection device based on a bone point according to this application;
  • FIG. 5 is a schematic diagram of the hardware structure of the first embodiment of the memory sharing device of the present application.
  • the fall detection method, device, computer equipment and storage medium provided by the present application are applicable to the field of machine vision technology, and provide a fall detection method and device for the elderly or disabled persons living alone to detect fall behavior in time.
  • This application trains the first feature extraction neural network through the image sample library to extract the key bone point information in the human body; trains the second behavior classification neural network through the video sample library, and judges the video based on the extracted key bone point information Whether the human movement in is a fall behavior.
  • the first feature extraction neural network and the second behavior classification neural network trained by this application can accurately extract the bone point information of the monitored object, and timely determine whether the monitored object has fallen down according to the bone point information, which is beneficial to Greatly improve people's quality of life.
  • a fall detection method based on bone points in this embodiment includes the following steps:
  • the first picture sample is selected from the picture sample library to train the first feature extraction neural network.
  • the first picture sample is preferably a full-body picture of the person.
  • the first image sample is divided into a training image sample and a test image sample, where the training image sample is used to train the first feature extraction neural network, and the test image sample is used to verify the first feature after the training image sample training The effect of extracting the neural network when extracting the feature information in the picture.
  • the above training picture samples and test picture samples may be subjected to data enhancement preprocessing, such as performing contrast transformation and brightness transformation on each sample, adding local random Gaussian noise, and performing uniform normalization processing, thereby The training image samples and the test image samples after data enhancement are obtained.
  • the structure of the first feature extraction neural network in this step will be described in detail below with a test picture sample as an example, as shown in FIG. 2.
  • the test picture sample first enters the feature extraction module to extract the features in the test picture sample.
  • the feature extraction module in this embodiment uses a ResNet residual network to ensure better feature extraction performance.
  • the test sample image passes through the ResNet residual network
  • the first extracted data D 1 is obtained , and then the first extracted data D 1 enters four convolution modules with different expansion coefficients respectively, to obtain four second extracted data D with different feature channels. 2 .
  • the four second extracted data D 2 with different feature channels are combined into the first convolutional layer stacked by the residual module to obtain four third extracted data D 3 with different perceptual fields.
  • it after fusing four third extracted data D 3 with different perceptual fields, it enters the second convolutional layer piled up by the residual module again, and finally outputs multiple first feature points representing key bone points on the human body .
  • the convolution module includes the following layers in sequence: a convolution layer, a batch normalization layer, a Relu activation function layer, a convolution layer, a batch normalization layer, a Relu activation function layer, and a pooling layer, each The convolutional layers have different expansion coefficients.
  • the feature information is the bone feature points on the human body, including the feature points at the main joints of the body, such as the elbow joint, shoulder joint, knee joint, hip joint, etc.
  • the target feature point associated with the preset behavior can be further selected from the bone feature points.
  • the preset behavior may be squatting, bending over, standing up, falling, etc.
  • the characteristic points of displacement in different behaviors may be different, so the one that best reflects the characteristics of this behavior can be selected according to the behavior to be detected Target feature point.
  • a total of 14 bone point information including head, neck, shoulders, elbows, hands, hips, knees, and feet are selected as targets Feature points.
  • the selection method of the present invention can make the number of bone feature points as small as possible to reduce the calculation amount in the subsequent behavior analysis process; on the other hand, the above-mentioned 14 target feature points are evenly distributed at major joints of the human body , Can reflect the basic trend of human behavior as a whole.
  • the positions of the bone points listed above are only used as examples, and are not used to limit specific feature point information.
  • the above bone point information may also be deleted or added, or specific feature points may be changed
  • the location of the acupuncture point in the human body can also be obtained. This application does not limit this.
  • the plurality of first feature points in this embodiment may preferably be the above bone point distribution information maps marked in the human body.
  • x p and y p represent the predicted coordinates of the first feature point extracted by the first feature extraction neural network
  • x g and y g represent the actual coordinates of the first feature point
  • S2 Input the second video sample into the trained first feature extraction neural network to obtain a plurality of second feature points that represent key bone points of the human body in the second video sample.
  • this step uses the trained first feature extraction neural network to extract the second feature point in the video sample.
  • the second feature point is the above The 14 bone feature points mentioned.
  • the present application is based on the video information of the monitored person collected by a common camera for fall detection. Therefore, the object of feature point extraction in this step is a continuous video rather than a simple picture. Since the video is formed by a series of picture frames changing with time, the video needs to be sampled first to extract the target picture. For example, the video is extracted according to the standard of 20 frames per second, with 3 seconds as a sample. At the same time, in order to generate diverse samples, the starting frame can be randomly selected near the starting point of the behavior in the video.
  • the feature point information in the target pictures can be extracted through the first feature extraction neural network, preferably the 14 bone feature points mentioned above.
  • S3 Encoding the plurality of second feature points to generate a predicted feature map.
  • This step is used to process the extracted second feature points to obtain a predicted feature map. Taking the above 14 bone feature points as an example, the following processing steps are included:
  • any two feature points from the 14 bone feature points are paired, and the calculation formula is as follows:
  • x it and y it respectively represent the horizontal and vertical coordinates of the i-th second feature point at time t; l xjt represents the Euler of the i-th second feature point and j-th second feature point at time t distance, v xit represents the i-th second feature points at time t in the x-direction velocity, v yit speed of the i-th representative of a second characteristic point in the y-direction.
  • the matrix diagram is the prediction feature diagram.
  • the purpose of this step is to train a second behavior classification neural network to classify the behavior represented in the prediction feature map to determine whether a fall behavior has occurred.
  • the structure of the second behavior classification neural network in this application is shown in FIG. 3, which will be described in detail below.
  • the prediction feature map first passes through a conventional convolution module to obtain first classification data R1. Then, the first classification data R1 respectively passes through four convolution modules with different expansion coefficients to obtain four second classification data R2 with different characteristic channels. Preferably, the expansion coefficients of the above four convolution modules are 1 respectively. , 3, 6 and 12. Next, the above-mentioned four second classification data R2 with different feature channels are combined and then sequentially passed through three conventional convolution modules, and finally output behavior classification, which is used to judge which behavior category the behavior represented in the above-mentioned predicted feature map belongs to.
  • the convolution module includes the following layers in sequence: a convolution layer, a batch normalization layer, a Relu activation function layer, a convolution layer, a batch normalization layer, a Relu activation function layer, and a pooling layer.
  • the second behavior classification neural network is trained by the loss function L H (X, Y), the specific expression is as follows:
  • x k represents the parameter value of the kth behavior category
  • z k represents the predicted probability of the kth behavior category.
  • the second behavior classification neural network can recognize the categories of squatting, standing, waving, bending, falling, lying down, etc., each behavior corresponds to its own parameter value, such as When the monitored person is falling, then x k represents the parameter value of the monitored person's falling behavior, and z k represents the predicted probability of the monitored person's falling behavior.
  • this embodiment adds an L2 regular term after the loss function to prevent overfitting.
  • the resulting cost function is as follows:
  • S5 Input the video data of the monitored object into the trained first feature extraction neural network and the second behavior classification neural network in order to output the behavior category of the monitored object.
  • the present application can detect the fall behavior of the actual surveillance video.
  • the video information of the monitored object is photographed in real time through a common camera, and the video information is sampled to extract a certain number of target images.
  • the target image first undergoes a trained first feature extraction neural network to extract multiple feature points in the target image, such as bone feature points.
  • Calculate and combine multiple bone feature points for example, calculate the Euler distance between each two bone feature points and the speed in the x and y directions, and arrange the vectors calculated above in the order of each frame of image , And finally get the predicted feature map.
  • you can obtain the category to which the behavior included in the prediction feature map belongs, for example, whether it is a fall behavior.
  • the fall detection device 10 may include or be divided into one or more program modules, and one or more program modules are stored in a storage medium. , And executed by one or more processors to complete this application, and can implement the above-mentioned fall detection method.
  • the program module referred to in this application refers to a series of computer program instruction segments capable of performing specific functions, and is more suitable for describing the execution process of the fall detection device 10 in the storage medium than the program itself. The following description will specifically introduce the functions of the program modules of this embodiment:
  • the first neural network training module 11 is adapted to train a first feature extraction neural network through a first picture sample.
  • the first feature extraction neural network is used to extract multiple first feature points in the first picture sample.
  • the first feature point represents the key bone point on the human body;
  • the feature point extraction module 12 is adapted to input the second video sample into the trained first feature extraction neural network to obtain a plurality of second feature points characterizing key bone points of the human body in the second video sample;
  • the feature map generation module 13 is adapted to encode the multiple second feature points to generate a predicted feature map characterizing the distribution of the multiple second feature points;
  • the second neural network training module 14 is adapted to train a second behavior classification neural network through the predicted feature map, and the second behavior classification neural network is used to classify the behavior represented in the predicted feature map;
  • the classification module 15 is adapted to sequentially input the video data of the monitored object into the trained first feature extraction neural network and the second behavior classification neural network to output the behavior category of the monitored object.
  • This embodiment also provides a computer device, such as a smartphone, a tablet computer, a notebook computer, a desktop computer, a rack server, a blade server, a tower server, or a rack server (including an independent server, or A server cluster composed of multiple servers), etc.
  • the computer device 20 of this embodiment includes at least but not limited to: a memory 21 and a processor 22 that can be communicatively connected to each other through a system bus, as shown in FIG. 5. It should be noted that FIG. 5 only shows the computer device 20 having components 21-22, but it should be understood that it is not required to implement all the components shown, and that more or fewer components may be implemented instead.
  • the memory 21 (ie, readable storage medium) includes flash memory, hard disk, multimedia card, card-type memory (for example, SD or DX memory, etc.), random access memory (RAM), static random access memory (SRAM), Read only memory (ROM), electrically erasable programmable read only memory (EEPROM), programmable read only memory (PROM), magnetic memory, magnetic disk, optical disk, etc.
  • the memory 21 may be an internal storage unit of the computer device 20, such as a hard disk or memory of the computer device 20.
  • the memory 21 may also be an external storage device of the computer device 20, such as a plug-in hard disk equipped on the computer device 20, a smart memory card (Smart Media, Card, SMC), and secure digital (Secure Digital, SD) card, flash card (Flash Card), etc.
  • the memory 21 may also include both the internal storage unit of the computer device 20 and its external storage device.
  • the memory 21 is generally used to store the operating system and various application software installed in the computer device 20, such as the program code of the fall detection device 10 of the first embodiment.
  • the memory 21 can also be used to temporarily store various types of data that have been output or will be output.
  • the processor 22 may be a central processing unit (CPU), controller, microcontroller, microprocessor, or other data processing chip in some embodiments.
  • the processor 22 is generally used to control the overall operation of the computer device 20.
  • the processor 22 is used to run the program code or process data stored in the memory 21, for example, to run the fall detection device 10, so as to implement the fall detection method of the first embodiment.
  • This embodiment also provides a computer-readable storage medium, such as flash memory, hard disk, multimedia card, card-type memory (for example, SD or DX memory, etc.), random access memory (RAM), static random access memory (SRAM), only Read memory (ROM), electrically erasable programmable read-only memory (EEPROM), programmable read-only memory (PROM), magnetic memory, magnetic disk, optical disk, server, App store, etc., which store computer programs, When the program is executed by the processor, the corresponding function is realized.
  • the computer-readable storage medium of this embodiment is used to store the fall detection device 10, and when executed by the processor, the fall detection method of the first embodiment is implemented.
  • Any process or method description in a flowchart or otherwise described herein may be understood to represent a module, segment, or portion of code that includes one or more executable instructions for implementing specific logical functions or steps of a process , And the scope of the preferred embodiment of the present application includes additional implementations, in which the functions may not be performed in the order shown or discussed, including performing functions in a substantially simultaneous manner or in reverse order according to the functions involved, which shall It is understood by those skilled in the art to which the embodiments of the present application belong.

Landscapes

  • Engineering & Computer Science (AREA)
  • Health & Medical Sciences (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • General Health & Medical Sciences (AREA)
  • General Physics & Mathematics (AREA)
  • Biomedical Technology (AREA)
  • Surgery (AREA)
  • Evolutionary Computation (AREA)
  • Evolutionary Biology (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Veterinary Medicine (AREA)
  • Public Health (AREA)
  • Animal Behavior & Ethology (AREA)
  • General Engineering & Computer Science (AREA)
  • Physiology (AREA)
  • Dentistry (AREA)
  • Oral & Maxillofacial Surgery (AREA)
  • Biophysics (AREA)
  • Pathology (AREA)
  • Artificial Intelligence (AREA)
  • Heart & Thoracic Surgery (AREA)
  • Medical Informatics (AREA)
  • Molecular Biology (AREA)
  • Multimedia (AREA)
  • Human Computer Interaction (AREA)
  • Social Psychology (AREA)
  • Psychiatry (AREA)
  • Nuclear Medicine, Radiotherapy & Molecular Imaging (AREA)
  • Radiology & Medical Imaging (AREA)
  • Image Analysis (AREA)

Abstract

The present application provides a bone point-based fall detection method and a fall detection device therefor. The method comprises: training a first feature extraction neural network by means of a first picture sample, the first feature extraction neural network being used for extracting a plurality of first feature points representing key bone points of a human body; inputting a second video sample to the trained first feature extraction neural network to obtain a plurality of second feature points representing the key bone points of the human body in the second video sample; encoding the plurality of second feature points to generate a prediction feature map; training a second behavior classification neural network by means of the prediction feature map, the second behavior classification neural network being used for classifying behaviors represented in the prediction feature map; and sequentially inputting video data of a monitored object into the trained first feature extraction neural network and the trained second behavior classification neural network to output a behavior category of the monitored object.

Description

基于骨骼点的跌倒检测方法及其跌倒检测装置Skeletal point-based fall detection method and fall detection device
相关申请的交叉引用Cross-reference of related applications
本申请申明享有2018年11月28日递交的申请号为CN201811433808.3、名称为“基于骨骼点的跌倒检测方法及其跌倒检测装置”的中国专利申请的优先权,该中国专利申请的整体内容以参考的方式结合在本申请中。This application declares to enjoy the priority of the Chinese patent application filed on November 28, 2018 with the application number CN201811433808.3 and titled "skeletal point-based fall detection method and fall detection device". The entire content of the Chinese patent application Incorporated by reference in this application.
技术领域Technical field
本申请涉及机器视觉深度学习技术领域,尤其涉及一种基于骨骼点的跌倒检测方法、装置、计算机设备及存储介质。The present application relates to the field of machine vision deep learning technology, and in particular to a fall detection method, device, computer equipment, and storage medium based on bone points.
背景技术Background technique
随着我国进入老龄化社会,养老问题日趋严峻。老年人的各项身体机能指标下降,活动能力降低,特别是平衡力、反应能力和协同能力的不足可能造成意外跌倒情况发生。当老人发生跌倒后,如果没有获得及时的援助甚至可能因此在家中身亡。因此,家庭或者其他环境中针对老人的跌倒检测是计算机视觉和机器学习领域中一个很有意义的研究问题。As my country enters an aging society, the problem of old-age care is getting more and more serious. The elderly's various physical function indexes decline, and their activity ability decreases, especially the lack of balance, reaction ability and coordination ability may cause accidental falls. When the old man falls, he may even die at home if he does not receive timely assistance. Therefore, the fall detection for the elderly in the family or other environments is a very meaningful research problem in the field of computer vision and machine learning.
目前现有的跌倒检测主要有三种方法,分别为基于穿戴式设备的跌倒检测、基于深度摄像头的跌倒检测和基于普通摄像头的跌倒检测。其中基于穿戴式设备的方法必须时刻携带,给使用者带来很大不便,实际应用价值不大;基于深度摄像头的方法由于成本昂贵,实际推广难度大;而基于普通摄像头的方法成本便宜、使用方便,但对算法的要求较高。At present, there are three main methods of fall detection, which are fall detection based on wearable devices, fall detection based on depth cameras and fall detection based on ordinary cameras. Among them, the method based on wearable devices must be carried at all times, which causes great inconvenience to users and has little practical application value; the method based on depth cameras is expensive and difficult to promote in practice; and the method based on ordinary cameras is cheap and easy to use Convenient, but requires higher algorithm.
由于普通摄像头能够覆盖各个地方,因此其硬件基础是成熟的。目前业内利用普通摄像头进行跌倒检测已经提出了很多方法。例如,直接利用图像序列的信息对跌倒行为进行分类,利用检测算法对人物的边框变化进行分类。但是目前跌倒检测的数据较少,场景单一,不能应用在各种实际场景中。对于利用 图像序列的分类方法,由于数据少,不能够训练出优秀的网络。对于利用检测算法对人物边框分类的方法,利用了大量其他数据集的信息,能够有效检测到人,但是在利用边框信息分类的时候,由于边框信息有限,不能得到泛化性好的网络。Since ordinary cameras can cover all places, their hardware foundation is mature. At present, many methods have been proposed in the industry for using ordinary cameras for fall detection. For example, the information of the image sequence is directly used to classify the fall behavior, and the detection algorithm is used to classify the change of the person's frame. However, at present, there are few data for fall detection, and the scene is single, so it cannot be applied to various actual scenes. For the classification method using image sequences, due to the lack of data, it is not possible to train an excellent network. For the method of using the detection algorithm to classify the person's frame, using a large amount of information from other data sets, people can be effectively detected, but when using the frame information to classify, because the frame information is limited, a network with good generalization cannot be obtained.
发明内容Summary of the invention
本申请的目的是提供一种基于骨骼点的跌倒检测方法、装置、计算机设备及存储介质,用于解决现有技术存在的问题。The purpose of this application is to provide a fall detection method, device, computer equipment and storage medium based on bone points, which are used to solve the problems in the prior art.
为实现上述目的,本申请提供一种基于骨骼点的跌倒检测方法,包括以下步骤:To achieve the above objective, the present application provides a bone point-based fall detection method, including the following steps:
通过第一图片样本训练第一特征提取神经网络,所述第一特征提取神经网络用于提取所述第一图片样本中的多个第一特征点,所述第一特征点表征人体上的关键骨骼点;Train a first feature extraction neural network through a first picture sample, the first feature extraction neural network is used to extract a plurality of first feature points in the first picture sample, the first feature points represent key points on the human body Bone point
将第二视频样本输入训练好的所述第一特征提取神经网络,得到表征所述第二视频样本中的人体的关键骨骼点的多个第二特征点;Input a second video sample into the trained first feature extraction neural network to obtain a plurality of second feature points characterizing key bone points of the human body in the second video sample;
对所述多个第二特征点进行编码生成预测特征图;Encoding the plurality of second feature points to generate a predicted feature map;
通过所述预测特征图训练第二行为分类神经网络,所述第二行为分类神经网络用于对所述预测特征图中表示的行为进行分类;Training a second behavior classification neural network through the predicted feature map, the second behavior classification neural network is used to classify the behavior represented in the predicted feature map;
将被监测对象的视频数据依次输入训练好的所述第一特征提取神经网络和所述第二行为分类神经网络,以输出所述被监测对象的行为类别。The video data of the monitored object is sequentially input into the trained first feature extraction neural network and the second behavior classification neural network to output the behavior category of the monitored object.
为实现上述目的,本申请还提出一种基于骨骼点的跌倒检测装置,包括:To achieve the above purpose, the present application also proposes a fall detection device based on bone points, including:
第一神经网络训练模块,适用于通过第一图片样本训练第一特征提取神经网络,所述第一特征提取神经网络用于提取所述第一图片样本中的多个第一特征点,所述第一特征点表征人体上的关键骨骼点;The first neural network training module is adapted to train a first feature extraction neural network through a first picture sample, the first feature extraction neural network is used to extract multiple first feature points in the first picture sample, the The first feature point represents the key bone point on the human body;
特征点提取模块,适用于将第二视频样本输入训练好的所述第一特征提取神经网络,得到表征所述第二视频样本中的人体的关键骨骼点的多个第二特征 点;The feature point extraction module is adapted to input the second video sample into the trained first feature extraction neural network to obtain a plurality of second feature points characterizing key bone points of the human body in the second video sample;
特征图生成模块,适用于对所述多个第二特征点进行编码生成表征所述多个第二特征点的分布情况的预测特征图;A feature map generation module, adapted to encode the multiple second feature points to generate a predicted feature map characterizing the distribution of the multiple second feature points;
第二神经网络训练模块,适用于通过所述预测特征图训练第二行为分类神经网络,所述第二行为分类神经网络用于对所述预测特征图中表示的行为进行分类;A second neural network training module, adapted to train a second behavior classification neural network through the predicted feature map, and the second behavior classification neural network is used to classify the behavior represented in the predicted feature map;
分类模块,适用于将被监测对象的视频数据依次输入训练好的所述第一特征提取神经网络和所述第二行为分类神经网络,以输出所述被监测对象的行为类别。The classification module is adapted to sequentially input the video data of the monitored object into the trained first feature extraction neural network and the second behavior classification neural network to output the behavior category of the monitored object.
为实现上述目的,本申请还提供一种计算机设备,包括存储器、处理器以及存储在存储器上并可在处理器上运行的计算机程序,所述处理器执行所述计算机程序时实现以下步骤:In order to achieve the above object, the present application also provides a computer device, including a memory, a processor, and a computer program stored on the memory and executable on the processor. When the processor executes the computer program, the following steps are implemented:
通过第一图片样本训练第一特征提取神经网络,所述第一特征提取神经网络用于提取所述第一图片样本中的多个第一特征点,所述第一特征点表征人体上的关键骨骼点;Train a first feature extraction neural network through a first picture sample, the first feature extraction neural network is used to extract a plurality of first feature points in the first picture sample, the first feature points represent key points on the human body Bone point
将第二视频样本输入训练好的所述第一特征提取神经网络,得到表征所述第二视频样本中的人体的关键骨骼点的多个第二特征点;Input a second video sample into the trained first feature extraction neural network to obtain a plurality of second feature points characterizing key bone points of the human body in the second video sample;
对所述多个第二特征点进行编码生成预测特征图;Encoding the plurality of second feature points to generate a predicted feature map;
通过所述预测特征图训练第二行为分类神经网络,所述第二行为分类神经网络用于对所述预测特征图中表示的行为进行分类;Training a second behavior classification neural network through the predicted feature map, the second behavior classification neural network is used to classify the behavior represented in the predicted feature map;
将被监测对象的视频数据依次输入训练好的所述第一特征提取神经网络和所述第二行为分类神经网络,以输出所述被监测对象的行为类别。The video data of the monitored object is sequentially input into the trained first feature extraction neural network and the second behavior classification neural network to output the behavior category of the monitored object.
为实现上述目的,本申请还提供计算机可读存储介质,其上存储有计算机程序,所述计算机程序被处理器执行时实现以下步骤:To achieve the above purpose, the present application also provides a computer-readable storage medium on which a computer program is stored, and when the computer program is executed by the processor, the following steps are realized:
通过第一图片样本训练第一特征提取神经网络,所述第一特征提取神经网 络用于提取所述第一图片样本中的多个第一特征点,所述第一特征点表征人体上的关键骨骼点;Train a first feature extraction neural network through a first picture sample, the first feature extraction neural network is used to extract a plurality of first feature points in the first picture sample, the first feature points represent key points on the human body Bone point
将第二视频样本输入训练好的所述第一特征提取神经网络,得到表征所述第二视频样本中的人体的关键骨骼点的多个第二特征点;Input a second video sample into the trained first feature extraction neural network to obtain a plurality of second feature points characterizing key bone points of the human body in the second video sample;
对所述多个第二特征点进行编码生成预测特征图;Encoding the plurality of second feature points to generate a predicted feature map;
通过所述预测特征图训练第二行为分类神经网络,所述第二行为分类神经网络用于对所述预测特征图中表示的行为进行分类;Training a second behavior classification neural network through the predicted feature map, the second behavior classification neural network is used to classify the behavior represented in the predicted feature map;
将被监测对象的视频数据依次输入训练好的所述第一特征提取神经网络和所述第二行为分类神经网络,以输出所述被监测对象的行为类别。The video data of the monitored object is sequentially input into the trained first feature extraction neural network and the second behavior classification neural network to output the behavior category of the monitored object.
本申请针对现有技术中跌倒检测数据不足的问题,利用其他数据训练人体骨骼点特征提取神经网络;针对利用边框信息不足以检测出跌倒行为的问题,利用骨骼点信息对跌倒行为进行分类。本申请通过图片样本库训练第一特征提取神经网络,用于提取人体中的关键骨骼点信息;通过视频样本库训练第二行为分类神经网络,在已提取关键骨骼点信息的基础上,判断视频中的人体动作是否属于跌倒行为。通过本申请训练出的第一特征提取神经网络和第二行为分类神经网络,能够准确地提取被监测对象的骨骼点信息,并根据骨骼点信息及时判断被监测对象是否发生了跌倒行为,能够为行动不便的老人、伤残人士等提供及时有效地看护,有利于提高人们的生活质量。This application addresses the problem of insufficient fall detection data in the prior art, and uses other data to train human bone point feature extraction neural networks; for the problem of using frame information to detect fall behavior, use bone point information to classify fall behavior. This application trains the first feature extraction neural network through the image sample library to extract the key bone point information in the human body; trains the second behavior classification neural network through the video sample library, and judges the video based on the extracted key bone point information Whether the human movement in is a fall behavior. Through the first feature extraction neural network and the second behavior classification neural network trained by this application, the bone point information of the monitored object can be accurately extracted, and according to the bone point information, it can be judged in time whether the monitored object has fallen down. The provision of timely and effective care for the handicapped elderly and disabled persons is conducive to improving people's quality of life.
附图说明BRIEF DESCRIPTION
图1为本申请基于骨骼点的跌倒检测方法实施例一的流程图;FIG. 1 is a flowchart of Embodiment 1 of a fall detection method based on bone points of the present application;
图2为本申请实施例一中的第一特征提取神经网络的结构示意图;2 is a schematic structural diagram of a first feature extraction neural network in Embodiment 1 of the present application;
图3为本申请实施例一中的第二行为分类神经网络的结构示意图;3 is a schematic structural diagram of a second behavior classification neural network in Embodiment 1 of the present application;
图4为本申请基于骨骼点的跌倒检测装置实施例一的程序模块示意图;4 is a schematic diagram of a program module of a first embodiment of a fall detection device based on a bone point according to this application;
图5为本申请内存共享装置实施例一的硬件结构示意图。5 is a schematic diagram of the hardware structure of the first embodiment of the memory sharing device of the present application.
具体实施方式detailed description
为了使本申请的目的、技术方案及优点更加清楚明白,以下结合附图及实施例,对本申请进行进一步详细说明。应当理解,此处所描述的具体实施例仅用以解释本申请,并不用于限定本申请。基于本申请中的实施例,本领域普通技术人员在没有做出创造性劳动前提下所获得的所有其他实施例,都属于本申请保护的范围。In order to make the purpose, technical solutions and advantages of the present application more clear, the present application will be described in further detail in conjunction with the accompanying drawings and embodiments. It should be understood that the specific embodiments described herein are only used to explain the present application, and are not used to limit the present application. Based on the embodiments in the present application, all other embodiments obtained by a person of ordinary skill in the art without creative work fall within the protection scope of the present application.
本申请提供的跌倒检测方法、装置、计算机设备及存储介质,适用于机器视觉技术领域,为独居情况下的老人或伤残人士等提供一种可及时发现跌倒行为的跌倒检测方法及其装置。本申请通过图片样本库训练第一特征提取神经网络,用于提取人体中的关键骨骼点信息;通过视频样本库训练第二行为分类神经网络,在已提取关键骨骼点信息的基础上,判断视频中的人体动作是否属于跌倒行为。通过本申请训练出的第一特征提取神经网络和第二行为分类神经网络,能够准确地提取被监测对象的骨骼点信息,并根据骨骼点信息及时判断被监测对象是否发生了跌倒行为,有利于极大提高人们的生活质量。The fall detection method, device, computer equipment and storage medium provided by the present application are applicable to the field of machine vision technology, and provide a fall detection method and device for the elderly or disabled persons living alone to detect fall behavior in time. This application trains the first feature extraction neural network through the image sample library to extract the key bone point information in the human body; trains the second behavior classification neural network through the video sample library, and judges the video based on the extracted key bone point information Whether the human movement in is a fall behavior. The first feature extraction neural network and the second behavior classification neural network trained by this application can accurately extract the bone point information of the monitored object, and timely determine whether the monitored object has fallen down according to the bone point information, which is beneficial to Greatly improve people's quality of life.
实施例1Example 1
请参阅图1,本实施例的一种基于骨骼点的跌倒检测方法,包括以下步骤:Referring to FIG. 1, a fall detection method based on bone points in this embodiment includes the following steps:
S1:通过第一图片样本训练第一特征提取神经网络,所述第一特征提取神经网络用于提取所述第一图片样本中的多个第一特征点,所述第一特征点表征人体上的关键骨骼点。S1: Train a first feature extraction neural network through a first picture sample, the first feature extraction neural network is used to extract a plurality of first feature points in the first picture sample, and the first feature points represent the human body Key bone points.
本步骤中,从图片样本库中选择第一图片样本来训练第一特征提取神经网络,第一图片样本优选为人物的全身图片。实施时,将第一图片样本分为训练图片样本和测试图片样本,其中训练图片样本用于对第一特征提取神经网络进行训练,测试图片样本用于验证经过训练图片样本训练后的第一特征提取神经网络在提取图片中的特征信息时的效果。优选的,可以对上述训练图片样本和测试图片样本进行数据增强的预处理,例如对每个样本进行对比度变换和亮度变换,再加上局部随机高斯噪声,并进行统一的归一化处理,从而得到数据增强后的训练图片样本和测试图片样本。In this step, the first picture sample is selected from the picture sample library to train the first feature extraction neural network. The first picture sample is preferably a full-body picture of the person. During implementation, the first image sample is divided into a training image sample and a test image sample, where the training image sample is used to train the first feature extraction neural network, and the test image sample is used to verify the first feature after the training image sample training The effect of extracting the neural network when extracting the feature information in the picture. Preferably, the above training picture samples and test picture samples may be subjected to data enhancement preprocessing, such as performing contrast transformation and brightness transformation on each sample, adding local random Gaussian noise, and performing uniform normalization processing, thereby The training image samples and the test image samples after data enhancement are obtained.
下面以某测试图片样本为例进行详细说明本步骤中的第一特征提取神经网络的结构,如图2所示。测试图片样本首先进入特征提取模块来提取该测试图片样本中的特征,本实施例中的特征提取模块采用ResNet残差网络,以保证较好的特征提取性能。测试样本图片经过ResNet残差网络后得到第一提取数据D 1,然后第一提取数据D 1分别进入四个含有不同膨胀系数的卷积模块,得到四个具有不同特征通道的第二提取数据D 2。接下来,四个具有不同特征通道的第二提取数据D 2经组合后进入以残差模块堆积起来的第一卷积层,得到四个具有不同感知野的第三提取数据D 3。最后,将四个具有不同感知野的第三提取数据D 3融合后,再次进入以残差模块堆积起来的第二卷积层,最终输出表征人体上的关键骨骼点的多个第一特征点。 The structure of the first feature extraction neural network in this step will be described in detail below with a test picture sample as an example, as shown in FIG. 2. The test picture sample first enters the feature extraction module to extract the features in the test picture sample. The feature extraction module in this embodiment uses a ResNet residual network to ensure better feature extraction performance. After the test sample image passes through the ResNet residual network, the first extracted data D 1 is obtained , and then the first extracted data D 1 enters four convolution modules with different expansion coefficients respectively, to obtain four second extracted data D with different feature channels. 2 . Next, the four second extracted data D 2 with different feature channels are combined into the first convolutional layer stacked by the residual module to obtain four third extracted data D 3 with different perceptual fields. Finally, after fusing four third extracted data D 3 with different perceptual fields, it enters the second convolutional layer piled up by the residual module again, and finally outputs multiple first feature points representing key bone points on the human body .
需要说明的是,本实施例中所公开的卷积模块个数以及膨胀系数的取值仅仅是作为示例性说明,并不以此为限。本领域普通技术人员可以根据实际需要任意改变上述卷积模块的个数以及膨胀系数的数值,均属于本申请的保护范围之内。It should be noted that the number of convolution modules and the expansion coefficient values disclosed in this embodiment are merely exemplary descriptions, and are not limited thereto. A person of ordinary skill in the art can arbitrarily change the number of the above-mentioned convolution modules and the value of the expansion coefficient according to actual needs, and all fall within the protection scope of the present application.
优选的,上述卷积模块依次包括以下层的组成:卷积层、批规范化层、Relu激活函数层、卷积层、批规范化层、Relu激活函数层和pool池化层,其中每个模块中的卷积层具有不同的膨胀系数。Preferably, the convolution module includes the following layers in sequence: a convolution layer, a batch normalization layer, a Relu activation function layer, a convolution layer, a batch normalization layer, a Relu activation function layer, and a pooling layer, each The convolutional layers have different expansion coefficients.
本步骤中,特征信息为人身体上的骨骼特征点,包括身体主要关节处的特征点,如肘关节、肩关节、膝关节、髋关节等。在此基础上,还可以进一步从骨骼特征点中选取与预设行为相关联的目标特征点。所述预设行为可以是蹲下、弯腰、站起、跌倒等等,不同的行为中发生位移的特征点可能各不相同,因此可以根据要检测的行为选取最能反映这种行为特点的目标特征点。在本发明的优选实施例中,通过综合考虑检测准确度和数据处理量,选择包括头、颈、两肩、两肘、两手、两臀、两膝以及两脚共14个骨骼点信息作为目标特征点。本发明的这种选取方式一方面可以使得骨骼特征点的数量尽可能少,以减少后续行为分析过程中的计算量;另一方面,上述14个目标特征点平均分布在人体各大主要关节处,能够从整体上反映出人体行为的基本趋势。本领域技术人 员能够理解,以上列举的骨骼点位置仅仅是用于举例,并不用来限制具体的特征点信息,根据具体情况,也可以对上述骨骼点信息进行删除或者增加,或者改变具体特征点的位置,例如也可获取人体中的穴位特征点信息,本申请对此不做限制。在此基础上,本实施例中的多个第一特征点优选的可以是在人体中标示出的上述骨骼点分布信息图。In this step, the feature information is the bone feature points on the human body, including the feature points at the main joints of the body, such as the elbow joint, shoulder joint, knee joint, hip joint, etc. On this basis, the target feature point associated with the preset behavior can be further selected from the bone feature points. The preset behavior may be squatting, bending over, standing up, falling, etc. The characteristic points of displacement in different behaviors may be different, so the one that best reflects the characteristics of this behavior can be selected according to the behavior to be detected Target feature point. In the preferred embodiment of the present invention, by comprehensively considering the detection accuracy and the amount of data processing, a total of 14 bone point information including head, neck, shoulders, elbows, hands, hips, knees, and feet are selected as targets Feature points. On the one hand, the selection method of the present invention can make the number of bone feature points as small as possible to reduce the calculation amount in the subsequent behavior analysis process; on the other hand, the above-mentioned 14 target feature points are evenly distributed at major joints of the human body , Can reflect the basic trend of human behavior as a whole. Those skilled in the art can understand that the positions of the bone points listed above are only used as examples, and are not used to limit specific feature point information. According to specific circumstances, the above bone point information may also be deleted or added, or specific feature points may be changed For example, the location of the acupuncture point in the human body can also be obtained. This application does not limit this. On this basis, the plurality of first feature points in this embodiment may preferably be the above bone point distribution information maps marked in the human body.
本步骤是采用以交叉熵为损失函数带动量的随机梯度下降法来训练上述第一特征提取神经网络的。具体损失函数的表达式如下:In this step, a random gradient descent method with cross entropy as the loss function and momentum is used to train the first feature extraction neural network. The expression of the specific loss function is as follows:
Figure PCTCN2019089500-appb-000001
Figure PCTCN2019089500-appb-000001
其中x p、y p代表所述第一特征提取神经网络提取到的第一特征点的预测坐标,x g、y g代表所述第一特征点的实际坐标。 Where x p and y p represent the predicted coordinates of the first feature point extracted by the first feature extraction neural network, and x g and y g represent the actual coordinates of the first feature point.
S2:将第二视频样本输入训练好的所述第一特征提取神经网络,得到表征所述第二视频样本中的人体的关键骨骼点的多个第二特征点。S2: Input the second video sample into the trained first feature extraction neural network to obtain a plurality of second feature points that represent key bone points of the human body in the second video sample.
在步骤S1已训练完成第一特征提取神经网络的基础上,本步骤利用训练好的第一特征提取神经网络来提取视频样本中的第二特征点,优选的,该第二特征点为上文中提到的14个骨骼特征点。On the basis that the first feature extraction neural network has been trained in step S1, this step uses the trained first feature extraction neural network to extract the second feature point in the video sample. Preferably, the second feature point is the above The 14 bone feature points mentioned.
本申请是以普通摄像头采集到的被监测者的视频信息为基础进行跌倒检测的,因此本步骤中进行特征点提取的对象是连续视频而不是简单的图片。由于视频是通过一系列的图片帧随时间变换而形成的,因此首先需要对视频采样以提取目标图片。例如,将视频按照20帧/秒的标准来提取图片,以3秒作为一个样本。同时为了产生多样化的样本,可以在视频中的行为起始点附近随机选择起始帧。The present application is based on the video information of the monitored person collected by a common camera for fall detection. Therefore, the object of feature point extraction in this step is a continuous video rather than a simple picture. Since the video is formed by a series of picture frames changing with time, the video needs to be sampled first to extract the target picture. For example, the video is extracted according to the standard of 20 frames per second, with 3 seconds as a sample. At the same time, in order to generate diverse samples, the starting frame can be randomly selected near the starting point of the behavior in the video.
提取到足够数量的目标图片之后,便可以通过第一特征提取神经网络来提取目标图片中的特征点信息,优选的可以是上文中提到的14个骨骼特征点。After a sufficient number of target pictures are extracted, the feature point information in the target pictures can be extracted through the first feature extraction neural network, preferably the 14 bone feature points mentioned above.
S3:对所述多个第二特征点进行编码生成预测特征图。S3: Encoding the plurality of second feature points to generate a predicted feature map.
本步骤用于对提取到的第二特征点进行处理,以得到预测特征图。仍以上文中的14个骨骼特征点为例,包括以下处理步骤:This step is used to process the extracted second feature points to obtain a predicted feature map. Taking the above 14 bone feature points as an example, the following processing steps are included:
S31:对上述骨骼特征点进行两两配对。S31: Pairwise pair the above bone feature points.
本实施例是从14个骨骼特征点中任选两个特征点进行配对,计算式如下:In this embodiment, any two feature points from the 14 bone feature points are paired, and the calculation formula is as follows:
C(14,2)=14!/(12!*2!)=91;C(14,2)=14! /(12!*2!)=91;
S32:计算每两个骨骼特征点之间的欧拉距离l xjt和方向速度v xit和v yit: S32: l xjt calculated Euclidean distance and direction of the velocity v and v Xit skeletal features between each two points yit:
Figure PCTCN2019089500-appb-000002
Figure PCTCN2019089500-appb-000002
v xit=x it-x i(t-1) v xit = x it -x i(t-1)
v yit=y it-y i(t-1) v yit = y it -y i(t-1)
上式中,x it、y it分别代表t时刻的第i个第二特征点的横、纵坐标;l xjt代表t时刻第i个第二特征点和第j个第二特征点的欧拉距离,v xit代表第i个第二特征点在t时刻在x方向上的速度,v yit代表第i个第二特征点在y方向上的速度。 In the above formula, x it and y it respectively represent the horizontal and vertical coordinates of the i-th second feature point at time t; l xjt represents the Euler of the i-th second feature point and j-th second feature point at time t distance, v xit represents the i-th second feature points at time t in the x-direction velocity, v yit speed of the i-th representative of a second characteristic point in the y-direction.
S33:将所有计算得到的欧拉距离和方向速度数据组合形成预测特征图。S33: Combine all calculated Euler distance and directional speed data to form a prediction feature map.
对于任一幅样本图而言,14个骨骼特征点进行两两配对可以得到91种组合方式,也就是可以计算得到91个欧拉距离;每个骨骼特征点分别具有一个x方向的速度和一个y方向的速度,也就是共有14个x方向的速度和14个y方向的速度,综合起来共得到91+14+14=119个特征向量。For any sample image, pairing 14 bone feature points into two pairs can obtain 91 combinations, that is, 91 Euler distances can be calculated; each bone feature point has a speed in the x direction and a The speed in the y direction, that is, a total of 14 speeds in the x direction and 14 speeds in the y direction, combined to obtain a total of 91+14+14=119 feature vectors.
假设本步骤中共有60帧图像需要处理,那么把每一帧图像中的特征向量按顺序排列,可以得到60×119的矩阵图。该矩阵图即为预测特征图。Assuming that a total of 60 frames of images need to be processed in this step, the feature vectors in each frame of images are arranged in order, and a 60×119 matrix map can be obtained. The matrix diagram is the prediction feature diagram.
S4:通过所述预测特征图训练第二行为分类神经网络,所述第二行为分类神经网络用于对所述预测特征图中表示的行为进行分类。S4: Train a second behavior classification neural network through the predicted feature map, and the second behavior classification neural network is used to classify the behavior represented in the predicted feature map.
在已经得到预测特征图的基础上,本步骤的目的在于训练第二行为分类神经网络,用于对表示在预测特征图中的行为进行分类,从而确定是否发生了跌倒行为。本申请中的第二行为分类神经网络的结构如图3所示,下面进行详细说明。On the basis that the prediction feature map has been obtained, the purpose of this step is to train a second behavior classification neural network to classify the behavior represented in the prediction feature map to determine whether a fall behavior has occurred. The structure of the second behavior classification neural network in this application is shown in FIG. 3, which will be described in detail below.
以步骤S3中得到的某预测特征图为例。所述预测特征图首先通过常规卷积模块,得到第一分类数据R1。然后,该第一分类数据R1分别通过四个具有不同膨胀系数的卷积模块,得到四个具有不同特征通道的第二分类数据R2, 优选的,上述四个卷积模块的膨胀系数分别为1,3,6和12。接下来,上述四个具有不同特征通道的第二分类数据R2组合后依次通过三个常规卷积模块,最终输出行为分类,用以判断表示在上述预测特征图中的行为属于哪种行为类别。Take a certain predicted feature map obtained in step S3 as an example. The prediction feature map first passes through a conventional convolution module to obtain first classification data R1. Then, the first classification data R1 respectively passes through four convolution modules with different expansion coefficients to obtain four second classification data R2 with different characteristic channels. Preferably, the expansion coefficients of the above four convolution modules are 1 respectively. , 3, 6 and 12. Next, the above-mentioned four second classification data R2 with different feature channels are combined and then sequentially passed through three conventional convolution modules, and finally output behavior classification, which is used to judge which behavior category the behavior represented in the above-mentioned predicted feature map belongs to.
需要说明的是,本实施例中所公开的卷积模块个数以及膨胀系数的取值仅仅是作为示例性说明,并不以此为限。本领域普通技术人员可以根据实际需要任意改变上述卷积模块的个数以及膨胀系数的数值,均属于本申请的保护范围之内。It should be noted that the number of convolution modules and the expansion coefficient values disclosed in this embodiment are merely exemplary descriptions, and are not limited thereto. A person of ordinary skill in the art can arbitrarily change the number of the above-mentioned convolution modules and the value of the expansion coefficient according to actual needs, and all fall within the protection scope of the present application.
优选的,上述卷积模块依次包括以下层的组成:卷积层、批规范化层、Relu激活函数层、卷积层、批规范化层、Relu激活函数层和pool池化层。Preferably, the convolution module includes the following layers in sequence: a convolution layer, a batch normalization layer, a Relu activation function layer, a convolution layer, a batch normalization layer, a Relu activation function layer, and a pooling layer.
本步骤中通过损失函数L H(X,Y)来对第二行为分类神经网络进行训练,具体表达式如下: In this step, the second behavior classification neural network is trained by the loss function L H (X, Y), the specific expression is as follows:
Figure PCTCN2019089500-appb-000003
Figure PCTCN2019089500-appb-000003
上式中,所述x k代表第k种行为类别的参数值,z k代表第k种行为类别的预测概率。例如,第二行为分类神经网络可以识别的类别有蹲下、站起、挥手、弯腰、跌倒、平躺等多种行为,每一种行为分别对应各自的参数值,例如当通过视频识别被监测人正在发生跌倒行为时,那么x k表示被监测人处于跌倒行为的参数值,z k表示被监测人正在发生跌倒行为的预测概率。 In the above formula, x k represents the parameter value of the kth behavior category, and z k represents the predicted probability of the kth behavior category. For example, the second behavior classification neural network can recognize the categories of squatting, standing, waving, bending, falling, lying down, etc., each behavior corresponds to its own parameter value, such as When the monitored person is falling, then x k represents the parameter value of the monitored person's falling behavior, and z k represents the predicted probability of the monitored person's falling behavior.
为了防止过拟合,本实施例在损失函数后又加上了一个L2正则项,用于防止发生过拟合的情况,得到的代价函数如下所示:In order to prevent overfitting, this embodiment adds an L2 regular term after the loss function to prevent overfitting. The resulting cost function is as follows:
L(X,Y)=L H(X,Y)+L2。 L(X,Y)=L H (X,Y)+L2.
S5:将被监测对象的视频数据依次输入训练好的所述第一特征提取神经网络和所述第二行为分类神经网络,以输出所述被监测对象的行为类别。S5: Input the video data of the monitored object into the trained first feature extraction neural network and the second behavior classification neural network in order to output the behavior category of the monitored object.
在已经完成对第一特征提取神经网络和第二行为分类神经网络的训练的基础上,本申请即可对实际的监控视频进行跌倒行为的检测。具体而言,本申请通过普通摄像头实时拍摄被监护对象的视频信息,该视频信息经过采样提取 到一定数量的目标图像。该目标图像首先经过训练好的第一特征提取神经网络,提取出目标图像中的多个特征点,例如是骨骼特征点。对多个骨骼特征点进行计算、组合,例如计算每两个骨骼特征点之间的欧拉距离和x方向、y方向上的速度,并对上述计算得到的矢量按照每帧图像的顺序进行排列,最终得到预测特征图。接下来,将预测特征图输入第二行为分类神经网络,便可以得到包含在预测特征图中的行为所述所属的类别,例如是否是跌倒行为。On the basis of having completed the training of the first feature extraction neural network and the second behavior classification neural network, the present application can detect the fall behavior of the actual surveillance video. Specifically, in this application, the video information of the monitored object is photographed in real time through a common camera, and the video information is sampled to extract a certain number of target images. The target image first undergoes a trained first feature extraction neural network to extract multiple feature points in the target image, such as bone feature points. Calculate and combine multiple bone feature points, for example, calculate the Euler distance between each two bone feature points and the speed in the x and y directions, and arrange the vectors calculated above in the order of each frame of image , And finally get the predicted feature map. Next, by inputting the prediction feature map into the second behavior classification neural network, you can obtain the category to which the behavior included in the prediction feature map belongs, for example, whether it is a fall behavior.
请继续参阅图4,示出了一种跌倒检测装置,在本实施例中,跌倒检测装置10可以包括或被分割成一个或多个程序模块,一个或者多个程序模块被存储于存储介质中,并由一个或多个处理器所执行,以完成本申请,并可实现上述跌倒检测方法。本申请所称的程序模块是指能够完成特定功能的一系列计算机程序指令段,比程序本身更适合于描述跌倒检测装置10在存储介质中的执行过程。以下描述将具体介绍本实施例各程序模块的功能:Please continue to refer to FIG. 4, which shows a fall detection device. In this embodiment, the fall detection device 10 may include or be divided into one or more program modules, and one or more program modules are stored in a storage medium. , And executed by one or more processors to complete this application, and can implement the above-mentioned fall detection method. The program module referred to in this application refers to a series of computer program instruction segments capable of performing specific functions, and is more suitable for describing the execution process of the fall detection device 10 in the storage medium than the program itself. The following description will specifically introduce the functions of the program modules of this embodiment:
第一神经网络训练模块11,适用于通过第一图片样本训练第一特征提取神经网络,所述第一特征提取神经网络用于提取所述第一图片样本中的多个第一特征点,所述第一特征点表征人体上的关键骨骼点;The first neural network training module 11 is adapted to train a first feature extraction neural network through a first picture sample. The first feature extraction neural network is used to extract multiple first feature points in the first picture sample. The first feature point represents the key bone point on the human body;
特征点提取模块12,适用于将第二视频样本输入训练好的所述第一特征提取神经网络,得到表征所述第二视频样本中的人体的关键骨骼点的多个第二特征点;The feature point extraction module 12 is adapted to input the second video sample into the trained first feature extraction neural network to obtain a plurality of second feature points characterizing key bone points of the human body in the second video sample;
特征图生成模块13,适用于对所述多个第二特征点进行编码生成表征所述多个第二特征点的分布情况的预测特征图;The feature map generation module 13 is adapted to encode the multiple second feature points to generate a predicted feature map characterizing the distribution of the multiple second feature points;
第二神经网络训练模块14,适用于通过所述预测特征图训练第二行为分类神经网络,所述第二行为分类神经网络用于对所述预测特征图中表示的行为进行分类;The second neural network training module 14 is adapted to train a second behavior classification neural network through the predicted feature map, and the second behavior classification neural network is used to classify the behavior represented in the predicted feature map;
分类模块15,适用于将被监测对象的视频数据依次输入训练好的所述第一特征提取神经网络和所述第二行为分类神经网络,以输出所述被监测对象的行为类别。The classification module 15 is adapted to sequentially input the video data of the monitored object into the trained first feature extraction neural network and the second behavior classification neural network to output the behavior category of the monitored object.
本实施例还提供一种计算机设备,如可以执行程序的智能手机、平板电脑、笔记本电脑、台式计算机、机架式服务器、刀片式服务器、塔式服务器或机柜式服务器(包括独立的服务器,或者多个服务器所组成的服务器集群)等。本实施例的计算机设备20至少包括但不限于:可通过系统总线相互通信连接的存储器21、处理器22,如图5所示。需要指出的是,图5仅示出了具有组件21-22的计算机设备20,但是应理解的是,并不要求实施所有示出的组件,可以替代的实施更多或者更少的组件。This embodiment also provides a computer device, such as a smartphone, a tablet computer, a notebook computer, a desktop computer, a rack server, a blade server, a tower server, or a rack server (including an independent server, or A server cluster composed of multiple servers), etc. The computer device 20 of this embodiment includes at least but not limited to: a memory 21 and a processor 22 that can be communicatively connected to each other through a system bus, as shown in FIG. 5. It should be noted that FIG. 5 only shows the computer device 20 having components 21-22, but it should be understood that it is not required to implement all the components shown, and that more or fewer components may be implemented instead.
本实施例中,存储器21(即可读存储介质)包括闪存、硬盘、多媒体卡、卡型存储器(例如,SD或DX存储器等)、随机访问存储器(RAM)、静态随机访问存储器(SRAM)、只读存储器(ROM)、电可擦除可编程只读存储器(EEPROM)、可编程只读存储器(PROM)、磁性存储器、磁盘、光盘等。在一些实施例中,存储器21可以是计算机设备20的内部存储单元,例如该计算机设备20的硬盘或内存。在另一些实施例中,存储器21也可以是计算机设备20的外部存储设备,例如该计算机设备20上配备的插接式硬盘,智能存储卡(Smart Media Card,SMC),安全数字(Secure Digital,SD)卡,闪存卡(Flash Card)等。当然,存储器21还可以既包括计算机设备20的内部存储单元也包括其外部存储设备。本实施例中,存储器21通常用于存储安装于计算机设备20的操作系统和各类应用软件,例如实施例一的跌倒检测装置10的程序代码等。此外,存储器21还可以用于暂时地存储已经输出或者将要输出的各类数据。In this embodiment, the memory 21 (ie, readable storage medium) includes flash memory, hard disk, multimedia card, card-type memory (for example, SD or DX memory, etc.), random access memory (RAM), static random access memory (SRAM), Read only memory (ROM), electrically erasable programmable read only memory (EEPROM), programmable read only memory (PROM), magnetic memory, magnetic disk, optical disk, etc. In some embodiments, the memory 21 may be an internal storage unit of the computer device 20, such as a hard disk or memory of the computer device 20. In other embodiments, the memory 21 may also be an external storage device of the computer device 20, such as a plug-in hard disk equipped on the computer device 20, a smart memory card (Smart Media, Card, SMC), and secure digital (Secure Digital, SD) card, flash card (Flash Card), etc. Of course, the memory 21 may also include both the internal storage unit of the computer device 20 and its external storage device. In this embodiment, the memory 21 is generally used to store the operating system and various application software installed in the computer device 20, such as the program code of the fall detection device 10 of the first embodiment. In addition, the memory 21 can also be used to temporarily store various types of data that have been output or will be output.
处理器22在一些实施例中可以是中央处理器(Central Processing Unit,CPU)、控制器、微控制器、微处理器、或其他数据处理芯片。该处理器22通常用于控制计算机设备20的总体操作。本实施例中,处理器22用于运行存储器21中存储的程序代码或者处理数据,例如运行跌倒检测装置10,以实现实施例一的跌倒检测方法。The processor 22 may be a central processing unit (CPU), controller, microcontroller, microprocessor, or other data processing chip in some embodiments. The processor 22 is generally used to control the overall operation of the computer device 20. In this embodiment, the processor 22 is used to run the program code or process data stored in the memory 21, for example, to run the fall detection device 10, so as to implement the fall detection method of the first embodiment.
本实施例还提供一种计算机可读存储介质,如闪存、硬盘、多媒体卡、卡 型存储器(例如,SD或DX存储器等)、随机访问存储器(RAM)、静态随机访问存储器(SRAM)、只读存储器(ROM)、电可擦除可编程只读存储器(EEPROM)、可编程只读存储器(PROM)、磁性存储器、磁盘、光盘、服务器、App应用商城等等,其上存储有计算机程序,程序被处理器执行时实现相应功能。本实施例的计算机可读存储介质用于存储跌倒检测装置10,被处理器执行时实现实施例一的跌倒检测方法。This embodiment also provides a computer-readable storage medium, such as flash memory, hard disk, multimedia card, card-type memory (for example, SD or DX memory, etc.), random access memory (RAM), static random access memory (SRAM), only Read memory (ROM), electrically erasable programmable read-only memory (EEPROM), programmable read-only memory (PROM), magnetic memory, magnetic disk, optical disk, server, App store, etc., which store computer programs, When the program is executed by the processor, the corresponding function is realized. The computer-readable storage medium of this embodiment is used to store the fall detection device 10, and when executed by the processor, the fall detection method of the first embodiment is implemented.
上述本申请实施例序号仅仅为了描述,不代表实施例的优劣。The sequence numbers of the above embodiments of the present application are for description only, and do not represent the advantages and disadvantages of the embodiments.
流程图中或在此以其它方式描述的任何过程或方法描述可以被理解为,表示包括一个或更多个用于实现特定逻辑功能或过程的步骤的可执行指令的代码的模块、片段或部分,并且本申请的优选实施方式的范围包括另外的实现,其中可以不按所示出或讨论的顺序,包括根据所涉及的功能按基本同时的方式或按相反的顺序,来执行功能,这应被本申请的实施例所属技术领域的技术人员所理解。Any process or method description in a flowchart or otherwise described herein may be understood to represent a module, segment, or portion of code that includes one or more executable instructions for implementing specific logical functions or steps of a process , And the scope of the preferred embodiment of the present application includes additional implementations, in which the functions may not be performed in the order shown or discussed, including performing functions in a substantially simultaneous manner or in reverse order according to the functions involved, which shall It is understood by those skilled in the art to which the embodiments of the present application belong.
本技术领域的普通技术人员可以理解,实现上述实施例方法携带的全部或部分步骤是可以通过程序来指令相关的硬件完成,所述的程序可以存储于一种计算机可读介质中,该程序在执行时,包括方法实施例的步骤之一或其组合。A person of ordinary skill in the art can understand that all or part of the steps carried in the method of the above embodiment can be completed by instructing relevant hardware through a program, and the program can be stored in a computer-readable medium. When executed, it includes one of the steps of the method embodiment or a combination thereof.
在本说明书的描述中,参考术语“一个实施例”、“一些实施例”、“示例”、“具体示例”或“一些示例”等的描述意指结合该实施例或示例描述的具体特征、结构、材料或者特点包含于本申请的至少一个实施例或示例中。在本说明书中,对上述术语的示意性表述不一定指的是相同的实施例或示例。而且,描述的具体特征、结构、材料或者特点可以在任何的一个或多个实施例或示例中以合适的方式结合。In the description of this specification, the description with reference to the terms "one embodiment", "some embodiments", "examples", "specific examples" or "some examples" means specific features described in conjunction with the embodiment or examples, The structure, material or characteristic is included in at least one embodiment or example of the present application. In this specification, the schematic expression of the above terms does not necessarily refer to the same embodiment or example. Moreover, the specific features, structures, materials, or characteristics described may be combined in any suitable manner in any one or more embodiments or examples.
通过以上的实施方式的描述,本领域的技术人员可以清楚地了解到上述实施例方法可借助软件加必需的通用硬件平台的方式来实现,当然也可以通过硬件,但很多情况下前者是更佳的实施方式。Through the description of the above embodiments, those skilled in the art can clearly understand that the methods in the above embodiments can be implemented by means of software plus a necessary general hardware platform, and of course, can also be implemented by hardware, but in many cases the former is better Implementation.
以上仅为本申请的优选实施例,并非因此限制本申请的专利范围,凡是利 用本申请说明书及附图内容所作的等效结构或等效流程变换,或直接或间接运用在其他相关的技术领域,均同理包括在本申请的专利保护范围内。The above are only the preferred embodiments of the present application, and do not limit the scope of the patent of the present application. Any equivalent structure or equivalent process transformation made by the description and drawings of this application, or directly or indirectly used in other related technical fields The same reason is included in the patent protection scope of this application.

Claims (27)

  1. 一种基于骨骼点的跌倒检测方法,其特征在于,包括以下步骤:A fall detection method based on bone points is characterized by the following steps:
    通过第一图片样本训练第一特征提取神经网络,所述第一特征提取神经网络用于提取表征人体上的关键骨骼点的多个第一特征点;Training a first feature extraction neural network through a first picture sample, the first feature extraction neural network is used to extract a plurality of first feature points representing key bone points on the human body;
    将第二视频样本输入训练好的所述第一特征提取神经网络,得到表征所述第二视频样本中的人体的关键骨骼点的多个第二特征点;Input a second video sample into the trained first feature extraction neural network to obtain a plurality of second feature points characterizing key bone points of the human body in the second video sample;
    对所述多个第二特征点进行编码生成预测特征图;Encoding the plurality of second feature points to generate a predicted feature map;
    通过所述预测特征图训练第二行为分类神经网络,所述第二行为分类神经网络用于对所述预测特征图中表示的行为进行分类;Training a second behavior classification neural network through the predicted feature map, the second behavior classification neural network is used to classify the behavior represented in the predicted feature map;
    将被监测对象的视频数据依次输入训练好的所述第一特征提取神经网络和所述第二行为分类神经网络,输出所述被监测对象的行为类别。The video data of the monitored object is sequentially input into the trained first feature extraction neural network and the second behavior classification neural network, and the behavior category of the monitored object is output.
  2. 根据权利要求1所述的跌倒检测方法,其特征在于,所述通过第一图片样本训练第一特征提取神经网络,包括:The fall detection method according to claim 1, wherein the training of the first feature extraction neural network through the first picture sample includes:
    所述第一图片样本输入Resnet残差网络,得到第一提取数据;The first picture sample is input to a Resnet residual network to obtain first extracted data;
    所述第一提取数据分别通过多个具有不同膨胀系数的卷积模块,得到多个具有不同特征通道的第二提取数据;The first extracted data respectively passes through multiple convolution modules with different expansion coefficients to obtain multiple second extracted data with different feature channels;
    所述多个具有不同特征通道的第二提取数据组合后进入以残差卷积堆积起来的第一卷积层,得到多个具有不同感知野的第三提取数据;The plurality of second extracted data with different feature channels are combined to enter a first convolution layer piled up by residual convolution, to obtain a plurality of third extracted data with different perception fields;
    对所述多个具有不同感知野的第三提取数据进行融合,然后进入以残差模块堆积起来的第二卷积层,最终输出表征人体上的关键骨骼点的多个第一特征点;Fuse the plurality of third extracted data with different perceptual fields, and then enter a second convolutional layer piled up with a residual module, and finally output a plurality of first feature points representing key bone points on the human body;
    通过第一损失函数对所述第一特征提取网络进行反向训练。Perform reverse training on the first feature extraction network through a first loss function.
  3. 根据权利要求1所述的跌倒检测方法,其特征在于,所述通过所述预测特征图训练第二行为分类神经网络,包括:The fall detection method according to claim 1, wherein the training of the second behavior classification neural network through the predicted feature map includes:
    所述预测特征图通过常规卷积模块,得到第一分类数据;The predicted feature map passes through a conventional convolution module to obtain first classification data;
    所述第一分类数据分别通过多个具有不同膨胀系数的卷积模块,得到多个具有不同特征通道的第二分类数据;The first classification data respectively passes through multiple convolution modules with different expansion coefficients to obtain multiple second classification data with different feature channels;
    所述多个具有不同特征通道的第二分类数据组合后依次通过三个常规卷积模块,最终输出行为分类。The plurality of second classification data with different feature channels are combined and sequentially passed through three conventional convolution modules to finally output behavior classification.
  4. 根据权利要求2所述的跌倒检测方法,其特征在于,所述第一损失函数F为:The fall detection method according to claim 2, wherein the first loss function F is:
    Figure PCTCN2019089500-appb-100001
    Figure PCTCN2019089500-appb-100001
    其中x p、y p代表所述第一特征提取神经网络提取到的第一特征点的预测坐标,x g、y g代表所述第一特征点的实际坐标。 Where x p and y p represent the predicted coordinates of the first feature point extracted by the first feature extraction neural network, and x g and y g represent the actual coordinates of the first feature point.
  5. 根据权利要求3所述的跌倒检测方法,其特征在于,所述第二损失函数L为:The fall detection method according to claim 3, wherein the second loss function L is:
    Figure PCTCN2019089500-appb-100002
    Figure PCTCN2019089500-appb-100002
    其中,所述x k代表第k种行为类别的参数值,z k代表第k中行为类别的预测概率,L2代表防止发生过拟合的正则项。 Wherein, x k represents the parameter value of the kth behavior category, z k represents the predicted probability of the kth behavior category, and L2 represents the regular term that prevents overfitting.
  6. 根据权利要求2所述的跌倒检测方法,其特征在于,所述卷积模块由以下层依次串联组成:卷积层、批次正则化层、Relu激活函数层、卷积层、批次正则化层、Relu激活函数层、池化层。The fall detection method according to claim 2, wherein the convolution module is composed of the following layers in series: a convolutional layer, a batch regularization layer, a Relu activation function layer, a convolutional layer, a batch regularization Layer, Relu activation function layer, pooling layer.
  7. 根据权利要求3所述的跌倒检测方法,其特征在于,所述卷积模块由以下层依次串联组成:卷积层、批次正则化层、Relu激活函数层、卷积层、批次正则化层、Relu激活函数层、池化层。The fall detection method according to claim 3, wherein the convolution module is composed of the following layers connected in series: a convolutional layer, a batch regularization layer, a Relu activation function layer, a convolutional layer, a batch regularization Layer, Relu activation function layer, pooling layer.
  8. 根据权利要求1所述的跌倒检测方法,其特征在于,所述对所述多个第二特征点进行编码生成预测特征图,包括:The fall detection method according to claim 1, wherein the encoding of the plurality of second feature points to generate a predicted feature map includes:
    对所述多个第二特征点进行两两配对;Pairwise pairing the plurality of second feature points;
    计算每两个第二特征点之间的距离和速度:Calculate the distance and speed between every two second feature points:
    Figure PCTCN2019089500-appb-100003
    Figure PCTCN2019089500-appb-100003
    上式中,x it、y it分别代表t时刻的第i个第二特征点的横、纵坐标;l xjt代表t时刻第i个第二特征点和第j个第二特征点的欧拉距离,v xit代表第i 个第二特征点在t时刻在x方向上的速度,v yit代表第i个第二特征点在y方向上的速度; In the above formula, x it and y it respectively represent the horizontal and vertical coordinates of the i-th second feature point at time t; l xjt represents the Euler of the i-th second feature point and j-th second feature point at time t distance, v xit represents the i-th second feature points at time t in the x-direction velocity, v yit speed of the i-th representative of a second characteristic point in the y direction;
    将所有计算得到的距离和速度数据组合形成预测特征图。Combine all calculated distance and speed data to form a prediction feature map.
  9. 根据权利要求8所述的跌倒检测方法,其特征在于,所述对所述多个第二特征点进行编码生成预测特征图的步骤包括:The fall detection method according to claim 8, wherein the step of encoding the plurality of second feature points to generate a predicted feature map includes:
    从所述多个第二特征点中选取与预设行为相关联的多个目标特征点;Selecting a plurality of target feature points associated with a preset behavior from the plurality of second feature points;
    对所述多个目标特征点进行编码生成预测特征图。Encoding the plurality of target feature points to generate a predicted feature map.
  10. 根据权利要求1所述的跌倒检测方法,其特征在于,所述通过第一图片样本训练第一特征提取神经网络的步骤之前,还包括:The fall detection method according to claim 1, wherein before the step of training the first feature extraction neural network through the first picture sample, further comprising:
    对所述第一图片样本进行数据增强的预处理。Data pre-processing is performed on the first picture sample.
  11. 一种基于骨骼点的跌倒检测装置,其特征在于,包括:A fall detection device based on bone points is characterized by comprising:
    第一神经网络训练模块,适用于通过第一图片样本训练第一特征提取神经网络,所述第一特征提取神经网络用于提取所述第一图片样本中的多个第一特征点,所述第一特征点表征人体上的关键骨骼点;The first neural network training module is adapted to train a first feature extraction neural network through a first picture sample, the first feature extraction neural network is used to extract multiple first feature points in the first picture sample, the The first feature point represents the key bone point on the human body;
    特征点提取模块,适用于将第二视频样本输入训练好的所述第一特征提取神经网络,得到表征所述第二视频样本中的人体的关键骨骼点的多个第二特征点;The feature point extraction module is adapted to input the second video sample into the trained first feature extraction neural network to obtain a plurality of second feature points characterizing key bone points of the human body in the second video sample;
    特征图生成模块,适用于对所述多个第二特征点进行编码生成预测特征图;A feature map generation module, adapted to encode the multiple second feature points to generate a predicted feature map;
    第二神经网络训练模块,适用于通过所述预测特征图训练第二行为分类神经网络,所述第二行为分类神经网络用于对所述预测特征图中表示的行为进行分类;A second neural network training module, adapted to train a second behavior classification neural network through the predicted feature map, and the second behavior classification neural network is used to classify the behavior represented in the predicted feature map;
    分类模块,适用于将被监测对象的视频数据依次输入训练好的所述第一特征提取神经网络和所述第二行为分类神经网络,以输出所述被监测对象的行为类别。The classification module is adapted to sequentially input the video data of the monitored object into the trained first feature extraction neural network and the second behavior classification neural network to output the behavior category of the monitored object.
  12. 一种计算机设备,包括存储器、处理器以及存储在存储器上并可在处理器上运行的计算机程序,其特征在于,所述处理器执行所述计算机程序时实现以下步骤:A computer device includes a memory, a processor, and a computer program stored on the memory and executable on the processor, characterized in that the processor implements the following steps when executing the computer program:
    通过第一图片样本训练第一特征提取神经网络,所述第一特征提取神经网络用于提取所述第一图片样本中的多个第一特征点,所述第一特征点表征人体上的关键骨骼点;Train a first feature extraction neural network through a first picture sample, the first feature extraction neural network is used to extract a plurality of first feature points in the first picture sample, the first feature points represent key points on the human body Bone point
    将第二视频样本输入训练好的所述第一特征提取神经网络,得到表征所述第二视频样本中的人体的关键骨骼点的多个第二特征点;Input a second video sample into the trained first feature extraction neural network to obtain a plurality of second feature points characterizing key bone points of the human body in the second video sample;
    对所述多个第二特征点进行编码生成预测特征图;Encoding the plurality of second feature points to generate a predicted feature map;
    通过所述预测特征图训练第二行为分类神经网络,所述第二行为分类神经网络用于对所述预测特征图中表示的行为进行分类;Training a second behavior classification neural network through the predicted feature map, the second behavior classification neural network is used to classify the behavior represented in the predicted feature map;
    将被监测对象的视频数据依次输入训练好的所述第一特征提取神经网络和所述第二行为分类神经网络,以输出所述被监测对象的行为类别。The video data of the monitored object is sequentially input into the trained first feature extraction neural network and the second behavior classification neural network to output the behavior category of the monitored object.
  13. 根据权利要求12所述的计算机设备,其特征在于,所述通过第一图片样本训练第一特征提取神经网络,包括:The computer device according to claim 12, wherein the training of the first feature extraction neural network through the first picture sample includes:
    所述第一图片样本输入Resnet残差网络,得到第一提取数据;The first picture sample is input to a Resnet residual network to obtain first extracted data;
    所述第一提取数据分别通过多个具有不同膨胀系数的卷积模块,得到多个具有不同特征通道的第二提取数据;The first extracted data respectively passes through multiple convolution modules with different expansion coefficients to obtain multiple second extracted data with different feature channels;
    所述多个具有不同特征通道的第二提取数据组合后进入以残差卷积堆积起来的第一卷积层,得到多个具有不同感知野的第三提取数据;The plurality of second extracted data with different feature channels are combined to enter a first convolution layer piled up by residual convolution, to obtain a plurality of third extracted data with different perception fields;
    对所述多个具有不同感知野的第三提取数据进行融合,然后进入以残差模块堆积起来的第二卷积层,最终输出表征人体上的关键骨骼点的多个第一特征点;Fuse the plurality of third extracted data with different perceptual fields, and then enter a second convolutional layer piled up with a residual module, and finally output a plurality of first feature points representing key bone points on the human body;
    通过第一损失函数对所述第一特征提取网络进行反向训练。Perform reverse training on the first feature extraction network through a first loss function.
  14. 根据权利要求12所述的计算机设备,其特征在于,所述通过所述预测特征图训练第二行为分类神经网络,包括:The computer device according to claim 12, wherein the training of the second behavior classification neural network through the predicted feature map comprises:
    所述预测特征图通过常规卷积模块,得到第一分类数据;The predicted feature map passes through a conventional convolution module to obtain first classification data;
    所述第一分类数据分别通过多个具有不同膨胀系数的卷积模块,得到多个具有不同特征通道的第二分类数据;The first classification data respectively passes through multiple convolution modules with different expansion coefficients to obtain multiple second classification data with different feature channels;
    所述多个具有不同特征通道的第二分类数据组合后依次通过三个常规卷 积模块,最终输出行为分类。The plurality of second classification data with different feature channels are combined and sequentially passed through three conventional convolution modules to finally output behavior classification.
  15. 根据权利要求13所述的计算机设备,其特征在于,所述第一损失函数F为:The computer device according to claim 13, wherein the first loss function F is:
    Figure PCTCN2019089500-appb-100004
    Figure PCTCN2019089500-appb-100004
    其中x p、y p代表所述第一特征提取神经网络提取到的第一特征点的预测坐标,x g、y g代表所述第一特征点的实际坐标。 Where x p and y p represent the predicted coordinates of the first feature point extracted by the first feature extraction neural network, and x g and y g represent the actual coordinates of the first feature point.
  16. 根据权利要求13所述的计算机设备,其特征在于,所述第二损失函数L为:The computer device according to claim 13, wherein the second loss function L is:
    Figure PCTCN2019089500-appb-100005
    Figure PCTCN2019089500-appb-100005
    其中,所述x k代表第k种行为类别的参数值,z k代表第k中行为类别的预测概率,L2代表防止发生过拟合的正则项。 Wherein, x k represents the parameter value of the kth behavior category, z k represents the predicted probability of the kth behavior category, and L2 represents the regular term that prevents overfitting.
  17. 根据权利要求13所述的计算机设备,其特征在于,所述卷积模块由以下层依次串联组成:卷积层、批次正则化层、Relu激活函数层、卷积层、批次正则化层、Relu激活函数层、池化层。The computer device according to claim 13, wherein the convolution module is composed of the following layers connected in series: a convolution layer, a batch regularization layer, a Relu activation function layer, a convolution layer, a batch regularization layer , Relu activation function layer, pooling layer.
  18. 根据权利要求14所述的计算机设备,其特征在于,所述卷积模块由以下层依次串联组成:卷积层、批次正则化层、Relu激活函数层、卷积层、批次正则化层、Relu激活函数层、池化层。The computer device according to claim 14, wherein the convolution module is composed of the following layers connected in series in sequence: a convolution layer, a batch regularization layer, a Relu activation function layer, a convolution layer, a batch regularization layer , Relu activation function layer, pooling layer.
  19. 根据权利要求12所述的计算机设备,其特征在于,所述对所述多个第二特征点进行编码生成预测特征图,包括:The computer device according to claim 12, wherein the encoding of the plurality of second feature points to generate a predicted feature map includes:
    对所述多个第二特征点进行两两配对;Pairwise pairing the plurality of second feature points;
    计算每两个第二特征点之间的距离和速度:Calculate the distance and speed between every two second feature points:
    Figure PCTCN2019089500-appb-100006
    Figure PCTCN2019089500-appb-100006
    上式中,x it、y it分别代表t时刻的第i个第二特征点的横、纵坐标;l xjt代表t时刻第i个第二特征点和第j个第二特征点的欧拉距离,v xit代表第i个第二特征点在t时刻在x方向上的速度,v yit代表第i个第二特征点在y方 向上的速度; In the above formula, x it and y it respectively represent the horizontal and vertical coordinates of the i-th second feature point at time t; l xjt represents the Euler of the i-th second feature point and j-th second feature point at time t distance, v xit represents the i-th second feature points at time t in the x-direction velocity, v yit speed of the i-th representative of a second characteristic point in the y direction;
    将所有计算得到的距离和速度数据组合形成预测特征图。Combine all calculated distance and speed data to form a prediction feature map.
  20. 一种计算机可读存储介质,其上存储有计算机程序,其特征在于,所述计算机程序被处理器执行时实现以下步骤:A computer-readable storage medium on which a computer program is stored, characterized in that when the computer program is executed by a processor, the following steps are realized:
    通过第一图片样本训练第一特征提取神经网络,所述第一特征提取神经网络用于提取所述第一图片样本中的多个第一特征点,所述第一特征点表征人体上的关键骨骼点;Train a first feature extraction neural network through a first picture sample, the first feature extraction neural network is used to extract a plurality of first feature points in the first picture sample, the first feature points represent key points on the human body Bone point
    将第二视频样本输入训练好的所述第一特征提取神经网络,得到表征所述第二视频样本中的人体的关键骨骼点的多个第二特征点;Input a second video sample into the trained first feature extraction neural network to obtain a plurality of second feature points characterizing key bone points of the human body in the second video sample;
    对所述多个第二特征点进行编码生成预测特征图;Encoding the plurality of second feature points to generate a predicted feature map;
    通过所述预测特征图训练第二行为分类神经网络,所述第二行为分类神经网络用于对所述预测特征图中表示的行为进行分类;Training a second behavior classification neural network through the predicted feature map, the second behavior classification neural network is used to classify the behavior represented in the predicted feature map;
    将被监测对象的视频数据依次输入训练好的所述第一特征提取神经网络和所述第二行为分类神经网络,以输出所述被监测对象的行为类别。The video data of the monitored object is sequentially input into the trained first feature extraction neural network and the second behavior classification neural network to output the behavior category of the monitored object.
  21. 根据权利要求20所述的计算机可读存储介质,其特征在于,所述通过第一图片样本训练第一特征提取神经网络,包括:The computer-readable storage medium of claim 20, wherein the training of the first feature extraction neural network through the first picture sample includes:
    所述第一图片样本输入Resnet残差网络,得到第一提取数据;The first picture sample is input to a Resnet residual network to obtain first extracted data;
    所述第一提取数据分别通过多个具有不同膨胀系数的卷积模块,得到多个具有不同特征通道的第二提取数据;The first extracted data respectively passes through multiple convolution modules with different expansion coefficients to obtain multiple second extracted data with different feature channels;
    所述多个具有不同特征通道的第二提取数据组合后进入以残差卷积堆积起来的第一卷积层,得到多个具有不同感知野的第三提取数据;The plurality of second extracted data with different feature channels are combined to enter a first convolution layer piled up by residual convolution, to obtain a plurality of third extracted data with different perception fields;
    对所述多个具有不同感知野的第三提取数据进行融合,然后进入以残差模块堆积起来的第二卷积层,最终输出表征人体上的关键骨骼点的多个第一特征点;Fuse the plurality of third extracted data with different perceptual fields, and then enter a second convolutional layer piled up with a residual module, and finally output a plurality of first feature points representing key bone points on the human body;
    通过第一损失函数对所述第一特征提取网络进行反向训练。Perform reverse training on the first feature extraction network through a first loss function.
  22. 根据权利要求20所述的计算机可读存储介质,其特征在于,所述通过所述预测特征图训练第二行为分类神经网络,包括:The computer-readable storage medium of claim 20, wherein the training of the second behavior classification neural network through the predicted feature map includes:
    所述预测特征图通过常规卷积模块,得到第一分类数据;The predicted feature map passes through a conventional convolution module to obtain first classification data;
    所述第一分类数据分别通过多个具有不同膨胀系数的卷积模块,得到多个具有不同特征通道的第二分类数据;The first classification data respectively passes through multiple convolution modules with different expansion coefficients to obtain multiple second classification data with different feature channels;
    所述多个具有不同特征通道的第二分类数据组合后依次通过三个常规卷积模块,最终输出行为分类。The plurality of second classification data with different feature channels are combined and sequentially passed through three conventional convolution modules to finally output behavior classification.
  23. 根据权利要求21所述的计算机可读存储介质,其特征在于,所述第一损失函数F为:The computer-readable storage medium of claim 21, wherein the first loss function F is:
    Figure PCTCN2019089500-appb-100007
    Figure PCTCN2019089500-appb-100007
    其中x p、y p代表所述第一特征提取神经网络提取到的第一特征点的预测坐标,x g、y g代表所述第一特征点的实际坐标。 Where x p and y p represent the predicted coordinates of the first feature point extracted by the first feature extraction neural network, and x g and y g represent the actual coordinates of the first feature point.
  24. 根据权利要求22所述的计算机可读存储介质,其特征在于,所述第二损失函数L为:The computer-readable storage medium of claim 22, wherein the second loss function L is:
    Figure PCTCN2019089500-appb-100008
    Figure PCTCN2019089500-appb-100008
    其中,所述x k代表第k种行为类别的参数值,z k代表第k中行为类别的预测概率,L2代表防止发生过拟合的正则项。 Wherein, x k represents the parameter value of the kth behavior category, z k represents the predicted probability of the kth behavior category, and L2 represents the regular term that prevents overfitting.
  25. 根据权利要求21所述的计算机可读存储介质,其特征在于,所述卷积模块由以下层依次串联组成:卷积层、批次正则化层、Relu激活函数层、卷积层、批次正则化层、Relu激活函数层、池化层。The computer-readable storage medium according to claim 21, wherein the convolution module is composed of the following layers connected in series: a convolution layer, a batch regularization layer, a Relu activation function layer, a convolution layer, a batch Regularization layer, Relu activation function layer, pooling layer.
  26. 根据权利要求22所述的计算机可读存储介质,其特征在于,所述卷积模块由以下层依次串联组成:卷积层、批次正则化层、Relu激活函数层、卷积层、批次正则化层、Relu激活函数层、池化层。The computer-readable storage medium of claim 22, wherein the convolution module is composed of the following layers connected in series: a convolution layer, a batch regularization layer, a Relu activation function layer, a convolution layer, a batch Regularization layer, Relu activation function layer, pooling layer.
  27. 根据权利要求20所述的计算机可读存储介质,其特征在于,所述对所述多个第二特征点进行编码生成预测特征图,包括:The computer-readable storage medium of claim 20, wherein the encoding of the plurality of second feature points to generate a predicted feature map includes:
    对所述多个第二特征点进行两两配对;Pairwise pairing the plurality of second feature points;
    计算每两个第二特征点之间的距离和速度:Calculate the distance and speed between every two second feature points:
    Figure PCTCN2019089500-appb-100009
    Figure PCTCN2019089500-appb-100009
    Figure PCTCN2019089500-appb-100010
    Figure PCTCN2019089500-appb-100010
    上式中,x it、y it分别代表t时刻的第i个第二特征点的横、纵坐标;l xjt代表t时刻第i个第二特征点和第j个第二特征点的欧拉距离,v xit代表第i个第二特征点在t时刻在x方向上的速度,v yit代表第i个第二特征点在y方向上的速度; In the above formula, x it and y it respectively represent the horizontal and vertical coordinates of the i-th second feature point at time t; l xjt represents the Euler of the i-th second feature point and j-th second feature point at time t distance, v xit represents the i-th second feature points at time t in the x-direction velocity, v yit speed of the i-th representative of a second characteristic point in the y direction;
    将所有计算得到的距离和速度数据组合形成预测特征图。Combine all calculated distance and speed data to form a prediction feature map.
PCT/CN2019/089500 2018-11-28 2019-05-31 Bone point-based fall detection method and fall detection device therefor WO2020107847A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN201811433808.3 2018-11-28
CN201811433808.3A CN109492612A (en) 2018-11-28 2018-11-28 Fall detection method and its falling detection device based on skeleton point

Publications (1)

Publication Number Publication Date
WO2020107847A1 true WO2020107847A1 (en) 2020-06-04

Family

ID=65698053

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2019/089500 WO2020107847A1 (en) 2018-11-28 2019-05-31 Bone point-based fall detection method and fall detection device therefor

Country Status (2)

Country Link
CN (1) CN109492612A (en)
WO (1) WO2020107847A1 (en)

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111860312A (en) * 2020-07-20 2020-10-30 上海汽车集团股份有限公司 Driving environment adjusting method and device
CN112364695A (en) * 2020-10-13 2021-02-12 杭州城市大数据运营有限公司 Behavior prediction method and device, computer equipment and storage medium
CN112541576A (en) * 2020-12-14 2021-03-23 四川翼飞视科技有限公司 Biological living body recognition neural network of RGB monocular image and construction method thereof
CN114882596A (en) * 2022-07-08 2022-08-09 深圳市信润富联数字科技有限公司 Behavior early warning method and device, electronic equipment and storage medium
CN115546491A (en) * 2022-11-28 2022-12-30 中南财经政法大学 Fall alarm method, system, electronic equipment and storage medium

Families Citing this family (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109492612A (en) * 2018-11-28 2019-03-19 平安科技(深圳)有限公司 Fall detection method and its falling detection device based on skeleton point
CN110378245B (en) * 2019-06-26 2023-07-21 平安科技(深圳)有限公司 Football match behavior recognition method and device based on deep learning and terminal equipment
CN110276332B (en) * 2019-06-28 2021-12-24 北京奇艺世纪科技有限公司 Video feature processing method and device
CN110532874B (en) * 2019-07-23 2022-11-11 深圳大学 Object attribute recognition model generation method, storage medium and electronic device
CN110633736A (en) * 2019-08-27 2019-12-31 电子科技大学 Human body falling detection method based on multi-source heterogeneous data fusion
SE1951443A1 (en) * 2019-12-12 2021-06-13 Assa Abloy Ab Improving machine learning for monitoring a person
CN111209848B (en) * 2020-01-03 2023-07-21 北京工业大学 Real-time falling detection method based on deep learning
CN113792595A (en) * 2021-08-10 2021-12-14 北京爱笔科技有限公司 Target behavior detection method and device, computer equipment and storage medium
CN113712538A (en) * 2021-08-30 2021-11-30 平安科技(深圳)有限公司 Fall detection method, device, equipment and storage medium based on WIFI signal
CN115661943B (en) * 2022-12-22 2023-03-31 电子科技大学 Fall detection method based on lightweight attitude assessment network
CN117238026B (en) * 2023-07-10 2024-03-08 中国矿业大学 Gesture reconstruction interactive behavior understanding method based on skeleton and image features

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107220604A (en) * 2017-05-18 2017-09-29 清华大学深圳研究生院 A kind of fall detection method based on video
US20170316578A1 (en) * 2016-04-29 2017-11-02 Ecole Polytechnique Federale De Lausanne (Epfl) Method, System and Device for Direct Prediction of 3D Body Poses from Motion Compensated Sequence
CN107392131A (en) * 2017-07-14 2017-11-24 天津大学 A kind of action identification method based on skeleton nodal distance
CN107784654A (en) * 2016-08-26 2018-03-09 杭州海康威视数字技术股份有限公司 Image partition method, device and full convolutional network system
CN108647776A (en) * 2018-05-08 2018-10-12 济南浪潮高新科技投资发展有限公司 A kind of convolutional neural networks convolution expansion process circuit and method
CN109492612A (en) * 2018-11-28 2019-03-19 平安科技(深圳)有限公司 Fall detection method and its falling detection device based on skeleton point

Family Cites Families (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105590099B (en) * 2015-12-22 2019-02-01 中国石油大学(华东) A kind of more people's Activity recognition methods based on improvement convolutional neural networks
CN108294759A (en) * 2017-01-13 2018-07-20 天津工业大学 A kind of Driver Fatigue Detection based on CNN Eye state recognitions
CN108280455B (en) * 2018-01-19 2021-04-02 北京市商汤科技开发有限公司 Human body key point detection method and apparatus, electronic device, program, and medium
CN108717569B (en) * 2018-05-16 2022-03-22 中国人民解放军陆军工程大学 Expansion full-convolution neural network device and construction method thereof
CN108776775B (en) * 2018-05-24 2020-10-27 常州大学 Old people indoor falling detection method based on weight fusion depth and skeletal features

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20170316578A1 (en) * 2016-04-29 2017-11-02 Ecole Polytechnique Federale De Lausanne (Epfl) Method, System and Device for Direct Prediction of 3D Body Poses from Motion Compensated Sequence
CN107784654A (en) * 2016-08-26 2018-03-09 杭州海康威视数字技术股份有限公司 Image partition method, device and full convolutional network system
CN107220604A (en) * 2017-05-18 2017-09-29 清华大学深圳研究生院 A kind of fall detection method based on video
CN107392131A (en) * 2017-07-14 2017-11-24 天津大学 A kind of action identification method based on skeleton nodal distance
CN108647776A (en) * 2018-05-08 2018-10-12 济南浪潮高新科技投资发展有限公司 A kind of convolutional neural networks convolution expansion process circuit and method
CN109492612A (en) * 2018-11-28 2019-03-19 平安科技(深圳)有限公司 Fall detection method and its falling detection device based on skeleton point

Cited By (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111860312A (en) * 2020-07-20 2020-10-30 上海汽车集团股份有限公司 Driving environment adjusting method and device
CN112364695A (en) * 2020-10-13 2021-02-12 杭州城市大数据运营有限公司 Behavior prediction method and device, computer equipment and storage medium
CN112541576A (en) * 2020-12-14 2021-03-23 四川翼飞视科技有限公司 Biological living body recognition neural network of RGB monocular image and construction method thereof
CN112541576B (en) * 2020-12-14 2024-02-20 四川翼飞视科技有限公司 Biological living body identification neural network construction method of RGB monocular image
CN114882596A (en) * 2022-07-08 2022-08-09 深圳市信润富联数字科技有限公司 Behavior early warning method and device, electronic equipment and storage medium
CN115546491A (en) * 2022-11-28 2022-12-30 中南财经政法大学 Fall alarm method, system, electronic equipment and storage medium
CN115546491B (en) * 2022-11-28 2023-03-10 中南财经政法大学 Fall alarm method, system, electronic equipment and storage medium

Also Published As

Publication number Publication date
CN109492612A (en) 2019-03-19

Similar Documents

Publication Publication Date Title
WO2020107847A1 (en) Bone point-based fall detection method and fall detection device therefor
Yang et al. Uncertainty-guided transformer reasoning for camouflaged object detection
CN109558832B (en) Human body posture detection method, device, equipment and storage medium
Wang et al. A deep network solution for attention and aesthetics aware photo cropping
WO2021129064A1 (en) Posture acquisition method and device, and key point coordinate positioning model training method and device
CN108205655B (en) Key point prediction method and device, electronic equipment and storage medium
CN110728209B (en) Gesture recognition method and device, electronic equipment and storage medium
WO2021227726A1 (en) Methods and apparatuses for training face detection and image detection neural networks, and device
WO2021017303A1 (en) Person re-identification method and apparatus, computer device and storage medium
WO2021218238A1 (en) Image processing method and image processing apparatus
WO2021051547A1 (en) Violent behavior detection method and system
CN111368672A (en) Construction method and device for genetic disease facial recognition model
WO2021238548A1 (en) Region recognition method, apparatus and device, and readable storage medium
CN112766186B (en) Real-time face detection and head posture estimation method based on multitask learning
CN110738650B (en) Infectious disease infection identification method, terminal device and storage medium
CN111754453A (en) Pulmonary tuberculosis detection method and system based on chest radiography image and storage medium
WO2022111387A1 (en) Data processing method and related apparatus
CN111738074B (en) Pedestrian attribute identification method, system and device based on weak supervision learning
CN116897012A (en) Method and device for identifying physique of traditional Chinese medicine, electronic equipment, storage medium and program
CN111340213B (en) Neural network training method, electronic device, and storage medium
Zhou et al. A study on attention-based LSTM for abnormal behavior recognition with variable pooling
Gupta et al. Advanced security system in video surveillance for COVID-19
CN114724251A (en) Old people behavior identification method based on skeleton sequence under infrared video
CN112381118B (en) College dance examination evaluation method and device
CN117037244A (en) Face security detection method, device, computer equipment and storage medium

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 19889552

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 19889552

Country of ref document: EP

Kind code of ref document: A1