CN114998229A - Non-contact sleep monitoring method based on deep learning and multi-parameter fusion - Google Patents

Non-contact sleep monitoring method based on deep learning and multi-parameter fusion Download PDF

Info

Publication number
CN114998229A
CN114998229A CN202210561402.3A CN202210561402A CN114998229A CN 114998229 A CN114998229 A CN 114998229A CN 202210561402 A CN202210561402 A CN 202210561402A CN 114998229 A CN114998229 A CN 114998229A
Authority
CN
China
Prior art keywords
sleep
tester
eye
video image
sleeping
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202210561402.3A
Other languages
Chinese (zh)
Other versions
CN114998229B (en
Inventor
张静
晏博赟
贺涛
杜晓辉
王祥舟
孙海鑫
刘娟秀
刘霖
刘永
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
University of Electronic Science and Technology of China
Original Assignee
University of Electronic Science and Technology of China
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by University of Electronic Science and Technology of China filed Critical University of Electronic Science and Technology of China
Priority to CN202210561402.3A priority Critical patent/CN114998229B/en
Publication of CN114998229A publication Critical patent/CN114998229A/en
Application granted granted Critical
Publication of CN114998229B publication Critical patent/CN114998229B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/0002Inspection of images, e.g. flaw detection
    • G06T7/0012Biomedical image inspection
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/02Detecting, measuring or recording pulse, heart rate, blood pressure or blood flow; Combined pulse/heart-rate/blood pressure determination; Evaluating a cardiovascular condition not otherwise provided for, e.g. using combinations of techniques provided for in this group with electrocardiography or electroauscultation; Heart catheters for measuring blood pressure
    • A61B5/0205Simultaneously evaluating both cardiovascular conditions and different types of body conditions, e.g. heart and respiratory condition
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/48Other medical applications
    • A61B5/4806Sleep evaluation
    • A61B5/4809Sleep detection, i.e. determining whether a subject is asleep or not
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/48Other medical applications
    • A61B5/4806Sleep evaluation
    • A61B5/4812Detecting sleep stages or cycles
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/48Other medical applications
    • A61B5/4806Sleep evaluation
    • A61B5/4815Sleep quality
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/72Signal processing specially adapted for physiological signals or for diagnostic purposes
    • A61B5/7235Details of waveform analysis
    • A61B5/7253Details of waveform analysis characterised by using transforms
    • A61B5/7257Details of waveform analysis characterised by using transforms using Fourier transforms
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/72Signal processing specially adapted for physiological signals or for diagnostic purposes
    • A61B5/7235Details of waveform analysis
    • A61B5/7264Classification of physiological signals or data, e.g. using neural networks, statistical classifiers, expert systems or fuzzy systems
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/02Detecting, measuring or recording pulse, heart rate, blood pressure or blood flow; Combined pulse/heart-rate/blood pressure determination; Evaluating a cardiovascular condition not otherwise provided for, e.g. using combinations of techniques provided for in this group with electrocardiography or electroauscultation; Heart catheters for measuring blood pressure
    • A61B5/024Detecting, measuring or recording pulse rate or heart rate
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/16Devices for psychotechnics; Testing reaction times ; Devices for evaluating the psychological state
    • A61B5/163Devices for psychotechnics; Testing reaction times ; Devices for evaluating the psychological state by tracking eye movement, gaze, or pupil change
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10016Video; Image sequence
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20081Training; Learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20084Artificial neural networks [ANN]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30196Human being; Person

Landscapes

  • Health & Medical Sciences (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Health & Medical Sciences (AREA)
  • Medical Informatics (AREA)
  • Molecular Biology (AREA)
  • Biophysics (AREA)
  • Biomedical Technology (AREA)
  • Animal Behavior & Ethology (AREA)
  • Pathology (AREA)
  • Heart & Thoracic Surgery (AREA)
  • Surgery (AREA)
  • Public Health (AREA)
  • Veterinary Medicine (AREA)
  • Physiology (AREA)
  • Artificial Intelligence (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Theoretical Computer Science (AREA)
  • Mathematical Physics (AREA)
  • Psychiatry (AREA)
  • Evolutionary Computation (AREA)
  • Signal Processing (AREA)
  • Cardiology (AREA)
  • General Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • Anesthesiology (AREA)
  • Nuclear Medicine, Radiotherapy & Molecular Imaging (AREA)
  • Software Systems (AREA)
  • General Engineering & Computer Science (AREA)
  • Computing Systems (AREA)
  • Radiology & Medical Imaging (AREA)
  • Computational Linguistics (AREA)
  • Quality & Reliability (AREA)
  • Pulmonology (AREA)
  • Fuzzy Systems (AREA)
  • Measuring And Recording Apparatus For Diagnosis (AREA)

Abstract

The invention discloses a non-contact sleep monitoring system based on deep learning and multi-parameter fusion, and belongs to the field of image processing and deep learning. The system firstly segments an acquired sleep video image, then builds a deep convolutional neural network for extracting and amplifying physiological signals, amplifies a heart rate signal of a forehead area and eye movement frequency of an eye area by setting different amplification factors of the network to obtain a forehead area video image for amplifying the heart rate signal and an eye area video image after amplifying the eye movement frequency, extracts corresponding frequency spectrum by utilizing fast Fourier transform, and finds out frequency corresponding to a frequency spectrum peak value as the monitored heart rate signal and eye movement frequency. And for the three-position body video images, building a deep learning-based sleeping posture monitoring neural network structure, inputting the sleeping posture characteristics extracted from the network into a full connection layer for six classification, wherein the classification results correspond to six sleeping postures of supine, prone, left-side straight lying, left-side lying, right-side straight lying and right-side lying, and counting the turn-over times by switching different sleeping postures of the six sleeping postures. And finally, comprehensively evaluating the sleep quality by integrating the monitored physiological signals. The invention has the characteristics of high comfort, multi-parameter fusion and high automation, and realizes the monitoring of physiological parameters such as heart rate, respiratory rate, eye movement rate, sleeping posture, turnover frequency and the like in a non-contact manner.

Description

Non-contact sleep monitoring method based on deep learning and multi-parameter fusion
Technical Field
The invention belongs to the field of image processing and the field of deep learning, and particularly relates to a non-contact sleep monitoring system for realizing multi-parameter fusion by combining a video image processing technology and a deep convolutional network.
Technical Field
In the sleeping process, a series of functions of the brain, muscles, eyes, the heart, breath and the like of a human body can change, and the judgment on the sleeping quality of the human body can be promoted by monitoring the changes. Sleep disorders generally refer to abnormalities in the quality or quantity of sleep or the occurrence of certain clinical symptoms during sleep, such as decreased or excessive sleep, sleep disordered breathing, disorders of rapid eye movement sleep behavior, and the like. Medical science proves that people with sleep disorder for a long time can induce various diseases, so that the timely diagnosis and treatment of the sleep disorder have important significance on human health.
Polysomnography is a golden method called sleep disorder diagnosis and treatment, and mainly monitors physiological signals of channels such as electroencephalograms, electrocardiograms, electrooculogram, oronasal airflow flux, blood oxygen saturation and the like of patients, and diagnoses according to the collected signals. The utility model discloses a measurement of sleep quality, including the person who is surveyed, need install multiple sensor on the person's body when utilizing polysomnography to monitor, bring very big discomfort for the person who is surveyed, in addition, even it monitors many parameters, the doctor still can be according to the past medical history of the person who is surveyed and the subjective impression during monitoring as one of the foundation of aassessment, and the interpretation result has stronger subjectivity. With the development of deep learning, many miniaturized sleep monitoring devices, such as intelligent pillows, mattresses, bracelets, etc., have appeared. The intelligent pillow and the mattress monitor the pressure change in the sleeping process through the pressure sensor, and the turn-over times of the tested person are counted. The intelligent bracelet can monitor the heart rate during sleep when worn. Although the existing sleep monitoring equipment reduces the uncomfortable feeling of wearing a sensor in the monitoring process, the existing sleep monitoring equipment can only monitor a single physiological parameter, so that the evaluation result of the sleep quality is inaccurate and incomplete.
Aiming at the problems in sleep monitoring, a non-contact sleep monitoring system based on deep learning and multi-parameter fusion is designed. For the heart rate and the eye movement frequency, a micro physiological signal is extracted from a sleep video through a deep convolutional neural network and amplified, meanwhile, the generation of artifacts is inhibited, the effective physiological signal amplification is realized, and finally, the amplified physiological signal is subjected to spectrum analysis. And for the sleeping postures and the turnover times, automatically extracting the sleeping posture characteristics of the video frames acquired by the three-position camera by using a convolutional neural network, and carrying out six classifications on the sleeping posture characteristics, wherein the turnover times in the sleeping process can be monitored according to the switching among different sleeping postures.
Disclosure of Invention
The invention designs a non-contact sleep monitoring system based on deep learning and multi-parameter fusion aiming at the problems that the contact monitoring of a polysomnogram brings discomfort and subjectivity of manual interpretation to a tested person and the singleness of monitoring physiological parameters of other sleep monitoring equipment and the like so as to realize the non-contact monitoring of multiple physiological parameters such as heart rate, eye movement frequency, sleeping posture, turning-over times and the like.
The technical scheme of the invention is a non-contact sleep monitoring method based on deep learning and multi-parameter fusion, which comprises the following steps:
step 1: building a sleep monitoring platform, and respectively placing three cameras at the upper, left and right positions of a tester body to obtain video images of the tester in the sleeping process;
step 2: performing image segmentation on the video image obtained by the camera positioned above the body of the tester in the step 1 to obtain a forehead area video image and an eye area video image of the tester;
and step 3: constructing a deep convolutional neural network for extracting and amplifying physiological signals, and extracting and amplifying tiny physiological signals in a video by using the deep convolutional neural network;
and 4, step 4: respectively inputting the forehead area video image and the eye area video image obtained in the step 2 into the deep convolutional neural network built in the step 3, respectively extracting and amplifying a heart rate signal of the forehead area and an eye movement signal of the eye area, and outputting a forehead area video image after amplifying the heart rate signal and an eye area video image after amplifying the eye movement frequency;
and 5: carrying out RGB three-channel separation on each frame of image in the forehead area video image obtained in the step 4 after the heart rate signal is amplified, averaging pixel points in R, G, B three channels, and then carrying out time sequence stacking to obtain a pulse wave signal;
step 6: performing fast Fourier transform on the pulse wave signals obtained in the step 5 to obtain a time series frequency spectrum of the human pulse waves;
and 7: performing spectrum analysis on the time series spectrum obtained in the step 6, and selecting the frequency corresponding to the peak value of the spectrum as a heart rate monitoring result;
and 8: stacking the eye region video images amplified by the eye movement frequency obtained in the step (4) according to a time sequence, and performing fast Fourier transform to obtain an eye region video image frequency spectrum;
and step 9: extracting a frequency corresponding to a frequency spectrum peak value of the eye region video image frequency spectrum obtained in the step 8 as a monitoring result of the eye movement frequency;
step 10: building a sleeping posture monitoring neural network structure based on deep learning, executing a step 11 if the sleeping posture monitoring neural network is not trained, and executing a step 13 if the sleeping posture monitoring neural network is trained;
step 11: collecting more than 1000 images of a tester in sleep in advance through three cameras arranged on the human body, the left camera and the right camera, carrying out feature marking on the obtained images, manually marking the sleeping posture state of the tester, and correspondingly sleeping in six sleeping postures, namely supine, prone, left-side straight lying, left-side crouching, right-side straight lying and right-side crouching;
step 12: sending the image data marked in the step 11 into a neural network for training, dividing a training set and a verification set according to the proportion of 8:2 to train the network until the accuracy of the verification set reaches more than 95%, and finishing the network training;
step 13: inputting the three-position body video images of the testee in the three cameras in the step 1 into a trained neural network, and carrying out six classification outputs on the images, wherein the classification results correspond to six sleeping positions of supine, prone, left-side lying straight, left-side lying crouch, right-side lying straight and right-side lying crouch;
step 14: if the tester switches any two sleeping postures of the six sleeping postures in the step 13, the tester records the switching as one-time turning, but the mutual conversion of the left-side lying and the mutual conversion of the right-side lying and the right-side lying do not take the statistics of the turning times;
step 15: and comprehensively evaluating the sleep quality result of the tester by combining the heart rate parameters obtained in the step 7, the eye movement frequency obtained in the step 9, the sleeping posture state obtained in the step 13 and the turn-over times obtained in the step 14.
Wherein, the step 2 specifically comprises the following steps:
step 2.1: calling a dlib library in python for a video image acquired by a camera positioned above the body of a tester, segmenting an area where a human face is positioned, and extracting a video image of the face of the tester;
step 2.2: performing face key point detection on the face video image obtained in the step 2.1, calling a dlib library in python to perform face key point detection, and obtaining the positions of 68 key points of the face of the tester;
step 2.3: carrying out image segmentation on the video image obtained in the step 2.1 again through the positions of the key points of the human face obtained in the step 2.2; segmenting a forehead area video image of a rectangular area by identifying key points of the centers of left and right eyebrows of the human face and the upper boundary of the human face identified by a dlib library;
step 2.4: finding out key points representing a left eye corner, a right eye corner, the uppermost part of an eye socket and the lowermost part of the eye socket by identifying key points of left and right eyes of the human face, and segmenting an eye area video image of a rectangular area through the four key points;
wherein, the step 3 specifically comprises the following steps:
step 3.1: the encoder structure of the deep convolutional neural network for extracting and amplifying the physiological signal is constructed by the following steps: after each convolution layer of the 2 convolution layers, a Relu activation function is used, then 3 residual error networks are used, physiological signals in the sleep video are extracted through one convolution layer, the step length of the physiological signals is set to be 2, and finally two residual error structures are connected for output;
step 3.2: building a modulation amplification structure of a deep convolutional neural network, performing convolutional operation on the difference of physiological signals of two frames of sleep images through a convolutional layer, enabling an activation function to be a Relu function, multiplying the activation function by an amplification factor alpha, and performing nonlinear change on the amplified characteristics by using the convolutional layer and a residual error structure to obtain amplified physiological signal difference characteristics;
step 3.3: and building a decoder structure of a deep convolutional neural network, superposing the amplified physiological signal differential characteristics to the initial sleep image, and decoding and outputting the amplified video through upsampling and two convolutional layers to realize the amplification of the physiological signal in the sleep video.
Wherein, the step 4 specifically comprises the following steps:
step 4.1: inputting the forehead area video image obtained in the step 2 into the deep convolutional neural network built in the step 3, setting an amplification factor alpha to be 15, extracting and amplifying the heart rate of the forehead area, and outputting the forehead area video image after amplifying the heart rate signal;
step 4.2: inputting the eye region video image obtained in the step 2 into the deep convolutional neural network constructed in the step 3, setting an amplification factor alpha to be 30, extracting and amplifying the eye movement frequency of the eye region, and outputting the eye region video image after amplifying the eye movement frequency;
wherein, the step 10 specifically comprises:
step 10.1: building a neural network structure, which comprises 4 convolutional layers, 3 maximum pooling layers, 1 full-connection layer and 1 classifier;
step 10.2: preventing the calculated amount from being too large, extracting the current-time key frames of the video every 1s, respectively extracting 1 frame of images from the cameras positioned on the left and right of the body of the tester, and forming three-channel image data to be input into the neural network in the step 10.1;
step 10.3: extracting image features of the three-channel image data in the step 10.2 through a convolution layer with convolution kernels of 10 x 10, a maximum pooling layer of 2 x 2, a convolution layer with convolution kernels of 10 x 10, a maximum pooling layer of 2 x 2 and a convolution layer with convolution kernels of 10 x 10 respectively;
step 10.4: inputting the extracted features into a full-connection layer for six classifications, wherein classification results correspond to six sleeping postures: supine, prone, left supine, left crouch, right supine, and right crouch.
Wherein, the step 15 specifically comprises:
step 15.1: the heart rate of a person in normal sleep is 60-100 times per minute, the heart rate of the person in deep sleep can be reduced to 50 times per minute, the heart rate of the tester in sleep is monitored, when the heart rate is obviously reduced, the tester is considered to enter a deep sleep period, when the heart rate is gradually increased, the tester is considered to exit the deep sleep period, finally, the proportion of the deep sleep period in sleep of the tester is counted, and the higher the proportion of the deep sleep period is, the higher the sleep quality is;
step 15.2: when the eye movement is in a rapid eye movement period in a sleep stage, the eyeballs can rotate rapidly, the eye movement of the tester is monitored, if the eye movement frequency is obviously increased, the heart rate is increased in the step 14.1, the tester can be considered to enter the rapid eye movement period, the duration time of the rapid eye movement period is counted, and if the rapid eye movement sleep is suddenly interrupted, the signal of the attack of diseases such as angina pectoris and asthma is often generated;
step 15.3: the sleeping posture is considered to be a better sleeping posture during sleeping, but the sleeping posture is not suitable for people suffering from respiratory diseases or people who frequently snore, and the sleeping posture of lying on the side is adopted; monitoring the sleeping posture of the tester, and if the tester has respiratory diseases or snores, carrying out sleeping posture adjustment suggestion when the tester adopts a non-side sleeping posture;
step 15.4: the turnover frequency of the tester is monitored, and if the turnover frequency is too high, the tester is prompted to possibly lack calcium ions or have high mental stress and poor sleep quality.
The invention relates to a non-contact sleep monitoring system based on deep learning and multi-parameter fusion, which comprises the steps of firstly segmenting an acquired video image, segmenting a forehead area video image, an eye area video image and a three-direction body video image, then building a deep convolutional neural network for extracting and amplifying physiological signals, amplifying a heart rate signal of a forehead area and an eye movement frequency of an eye area by setting different amplification factors of the network to obtain a forehead area video image for amplifying the heart rate signal and an eye area video image after amplifying the eye movement frequency, extracting corresponding frequency spectrums by utilizing fast Fourier transform, and finding out frequencies corresponding to frequency spectrum peaks as the monitored heart rate signal and eye movement frequency. For three-direction body video images, a sleeping posture monitoring neural network structure based on deep learning is built, sleeping posture features extracted from the network are input into a full connection layer for six classification, classification results correspond to six sleeping postures of supine, prone, left-side straight lying, left-side crouching, right-side straight lying and right-side crouching, and d number of turn-over times is counted through switching of different sleeping postures of the six sleeping postures. And finally, comprehensively evaluating the sleep quality by integrating the monitored physiological signals. The invention provides a sleep monitoring system with high comfort, multi-parameter fusion and high automation for testers, realizes the monitoring of physiological parameters such as heart rate, respiratory rate, eye movement frequency, sleeping posture, turning-over times and the like in a non-contact manner, improves the monitoring reliability, and has key effects on clinical diagnosis of sleep quality and clinical treatment and early intervention on sleep disorder patients or potential patients.
Drawings
FIG. 1 is a deep convolutional neural network diagram of physiological signal extraction and amplification
FIG. 2 is a flow chart of heart rate monitoring
FIG. 3 is a flow chart of eye movement monitoring
FIG. 4 is a flow chart for monitoring sleeping posture and turning-over times
FIG. 5 is a diagram of a sleeping posture monitoring neural network structure
Detailed Description
The following describes a non-contact sleep monitoring system based on deep learning and multi-parameter fusion in detail with reference to the accompanying drawings:
step 1: and (5) building a sleep monitoring platform. Three cameras are respectively arranged at the upper, left and right positions of the body of a tester to acquire video images of the tester in the sleeping process;
step 2: performing image segmentation on the video image obtained by the camera positioned above the body of the tester in the step 1 to obtain a forehead area video image and an eye area video image of the tester;
step 2.1: calling a dlib library in python for a video image acquired by a camera positioned above the body of a tester, segmenting an area where a human face is positioned, and extracting a video image of the face of the tester;
step 2.2: performing face key point detection on the face video image obtained in the step 2.1, calling a dlib library in python to perform face key point detection, and obtaining positions of 68 key points of the face of the tester;
step 2.3: and (4) carrying out image segmentation on the video image obtained in the step (2.1) again through the positions of the key points of the human face obtained in the step (2.2). Segmenting a forehead area video image of a rectangular area by identifying key points of the centers of left and right eyebrows of the human face and the upper boundary of the human face identified by a dlib library;
step 2.4: and finding out key points representing the left eye corner, the right eye corner, the uppermost part of the eye socket and the lowermost part of the eye socket by identifying the key points of the left eye and the right eye of the human face, and segmenting an eye area video image of a rectangular area by the four key points.
And step 3: constructing a deep convolutional neural network for extracting and amplifying physiological signals, and extracting and amplifying tiny physiological signals in a video by using the deep convolutional neural network;
step 3.1: constructing an encoder structure of a deep convolutional neural network for extracting and amplifying physiological signals, wherein the encoder structure comprises 2 convolutional layers and 3 residual error networks, a Relu activation function is used after each convolutional layer, then the physiological signals in the sleep video are extracted through one convolutional layer, the step length of the physiological signals is set to be 2, and finally the two residual error structures are connected for output;
step 3.2: building a modulation amplification structure of a deep convolutional neural network, performing convolutional operation on the difference of physiological signals of two frames of sleep images through a convolutional layer, enabling an activation function to be a Relu function, multiplying the activation function by an amplification factor alpha, and performing nonlinear change on the amplified characteristics by using the convolutional layer and a residual error structure to obtain amplified physiological signal difference characteristics;
step 3.3: and building a decoder structure of a deep convolutional neural network, superposing the amplified physiological signal differential characteristics to an initial sleep image, and decoding and outputting the amplified video through upsampling and two convolutional layers to realize amplification of the physiological signals in the sleep video.
And 4, step 4: respectively inputting the forehead area video image and the eye area video image obtained in the step 2 into the deep convolutional neural network built in the step 3, respectively extracting and amplifying a heart rate signal of the forehead area and an eye movement signal of the eye area, and outputting a forehead area video image after amplifying the heart rate signal and an eye area video image after amplifying the eye movement frequency;
step 4.1: inputting the forehead area video image obtained in the step 2 into the deep convolutional neural network built in the step 3, setting an amplification factor alpha to be 15, extracting and amplifying the heart rate of the forehead area, and outputting the forehead area video image after amplifying the heart rate signal;
and 4.2: inputting the eye region video image obtained in the step 2 into the deep convolutional neural network constructed in the step 3, setting an amplification factor alpha to be 30, extracting and amplifying the eye movement frequency of the eye region, and outputting the eye region video image after amplifying the eye movement frequency;
and 5: carrying out RGB three-channel separation on each frame of image in the forehead area video image obtained in the step 4 after the heart rate signal is amplified, averaging pixel points in R, G, B three channels, and then carrying out time sequence stacking to obtain a pulse wave signal;
step 6: performing fast Fourier transform on the pulse wave signals obtained in the step 5 to obtain a time series frequency spectrum of the human pulse wave;
and 7: performing spectrum analysis on the time series spectrum obtained in the step 6, and selecting the frequency corresponding to the peak value of the spectrum as a heart rate monitoring result;
and step 8: stacking the eye region video images amplified by the eye movement frequency obtained in the step (4) according to a time sequence, and performing fast Fourier transform to obtain an eye region video image frequency spectrum;
and step 9: extracting the frequency corresponding to the frequency spectrum peak value of the eye region video image frequency spectrum obtained in the step 8 as the monitoring result of the eye movement frequency;
step 10: and (3) building a sleeping posture monitoring neural network structure based on deep learning, and executing the step 11 if the sleeping posture monitoring neural network is not trained. If the sleeping posture monitoring neural network is trained, executing step 13;
step 10.1: building a neural network structure, which comprises 4 convolutional layers, 3 maximum pooling layers, 1 full-connection layer and 1 classifier;
step 10.2: preventing the calculated amount from being too large, extracting the current-time key frames of the video every 1s, respectively extracting 1 frame of images from the cameras positioned on the left and right of the body of the tester, and forming three-channel image data to be input into the neural network in the step 10.1;
step 10.3: extracting image features of the three-channel image data in the step 10.2 through a convolution layer with convolution kernels of 10 x 10, a maximum pooling layer of 2 x 2, a convolution layer with convolution kernels of 10 x 10, a maximum pooling layer of 2 x 2 and a convolution layer with convolution kernels of 10 x 10 respectively;
step 10.4: inputting the extracted features into a full-connection layer for six classifications, wherein classification results correspond to six sleeping postures: supine, prone, left supine, left crouch, right supine, and right crouch.
Step 11: collecting more than 1000 images of a tester in sleep in advance through three cameras arranged on the human body, the left camera and the right camera, carrying out feature marking on the obtained images, manually marking the sleeping posture state of the tester, and correspondingly sleeping in six sleeping postures, namely supine, prone, left-side straight lying, left-side crouching, right-side straight lying and right-side crouching;
step 12: sending the image data marked in the step 11 into a neural network for training, dividing a training set and a verification set according to the proportion of 8:2 to train the network until the accuracy of the verification set reaches more than 95%, and finishing the network training;
step 13: inputting the three-direction body video images of the testee in the three cameras in the step 1 into a trained neural network, and carrying out six classification outputs on the images, wherein the classification results correspond to six sleeping postures of supine, prone, left-side lying straight, left-side lying crouch, right-side lying straight and right-side lying crouch;
step 14: if the tester switches any two of the six sleeping postures in the step 13, the tester records the switching as one-time turning, but the mutual conversion between the left-side lying posture and the mutual conversion between the right-side lying posture and the right-side lying posture are not counted into the statistics of the turning times;
step 15: and comprehensively evaluating the sleep quality result of the tester by combining the heart rate parameters obtained in the step 7, the eye movement frequency obtained in the step 9, the sleeping posture state obtained in the step 13 and the turn-over times obtained in the step 14.
Step 15.1: the heart rate of a person in normal sleep is 60-100 times per minute, and the heart rate can be reduced to 50 times per minute in deep sleep. The heart rate of the tester during sleeping is monitored, when the heart rate is obviously reduced, the tester is considered to enter the deep sleep period, and when the heart rate is gradually increased, the tester is considered to exit the deep sleep period. Finally, counting the proportion of deep sleep periods in the sleep of the testers, wherein the larger the proportion of the deep sleep periods is, the higher the sleep quality is;
step 15.2: in the rapid eye movement stage of the sleep stage, the eyeball rotates rapidly. The invention monitors the eye movement of the tester, if the eye movement frequency is obviously increased, the tester can be considered to enter the rapid eye movement period by combining the heart rate increase in the step 14.1, the duration time of the rapid eye movement period is counted, and if the rapid eye movement sleep is suddenly interrupted, the rapid eye movement sleep is often a signal for the attack of diseases such as angina, asthma and the like;
step 15.3: lying is considered to be a better sleeping position during sleeping, but is not suitable for people suffering from respiratory diseases or who frequently snore, and a side sleeping position is adopted. The invention monitors the sleeping posture of a tester, and if the tester has respiratory tract diseases or snores, the tester carries out sleeping posture adjustment suggestion when the tester adopts a non-side sleeping posture;
step 15.4: the invention monitors the turnover frequency of the tester, and if the turnover frequency is too high, the tester is prompted to possibly lack calcium ions or have high mental stress and poor sleep quality.

Claims (6)

1. A non-contact sleep monitoring system based on deep learning and multi-parameter fusion comprises the following steps:
step 1: and (5) building a sleep monitoring platform. Three cameras are respectively arranged at the upper, left and right positions of the body of a tester to acquire video images of the tester in the sleeping process;
step 2: performing image segmentation on the video image obtained by the camera positioned above the body of the tester in the step 1 to obtain a forehead area video image and an eye area video image of the tester;
and step 3: constructing a deep convolutional neural network for extracting and amplifying physiological signals, and extracting and amplifying tiny physiological signals in a video by using the deep convolutional neural network;
and 4, step 4: respectively inputting the forehead area video image and the eye area video image obtained in the step 2 into the deep convolutional neural network built in the step 3, respectively extracting and amplifying a heart rate signal of the forehead area and an eye movement signal of the eye area, and outputting a forehead area video image after amplifying the heart rate signal and an eye area video image after amplifying the eye movement frequency;
and 5: carrying out RGB three-channel separation on each frame image in the forehead area video image obtained in the step 4 after the heart rate signal is amplified, averaging pixel points in R, G, B three channels, and then carrying out time sequence stacking to obtain a pulse wave signal;
step 6: performing fast Fourier transform on the pulse wave signals obtained in the step 5 to obtain a time series frequency spectrum of the human pulse waves;
and 7: carrying out spectrum analysis on the time series frequency spectrum obtained in the step 6, and selecting the frequency corresponding to the peak value of the frequency spectrum as a heart rate monitoring result;
and 8: stacking the eye region video images amplified by the eye movement frequency obtained in the step (4) according to a time sequence, and performing fast Fourier transform to obtain an eye region video image frequency spectrum;
and step 9: extracting the frequency corresponding to the frequency spectrum peak value of the eye region video image frequency spectrum obtained in the step 8 as the monitoring result of the eye movement frequency;
step 10: and (3) building a sleeping posture monitoring neural network structure based on deep learning, and executing the step 11 if the sleeping posture monitoring neural network is not trained. If the sleeping posture monitoring neural network is trained, executing the step 13;
step 11: collecting more than 1000 images of a tester in sleep in advance through three cameras arranged on the human body, the left camera and the right camera, carrying out feature marking on the obtained images, manually marking the sleeping posture state of the tester, and correspondingly sleeping in six sleeping postures, namely supine, prone, left-side straight lying, left-side crouching, right-side straight lying and right-side crouching;
step 12: sending the image data marked in the step 11 into a neural network for training, dividing a training set and a verification set according to the proportion of 8:2 to train the network until the accuracy of the verification set reaches more than 95%, and finishing the network training;
step 13: inputting the three-position body video images of the testee in the three cameras in the step 1 into a trained neural network, and carrying out six classification outputs on the images, wherein the classification results correspond to six sleeping positions of supine, prone, left-side lying straight, left-side lying crouch, right-side lying straight and right-side lying crouch;
step 14: if the tester switches any two of the six sleeping postures in the step 13, the tester records the switching as one-time turning, but the mutual conversion between the left-side lying posture and the mutual conversion between the right-side lying posture and the right-side lying posture are not counted into the statistics of the turning times;
step 15: and (3) comprehensively evaluating the sleep quality result of the tester by combining the heart rate parameters obtained in the step (7), the eye movement frequency obtained in the step (9), the sleeping posture state obtained in the step (13) and the turning times obtained in the step (14).
2. The system for monitoring sleep based on deep learning and multi-parameter fusion as claimed in claim 1, wherein the step 2 is specifically:
step 2.1: calling a dlib library in python for a video image acquired by a camera positioned above the body of the tester, segmenting a region where a face is positioned, and extracting a video image of the face of the tester;
step 2.2: performing face key point detection on the face video image obtained in the step 2.1, calling a dlib library in python to perform face key point detection, and obtaining positions of 68 key points of the face of the tester;
step 2.3: and (4) carrying out image segmentation on the video image obtained in the step (2.1) again through the positions of the key points of the human face obtained in the step (2.2). Segmenting a forehead area video image of a rectangular area by identifying key points of the centers of left and right eyebrows of the human face and the upper boundary of the human face identified by a dlib library;
step 2.4: and finding out key points representing the left eye corner, the right eye corner, the uppermost part of the eye socket and the lowermost part of the eye socket by identifying the key points of the left eye and the right eye of the human face, and segmenting an eye area video image of a rectangular area by the four key points.
3. The system for monitoring sleep based on deep learning and multi-parameter fusion as claimed in claim 1, wherein the step 3 is specifically:
step 3.1: constructing an encoder structure of a deep convolutional neural network for extracting and amplifying physiological signals, wherein the encoder structure comprises 2 convolutional layers and 3 residual error networks, a Relu activation function is used after each convolutional layer, then the physiological signals in the sleep video are extracted through one convolutional layer, the step length of the physiological signals is set to be 2, and finally the two residual error structures are connected for output;
step 3.2: building a modulation amplification structure of a deep convolutional neural network, performing convolution operation on the difference of physiological signals of two frames of sleep images through a convolutional layer, wherein an activation function is a Relu function, multiplying the activation function by an amplification factor alpha, and performing nonlinear change on the amplified characteristics by using the convolutional layer and a residual structure to obtain amplified physiological signal difference characteristics;
step 3.3: and building a decoder structure of a deep convolutional neural network, superposing the amplified physiological signal differential characteristics to an initial sleep image, and decoding and outputting the amplified video through upsampling and two convolutional layers to realize amplification of the physiological signals in the sleep video.
4. The system for monitoring sleep based on deep learning and multi-parameter fusion as claimed in claim 1, wherein the step 4 is specifically:
step 4.1: inputting the forehead area video image obtained in the step 2 into the deep convolutional neural network built in the step 3, setting an amplification factor alpha to be 15, extracting and amplifying the heart rate of the forehead area, and outputting the forehead area video image after amplifying the heart rate signal;
step 4.2: and (3) inputting the eye region video image obtained in the step (2) into the deep convolutional neural network built in the step (3), setting an amplification factor alpha to be 30, extracting and amplifying the eye movement frequency of the eye region, and outputting the eye region video image after the eye movement frequency is amplified.
5. The system according to claim 1, wherein the step 10 is specifically as follows:
step 10.1: building a neural network structure, which comprises 4 convolutional layers, 3 maximum pooling layers, 1 full-connection layer and 1 classifier;
step 10.2: preventing the calculated amount from being too large, extracting the current-time key frames of the video every 1s, respectively extracting 1 frame of images from the cameras positioned on the left and right of the body of the tester, and forming three-channel image data to be input into the neural network in the step 10.1;
step 10.3: extracting image features of the three-channel image data in the step 10.2 through a convolution layer with convolution kernels of 10 x 10, a maximum pooling layer of 2 x 2, a convolution layer with convolution kernels of 10 x 10, a maximum pooling layer of 2 x 2 and a convolution layer with convolution kernels of 10 x 10 respectively;
step 10.4: inputting the extracted features into a full-connection layer for six classifications, wherein classification results correspond to six sleeping postures: supine, prone, left supine, left crouch, right supine, and right crouch.
6. The system according to claim 1, wherein the step 15 is specifically as follows:
step 15.1: the heart rate of a person in normal sleep is 60-100 times per minute, and the heart rate can be reduced to 50 times per minute in deep sleep. The invention monitors the heart rate of a tester during sleeping, and considers that the tester enters a deep sleep period when the heart rate is obviously reduced, and considers that the tester exits the deep sleep period when the heart rate is gradually increased. Finally, counting the proportion of deep sleep periods in the sleep of the testers, wherein the larger the proportion of the deep sleep periods is, the higher the sleep quality is;
step 15.2: during the rapid eye movement phase of the sleep cycle, the eye ball rotates rapidly. The invention monitors the eye movement of the tester, if the eye movement frequency is obviously increased, the tester can be considered to enter the rapid eye movement period by combining the heart rate increase in the step 14.1, the duration time of the rapid eye movement period is counted, and if the rapid eye movement sleep is suddenly interrupted, the rapid eye movement sleep is often a signal for the attack of diseases such as angina, asthma and the like;
step 15.3: lying is considered to be a better sleeping position during sleeping, but is not suitable for people suffering from respiratory diseases or who frequently snore, and a side sleeping position is adopted. The invention monitors the sleeping posture of a tester, and if the tester has respiratory tract diseases or snores, the tester carries out sleeping posture adjustment suggestion when the tester adopts a non-side sleeping posture;
step 15.4: the invention monitors the turnover frequency of the tester, and if the turnover frequency is too high, the tester is prompted to possibly lack calcium ions or have high mental stress and poor sleep quality.
CN202210561402.3A 2022-05-23 2022-05-23 Non-contact sleep monitoring method based on deep learning and multi-parameter fusion Active CN114998229B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202210561402.3A CN114998229B (en) 2022-05-23 2022-05-23 Non-contact sleep monitoring method based on deep learning and multi-parameter fusion

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202210561402.3A CN114998229B (en) 2022-05-23 2022-05-23 Non-contact sleep monitoring method based on deep learning and multi-parameter fusion

Publications (2)

Publication Number Publication Date
CN114998229A true CN114998229A (en) 2022-09-02
CN114998229B CN114998229B (en) 2024-04-12

Family

ID=83027622

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202210561402.3A Active CN114998229B (en) 2022-05-23 2022-05-23 Non-contact sleep monitoring method based on deep learning and multi-parameter fusion

Country Status (1)

Country Link
CN (1) CN114998229B (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN116563887A (en) * 2023-04-21 2023-08-08 华北理工大学 Sleeping posture monitoring method based on lightweight convolutional neural network

Citations (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2004187961A (en) * 2002-12-12 2004-07-08 Toshiba Corp Sleeping condition detector and sleeping condition management system
CN105997004A (en) * 2016-06-17 2016-10-12 美的集团股份有限公司 Sleep reminding method and sleep monitoring device
US9993166B1 (en) * 2013-06-21 2018-06-12 Fitbit, Inc. Monitoring device using radar and measuring motion with a non-contact device
CN108836269A (en) * 2018-05-10 2018-11-20 电子科技大学 It is a kind of to merge the dynamic sleep mode automatically of heart rate breathing body method by stages
CN109431681A (en) * 2018-09-25 2019-03-08 吉林大学 A kind of intelligent eyeshade and its detection method detecting sleep quality
US20190097865A1 (en) * 2012-12-05 2019-03-28 Origin Wireless, Inc. Apparatus, systems and methods for event recognition based on a wireless signal
CN110957030A (en) * 2019-12-04 2020-04-03 中国人民解放军第二军医大学 Sleep quality monitoring and interaction system
CN111248868A (en) * 2020-02-20 2020-06-09 长沙湖湘医疗器械有限公司 Quick eye movement sleep analysis method, system and equipment
US20200397365A1 (en) * 2015-07-17 2020-12-24 Feng Zhang Method, apparatus, and system for wireless sleep monitoring
CN112451834A (en) * 2020-11-24 2021-03-09 珠海格力电器股份有限公司 Sleep quality management method, device, system and storage medium
CN112806975A (en) * 2021-02-01 2021-05-18 深圳益卡思科技发展有限公司 Sleep monitoring device, method and medium based on millimeter wave radar

Patent Citations (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2004187961A (en) * 2002-12-12 2004-07-08 Toshiba Corp Sleeping condition detector and sleeping condition management system
US20190097865A1 (en) * 2012-12-05 2019-03-28 Origin Wireless, Inc. Apparatus, systems and methods for event recognition based on a wireless signal
US9993166B1 (en) * 2013-06-21 2018-06-12 Fitbit, Inc. Monitoring device using radar and measuring motion with a non-contact device
US20200397365A1 (en) * 2015-07-17 2020-12-24 Feng Zhang Method, apparatus, and system for wireless sleep monitoring
CN105997004A (en) * 2016-06-17 2016-10-12 美的集团股份有限公司 Sleep reminding method and sleep monitoring device
CN108836269A (en) * 2018-05-10 2018-11-20 电子科技大学 It is a kind of to merge the dynamic sleep mode automatically of heart rate breathing body method by stages
CN109431681A (en) * 2018-09-25 2019-03-08 吉林大学 A kind of intelligent eyeshade and its detection method detecting sleep quality
CN110957030A (en) * 2019-12-04 2020-04-03 中国人民解放军第二军医大学 Sleep quality monitoring and interaction system
CN111248868A (en) * 2020-02-20 2020-06-09 长沙湖湘医疗器械有限公司 Quick eye movement sleep analysis method, system and equipment
CN112451834A (en) * 2020-11-24 2021-03-09 珠海格力电器股份有限公司 Sleep quality management method, device, system and storage medium
CN112806975A (en) * 2021-02-01 2021-05-18 深圳益卡思科技发展有限公司 Sleep monitoring device, method and medium based on millimeter wave radar

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
NIRANJAN SRIDHAR等: "Deep learning for automated sleep staging using instantaneous heart rate", 《NPJ DIGITAL MEDICINE》, 20 August 2020 (2020-08-20) *
冯静达;焦学军;李启杰;郭娅美;杨涵钧;楚洪祚;: "基于心率和呼吸特征结合的睡眠分期研究", 航天医学与医学工程, no. 02, 15 April 2020 (2020-04-15) *
张煜;刘保真;单聪淼;牟锴钰;: "基于IPPG技术的生理参数检测综述", 医疗卫生装备, no. 02, 15 February 2020 (2020-02-15) *

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN116563887A (en) * 2023-04-21 2023-08-08 华北理工大学 Sleeping posture monitoring method based on lightweight convolutional neural network
CN116563887B (en) * 2023-04-21 2024-03-12 华北理工大学 Sleeping posture monitoring method based on lightweight convolutional neural network

Also Published As

Publication number Publication date
CN114998229B (en) 2024-04-12

Similar Documents

Publication Publication Date Title
CN104834946B (en) A kind of contactless sleep monitor method and system
JP5961235B2 (en) Sleep / wake state evaluation method and system
JP5859979B2 (en) Health indicators based on multivariate residuals for human health monitoring
CN102973273B (en) Sleep respiratory function monitoring system based on infrared radiation detection
US9443289B2 (en) Compensating for motion induced artifacts in a physiological signal extracted from multiple videos
US20140275829A1 (en) Sleep stage annotation device
CN101365373A (en) Techniques for prediction and monitoring of clinical episodes
Waltisberg et al. Detecting disordered breathing and limb movement using in-bed force sensors
CN112806966B (en) Non-interference type early warning system for sleep apnea
WO2008116318A1 (en) Breathing sound analysis for estimation of air flow rate
CN103561648A (en) Method and apparatus for monitoring movement and breathing of multiple subjects in a common bed
CN111027485A (en) Heart rate detection method based on face video detection and chrominance model
Al-Naji et al. Simultaneous tracking of cardiorespiratory signals for multiple persons using a machine vision system with noise artifact removal
CN111544001A (en) Non-contact apnea detection device and method
CN109567756A (en) A kind of sleep state detection method based on artificial intelligence mattress
CN109044275B (en) Non-invasive sensing sleep quality analysis system and method based on fuzzy inference system
Dong et al. Wearable diet monitoring through breathing signal analysis
Massaroni et al. Comparison of two methods for estimating respiratory waveforms from videos without contact
CN114998229B (en) Non-contact sleep monitoring method based on deep learning and multi-parameter fusion
Kau et al. Pressure-sensor-based sleep status and quality evaluation system
CN107280673B (en) A kind of infrared imaging breath signal detection method based on key-frame extraction technique
Suriani et al. Non-contact facial based vital sign estimation using convolutional neural network approach
CN210990210U (en) Non-inductive sleep monitoring device and monitoring system
CN202942113U (en) Sleep respiratory function monitoring system based on infrared radiation detection
TWI756793B (en) A channel information processing system

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant