WO2021120007A1 - 一种基于红外图像序列的睡眠质量评估系统和方法 - Google Patents
一种基于红外图像序列的睡眠质量评估系统和方法 Download PDFInfo
- Publication number
- WO2021120007A1 WO2021120007A1 PCT/CN2019/126038 CN2019126038W WO2021120007A1 WO 2021120007 A1 WO2021120007 A1 WO 2021120007A1 CN 2019126038 W CN2019126038 W CN 2019126038W WO 2021120007 A1 WO2021120007 A1 WO 2021120007A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- infrared image
- respiratory
- tensor
- image sequence
- sleep quality
- Prior art date
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/0002—Inspection of images, e.g. flaw detection
- G06T7/0012—Biomedical image inspection
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/0059—Measuring for diagnostic purposes; Identification of persons using light, e.g. diagnosis by transillumination, diascopy, fluorescence
- A61B5/0082—Measuring for diagnostic purposes; Identification of persons using light, e.g. diagnosis by transillumination, diascopy, fluorescence adapted for particular medical purposes
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/08—Detecting, measuring or recording devices for evaluating the respiratory organs
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/48—Other medical applications
- A61B5/4806—Sleep evaluation
- A61B5/4815—Sleep quality
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/72—Signal processing specially adapted for physiological signals or for diagnostic purposes
- A61B5/7235—Details of waveform analysis
- A61B5/7264—Classification of physiological signals or data, e.g. using neural networks, statistical classifiers, expert systems or fuzzy systems
- A61B5/7267—Classification of physiological signals or data, e.g. using neural networks, statistical classifiers, expert systems or fuzzy systems involving training the classification device
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/764—Arrangements for image or video recognition or understanding using pattern recognition or machine learning using classification, e.g. of video objects
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/77—Processing image or video features in feature spaces; using data integration or data reduction, e.g. principal component analysis [PCA] or independent component analysis [ICA] or self-organising maps [SOM]; Blind source separation
- G06V10/774—Generating sets of training patterns; Bootstrap methods, e.g. bagging or boosting
- G06V10/7753—Incorporation of unlabelled data, e.g. multiple instance learning [MIL]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/82—Arrangements for image or video recognition or understanding using pattern recognition or machine learning using neural networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/20—Movements or behaviour, e.g. gesture recognition
-
- G—PHYSICS
- G16—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
- G16H—HEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
- G16H50/00—ICT specially adapted for medical diagnosis, medical simulation or medical data mining; ICT specially adapted for detecting, monitoring or modelling epidemics or pandemics
- G16H50/30—ICT specially adapted for medical diagnosis, medical simulation or medical data mining; ICT specially adapted for detecting, monitoring or modelling epidemics or pandemics for calculating health indices; for individual health risk assessment
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10016—Video; Image sequence
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10048—Infrared image
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20081—Training; Learning
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20084—Artificial neural networks [ANN]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30196—Human being; Person
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30232—Surveillance
Definitions
- This application belongs to the field of deep learning technology, and in particular relates to a sleep quality assessment system and method based on infrared image sequences.
- Sleep as a complex life behavior, is closely related to human health. Sleep occupies about one-third of human life. However, the pace of modern life is generally accelerating, the pressure of life and work continues to increase, and sleep loss is becoming more common, which seriously affects the quality of people's daily life and even their health. Therefore, it is of great significance to detect sleep quality, evaluate people's sleep status, and promptly guide treatment for poor sleep.
- the subject needs to use the relevant equipment to monitor for about 8 hours in the hospital, which is expensive, time cost and price
- the cost is relatively high; when the wearable device captures physiological data, when analyzing the quality of sleep, it needs to directly contact the human body, which brings inconvenience and psychological burden to the testee, interferes with the sleep process of the testee, and affects the testee’s sleep quality. Sleeping habits ultimately affect the accuracy of the assessment of the sleep quality of the subjects.
- the first aspect of the embodiments of the present application provides a sleep quality assessment method based on an infrared image sequence, and the sleep quality assessment method includes:
- the image sequence acquisition module is used to acquire multiple respiratory infrared image sequences to be evaluated, and one respiratory infrared image sequence to be evaluated includes multiple frames of respiratory infrared images to be evaluated;
- the sleep quality evaluation module is configured to evaluate the sleep quality of each of the plurality of respiratory infrared image sequences to be evaluated by using a classifier to obtain the sleep quality corresponding to each of the respiratory infrared image sequences to be evaluated evaluation result;
- the third aspect of the embodiments of the present application provides a terminal device, including a memory, a processor, and a computer program stored in the memory and running on the processor.
- the processor executes the computer program, The steps of the sleep quality evaluation method as described in the first aspect above are implemented.
- a fourth aspect of the embodiments of the present application provides a computer-readable storage medium, the computer-readable storage medium stores a computer program, and when the computer program is executed by a processor, the sleep quality evaluation as described in the first aspect is implemented. Method steps.
- the fifth aspect of the present application provides a computer program product, which when the computer program product runs on a terminal device, causes the terminal device to execute the steps of the sleep quality assessment method described in the first aspect.
- this solution uses the infrared camera device to obtain multiple infrared image sequences of the user’s breathing to be evaluated during sleep, which can realize the non-contact sleep monitoring of the user and reduce the monitoring cost at the same time.
- Evaluate sleep quality as a whole for each of the multiple respiratory infrared image sequences to be evaluated which can effectively retain the temporal and spatial continuity between the respiratory infrared images to be evaluated in each respiratory infrared image sequence to be evaluated sexual information, improve the accuracy of the sleep quality evaluation results for each breathing infrared image sequence to be evaluated, and use the sleep quality evaluation result that accounts for the most among multiple sleep quality evaluation results as the user’s sleep quality evaluation result, which improves The accuracy of the assessment of the user's sleep quality.
- FIG. 2 is a schematic diagram of the implementation process of the method for evaluating sleep quality based on infrared image sequences provided in the second embodiment of the present application;
- FIG. 4 is a schematic structural diagram of a sleep quality assessment system based on infrared image sequences provided in Embodiment 3 of the present application;
- FIG. 5 is a schematic structural diagram of a terminal device provided in Embodiment 4 of the present application.
- FIG. 1 is a schematic diagram of the implementation process of the sleep quality assessment method based on infrared image sequence provided in the first embodiment of the present application.
- the sleep quality assessment method is applied to a terminal device. As shown in the figure, the sleep quality assessment method may include the following steps:
- Step S101 Acquire multiple respiratory infrared image sequences to be evaluated, and one respiratory infrared image sequence to be evaluated includes multiple frames of respiratory infrared images to be evaluated.
- multiple infrared image sequences of breathing to be evaluated can be acquired through an infrared camera device when the user is sleeping.
- the infrared camera device can be integrated in the terminal device, and It can be independent of the terminal equipment (that is, the infrared camera is not integrated in the terminal equipment).
- the connection and communication between the infrared camera and the terminal equipment can be established through wireless or wired methods, and the infrared camera can be connected to the terminal equipment.
- the multiple respiratory infrared image sequences to be evaluated obtained by the camera device are transmitted to the terminal device.
- the breathing infrared image to be evaluated may refer to the image taken by the infrared camera device of the user (ie, the user whose sleep quality is evaluated) in the nose and mouth area.
- the infrared camera device does not need to be in contact with the user and can realize non-contact sleep monitoring of the user. , To avoid interference to the user's daily sleep, while reducing the cost of monitoring.
- the continuous acquisition of multiple frames of respiratory infrared images to be evaluated can capture the temperature changes in the mouth and nose area when the user breathes, so as to extract features such as respiratory frequency and respiratory depth.
- a breathing infrared image sequence to be evaluated may refer to a breathing infrared image sequence waiting to be evaluated for sleep quality.
- the infrared camera device When the infrared camera device acquires multiple infrared image sequences of breathing to be evaluated while the user is sleeping, it can use a sliding window method to collect one infrared image sequence of breathing to be evaluated while the user is sleeping with a preset duration as a basic unit, for example, one minute as one Basic unit, continuous collection of five minutes of infrared images to be evaluated, multiple frames of infrared images to be evaluated within one minute compose a sequence of infrared images to be evaluated, five minutes includes five to one minute, that is, five minutes corresponds to five to be evaluated Breathing infrared image sequence.
- a sliding window method to collect one infrared image sequence of breathing to be evaluated while the user is sleeping with a preset duration as a basic unit, for example, one minute as one Basic unit, continuous collection of five minutes of infrared images to be evaluated, multiple frames of infrared images to be evaluated within one minute compose a sequence of infrared images to be evaluated, five minutes includes five to
- each respiratory infrared image sequence to be evaluated may also be preprocessed.
- the preprocessing includes, but is not limited to, adding each respiratory infrared image sequence to be evaluated Adjust the size of the multiple frames of respiratory infrared images to be evaluated in the sequence to be the same (for example, adjust to a preset size) and/or adjust the pixel values of the multiple frames of respiratory infrared images to be evaluated in each respiratory infrared image sequence to be evaluated to the preset size. Set within the range.
- Step S102 Perform sleep quality assessment on each of the plurality of respiratory infrared image sequences to be evaluated by a classifier to obtain sleep quality assessment results corresponding to each of the respiratory infrared image sequences to be evaluated.
- the above-mentioned classifier can evaluate a breathing infrared image sequence to be evaluated as a whole for sleep quality evaluation, that is, the above-mentioned classifier can directly evaluate a breathing infrared image sequence to be evaluated for sleep quality, and can effectively retain a breathing infrared image sequence to be evaluated.
- the spatio-temporal continuity information between multiple frames of respiratory infrared images to be evaluated in the sequence improves the accuracy of the sleep quality evaluation result of a respiratory infrared image sequence to be evaluated.
- the sleep quality evaluation result is used to indicate whether the sleep quality is good or bad.
- the sleep quality evaluation result includes but is not limited to the first sleep quality evaluation result and the second sleep quality evaluation result.
- the first sleep quality evaluation result can mean that the sleep quality is good. 2.
- the sleep quality assessment result may refer to poor sleep quality, that is, the sleep quality indicated by the first sleep quality assessment result is better than the sleep quality indicated by the second sleep quality assessment result.
- the content of the sleep quality evaluation results can also be re-divided according to actual needs.
- the sleep quality evaluation results can be good sleep quality, good sleep quality, poor sleep quality, etc., which are not limited here.
- the classifier performs sleep quality assessment on a breathing infrared image sequence to be evaluated is to classify a breathing infrared image sequence to be evaluated.
- the classification category is the sleep quality assessment result.
- the classification categories include good sleep quality and If the sleep quality is bad, a classifier is used to determine whether the category of a breathing infrared image sequence to be assessed belongs to good sleep quality or bad sleep quality.
- the sleep quality assessment when the sleep quality assessment is performed on each of the multiple breathing infrared image sequences to be evaluated by the classifier, it may be obtained by classifying the multiple breathing infrared image sequences to be evaluated.
- the device separately evaluates the sleep quality of multiple breathing infrared image sequences to be evaluated; it can also evaluate the sleep quality of the first breathing infrared image sequence to be evaluated through the classifier when the first breathing infrared image sequence to be evaluated is obtained , Obtain the sleep quality assessment result of the first respiratory infrared image sequence to be evaluated (that is, the sleep quality assessment result corresponding to the first respiratory infrared image sequence to be assessed), and when the second respiratory infrared image sequence to be evaluated is obtained, pass The classifier evaluates the sleep quality of the second breathing infrared image sequence to be evaluated, and obtains the sleep quality evaluation result of the second breathing infrared image sequence to be evaluated, and so on, until the last one of the multiple breathing infrared image sequences to be evaluated is obtained.
- the sleep quality assessment is performed on each of the multiple breathing infrared image sequences to be evaluated by the classifier to obtain the sleep quality assessment corresponding to each breathing infrared image sequence to be evaluated
- the results include:
- the target feature map and the fully connected layer based on tensor decomposition in the classifier perform sleep quality assessment on each of the breathing infrared image sequences to be evaluated, and obtain the corresponding to each breathing infrared image sequence to be evaluated Results of sleep quality assessment.
- the classifier when it receives a respiratory infrared image sequence to be evaluated, it can use the second-order pooling block to use the second-order information of the respiratory infrared image sequence to be evaluated to automatically extract and Sleep-related breathing characteristics improve the accuracy of the assessment of the sleep quality of the breathing infrared image sequence to be assessed.
- the network layer based on tensor decomposition includes a convolutional layer based on tensor decomposition and two densely connected blocks based on tensor decomposition.
- the classifier uses the densely connected mechanism of densely connected blocks to effectively solve the problem of gradient disappearance.
- the densely connected blocks in the classifier may refer to a dense convolutional neural network, such as a residual network.
- a densely connected block usually includes multiple convolutional layers.
- the convolution layer based on tensor decomposition refers to the tensor decomposition of the convolution kernel of the convolution layer.
- the convolution layer based on tensor decomposition is a 3D convolution layer, and the convolution kernel of the 3D convolution layer is tensor decomposed.
- the convolution kernel of the 3D convolution layer can be decomposed into the product of two matrices and a three-dimensional tensor.
- the first and third orders are matrices, and the second order is a three-dimensional tensor.
- the fully connected layer based on tensor decomposition refers to the tensor decomposition of the weight of the fully connected layer.
- the weight of the fully connected layer can be decomposed into the product of two matrices and a three-dimensional tensor.
- the first and third orders are matrices.
- the second order is a three-dimensional tensor.
- the breathing infrared image sequence to be evaluated is first taken as a tensor as a whole, and convolved with the convolution kernel after tensor decomposition in the convolutional layer Calculate, and then perform convolution calculation with the convolution kernel after tensor decomposition in a densely connected block, and then pass the second-order pooling block, and then perform convolution calculation with the convolution sum after tensor decomposition in a densely connected block, Finally, the weighted fully connected layer with tensor decomposition is used to obtain the sleep quality evaluation result of the breathing infrared image sequence to be evaluated.
- the target feature map is the feature map output by the last densely connected block in the classifier. It is a high-order feature map generated after multi-layer convolution, because the number of feature maps output by the last densely connected block in the classifier is multiple. Therefore, it can be called a high-order feature map.
- the network layer and the fully connected layer in the classifier are represented by tensor decomposition, which can reduce the number of parameters in the classifier, and solve the loss of internal structure information of the tensor data and the number of parameters in the classifier caused by the vectorization calculation of tensor form data.
- the aforementioned tensor decomposition may refer to Tensor-Train tensor decomposition.
- Step S103 Count the number of different sleep quality evaluation results according to the sleep quality evaluation results corresponding to the plurality of breathing infrared image sequences to be evaluated, and determine the sleep quality evaluation result with the largest number as the sleep quality evaluation result of the user.
- one breathing infrared image sequence to be evaluated corresponds to one sleep quality assessment result
- multiple breathing infrared image sequences to be evaluated correspond to multiple sleep quality assessment results
- the same sleep quality assessment results may exist in the multiple sleep quality assessment results.
- the sleep quality of five respiratory infrared image sequences to be evaluated, the first respiratory infrared image sequence to be evaluated, the second respiratory infrared image sequence to be evaluated, and the fifth respiratory infrared image sequence to be evaluated are exemplified by the classifier.
- the evaluation results are all good sleep quality.
- the sleep quality evaluation results of the third breathing infrared image sequence to be evaluated and the fourth breathing infrared image sequence to be evaluated are bad sleep quality, and three of the five sleep quality evaluation results are counted. Good sleep quality and two bad sleep quality, the number of good sleep quality is the largest, then it can be determined that the user’s sleep quality evaluation result is good sleep quality.
- the infrared camera device obtains the user’s breathing infrared image sequence to be evaluated during sleep, and uses a tensor decomposition-based classifier to evaluate the sleep quality of each breathing infrared image sequence to be evaluated as a tensor as a whole.
- a contactless sleep quality evaluation can be realized, and the accuracy of the evaluation of the user's sleep quality can be improved.
- FIG. 2 is a schematic diagram of the implementation process of the sleep quality assessment method based on infrared image sequence provided in the second embodiment of the present application.
- the sleep quality assessment method is applied to a terminal device. As shown in the figure, the sleep quality assessment method may include the following steps:
- step S201 the classifier is trained through Zhang's quantified ternary generative confrontation network.
- the quantified ternary generative confrontation network includes a generator, a classifier, and a discriminator.
- the generator, classifier, and discriminator all use tensor decomposition, which effectively reduces the number of generators, classifiers, and discriminators.
- the number of parameters, and the overall processing of the respiratory infrared image sequence can be realized.
- Figure 3a shows an example of the structure of Zhang Quan’s ternary generative confrontation network.
- G represents the generator
- C represents the classifier
- D represents the discrimination.
- the unmarked respiratory infrared image sequence X c is the respiratory infrared image sequence without the tag
- the marked infrared image sequence (X l , Y l ) is the respiratory infrared image sequence with the tag
- Y c is the tag of the unmarked infrared image sequence
- X g is the respiratory infrared image sequence generated by the generator.
- the tensor decomposition algorithm Tensor-Train decomposes a d-order tensor, which can be expressed as the product of two matrices and d-2 three-dimensional tensors.
- the first and dth are matrices, and the remaining d-2 are Three-dimensional tensor, d is an integer greater than 2.
- the training of the classifier through the quantified ternary generative confrontation network includes:
- the second respiratory infrared image sequence is input to the classifier, and the third respiratory infrared image sequence is obtained through the second-order pooling block, the network layer based on tensor decomposition, and the fully connected layer in the classifier.
- the third breath infrared image sequence refers to the second breath infrared image sequence carrying the tag;
- the discriminator is trained, and the discriminator's response to the third respiratory infrared image sequence is obtained. Judgment result;
- the generator adopts the idea of conditional generation against the network, taking one-dimensional random noise that obeys a normal distribution as the input of the generator, and taking sleep quality as the target label as the conditional input, and the intermediate network layer adopts 3D reaction Convolutional layer, and then use Leaky ReLU as the activation function, and Batch Norm for batch regularization.
- the last 3D deconvolution layer of the generator is followed by a tanh activation layer, and the Tensor-Train tensor decomposition auxiliary generator is used to generate and carry
- the breathing infrared image sequence of the sleep quality label reduces the need for a real breathing infrared image sequence carrying the label.
- Figure 3b shows an example of the structure of the generator.
- the one-dimensional random noise carrying the target label is sequentially deconvolved through the 3D deconvolution layer, the Leaky ReLU activation function and batch regularization to perform the gradual deconvolution of the feature map, which can generate the first breath infrared image that is close to the real one carrying the target label. sequence.
- D(x g ,y g ) represents the discrimination result of the discriminator, if the discrimination result is true, then D(x g ,y g ) is 1, if the discrimination result is false, then D(x g ,y g ) Is 0;
- ⁇ is the weight parameter (users can set it according to actual needs);
- x label is the real respiratory infrared image sequence, and
- x g is the generated respiratory infrared image sequence (ie, the first respiratory infrared image sequence), It represents the L 1 loss between the real breathing infrared image sequence and the generated breathing infrared image sequence, so that the generated breathing infrared image sequence is more close to the
- the unlabeled respiratory infrared image sequence collected by the infrared camera into two parts: one part is used as the second respiratory infrared image sequence, the sleep quality is evaluated by the classifier, and the sleep quality evaluation result output by the classifier is used as the label , So as to obtain the third breath infrared image sequence; the other part uses sleep experts to evaluate the sleep quality, and use the sleep quality assessment result as a label to obtain the fourth breath infrared image sequence carrying the label; the first breath infrared image sequence, the first breath The three-breath infrared image sequence and the fourth breath infrared image sequence are respectively input to the discriminator, and the discriminator’s discrimination results of the first, third and fourth breath infrared image sequences are obtained.
- the discriminant results of a sequence of breathing infrared images obtain the loss function of the discriminator, and train the discriminator according to the loss function.
- the loss function Loss l when the breathing infrared image sequence trains the classifier, and the Loss g and Loss l are integrated into Loss supervised Loss g + ⁇ Loss l , and the classification is obtained according to the discriminator’s discriminating result of the third breathing infrared image sequence
- Zhang Quan’s ternary generation confrontation network uses a large number of second respiratory infrared image sequences (ie, respiratory infrared image sequences that do not carry tags) and a small number of fourth respiratory infrared image sequences (ie, tag-bearing infrared image sequences). Breathing infrared image sequence) training the ternary generation of Zhang quantified against the network can solve the problem of less data in the respiratory infrared image sequence with tags, and at the same time make full use of the respiratory infrared image sequence without tags, which is conducive to improving the quantification of Zhang. The robustness of the ternary generation against the network.
- the obtaining the first respiratory infrared image sequence carrying the target label through the deconvolution layer based on tensor decomposition in the generator includes:
- the deconvolution kernel of the 3D deconvolution layer in the generator after the deconvolution kernel of the 3D deconvolution layer in the generator is decomposed into a tensor, it can be decomposed into the product of two matrices and a three-dimensional tensor (where the first order and the third order Is a matrix, the second order is a three-dimensional tensor), and then the input one-dimensional random noise and the first tensor (that is, the product of two matrices and a three-dimensional tensor) are subjected to multi-layer deconvolution calculations, and then through the graph
- the activation function Leaky ReLU, batch regularization, and tanh activation layer in the structural example shown in 3b generate a sequence of first breath infrared images that are close to real and carry target tags.
- the network layer based on tensor decomposition in the discriminator includes a convolutional layer based on tensor decomposition, a first dense connection block, and a second dense connection block;
- obtaining the discrimination result of the first respiratory infrared image sequence by the discriminator includes:
- Tensor decomposition is performed on the convolution kernel of the first densely connected block in the discriminator to obtain a third tensor
- Tensor decomposition is performed on the convolution kernel of the second densely connected block in the discriminator to obtain a fourth tensor
- the first feature map, the second feature map, and the third feature map are all feature maps of the first respiratory infrared image sequence.
- the tensor decomposition of the convolution kernel of the 3D convolution layer in the discriminator after the tensor decomposition of the convolution kernel of the 3D convolution layer in the discriminator, it can be decomposed into the product of two matrices and a three-dimensional tensor (where the first order and the third order are matrix , The second order is a three-dimensional tensor), the product of the two matrices and a three-dimensional tensor obtained after decomposition is the second tensor, and the first breathing infrared image sequence is calculated by multi-layer convolution through the second tensor
- the resulting feature map is the first feature map; after tensor decomposition of the convolution kernel of the first densely connected block in the discriminator, it can be decomposed into the product of two matrices and a three-dimensional tensor (where the first order and The third order is a matrix, the second order is a three-dimensional tensor), the product of the
- the discriminator further includes a first transition layer and a second transition layer, the first transition layer and the second transition layer are both 1 ⁇ 1 ⁇ 1 convolution kernels; the first transition layer The layer is located between the first densely connected block and the second densely connected block in the discriminator, and is used to reduce the number of the second feature maps; the second transition layer is located in the second densely connected block in the discriminator Between it and the fully connected layer, it is used to reduce the number of the third feature maps.
- the number of second feature maps obtained increases, and the convolution calculation is performed through the 1 ⁇ 1 ⁇ 1 3D convolution kernel, which can reduce The number of second feature maps is to reduce the number of channels.
- the number of third feature maps obtained increases, and the convolution calculation is performed through the 1 ⁇ 1 ⁇ 1 3D convolution kernel, which can reduce the number of third feature maps. That is, reduce the number of channels.
- the tag-carrying infrared image sequence of breathing input to the discriminator (can be any tag-carrying infrared image sequence input to the discriminator) is convolution based on tensor decomposition Layer (ie 3D quantized convolutional layer), densely connected block based on tensor decomposition (ie 3D quantized densely connected block), transition layer and deep neural network composed of fully connected layer based on tensor decomposition, for breathing infrared images
- the sequence performs feature extraction to obtain the respiratory feature information that retains the respiratory infrared image sequence in time and space.
- the extracted respiratory feature information is passed through a fully connected layer based on tensor decomposition to realize the trueness of the respiratory infrared image sequence with tags. False discrimination.
- the discriminator also includes Leaky ReLU activation function, Batch Norm regularization and sigmoid function.
- the embodiment of the present application uses densely connected blocks to directly input the feature maps extracted by the network layer before the densely connected blocks to the subsequent network layers for cascading, reducing the feature loss in the gradient transfer process, and solving the deep neural network reverse
- the problem of the disappearance of gradients in the propagation process stabilizes the training of the confrontation generation network, and improves the discriminator's performance in discriminating between the generated samples and the real samples.
- the network layer based on tensor decomposition in the classifier includes a convolutional layer based on tensor decomposition, a third densely connected block, and a fourth densely connected block;
- the second-order pooling in the classifier Block, network layer and fully connected layer based on tensor decomposition, to obtain the third infrared absorption image sequence includes:
- Tensor decomposition is performed on the convolution kernel of the third densely connected block in the classifier to obtain a seventh tensor
- Tensor decomposition is performed on the convolution kernel of the fourth densely connected block in the classifier to obtain an eighth tensor
- the fourth feature map, the fifth feature map, the sixth feature map, and the seventh feature map are all feature maps of the second respiratory infrared image sequence.
- the convolution kernel of the 3D convolution layer in the classifier can be decomposed into the product of two matrices and a three-dimensional tensor (where the first order and the third order are matrix ,
- the second order is a three-dimensional tensor
- the product of the two matrices and a three-dimensional tensor obtained after decomposition is the sixth tensor
- the second breathing infrared image sequence is calculated by multi-layer convolution through the sixth tensor
- the resulting feature map is the fourth feature map; after tensor decomposition of the convolution kernel of the third densely connected block in the classifier, it can be decomposed into the product of two matrices and a three-dimensional tensor (where the first order and The third order is a matrix, the second order is a three-dimensional tensor), the product of the two matrices and a three-dimensional tensor obtained after decomposition is the seventh tensor, and the fourth feature
- the classifier further includes a third transition layer and a fourth transition layer, the third transition layer and the fourth transition layer are both 1 ⁇ 1 ⁇ 1 convolution kernels; the third transition layer The layer is located between the second-order pooling block and the fourth densely connected block in the classifier, and is used to reduce the number of the sixth feature maps; the fourth transition layer is located in the fourth densely connected block in the classifier Between it and the fully connected layer, it is used to reduce the number of the seventh feature maps.
- the number of fifth feature maps obtained increases, which in turn leads to a larger number of sixth feature maps processed by the second-level pooling block.
- the number of sixth feature maps can be reduced, that is, the number of channels can be reduced.
- the number of seventh feature maps obtained increases.
- the convolution calculation is performed through the 1 ⁇ 1 ⁇ 1 3D convolution kernel, which can reduce the number of seventh feature maps. That is, reduce the number of channels.
- Figure 3d shows an example of the structure of the classifier.
- the respiratory infrared image sequence input to the classifier passes through a convolutional layer based on tensor decomposition (i.e. 3D quantized convolutional layer) and densely connected blocks based on tensor decomposition (i.e. 3D quantization dense connection block) performs multi-layer convolution calculation, and reduces the number of feature maps through the transition layer, and then through a fully connected layer based on tensor decomposition, the probability value of the corresponding category can be obtained, thereby obtaining the breathing infrared image Sequence of sleep quality assessment results.
- the classifier also includes Leaky ReLU activation function, Batch Norm regularization and sigmoid function.
- the embodiment of the present application uses the self-attention mechanism based on second-order pooling densely connected blocks to weight the feature map in the dimension of the feature channel according to the correlation of different regions, so that the important channel weight is significant and not important.
- the channel weight is small, which effectively extracts the respiratory features in the infrared image sequence, and improves the accuracy of the assessment of sleep quality.
- Step S202 Acquire multiple respiratory infrared image sequences to be evaluated, and one respiratory infrared image sequence to be evaluated includes multiple frames of respiratory infrared images to be evaluated.
- step S101 This step is the same as step S101.
- step S101 For details, please refer to the related description of step S101, which will not be repeated here.
- Step S203 Perform sleep quality assessment on each of the plurality of respiratory infrared image sequences to be evaluated by a classifier to obtain sleep quality assessment results corresponding to each of the respiratory infrared image sequences to be evaluated.
- step S102 This step is the same as step S102.
- step S102 For details, please refer to the related description of step S102, which will not be repeated here.
- Step S204 According to the sleep quality evaluation results corresponding to the multiple breathing infrared image sequences to be evaluated, the number of different sleep quality evaluation results is counted, and the sleep quality evaluation result with the largest number is determined as the sleep quality evaluation result of the user.
- step S103 This step is the same as step S103.
- step S103 For details, please refer to the related description of step S103, which will not be repeated here.
- the entire ternary generation confrontation network is quantified, so that the network has a regularization effect, reduces the possibility of network overfitting, and enhances the prediction generalization ability of the network.
- Zhang Quan reduced the network parameters, accelerated the network training speed, and increased the network operation efficiency.
- the embodiment of the present application quantizes the respiratory infrared image sequence and replaces the 2D convolution with the 3D convolution, which effectively extracts the timing feature information, removes noise and unnecessary redundant information, and retains The feature relationship between the respiratory infrared image sequences reduces the loss of temporal feature information and improves the classification ability and classification accuracy of the classifier.
- FIG. 4 is a schematic diagram of a sleep quality assessment system based on infrared image sequences provided in the third embodiment of the present application. For ease of description, only the parts related to the embodiment of the present application are shown.
- the sleep quality assessment system includes:
- the image sequence acquisition module 41 is configured to acquire a plurality of respiratory infrared image sequences to be evaluated, and one respiratory infrared image sequence to be evaluated includes multiple frames of respiratory infrared images to be evaluated;
- the sleep quality assessment module 42 is configured to perform sleep quality assessment on each of the plurality of respiratory infrared image sequences to be evaluated by a classifier, and obtain the sleep corresponding to each of the respiratory infrared image sequences to be evaluated. Quality assessment results;
- the sleep quality determination module 43 is configured to count the number of different sleep quality evaluation results according to the sleep quality evaluation results corresponding to the plurality of breathing infrared image sequences to be evaluated, and determine that the sleep quality evaluation result with the largest number is the sleep quality of the user evaluation result.
- the sleep quality assessment module 42 includes:
- a sequence input unit configured to input each of the breathing infrared image sequences to be evaluated to the classifier
- the target feature map acquiring unit is configured to acquire the target feature map of each respiratory infrared image sequence to be evaluated according to the second-order pooling block in the classifier and the network layer based on tensor decomposition;
- the evaluation result acquisition unit is configured to evaluate the sleep quality of each breathing infrared image sequence to be evaluated according to the target feature map and the fully connected layer based on tensor decomposition in the classifier, to obtain each of the to-be-evaluated respiratory infrared image sequences. Evaluate the sleep quality evaluation results corresponding to the breathing infrared image sequence.
- the sleep quality assessment system further includes:
- the classifier training module is used to train the classifier through the quantified ternary generative confrontation network.
- the quantified ternary generative confrontation network includes a generator, the classifier, and a discriminator;
- the classifier training module includes:
- the first processing unit is configured to input one-dimensional random noise and a target label to the generator, and obtain a first breathing infrared image carrying the target label through the deconvolution layer based on tensor decomposition in the generator sequence;
- the second processing unit is configured to input the first respiratory infrared image sequence to the discriminator, and obtain the discriminator’s response to the first tensor decomposition-based network layer and the fully connected layer in the discriminator.
- the first training unit is used to train the generator according to the discrimination result
- the first acquiring unit is configured to acquire a second respiratory infrared image sequence without a tag
- the third processing unit is configured to input the second breathing infrared image sequence to the classifier, and obtain the third order through the second-order pooling block, the network layer based on tensor decomposition, and the fully connected layer in the classifier.
- a breathing infrared image sequence where the third breathing infrared image sequence refers to a second breathing infrared image sequence carrying a tag;
- the second acquiring unit is used to acquire the fourth respiratory infrared image sequence carrying the tag
- the second training unit is used to train the discriminator according to the first respiratory infrared image sequence, the third respiratory infrared image sequence, and the fourth respiratory infrared image sequence, and to obtain the discriminator’s response to the The discrimination result of the third breath infrared image sequence;
- the third training unit is configured to train the classifier according to the first respiratory infrared image sequence, the discrimination result of the third respiratory infrared image sequence by the discriminator, and the fourth respiratory infrared image sequence.
- the first processing unit is specifically configured to:
- the network layer based on tensor decomposition in the discriminator includes a convolutional layer based on tensor decomposition, a first dense connection block, and a second dense connection block; the second processing unit is specifically configured to:
- Tensor decomposition is performed on the convolution kernel of the first densely connected block in the discriminator to obtain a third tensor
- Tensor decomposition is performed on the convolution kernel of the second densely connected block in the discriminator to obtain a fourth tensor
- the first feature map, the second feature map, and the third feature map are all feature maps of the first respiratory infrared image sequence.
- the discriminator further includes a first transition layer and a second transition layer, the first transition layer and the second transition layer are both 1 ⁇ 1 ⁇ 1 convolution kernels; the first transition layer The layer is located between the first densely connected block and the second densely connected block in the discriminator, and is used to reduce the number of the second feature maps; the second transition layer is located in the second densely connected block in the discriminator Between it and the fully connected layer, it is used to reduce the number of the third feature maps.
- the network layer based on tensor decomposition in the classifier includes a convolutional layer based on tensor decomposition, a third dense connection block, and a fourth dense connection block; the third processing unit is specifically configured to:
- Tensor decomposition is performed on the convolution kernel of the third densely connected block in the classifier to obtain a seventh tensor
- Tensor decomposition is performed on the convolution kernel of the fourth densely connected block in the classifier to obtain an eighth tensor
- the fourth feature map, the fifth feature map, the sixth feature map, and the seventh feature map are all feature maps of the second respiratory infrared image sequence.
- the classifier further includes a third transition layer and a fourth transition layer, the third transition layer and the fourth transition layer are both 1 ⁇ 1 ⁇ 1 convolution kernels; the third transition layer The layer is located between the second-order pooling block and the fourth densely connected block in the classifier, and is used to reduce the number of the sixth feature maps; the fourth transition layer is located in the fourth densely connected block in the classifier Between it and the fully connected layer, it is used to reduce the number of the seventh feature maps.
- the sleep quality evaluation system provided by the embodiment of the present application can be applied in the foregoing method embodiment 1 and embodiment 2.
- FIG. 5 is a schematic diagram of a terminal device provided in Embodiment 4 of the present application.
- the terminal device 5 of this embodiment includes a processor 50, a memory 51, and a computer program 52 stored in the memory 51 and running on the processor 50.
- the processor 50 executes the computer program 52, the steps in the foregoing sleep quality assessment method embodiments are implemented.
- the processor 50 executes the computer program 52, the functions of the modules/units in the foregoing system embodiments are realized.
- the terminal device 5 may be a computing device such as a desktop computer, a notebook, a palmtop computer, and a cloud server.
- the terminal device may include, but is not limited to, a processor 50 and a memory 51.
- FIG. 5 is only an example of the terminal device 5, and does not constitute a limitation on the terminal device 5. It may include more or less components than those shown in the figure, or a combination of certain components, or different components.
- the terminal device may also include input and output devices, network access devices, buses, and so on.
- the so-called processor 50 may be a central processing unit (Central Processing Unit, CPU), other general-purpose processors, digital signal processors (Digital Signal Processor, DSP), application specific integrated circuits (Application Specific Integrated Circuit, ASIC), Ready-made programmable gate array (Field-Programmable Gate Array, FPGA) or other programmable logic devices, discrete gates or transistor logic devices, discrete hardware components, etc.
- the general-purpose processor may be a microprocessor or the processor may also be any conventional processor or the like.
- the memory 51 may be an internal storage unit of the terminal device 5, such as a hard disk or a memory of the terminal device 5.
- the memory 51 may also be an external storage device of the terminal device 5, such as a plug-in hard disk equipped on the terminal device 5, a smart memory card (Smart Media Card, SMC), or a Secure Digital (SD). Card, Flash Card, etc. Further, the memory 51 may also include both an internal storage unit of the terminal device 5 and an external storage device.
- the memory 51 is used to store the computer program and other programs and data required by the terminal device.
- the memory 51 can also be used to temporarily store data that has been output or will be output.
- system/terminal device and method may be implemented in other ways.
- the system/terminal device embodiments described above are only illustrative.
- the division of the modules or units is only a logical function division, and there may be other divisions in actual implementation, such as multiple units.
- components can be combined or integrated into another system, or some features can be omitted or not implemented.
- the displayed or discussed mutual coupling or direct coupling or communication connection may be indirect coupling or communication connection through some interfaces, devices or units, and may be in electrical, mechanical or other forms.
- the units described as separate components may or may not be physically separated, and the components displayed as units may or may not be physical units, that is, they may be located in one place, or they may be distributed on multiple network units. Some or all of the units may be selected according to actual needs to achieve the objectives of the solutions of the embodiments.
- the functional units in the various embodiments of the present application may be integrated into one processing unit, or each unit may exist alone physically, or two or more units may be integrated into one unit.
- the above-mentioned integrated unit can be implemented in the form of hardware or software functional unit.
- the integrated module/unit is implemented in the form of a software functional unit and sold or used as an independent product, it can be stored in a computer readable storage medium.
- the present application implements all or part of the processes in the above-mentioned embodiments and methods, and can also be completed by instructing relevant hardware through a computer program.
- the computer program can be stored in a computer-readable storage medium. When the program is executed by the processor, it can implement the steps of the foregoing method embodiments.
- the computer program includes computer program code, and the computer program code may be in the form of source code, object code, executable file, or some intermediate forms.
- the computer-readable medium may include: any entity or device capable of carrying the computer program code, recording medium, U disk, mobile hard disk, magnetic disk, optical disk, computer memory, read-only memory (ROM, Read-Only Memory) , Random Access Memory (RAM, Random Access Memory), electrical carrier signal, telecommunications signal, and software distribution media, etc.
- ROM Read-Only Memory
- RAM Random Access Memory
- electrical carrier signal telecommunications signal
- software distribution media etc.
- the content contained in the computer-readable medium can be appropriately added or deleted according to the requirements of the legislation and patent practice in the jurisdiction.
- the computer-readable medium Does not include electrical carrier signals and telecommunication signals.
- This application implements all or part of the processes in the above-mentioned embodiments and methods, and can also be completed by a computer program product.
- the computer program product When the computer program product is run on a terminal device, the terminal device can realize all of the foregoing when the terminal device is executed. Steps in the method embodiment.
Landscapes
- Engineering & Computer Science (AREA)
- Health & Medical Sciences (AREA)
- Life Sciences & Earth Sciences (AREA)
- Physics & Mathematics (AREA)
- General Health & Medical Sciences (AREA)
- Medical Informatics (AREA)
- Theoretical Computer Science (AREA)
- Artificial Intelligence (AREA)
- Public Health (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Pathology (AREA)
- Biomedical Technology (AREA)
- Evolutionary Computation (AREA)
- Animal Behavior & Ethology (AREA)
- Surgery (AREA)
- Biophysics (AREA)
- Veterinary Medicine (AREA)
- Heart & Thoracic Surgery (AREA)
- Molecular Biology (AREA)
- General Physics & Mathematics (AREA)
- Databases & Information Systems (AREA)
- Multimedia (AREA)
- Software Systems (AREA)
- Computing Systems (AREA)
- Physiology (AREA)
- Psychiatry (AREA)
- Radiology & Medical Imaging (AREA)
- Pulmonology (AREA)
- Quality & Reliability (AREA)
- Fuzzy Systems (AREA)
- Mathematical Physics (AREA)
- Nuclear Medicine, Radiotherapy & Molecular Imaging (AREA)
- Signal Processing (AREA)
- Social Psychology (AREA)
- Human Computer Interaction (AREA)
- Data Mining & Analysis (AREA)
- Epidemiology (AREA)
- Primary Health Care (AREA)
- Measurement Of The Respiration, Hearing Ability, Form, And Blood Characteristics Of Living Organisms (AREA)
Abstract
Description
Claims (20)
- 一种基于红外图像序列的睡眠质量评估方法,其特征在于,所述睡眠质量评估方法包括:获取多个待评估呼吸红外图像序列,一个待评估呼吸红外图像序列包括多帧待评估呼吸红外图像;通过分类器对所述多个待评估呼吸红外图像序列中每个待评估呼吸红外图像序列进行睡眠质量评估,获得所述每个待评估呼吸红外图像序列对应的睡眠质量评估结果;根据所述多个待评估呼吸红外图像序列分别对应的睡眠质量评估结果,统计不同睡眠质量评估结果的数量,并确定数量最多的睡眠质量评估结果为用户的睡眠质量评估结果。
- 如权利要求1所述的睡眠质量评估方法,其特征在于,所述通过分类器对所述多个待评估呼吸红外图像序列中每个待评估呼吸红外图像序列进行睡眠质量评估,获得所述每个待评估呼吸红外图像序列对应的睡眠质量评估结果包括:将所述每个待评估呼吸红外图像序列输入至所述分类器;根据所述分类器中二阶池化块和基于张量分解的网络层,获取所述每个待评估呼吸红外图像序列的目标特征图;根据所述目标特征图和所述分类器中基于张量分解的全连接层,对所述每个待评估呼吸红外图像序列进行睡眠质量评估,获得所述每个待评估呼吸红外图像序列对应的睡眠质量评估结果。
- 如权利要求1所述的睡眠质量评估方法,其特征在于,在通过分类器对所述多个待评估呼吸红外图像序列中每个待评估呼吸红外图像序列进行睡眠质量评估之前,还包括:通过张量化的三元生成对抗网络对所述分类器进行训练。
- 如权利要求3所述的睡眠质量评估方法,其特征在于,所述张量化的三元生成对抗网络包括生成器、所述分类器和判别器;所述通过张量化的三元生成对抗网络对所述分类器进行训练包括:将一维随机噪声和目标标签输入至所述生成器,通过所述生成器中基于张量分解的反卷积层,获得携带所述目标标签的第一呼吸红外图像序列;将所述第一呼吸红外图像序列输入至所述判别器,通过所述判别器中基于张量分解的网络层和全连接层,获得所述判别器对所述第一呼吸红外图像序列的判别结果;根据所述判别结果训练所述生成器;获取未携带标签的第二呼吸红外图像序列;将所述第二呼吸红外图像序列输入至所述分类器,通过所述分类器中二阶池化块、基于张量分解的网络层和全连接层,获得第三呼吸红外图像序列,所述第三呼吸红外图像序列是指携带标签的第二呼吸红外图像序列;获取携带标签的第四呼吸红外图像序列;根据所述第一呼吸红外图像序列、所述第三呼吸红外图像序列以及所述第四呼吸红外图像序列,训练所述判别器,并获取所述判别器对所述第三呼吸红外图像序列的判别结果;根据所述第一呼吸红外图像序列、所述判别器对所述第三呼吸红外图像序列的判别结果以及所述第四呼吸红外图像序列,训练所述分类器。
- 如权利要求4所述的睡眠质量评估方法,其特征在于,所述通过生成器中基于张量分解的反卷积层,获得携带所述目标标签的第一呼吸红外图像序列包括:对所述生成器中反卷积层的反卷积核进行张量分解,获得第一张量;将所述一维随机噪声与所述第一张量进行反卷积计算,获得携带所述目标标签的第一呼吸红外图像序列。
- 如权利要求4所述的睡眠质量评估方法,其特征在于,所述判别器中基于张量分解的网络层包括基于张量分解的卷积层、第一密集连接块以及第二密集连接块;所述通过所述判别器中基于张量分解的网络层和全连接层,获得所述判别器对所述第一呼吸红外图像序列的判别结果包括:对所述判别器中卷积层的卷积核进行张量分解,获得第二张量;将所述第一呼吸红外图像序列与所述第二张量进行卷积计算,获得第一特征图;对所述判别器中第一密集连接块的卷积核进行张量分解,获得第三张量;将所述第一特征图与所述第三张量进行卷积计算,获得第二特征图;对所述判别器中第二密集连接块的卷积核进行张量分解,获得第四张量;将所述第二特征图与所述第四张量进行卷积计算,获得第三特征图;对所述判别器中全连接层的权重进行张量分解,获得第五张量;根据所述第三特征图和所述第五张量,获得所述判别器对所述第一呼吸红外图像序列的判别结果;其中,所述第一特征图、所述第二特征图以及所述第三特征图均为所述第一呼吸红外图像序列的特征图。
- 如权利要求6所述的睡眠质量评估方法,其特征在于,所述判别器还包括第一过渡层和第二过渡层,所述第一过渡层、所述第二过渡层均为1×1×1的卷积核;所述第一过渡层位于所述判别器中第一密集连接块与第二密集连接块之间,用于减少所述第二特征图的数量;所述第二过渡层位于所述判别器中第二密集连接块与全连接层之间,用于减少所述第三特征图的数量。
- 如权利要求4所述的睡眠质量评估方法,其特征在于,所述分类器中基于张量分解的网络层包括基于张量分解的卷积层、第三密集连接块以及第四密集连接块;所述通过所述分类器中二阶池化块、基于张量分解的网络层和全连接层,获得第三吸红外图像序列包括:对所述分类器中卷积层的卷积核进行张量分解,获得第六张量;将所述第二呼吸红外图像序列与所述第六张量进行卷积计算,获得第四特征图;对所述分类器中第三密集连接块的卷积核进行张量分解,获得第七张量;将所述第四特征图和所述第七张量进行卷积计算,获得第五特征图;通过所述分类器中二阶池化块处理所述第五特征图,获得第六特征图;对所述分类器中第四密集连接块的卷积核进行张量分解,获得第八张量;将所述第六特征图与所述第八张量进行卷积计算,获得第七特征图;对所述分类器中全连接层的权重进行张量分解,获得第九张量;根据所述第七特征图和所述第九张量,获得所述第三呼吸红外图像序列;其中,所述第四特征图、所述第五特征图、所述第六特征图以及所述第七特征图均为所述第二呼吸红外图像序列的特征图。
- 如权利要求8所述的睡眠质量评估方法,其特征在于,所述分类器还包括第三过渡层和第四过渡层,所述第三过渡层和所述第四过渡层均为1×1×1的卷积核;所述第三过渡层位于所述分类器中二阶池化块与第四密集连接块之间,用于减少所述第六特征图的数量;所述第四过渡层位于所述分类器中第四密集连接块与全连接层之间,用于减少所述第七特征图的数量。
- 一种基于红外图像序列的睡眠质量评估系统,其特征在于,所述睡眠质量评估系统包括:图像序列获取模块,用于获取多个待评估呼吸红外图像序列,一个待评估呼吸红外图像序列包括多帧待评估呼吸红外图像;睡眠质量评估模块,用于通过分类器对所述多个待评估呼吸红外图像序列中每个待评估呼吸红外图像序列进行睡眠质量评估,获得所述每个待评估呼吸红外图像序列 对应的睡眠质量评估结果;睡眠质量确定模块,用于根据所述多个待评估呼吸红外图像序列分别对应的睡眠质量评估结果,统计不同睡眠质量评估结果的数量,并确定数量最多的睡眠质量评估结果为用户的睡眠质量评估结果。
- 一种终端设备,包括存储器、处理器以及存储在所述存储器中并可在所述处理器上运行的计算机程序,其特征在于,所述处理器执行所述计算机程序时实现如下步骤:获取多个待评估呼吸红外图像序列,一个待评估呼吸红外图像序列包括多帧待评估呼吸红外图像;通过分类器对所述多个待评估呼吸红外图像序列中每个待评估呼吸红外图像序列进行睡眠质量评估,获得所述每个待评估呼吸红外图像序列对应的睡眠质量评估结果;根据所述多个待评估呼吸红外图像序列分别对应的睡眠质量评估结果,统计不同睡眠质量评估结果的数量,并确定数量最多的睡眠质量评估结果为用户的睡眠质量评估结果。
- 如权利要求11所述的终端设备,其特征在于,所述处理器执行所述计算机程序时,所述通过分类器对所述多个待评估呼吸红外图像序列中每个待评估呼吸红外图像序列进行睡眠质量评估,获得所述每个待评估呼吸红外图像序列对应的睡眠质量评估结果包括:将所述每个待评估呼吸红外图像序列输入至所述分类器;根据所述分类器中二阶池化块和基于张量分解的网络层,获取所述每个待评估呼吸红外图像序列的目标特征图;根据所述目标特征图和所述分类器中基于张量分解的全连接层,对所述每个待评估呼吸红外图像序列进行睡眠质量评估,获得所述每个待评估呼吸红外图像序列对应的睡眠质量评估结果。
- 如权利要求11所述的终端设备,其特征在于,所述处理器执行所述计算机程序时还实现如下步骤:在通过分类器对所述多个待评估呼吸红外图像序列中每个待评估呼吸红外图像序列进行睡眠质量评估之前,通过张量化的三元生成对抗网络对所述分类器进行训练。
- 如权利要求13所述的终端设备,其特征在于,所述张量化的三元生成对抗网络包括生成器、所述分类器和判别器;所述处理器执行所述计算机程序时,所述通过张量化的三元生成对抗网络对所述分类器进行训练包括:将一维随机噪声和目标标签输入至所述生成器,通过所述生成器中基于张量分解的反卷积层,获得携带所述目标标签的第一呼吸红外图像序列;将所述第一呼吸红外图像序列输入至所述判别器,通过所述判别器中基于张量分解的网络层和全连接层,获得所述判别器对所述第一呼吸红外图像序列的判别结果;根据所述判别结果训练所述生成器;获取未携带标签的第二呼吸红外图像序列;将所述第二呼吸红外图像序列输入至所述分类器,通过所述分类器中二阶池化块、基于张量分解的网络层和全连接层,获得第三呼吸红外图像序列,所述第三呼吸红外图像序列是指携带标签的第二呼吸红外图像序列;获取携带标签的第四呼吸红外图像序列;根据所述第一呼吸红外图像序列、所述第三呼吸红外图像序列以及所述第四呼吸红外图像序列,训练所述判别器,并获取所述判别器对所述第三呼吸红外图像序列的判别结果;根据所述的第一呼吸红外图像序列、所述判别器对所述第三呼吸红外图像序列的判别结果以及所述第四呼吸红外图像序列,训练所述分类器。
- 如权利要求14所述的终端设备,其特征在于,所述处理器执行所述计算机程序时,所述通过生成器中基于张量分解的反卷积层,获得携带所述目标标签的第一呼吸红外图像序列包括:对所述生成器中反卷积层的反卷积核进行张量分解,获得第一张量;将所述一维随机噪声与所述第一张量进行反卷积计算,获得携带所述目标标签的第一呼吸红外图像序列。
- 如权利要求14所述的终端设备,其特征在于,所述判别器中基于张量分解的网络层包括基于张量分解的卷积层、第一密集连接块以及第二密集连接块;所述处理器执行所述计算机程序时,所述通过所述判别器中基于张量分解的网络层和全连接层,获得所述判别器对所述第一呼吸红外图像序列的判别结果包括:对所述判别器中卷积层的卷积核进行张量分解,获得第二张量;将所述第一呼吸红外图像序列与所述第二张量进行卷积计算,获得第一特征图;对所述判别器中第一密集连接块的卷积核进行张量分解,获得第三张量;将所述第一特征图与所述第三张量进行卷积计算,获得第二特征图;对所述判别器中第二密集连接块的卷积核进行张量分解,获得第四张量;将所述第二特征图与所述第四张量进行卷积计算,获得第三特征图;对所述判别器中全连接层的权重进行张量分解,获得第五张量;根据所述第三特征图和所述第五张量,获得所述判别器对所述第一呼吸红外图像序列的判别结果;其中,所述第一特征图、所述第二特征图以及所述第三特征图均为所述第一呼吸红外图像序列的特征图。
- 如权利要求16所述的终端设备,其特征在于,所述判别器还包括第一过渡层和第二过渡层,所述第一过渡层、所述第二过渡层均为1×1×1的卷积核;所述第一过渡层位于所述判别器中第一密集连接块与第二密集连接块之间,用于减少所述第二特征图的数量;所述第二过渡层位于所述判别器中第二密集连接块与全连接层之间,用于减少所述第三特征图的数量。
- 如权利要求14所述的终端设备,其特征在于,所述分类器中基于张量分解的网络层包括基于张量分解的卷积层、第三密集连接块以及第四密集连接块;所述处理器执行所述计算机程序时,所述通过所述分类器中二阶池化块、基于张量分解的网络层和全连接层,获得第三吸红外图像序列包括:对所述分类器中卷积层的卷积核进行张量分解,获得第六张量;将所述第二呼吸红外图像序列与所述第六张量进行卷积计算,获得第四特征图;对所述分类器中第三密集连接块的卷积核进行张量分解,获得第七张量;将所述第四特征图和所述第七张量进行卷积计算,获得第五特征图;通过所述分类器中二阶池化块处理所述第五特征图,获得第六特征图;对所述分类器中第四密集连接块的卷积核进行张量分解,获得第八张量;将所述第六特征图与所述第八张量进行卷积计算,获得第七特征图;对所述分类器中全连接层的权重进行张量分解,获得第九张量;根据所述第七特征图和所述第九张量,获得所述第三呼吸红外图像序列;其中,所述第四特征图、所述第五特征图、所述第六特征图以及所述第七特征图均为所述第二呼吸红外图像序列的特征图。
- 如权利要求18所述的终端设备,其特征在于,所述分类器还包括第三过渡层和第四过渡层,所述第三过渡层和所述第四过渡层均为1×1×1的卷积核;所述第三过渡层位于所述分类器中二阶池化块与第四密集连接块之间,用于减少所述第六特征图的数量;所述第四过渡层位于所述分类器中第四密集连接块与全连接层之间,用于减少所述第七特征图的数量。
- 一种计算机可读存储介质,所述计算机可读存储介质存储有计算机程序,其特征在于,所述计算机程序被处理器执行时实现如权利求1至9任一项所述的睡眠质量评估方法。
Priority Applications (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
PCT/CN2019/126038 WO2021120007A1 (zh) | 2019-12-17 | 2019-12-17 | 一种基于红外图像序列的睡眠质量评估系统和方法 |
US17/786,840 US20230022206A1 (en) | 2019-12-17 | 2019-12-17 | Infrared image sequence-based sleep quality evaluation system and method |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
PCT/CN2019/126038 WO2021120007A1 (zh) | 2019-12-17 | 2019-12-17 | 一种基于红外图像序列的睡眠质量评估系统和方法 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2021120007A1 true WO2021120007A1 (zh) | 2021-06-24 |
Family
ID=76476983
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/CN2019/126038 WO2021120007A1 (zh) | 2019-12-17 | 2019-12-17 | 一种基于红外图像序列的睡眠质量评估系统和方法 |
Country Status (2)
Country | Link |
---|---|
US (1) | US20230022206A1 (zh) |
WO (1) | WO2021120007A1 (zh) |
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN113610153A (zh) * | 2021-08-06 | 2021-11-05 | 长沙理工大学 | 人体红外图像识别方法、装置、计算机设备及存储介质 |
CN114699040A (zh) * | 2022-02-21 | 2022-07-05 | 华南师范大学 | 基于生理信号的醒睡检测方法、装置、设备以及存储介质 |
CN117351049A (zh) * | 2023-12-04 | 2024-01-05 | 四川金信石信息技术有限公司 | 热成像与可见光融合的测点配准引导方法、设备和介质 |
Families Citing this family (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US12073538B2 (en) * | 2021-04-08 | 2024-08-27 | Canon Medical Systems Corporation | Neural network for improved performance of medical imaging systems |
CN116386120B (zh) * | 2023-05-24 | 2023-08-18 | 杭州企智互联科技有限公司 | 一种用于智慧校园宿舍的无感监控管理系统 |
Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5704367A (en) * | 1995-03-28 | 1998-01-06 | Nihon Kohden Corporation | Respiration monitor for monitoring respiration based upon an image signal of a facial region |
CN101489478A (zh) * | 2006-06-01 | 2009-07-22 | 必安康医疗有限公司 | 用于监视生理症状的装置、系统和方法 |
CN102973273A (zh) * | 2012-11-29 | 2013-03-20 | 中国人民解放军第四军医大学 | 一种基于红外辐射检测的睡眠呼吸功能监测系统 |
CN104834946A (zh) * | 2015-04-09 | 2015-08-12 | 清华大学 | 一种非接触式睡眠监测方法及系统 |
CN106236013A (zh) * | 2016-06-22 | 2016-12-21 | 京东方科技集团股份有限公司 | 一种睡眠监测方法及装置 |
CN107280673A (zh) * | 2017-06-02 | 2017-10-24 | 南京理工大学 | 一种基于关键帧提取技术的红外成像呼吸信号检测方法 |
Family Cites Families (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US11734955B2 (en) * | 2017-09-18 | 2023-08-22 | Board Of Trustees Of Michigan State University | Disentangled representation learning generative adversarial network for pose-invariant face recognition |
CN110647973A (zh) * | 2018-06-27 | 2020-01-03 | 北京中科寒武纪科技有限公司 | 运算方法及相关方法和产品 |
WO2021112335A1 (ko) * | 2019-12-06 | 2021-06-10 | 주식회사 애자일소다 | 생성적 적대 신경망 기반의 분류 시스템 및 방법 |
-
2019
- 2019-12-17 US US17/786,840 patent/US20230022206A1/en active Pending
- 2019-12-17 WO PCT/CN2019/126038 patent/WO2021120007A1/zh active Application Filing
Patent Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5704367A (en) * | 1995-03-28 | 1998-01-06 | Nihon Kohden Corporation | Respiration monitor for monitoring respiration based upon an image signal of a facial region |
CN101489478A (zh) * | 2006-06-01 | 2009-07-22 | 必安康医疗有限公司 | 用于监视生理症状的装置、系统和方法 |
CN102973273A (zh) * | 2012-11-29 | 2013-03-20 | 中国人民解放军第四军医大学 | 一种基于红外辐射检测的睡眠呼吸功能监测系统 |
CN104834946A (zh) * | 2015-04-09 | 2015-08-12 | 清华大学 | 一种非接触式睡眠监测方法及系统 |
CN106236013A (zh) * | 2016-06-22 | 2016-12-21 | 京东方科技集团股份有限公司 | 一种睡眠监测方法及装置 |
CN107280673A (zh) * | 2017-06-02 | 2017-10-24 | 南京理工大学 | 一种基于关键帧提取技术的红外成像呼吸信号检测方法 |
Cited By (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN113610153A (zh) * | 2021-08-06 | 2021-11-05 | 长沙理工大学 | 人体红外图像识别方法、装置、计算机设备及存储介质 |
CN114699040A (zh) * | 2022-02-21 | 2022-07-05 | 华南师范大学 | 基于生理信号的醒睡检测方法、装置、设备以及存储介质 |
CN117351049A (zh) * | 2023-12-04 | 2024-01-05 | 四川金信石信息技术有限公司 | 热成像与可见光融合的测点配准引导方法、设备和介质 |
CN117351049B (zh) * | 2023-12-04 | 2024-02-13 | 四川金信石信息技术有限公司 | 热成像与可见光融合的测点配准引导方法、设备和介质 |
Also Published As
Publication number | Publication date |
---|---|
US20230022206A1 (en) | 2023-01-26 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2021120007A1 (zh) | 一种基于红外图像序列的睡眠质量评估系统和方法 | |
CN107392109A (zh) | 一种基于深度神经网络的新生儿疼痛表情识别方法 | |
Hasan et al. | SmartHeLP: Smartphone-based hemoglobin level prediction using an artificial neural network | |
Muhaba et al. | Automatic skin disease diagnosis using deep learning from clinical image and patient information | |
WO2021068781A1 (zh) | 一种疲劳状态识别方法、装置和设备 | |
Yang | Medical multimedia big data analysis modeling based on DBN algorithm | |
CN104636580A (zh) | 一种基于人脸的健康监控手机 | |
CN117542474A (zh) | 基于大数据的远程护理监测系统及方法 | |
CN112509696A (zh) | 基于卷积自编码器高斯混合模型的健康数据检测方法 | |
CN111954250A (zh) | 一种轻量级Wi-Fi行为感知方法和系统 | |
US11244456B2 (en) | System and method for image segmentation and digital analysis for clinical trial scoring in skin disease | |
Joshi et al. | Deep learning based person authentication using hand radiographs: A forensic approach | |
Hajabdollahi et al. | Segmentation of bleeding regions in wireless capsule endoscopy images an approach for inside capsule video summarization | |
CN111063438B (zh) | 一种基于红外图像序列的睡眠质量评估系统和方法 | |
Elshwemy et al. | A New Approach for Thermal Vision based Fall Detection Using Residual Autoencoder. | |
CN116469148A (zh) | 一种基于面部结构识别的概率预测系统及预测方法 | |
Bharathi et al. | Diabetes diagnostic method based on tongue image classification using machine learning algorithms | |
Soundrapandiyan et al. | AI-based wavelet and stacked deep learning architecture for detecting coronavirus (COVID-19) from chest X-ray images | |
Ramkumar et al. | Attention induced multi-head convolutional neural network organization with MobileNetv1 transfer learning and COVID-19 diagnosis using jellyfish search optimization process on chest X-ray images | |
Arai et al. | Human gait gender classification using 2D discrete wavelet transforms energy | |
CN111700592A (zh) | 一种癫痫脑电自动分类模型的获取方法、系统及分类系统 | |
CN114694234B (zh) | 情绪识别方法、系统、电子设备和存储介质 | |
Oz et al. | Efficacy of biophysiological measurements at FTFPs for facial expression classification: A validation | |
CN111798455A (zh) | 一种基于全卷积密集空洞网络的甲状腺结节实时分割方法 | |
Yudhana et al. | Glucose Content Analysis using Image Processing and Machine Learning Techniques |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 19956493 Country of ref document: EP Kind code of ref document: A1 |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 19956493 Country of ref document: EP Kind code of ref document: A1 |
|
32PN | Ep: public notification in the ep bulletin as address of the adressee cannot be established |
Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205A DATED 16.01.2023) |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 19956493 Country of ref document: EP Kind code of ref document: A1 |