CN113191171B - Pain intensity evaluation method based on feature fusion - Google Patents
Pain intensity evaluation method based on feature fusion Download PDFInfo
- Publication number
- CN113191171B CN113191171B CN202010034499.3A CN202010034499A CN113191171B CN 113191171 B CN113191171 B CN 113191171B CN 202010034499 A CN202010034499 A CN 202010034499A CN 113191171 B CN113191171 B CN 113191171B
- Authority
- CN
- China
- Prior art keywords
- cnn
- pain intensity
- features
- face
- image
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
- G06V40/174—Facial expression recognition
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/25—Fusion techniques
- G06F18/253—Fusion techniques of extracted features
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
Abstract
The invention provides a pain intensity assessment method based on feature fusion, and mainly relates to pain intensity regression of face image data by using a three-dimensional convolutional neural network (3D CNN) and a two-dimensional convolutional neural network (2D CNN). The method comprises the following steps: and carrying out affine transformation on the face data in the original data set to obtain image data with a fixed shape and without background information. And then, extracting dynamic time sequence features in the image sequence by using the 3D CNN, extracting static space features in the image by using the 2D CNN, fusing the features extracted in the two modes by a feature map splicing mode to serve as the input of the subsequent 3D CNN, and finally predicting the pain intensity by a regression mode. The method gives full play to the advantages of deep learning, and utilizes the convolutional neural networks with different dimensions to respectively extract the dynamic time sequence information and the static space information of the input data, thereby improving the accuracy of pain intensity regression.
Description
Technical Field
The invention relates to a pain intensity assessment problem in the field of deep learning, in particular to a pain intensity assessment method based on feature fusion.
Background
The advent of visual sensor data has created opportunities in many areas. Particularly in the field of clinical medicine, the health state information of the patient can be obtained by analyzing the visual sensing image data, and scientific basis is provided for the design of a related prototype system. By estimating the pain intensity of the patient from the visual sensing image data by using a deep learning method, a quantitative value of the pain intensity of the patient can be provided for medical staff.
The evaluation of pain intensity is an important research content in the field of computer vision, and is widely concerned by researchers at home and abroad, and the evaluation of pain intensity is also an important component of facial emotion recognition, so that the evaluation method has great research value. Currently, recent advances in deep learning provide opportunities for processing large volumes of medical visual sensory data. The patient image data acquired by the visual sensor is analyzed by utilizing a deep learning method, so that a foundation can be provided for constructing a corresponding intelligent system. Whereas existing pain intensity assessment datasets are video sequences taken from the facial expression of the patient. Thus, the patent uses a dynamic video sequence to assess the patient's facial pain intensity.
Deep Learning (Deep Learning) is a research field which is concerned in recent years, and combines a plurality of abstract data processing layers to form a calculation model to replace a traditional method for manually selecting features, so that a machine can autonomously learn the features of a data sample, and the defect of manually selecting the features is effectively avoided. Compared with the manual feature selection, the deep learning method utilizes a large amount of data to learn the features, and feature information describing the data can be described more abundantly. In short, deep learning, both in terms of recognition time and accuracy, is a great improvement over conventional methods.
Disclosure of Invention
The invention aims to provide a pain intensity evaluation method based on feature fusion, which introduces a three-dimensional convolutional neural network (3D CNN) and a two-dimensional convolutional neural network (2D CNN) in deep learning, respectively extracts dynamic time sequence features and static space features of input data, and then fuses the features extracted in the two modes. The problems that parameter adjustment of current shallow learning is difficult, accuracy is low and the like are effectively solved.
For convenience of explanation, the following concepts are first introduced:
two-dimensional Convolutional Neural Network (2-proportional Convolutional Neural Network,2D CNN): the convolutional neural network is designed based on the inspiration of a visual neural mechanism, is a multilayer feedforward neural network, each layer is composed of a plurality of two-dimensional planes, each neuron on each plane works independently, and the convolutional neural network mainly comprises a feature extraction layer and a feature mapping layer. The method is mainly used for feature extraction of a single image.
Three-dimensional Convolutional Neural Network (3-proportional Convolutional Neural Network,3D CNN): similar to a two-dimensional convolutional neural network, the method is mainly used for feature extraction of image sequences.
Feature Map (Feature Map): after the input data passes through the convolutional neural network, a multi-dimensional output is generated, and the output is called a feature map.
Feature fusion: and fusing the static spatial features extracted by the 2D CNN and the dynamic time sequence features extracted by the 3D CNN in a feature map splicing mode.
The invention specifically adopts the following technical scheme:
a pain intensity evaluation method based on feature fusion is provided, and the method is mainly characterized in that:
a. carrying out affine transformation on original image data and cutting to obtain a face image sequence without background information;
b. inputting the face image sequence obtained in the step a into a 3D CNN by taking 16 frames as a sample to extract time sequence features in the image sequence, and inputting the last frame of the sample into a 2D CNN to extract static spatial features of the image;
c. fusing the two characteristics obtained in the step b, inputting the obtained characteristics into a 3D CNN for higher-level characteristic learning, and obtaining a final pain intensity evaluation network model;
the method mainly comprises the following steps:
(1) preprocessing the used pain data set, wherein all faces in an original image sequence are fixed to the same shape through face affine transformation, background information in the original image data is removed, an image only containing face information is obtained, a face area is cut off, and the resolution ratio is set to be 112 x 112, so that a preprocessed pain data set is obtained;
(2) respectively extracting dynamic time sequence characteristics and static space characteristics in an input sample by using the 3D CNN and the 2D CNN;
(3) and (3) fusing the dynamic time sequence characteristics and the static spatial characteristics respectively extracted by the 3D CNN and the 2D CNN in the step (2), inputting the fused characteristics into the 3D CNN for higher-level characteristic learning, and finally performing regression prediction to obtain a final pain intensity prediction regression network model.
The invention has the beneficial effects that:
(1) the advantage of self-learning in the deep learning is fully developed, the machine can automatically learn the image characteristics, the problem of deviation and low efficiency of manually selecting the characteristics is effectively avoided, and the adaptive capacity is stronger.
(2) The dynamic time sequence characteristic of the input sample is fully extracted by using the 3D CNN, the static space characteristic of the sample is fully extracted by using the 2D CNN, the complementary advantages of the static characteristic and the dynamic characteristic are effectively combined, and the accuracy of the training effect is improved.
(3) The two features are fused together by means of feature map splicing and input into a subsequent 3D CNN for higher-level feature learning, and the final classification effect is improved.
(4) The deep learning and the pain intensity evaluation are combined, the problem that the accuracy rate of the traditional method is low is solved, and the research value is improved.
Drawings
FIG. 1 is an affine transformation process of a face image sequence in the present invention.
Fig. 2 shows the complete network structure of the present invention.
FIG. 3 illustrates the fusion of two different features of the present invention.
Detailed Description
The present invention is further described in detail with reference to the drawings and examples, it should be noted that the following examples are only for illustrating the present invention and should not be construed as limiting the scope of the present invention, and those skilled in the art should be able to make certain insubstantial modifications and adaptations to the present invention based on the above disclosure and should still fall within the scope of the present invention.
In fig. 2, the pain intensity assessment method based on feature fusion specifically includes the following steps:
(1) the original pain data set was affine transformed in the manner shown in fig. 1, and the resulting face image resolution was set to 112 × 112.
(2) The method comprises the steps of respectively extracting dynamic time sequence characteristics and static space characteristics of an image by using 3D CNN and 2D CNN, using 16 continuous frames in preprocessed data as a sample, extracting the dynamic time sequence characteristics in the sample by using the 3D CNN and extracting the static space characteristics of the last frame in the sample by using the 2D CNN, fusing two different characteristics together by using a characteristic diagram fusion mode of FIG. 3, sending the fused characteristics into the subsequent 3D CNN for higher-level characteristic learning, and finally realizing the evaluation of pain intensity.
(3) Training: the network structure given in fig. 2 is used for training. As used herein, a data set contains 25 objects in total, and thus the training strategy is to use the data contained in one of the objects as a test set, the data contained in the other 24 objects as a training set, and so on until all the objects have been used as test sets.
(4) After step (3), 25 sets of experimental results are obtained, and the final predicted result is obtained after averaging.
Claims (4)
1. A pain intensity assessment method based on feature fusion is characterized in that:
a. carrying out affine transformation on the original image data and cutting to obtain a face image sequence without background information;
b. inputting the face image sequence obtained in the step a into a 3D CNN by taking 16 frames as a sample to extract time sequence features in the image sequence, and inputting the last frame of the sample into a 2D CNN to extract static spatial features of the image;
c. fusing the two characteristics obtained in the step b, inputting the obtained characteristics into a 3D CNN for higher-level characteristic learning, and obtaining a final pain intensity evaluation network model;
the method mainly comprises the following steps:
(1) preprocessing the used pain data set, wherein all faces in an original image sequence are fixed to the same shape through face affine transformation, background information in the original image data is removed, an image only containing face information is obtained, a face area is cut off, and the resolution ratio is set to be 112 x 112, so that a preprocessed pain data set is obtained;
(2) respectively extracting dynamic time sequence characteristics and static space characteristics in an input sample by using the 3D CNN and the 2D CNN;
(3) and (3) fusing the dynamic time sequence characteristics and the static spatial characteristics respectively extracted by the 3D CNN and the 2D CNN in the step (2), inputting the fused characteristics into the 3D CNN for higher-level characteristic learning, and finally performing regression prediction to obtain a final pain intensity prediction regression network model.
2. The method according to claim 1, wherein the step (1) comprises fixing all images in the original image sequence to the same shape by using human face affine transformation and removing background information in the original image data to obtain an image containing only human face information.
3. The method according to claim 1, wherein the 3D CNN and the 2D CNN are used to extract dynamic temporal features and static spatial features in the input sample in step (2).
4. The method according to claim 1, wherein the dynamic temporal features and the static spatial features extracted from the 3D CNN and the 2D CNN respectively are fused in step (3), and the fused features are input into the subsequent 3D CNN for higher-level feature learning.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202010034499.3A CN113191171B (en) | 2020-01-14 | 2020-01-14 | Pain intensity evaluation method based on feature fusion |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202010034499.3A CN113191171B (en) | 2020-01-14 | 2020-01-14 | Pain intensity evaluation method based on feature fusion |
Publications (2)
Publication Number | Publication Date |
---|---|
CN113191171A CN113191171A (en) | 2021-07-30 |
CN113191171B true CN113191171B (en) | 2022-06-17 |
Family
ID=76972291
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202010034499.3A Active CN113191171B (en) | 2020-01-14 | 2020-01-14 | Pain intensity evaluation method based on feature fusion |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN113191171B (en) |
Citations (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN106682616A (en) * | 2016-12-28 | 2017-05-17 | 南京邮电大学 | Newborn-painful-expression recognition method based on dual-channel-characteristic deep learning |
CN107679491A (en) * | 2017-09-29 | 2018-02-09 | 华中师范大学 | A kind of 3D convolutional neural networks sign Language Recognition Methods for merging multi-modal data |
WO2018126275A1 (en) * | 2016-12-30 | 2018-07-05 | Dirk Schneemann, LLC | Modeling and learning character traits and medical condition based on 3d facial features |
CN108898606A (en) * | 2018-06-20 | 2018-11-27 | 中南民族大学 | Automatic division method, system, equipment and the storage medium of medical image |
CN109815785A (en) * | 2018-12-05 | 2019-05-28 | 四川大学 | A kind of face Emotion identification method based on double-current convolutional neural networks |
CN109903255A (en) * | 2019-03-04 | 2019-06-18 | 北京工业大学 | A kind of high spectrum image Super-Resolution method based on 3D convolutional neural networks |
CN109993151A (en) * | 2019-04-15 | 2019-07-09 | 方玉明 | A kind of 3 D video visual attention detection method based on the full convolutional network of multimode |
Family Cites Families (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9031293B2 (en) * | 2012-10-19 | 2015-05-12 | Sony Computer Entertainment Inc. | Multi-modal sensor based emotion recognition and emotional interface |
-
2020
- 2020-01-14 CN CN202010034499.3A patent/CN113191171B/en active Active
Patent Citations (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN106682616A (en) * | 2016-12-28 | 2017-05-17 | 南京邮电大学 | Newborn-painful-expression recognition method based on dual-channel-characteristic deep learning |
WO2018126275A1 (en) * | 2016-12-30 | 2018-07-05 | Dirk Schneemann, LLC | Modeling and learning character traits and medical condition based on 3d facial features |
CN107679491A (en) * | 2017-09-29 | 2018-02-09 | 华中师范大学 | A kind of 3D convolutional neural networks sign Language Recognition Methods for merging multi-modal data |
CN108898606A (en) * | 2018-06-20 | 2018-11-27 | 中南民族大学 | Automatic division method, system, equipment and the storage medium of medical image |
CN109815785A (en) * | 2018-12-05 | 2019-05-28 | 四川大学 | A kind of face Emotion identification method based on double-current convolutional neural networks |
CN109903255A (en) * | 2019-03-04 | 2019-06-18 | 北京工业大学 | A kind of high spectrum image Super-Resolution method based on 3D convolutional neural networks |
CN109993151A (en) * | 2019-04-15 | 2019-07-09 | 方玉明 | A kind of 3 D video visual attention detection method based on the full convolutional network of multimode |
Non-Patent Citations (4)
Title |
---|
Facial Expression Recognition Based on Group Domain Random Frame Extraction;Wenjun Zhou 等;《International Conference on Image and Graphics》;20191231;467-479 * |
Mei Chee Leong 等.Semi-CNN Architecture for Effective Spatio-Temporal Learning in Action Recognition.《applied sciences》.2020,1-14. * |
二维和三维卷积神经网络相结合的CT图像肺结节检测方法;苗光 等;《激光与光电子学进展》;20171201;051006-1-9 * |
结合注意力的3D卷积网络脑胶质瘤分割算法;胡睿 等;《计算机工程与应用》;20190904;第56卷(第12期);187-192 * |
Also Published As
Publication number | Publication date |
---|---|
CN113191171A (en) | 2021-07-30 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2020215985A1 (en) | Medical image segmentation method and device, electronic device and storage medium | |
CN107679522B (en) | Multi-stream LSTM-based action identification method | |
CN109846469B (en) | Non-contact heart rate measurement method based on convolutional neural network | |
CN109902558B (en) | CNN-LSTM-based human health deep learning prediction method | |
CN109920538B (en) | Zero sample learning method based on data enhancement | |
CN113724206B (en) | Fundus image blood vessel segmentation method and system based on self-supervision learning | |
CN109063643B (en) | Facial expression pain degree identification method under condition of partial hiding of facial information | |
CN110859624A (en) | Brain age deep learning prediction system based on structural magnetic resonance image | |
Wei et al. | Real-time facial expression recognition for affective computing based on Kinect | |
CN110477907B (en) | Modeling method for intelligently assisting in recognizing epileptic seizures | |
CN113781640A (en) | Three-dimensional face reconstruction model establishing method based on weak supervised learning and application thereof | |
Liu et al. | PRA-Net: Part-and-Relation Attention Network for depression recognition from facial expression | |
WO2023173804A1 (en) | Brain-computer information fusion classification method and system for shared subspace learning | |
CN112288749A (en) | Skull image segmentation method based on depth iterative fusion depth learning model | |
CN110889335B (en) | Human skeleton double interaction behavior identification method based on multichannel space-time fusion network | |
CN113486700A (en) | Facial expression analysis method based on attention mechanism in teaching scene | |
CN113191171B (en) | Pain intensity evaluation method based on feature fusion | |
CN103235943A (en) | Principal component analysis-based (PCA-based) three-dimensional (3D) face recognition system | |
CN112634293A (en) | Temporal bone inner ear bone cavity structure automatic segmentation method based on coarse-to-fine dense coding and decoding network | |
CN116993699A (en) | Medical image segmentation method and system under eye movement auxiliary training | |
CN116091793A (en) | Light field significance detection method based on optical flow fusion | |
CN116092189A (en) | Bimodal human behavior recognition method based on RGB data and bone data | |
CN115601840A (en) | Behavior disorder detection method considering vision and plantar pressure multi-mode sensing | |
CN116115240A (en) | Electroencephalogram emotion recognition method based on multi-branch chart convolution network | |
Yu et al. | Research on 3D Medical Image Segmentation based on improved 3D-Unet |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |