CN111160443B - Activity and user identification method based on deep multitasking learning - Google Patents
Activity and user identification method based on deep multitasking learning Download PDFInfo
- Publication number
- CN111160443B CN111160443B CN201911355355.1A CN201911355355A CN111160443B CN 111160443 B CN111160443 B CN 111160443B CN 201911355355 A CN201911355355 A CN 201911355355A CN 111160443 B CN111160443 B CN 111160443B
- Authority
- CN
- China
- Prior art keywords
- user
- activity
- network
- recognition
- representing
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
- 230000000694 effects Effects 0.000 title claims abstract description 109
- 238000000034 method Methods 0.000 title claims abstract description 39
- 230000007246 mechanism Effects 0.000 claims abstract description 5
- 238000005457 optimization Methods 0.000 claims abstract description 3
- 238000007781 pre-processing Methods 0.000 claims abstract description 3
- 238000012549 training Methods 0.000 claims description 33
- 239000013598 vector Substances 0.000 claims description 22
- 230000002457 bidirectional effect Effects 0.000 claims description 11
- 230000006870 function Effects 0.000 claims description 11
- 238000012545 processing Methods 0.000 claims description 9
- 230000008569 process Effects 0.000 claims description 8
- 238000013507 mapping Methods 0.000 claims description 6
- 238000013528 artificial neural network Methods 0.000 claims description 5
- 238000013527 convolutional neural network Methods 0.000 claims description 5
- 238000004364 calculation method Methods 0.000 claims description 4
- 238000000354 decomposition reaction Methods 0.000 claims description 4
- 238000009826 distribution Methods 0.000 claims description 3
- ORILYTVJVMAKLC-UHFFFAOYSA-N Adamantane Natural products C1C(C2)CC3CC1CC2C3 ORILYTVJVMAKLC-UHFFFAOYSA-N 0.000 claims description 2
- 239000011159 matrix material Substances 0.000 claims description 2
- 239000010410 layer Substances 0.000 claims 18
- 230000000875 corresponding effect Effects 0.000 claims 4
- 239000011229 interlayer Substances 0.000 claims 1
- 238000010586 diagram Methods 0.000 description 7
- 238000010276 construction Methods 0.000 description 5
- 238000011176 pooling Methods 0.000 description 5
- 230000006399 behavior Effects 0.000 description 4
- 238000010606 normalization Methods 0.000 description 3
- 230000002159 abnormal effect Effects 0.000 description 2
- 230000008859 change Effects 0.000 description 2
- 238000001514 detection method Methods 0.000 description 2
- 230000008030 elimination Effects 0.000 description 2
- 238000003379 elimination reaction Methods 0.000 description 2
- 230000036541 health Effects 0.000 description 2
- 230000004913 activation Effects 0.000 description 1
- 230000006978 adaptation Effects 0.000 description 1
- 238000007792 addition Methods 0.000 description 1
- 238000013459 approach Methods 0.000 description 1
- 230000015556 catabolic process Effects 0.000 description 1
- 238000012512 characterization method Methods 0.000 description 1
- 238000006243 chemical reaction Methods 0.000 description 1
- 238000006731 degradation reaction Methods 0.000 description 1
- 239000006185 dispersion Substances 0.000 description 1
- 230000002708 enhancing effect Effects 0.000 description 1
- 238000011156 evaluation Methods 0.000 description 1
- 230000003993 interaction Effects 0.000 description 1
- 238000002372 labelling Methods 0.000 description 1
- 239000000203 mixture Substances 0.000 description 1
- 230000008447 perception Effects 0.000 description 1
- 238000006467 substitution reaction Methods 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/24—Classification techniques
- G06F18/241—Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
- G06F18/2415—Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches based on parametric or probabilistic models, e.g. based on likelihood ratio or false acceptance rate versus a false rejection rate
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/044—Recurrent networks, e.g. Hopfield networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y02—TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
- Y02D—CLIMATE CHANGE MITIGATION TECHNOLOGIES IN INFORMATION AND COMMUNICATION TECHNOLOGIES [ICT], I.E. INFORMATION AND COMMUNICATION TECHNOLOGIES AIMING AT THE REDUCTION OF THEIR OWN ENERGY USE
- Y02D10/00—Energy efficient computing, e.g. low power processors, power management or thermal management
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- Data Mining & Analysis (AREA)
- Evolutionary Computation (AREA)
- Life Sciences & Earth Sciences (AREA)
- Artificial Intelligence (AREA)
- General Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- Computing Systems (AREA)
- Software Systems (AREA)
- Molecular Biology (AREA)
- Computational Linguistics (AREA)
- Biophysics (AREA)
- Biomedical Technology (AREA)
- Mathematical Physics (AREA)
- General Health & Medical Sciences (AREA)
- Health & Medical Sciences (AREA)
- Probability & Statistics with Applications (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Bioinformatics & Computational Biology (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Evolutionary Biology (AREA)
- Image Analysis (AREA)
Abstract
The invention discloses an activity and user identification method based on deep multitask learning, which comprises the following steps: (1) Collecting sensor data of each user during activities by using a wearable sensor, and preprocessing; (2) Jointly constructing an activity recognition network and a user recognition network for predicting activities and users, wherein the activity recognition network and the user recognition network share part hidden parameters and introduce a mutual attention mechanism, and weighting each part of the characteristics by utilizing knowledge learned by the other party; (3) Constructing joint loss to perform collaborative optimization on the activity recognition network and the user recognition network to obtain an activity recognition model and a user recognition model with determined parameters; (4) And inputting the preprocessed sensor data into the activity recognition model and the user recognition model, obtaining an activity recognition result by using the activity recognition model, and obtaining a user recognition result by using the user recognition model. The method improves activity and user recognition capability.
Description
Technical Field
The invention relates to the field of activity recognition and user recognition, in particular to an activity and user recognition method based on deep multitask learning.
Background
The method for deducing the activity and the user information based on the activity of the wearable sensor and the user identification by utilizing the data acquired by the sensor worn by the user is two key tasks in the fields of pervasive and mobile computing, and has important significance for realizing the computation with human centers. The wearable sensor-based activity and user identification can provide support for health support, skill assessment, biometric identification, and other applications.
Most wearable sensor-based activity recognition methods mix labeling samples of all training users, build an activity classifier using supervised learning methods, and use it directly for new users, ignoring differences in different user behavior patterns. Related studies have shown that different users have different patterns of behavior, which means that there is a distribution difference between the sensor data acquired from different users, and therefore, when an activity recognition model that performs well on a training user is directly applied to a new user, a significant performance degradation tends to occur. How to guarantee the recognition performance of new users becomes an important challenge for activity recognition based on wearable sensors. To cope with this challenge, several methods have been proposed at home and abroad, which can be roughly divided into two categories: the first type of method builds user-independent features for modeling, so that the generalization capability of the model on new users is guaranteed. Such methods have difficulty in fully utilizing the user information in the training data, and impair their activity characterization capabilities while enhancing feature user independence. The second class of methods builds a custom model for each user for its behavior pattern. Such methods often require data acquisition and model adaptation for each new user, which is limited in use at a high cost.
Most of the user identification methods based on the wearable sensors only support the identification of users in the walking activity scene, and although the methods can achieve good identification effect, the limited applicable scene limits the use of the methods in daily life. Extending the activity scenarios supported by wearable sensor-based user identification from walking to other daily activities is valuable, but also challenging, because there are significant differences in sensor data from activity scenario to activity scenario.
Existing approaches typically model activity recognition or user recognition tasks alone, ignoring the correlation between the two tasks. The activity related information acquired by the activity recognition task is helpful for the user to recognize task perception and adapt to different activity scenes; the user-related information learned by the user identification task enables the activity identification model to take into account the current user's behavior pattern when identifying human activity.
Disclosure of Invention
The invention aims to solve the technical problem of improving the generalization capability of an identification model by utilizing the correlation between two tasks of activity identification and user identification.
In order to solve the above problems, the present invention provides an activity and user identification method based on deep multitasking learning, comprising the following steps:
(1) Collecting sensor data of each user during activities by using a wearable sensor, and preprocessing;
(2) Jointly constructing an activity recognition network and a user recognition network for predicting activities and users, wherein the activity recognition network and the user recognition network share part hidden parameters and introduce a mutual attention mechanism, and weighting each part of the characteristics by utilizing knowledge learned by the other party;
(3) Constructing joint loss to perform collaborative optimization on the activity recognition network and the user recognition network to obtain an activity recognition model and a user recognition model with determined parameters;
(4) And inputting the preprocessed sensor data into the activity recognition model and the user recognition model, obtaining an activity recognition result by using the activity recognition model, and obtaining a user recognition result by using the user recognition model.
Compared with the prior method, the method has the advantages that:
1) And jointly constructing an activity recognition model and a user recognition model, and utilizing hidden parameters to carry out cross-task knowledge sharing so as to realize interaction promotion by utilizing the commonality and the difference between the two tasks.
2) A mutual attention mechanism is introduced between the activity recognition model and the user recognition model, so that the two models can mutually utilize knowledge learned by each other to give weight to each part of the feature, thereby adapting to the change of the user and the activity scene.
Drawings
In order to more clearly illustrate the embodiments of the present invention or the technical solutions in the prior art, the drawings that are required in the embodiments or the description of the prior art will be briefly described, and it is obvious that the drawings in the following description are only some embodiments of the present invention, and other drawings may be obtained according to these drawings without inventive effort for a person skilled in the art.
FIG. 1 is a general flow diagram of an activity and user identification method based on deep multitasking learning provided by an embodiment;
FIG. 2 is a general architecture diagram of an activity and user identification method based on deep multitasking learning provided by an embodiment;
FIG. 3 is a schematic diagram of a convolutional neural network architecture provided by an embodiment;
fig. 4 is a schematic diagram of a bidirectional long and short time memory network according to an embodiment.
Detailed Description
The present invention will be described in further detail with reference to the drawings and examples, in order to make the objects, technical solutions and advantages of the present invention more apparent. It should be understood that the detailed description is presented by way of example only and is not intended to limit the scope of the invention.
FIG. 1 is a general flow diagram of an activity and user identification method based on deep multitasking learning provided by an embodiment; fig. 2 is an overall architecture diagram of an activity and user recognition method based on deep multitasking learning provided in an embodiment, in fig. 2, conv1D and Pool1D represent a one-dimensional convolutional network layer and a one-dimensional max-pooling layer, respectively, U represents a bidirectional long-short-time memory network element, attNet represents an attention network, which outputs a weight vector w to distinguish the importance of a feature r,representing Hadamard sum, ">Representing the hadamard product.
Referring to fig. 1 and 2, the activity and user recognition method based on deep multitasking learning provided in the embodiment includes a data set construction stage, a hidden parameter initialization stage, and a model training stage.
Data set construction phase
The data set construction stage mainly collects sensor data, preprocesses the collected sensor data, and constructs a training data set, and the specific process is as follows:
step 1-1, acquiring sensor data of each user when different activities are performed by using a wearable sensor;
in this step, wearable sensor data is recorded as the user performs different activities. Common wearable sensors include accelerometers, gyroscopes, magnetometers, and the like. In order to ensure the recognition accuracy, the number, the type and the placement orientation of the sensors are required to be consistent when the training data set is constructed and when the training data set is actually used.
And step 1-2, carrying out outlier elimination and normalization processing on the acquired data, and dividing the processed data by utilizing a sliding window to obtain a data sample.
In this step, the processing of the acquired data mainly includes:
a) Abnormal value detection is carried out on the original sensor data, and invalid values (such as values exceeding a normal range and missing values) in the abnormal value detection are removed or mean filling is carried out.
b) Carrying out dispersion standardization processing on sensor data according to channels, so that the processed data are standardized to be within an [ -1,1] interval, and the conversion formula is as follows:
wherein X is the original value, X min To be the minimum value of the channel where the value is located, X max For the maximum value of the channel in which the value is located, x' is the value after normalization.
c) The processed data was divided using a fixed length sliding window, the sliding window length was set manually according to experience, and the overlap was set to 50%.
And step 1-3, taking each data sample, the corresponding active label and the user label as a training sample, and constructing a training data set. One training sample may be denoted (x, a, u), where x represents the data sample, a represents the active tag, and u represents the user tag.
Step 1-4, the training data set is batched according to a fixed size, and the total number of batches is N.
In this step, the batch size is manually set according to experience, and the total number of batches N is calculated as follows:
wherein M is total To train the total number of samples in the dataset, M batch Is of a batch size.
Hidden parameter initialization stage
The specific process is as follows:
step 2-1, sequentially selecting a batch of training samples from the training dataset, and repeating steps 2-2 to 2-5 for each training sample (x, a, u) in the batch.
Step 2-2, respectively processing the data sample x by using two independent convolutional neural networks in the active recognition network and the user recognition network to obtain an intermediate layer characteristic representation r of the sample a And r u 。
In this step, the two convolutional neural networks have the same network architecture and are respectively composed of three convolutional network layers and one max-pooling layer, as shown in fig. 3, wherein the convolutional operation and the max-pooling operation are performed along the time dimension. In fig. 3, conv1D and Pool1D represent a one-dimensional convolutional network layer and a one-dimensional max-pooling layer, respectively, the numbers before @ represent the size of the convolutional kernel or pooling region, the numbers after @ represent the number of generated feature maps, and ReLU represents the activation function.
Step 2-3, representing the intermediate layer characteristics as r a And r u The vector sequence with the length of l is expanded and is respectively input into two independent bidirectional long-short-time memory networks in the activity recognition network and the user recognition network to extract the time dependency relationship to obtain the characteristic representation vector sequenceAnd->
In this step, the intermediate layer features represent r a And r u The time dimension of (c) remains unchanged and the other dimensions are expanded. The two bidirectional long-short-time memory networks have the same network architecture, and each bidirectional long-short-time memory network layer comprises a bidirectional long-short-time memory network layer which is shown in a time expansion schematic diagram in fig. 4. In FIG. 4, L f And L b Respectively representing a forward long-short time memory network unit and a backward long-short time memory network unit, wherein the two long-short time memory network layers are opposite in directionThe short-time memory network layer is connected to the same output, one long-time memory network layer processes the input sequence in the forward direction, the other one processes in the backward direction, and the output can acquire the time dependence relationship between the forward direction and the backward direction at the same time.
Step 2-4, using two independent attention networks in the activity recognition network and the user recognition network to represent the feature vector sequence e a Mapping to weight vectorsRepresenting the vector sequence e for the feature u Each part is given weight; representing the feature by a vector sequence e u Mapping to weight vector->Representing the vector sequence e for the feature a Each portion is given a weight.
In this step, the two attention networks have the same network architecture, and each includes a fully connected network layer, and the output of the fully connected network layer is normalized by a softmax function to obtain a weight vector, and the process is as shown in formulas (3) - (8):
ω a =AttNet a (e u ) (3)
ω u =AttNet u (e a ) (6)
wherein the method comprises the steps of,AttNet u (. Cndot.) and AttNet a (. Cndot.) represents the attention network, ω u And omega a Representing non-normalized weight vector, alpha a And alpha u Is a weighted sum of the feature representations.
Step 2-5, utilizing an activity classification output layer formed from fully-connected network layer and softmax function in the activity recognition network, according to the weighted sum alpha of characteristic representation a Obtaining an activity prediction result a'; using a user classification output layer in the user identification network, which consists of a fully connected network layer and a softmax function, a weighted sum alpha of feature representations is used u Obtaining a user prediction result u'.
In this step, the activity and user classification layer outputs probability distributions of different activity labels and different user labels, and the calculation formula is as follows:
wherein alpha is a And alpha u Representing the input of the last step, θ a And theta u Parameters representing the fully connected network layer, |a| represents the number of active labels, |u| represents the number of user labels, a ′ i Representing the probability of predicting an active tag as i, u ′ i Representing the probability of predicting the user tag as i.
And 2-6, calculating activity classification loss according to the activity prediction and the activity label for all training samples of the batch, calculating user classification loss according to the user prediction and the user label, and updating parameters of each part of the network according to the loss.
In this step, the loss is calculated using a cross entropy loss function, the calculation formula is as follows:
wherein L is a And L u All representing loss of a single sample, M batch Indicating the batch size, loss indicates the joint Loss of a batch of samples. Adam algorithm is used for network parameter update.
Step 2-7, if the specified iteration times are reached, entering step 2-8, otherwise returning to step 2-1.
And 2-8, stacking convolution kernels of the updated two convolution neural networks corresponding to the convolution network layers to form a high-order tensor, and decomposing by a Tucker to obtain an initial hidden parameter.
In this step, the convolution kernel of each convolution network layer may be represented as a d 1 ×d 2 ×d 3 The convolution kernels of the corresponding convolution network layers of the two convolution neural networks are stacked to form d by the dimension third-order tensor 1 ×d 2 ×d 3 ×d 4 (d 4 =2) a four-dimensional tensor T, which is subjected to a Tucker decomposition to obtain the initial hidden parameters, as shown in equation (14):
T=S· (1,2) R (1) · (1,2) R (2) · (1,2) R (3) · (1,2) R (4) (14)
wherein S is a v 1 ×v 2 ×v 3 ×v 4 Dimension fourth-order tensor, R (i) Is d i ×v i The dimension matrix, S and R are initial hidden parameters, R (4) Hidden parameters, S, R, specific to the identification network of the activity and the identification network of the user (1) ,R (2) And R is (3) Hidden parameters representing the sharing of an active identification network and a user identification network, subscript · (i,j) Representing coordinate axes for performing tensor dot product operations。
Model training stage
Step 3-1, sequentially selecting a batch of training samples from the training dataset, and repeating steps 3-2 to 3-3 for each training sample (x, a, u) in the batch.
Step 3-2, respectively processing the data sample x by using two convolutional neural networks sharing part of hidden parameters in the active recognition network and the user recognition network to obtain an intermediate layer characteristic representation r of the sample a And r u . The hidden parameters are stored and updated instead of the convolution kernels, specifically, in forward propagation, the hidden parameters are firstly used for generating the convolution kernels, and then the generated convolution kernels are used for carrying out convolution operation; in back propagation, hidden parameters are updated.
In this step, a corresponding convolution kernel is generated from the hidden parameter by inverse operation of the Tucker decomposition, and a convolution operation is performed using the generated convolution kernel. Wherein, part of hidden parameters are shared by the activity recognition network and the user recognition network, and when one network is trained, the updating of the shared hidden parameters also affects the convolution kernel generated by the other network, thereby realizing the cross-task sharing of knowledge.
Step 3-3, obtaining an activity prediction result a 'and a user prediction result u' according to the steps 2-3-2-5; namely, specifically comprises:
step 3-3-1, r a And r u Expanded into a vector sequence with the length of l, and respectively input into two independent bidirectional long-short-time memory networks in the activity recognition network and the user recognition network to obtain a characteristic representation vector sequenceAnd
the implementation details of the step are consistent with those of the step 2-3 in the hidden parameter initialization stage.
Step 3-3-2, using two independent attention networks of the activity recognition network and the user recognition network, e a Mapping to weight vectorsE is u Each part is given weight; will e u Mapping to weight vector->E is a Each portion is given a weight.
The implementation details of the step are consistent with those of the step 2-4 in the hidden parameter initialization stage.
Step 3-3-3, obtaining an activity prediction result a' by using an activity classification output layer which is formed by a fully connected network layer and a softmax function in an activity recognition network; and obtaining a user prediction result u' by using a user classification output layer in the user identification network, wherein the user classification output layer consists of a fully connected network layer and a softmax function.
The implementation details of the step are consistent with those of the step 2-5 in the hidden parameter initialization stage.
And 3-4, updating parameters of each part of the network according to the steps 2-6.
In this step, for all training samples of the batch, activity classification losses are calculated according to the activity predictions and the activity labels, user classification losses are calculated according to the user predictions and the user labels, and parameters of each part of the network are updated according to the losses.
The implementation details of the step are consistent with those of the step 2-6 in the hidden parameter initialization stage.
And 3-5, if the specified training iteration times are reached, finishing training, obtaining an active recognition model and a user recognition model with determined parameters, and otherwise, returning to the step 3-1.
After the network parameters of the activity recognition model and the user recognition model are determined through the data set construction stage, the hidden parameter initialization stage and the model training stage, the activity recognition model and the user recognition model can be utilized for carrying out activity and user recognition, and the specific process is as follows:
for the acquired sensor data, performing outlier elimination and normalization processing on the data according to the steps 1-2 in the data set construction stage, and dividing the processed data to obtain a data sample;
processing the data sample by using an activity recognition model determined by the network parameters to obtain an activity recognition result;
and processing the data sample by using the user identification model determined by the network parameters to obtain a user identification result.
The invention jointly builds the activity recognition and user recognition models, utilizes hidden parameters to share information between the two models, further utilizes the commonality and the difference between the two tasks to realize mutual promotion, introduces a mutual attention mechanism, enables the two models to mutually utilize the knowledge learned by the other party to give weight to each part of the feature, thereby adapting to the change of the user and the activity scene, and has wide application prospect in the fields of health support, skill evaluation, biological recognition and the like.
The foregoing detailed description of the preferred embodiments and advantages of the invention will be appreciated that the foregoing description is merely illustrative of the presently preferred embodiments of the invention, and that no changes, additions, substitutions and equivalents of those embodiments are intended to be included within the scope of the invention.
Claims (5)
1. An activity and user identification method based on deep multitasking learning, comprising the following steps:
(1) Acquiring sensor data of each user during activities by using a wearable sensor, preprocessing to obtain a data sample x, taking the data sample x, a corresponding activity label a and a corresponding user label u as training samples, and constructing a training sample set;
(2) Selecting a portion of the training samples from the training sample set to jointly construct an activity recognition network and a user recognition network for predicting activities and users, wherein the activity recognition network and the user recognition network share a portion of hidden parameters and introduce a mutual attention mechanism, and weighting each portion of the features by utilizing knowledge learned by the other party, comprising:
(2-1) selecting a portion of the training samples from the training sample set;
(2-2) Using two independent volumes in an Activity recognition network and a subscriber recognition networkThe data samples x are processed by the neural network respectively to obtain an intermediate layer characteristic representation r of the samples a And r u ;
(2-3) representing the intermediate layer characteristics r a And r u The vector sequence with the length of l is expanded and is respectively input into two independent bidirectional long-short-time memory networks in the activity recognition network and the user recognition network to extract the time dependency relationship to obtain the characteristic representation vector sequenceAnd->Wherein the interlayer features represent r a And r u The time dimension of the memory is kept unchanged, other dimensions are expanded, the two bidirectional long-short-time memory networks have the same network architecture and both comprise a bidirectional long-short-time memory network layer, the bidirectional long-short-time memory network layer connects the two long-short-time memory network layers with opposite directions to the same output, one long-short-time memory network layer processes an input sequence in a forward direction, the other long-short-time memory network layer processes in a backward direction, and the output can acquire the time dependence relationship of the forward direction and the backward direction at the same time;
(2-4) representing the feature vector sequence e with two independent attention networks of the activity recognition network and the user recognition network a Mapping to weight vectorsRepresenting the vector sequence e for the feature u Each part is given weight; representing the feature by a vector sequence e u Mapping to weight vector->Representing the vector sequence e for the feature a Each part is given weight;
(2-5) utilizing an activity classification output layer in the activity recognition network, which consists of a fully connected network layer and a softmax function, according to a weighted sum alpha of feature representations a Obtaining an activity prediction result a'; utilization ofUser classification output layer in user identification network composed of fully connected network layer and softmax function according to weighted sum alpha of characteristic representation u Obtaining a user prediction result u';
(2-6) calculating an activity classification loss according to the activity prediction and the activity label, calculating a user classification loss according to the user prediction and the user label, and updating parameters of each part of the network according to the loss;
(2-7) stacking convolution kernels of the updated two convolution neural networks corresponding to the convolution network layers to form a higher-order tensor, and obtaining an initial hidden parameter through a Tucker decomposition, wherein the method comprises the following steps:
the convolution kernel of each convolution network layer may be represented as a d 1 ×d 2 ×d 3 The convolution kernels of the corresponding convolution network layers of the two convolution neural networks are stacked to form d by the dimension third-order tensor 1 ×d 2 ×d 3 ×d 4 A dimension fourth-order tensor T, where d 4 =2, performing a Tucker decomposition on the tensor T to obtain an initial hidden parameter, as shown in equation (1):
T=S· (1,2) R (1) · (1,2) R (2) · (1,2) R (3) · (1,2) R (4) (1)
wherein S is a v 1 ×v 2 ×v 3 ×v 4 Dimension fourth-order tensor, R (i) Is d i ×v i The dimension matrix, S and R are initial hidden parameters, R (4) Hidden parameters, S, R, specific to the identification network of the activity and the identification network of the user (1) ,R (2) And R is (3) Hidden parameters representing the sharing of an active identification network and a user identification network, subscript · (i,j) Representing coordinate axes for performing tensor dot product operations;
(3) Selecting part of training samples from the training sample set to construct a joint loss to carry out collaborative optimization on the active recognition network and the user recognition network, so as to obtain an active recognition model and a user recognition model with determined parameters;
(4) And inputting the preprocessed sensor data into the activity recognition model and the user recognition model, obtaining an activity recognition result by using the activity recognition model, and obtaining a user recognition result by using the user recognition model.
2. The deep multitasking learning-based activity and user recognition method of claim 1 in which in step (2-4) both attention networks have the same network architecture and each comprise a fully connected network layer whose outputs are normalized by a softmax function to obtain weight vectors as shown in equations (2) - (7):
ω a =AttNet a (e u ) (2)
ω u =AttNet u (e a ) (5)
wherein, attNet u (. Cndot.) and AttNet a (. Cndot.) represents the attention network, ω u And omega a Representing non-normalized weight vector, alpha a And alpha u Is a weighted sum of the feature representations.
3. The activity and user identification method based on deep multitasking learning of claim 1, characterized in that in step (2-5), the activity and user classification layer outputs probability distributions of different activity labels and different user labels, the calculation formula is as follows:
wherein alpha is a And alpha u Representing the input of the last step, θ a And theta u Parameters representing the fully connected network layer, |a| represents the number of active labels, |u| represents the number of user labels, a' i Representing the probability of predicting an active tag as i, u' i Representing the probability of predicting the user tag as i.
4. The deep multitasking learning based activity and user identification method of claim 1 in which step (3) specifically comprises:
(3-1) selecting a portion of the training samples from the training sample set;
(3-2) processing the data samples x respectively using two convolutional neural networks sharing part of the hidden parameters to obtain an intermediate layer characteristic representation r of the samples a And r u The hidden parameters are stored and updated instead of the convolution kernels, specifically, in forward propagation, the hidden parameters are firstly used for generating the convolution kernels, and then the generated convolution kernels are used for carrying out convolution operation; in back propagation, hidden parameters are updated;
(3-3) obtaining an activity prediction result a 'and a user prediction result u' according to the steps (2-3) to (2-5);
(3-4) updating the parameters of the network parts according to the step (2-6) to determine the network parameters, and obtaining the active recognition model and the user recognition model determined by the parameters.
5. The activity and user identification method based on deep multitasking learning of claim 1, characterized in that in step (2-6) the loss is calculated using a cross entropy loss function, the calculation formula is as follows:
wherein L is a And L u All representing loss of a single sample, M batch Representing the batch size, loss represents the joint Loss of a batch of samples, adam algorithm is used for network parameter update, |a| represents the number of active labels, |u| represents the number of user labels, a' i Representing the probability of predicting an active tag as i, u' i Representing the probability of predicting the user tag as i.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201911355355.1A CN111160443B (en) | 2019-12-25 | 2019-12-25 | Activity and user identification method based on deep multitasking learning |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201911355355.1A CN111160443B (en) | 2019-12-25 | 2019-12-25 | Activity and user identification method based on deep multitasking learning |
Publications (2)
Publication Number | Publication Date |
---|---|
CN111160443A CN111160443A (en) | 2020-05-15 |
CN111160443B true CN111160443B (en) | 2023-05-23 |
Family
ID=70558071
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201911355355.1A Active CN111160443B (en) | 2019-12-25 | 2019-12-25 | Activity and user identification method based on deep multitasking learning |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN111160443B (en) |
Families Citing this family (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111709243B (en) * | 2020-06-19 | 2023-07-07 | 南京优慧信安科技有限公司 | Knowledge extraction method and device based on deep learning |
CN113472484B (en) * | 2021-06-29 | 2022-08-05 | 哈尔滨工业大学 | Internet of things equipment user feature code identification method based on cross entropy iterative learning |
CN114462557B (en) * | 2022-04-13 | 2022-07-01 | 北京大学 | Physiological state identification and analysis method based on multi-source information fusion |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN105404877A (en) * | 2015-12-08 | 2016-03-16 | 商汤集团有限公司 | Human face attribute prediction method and apparatus based on deep study and multi-task study |
CN108846384A (en) * | 2018-07-09 | 2018-11-20 | 北京邮电大学 | Merge the multitask coordinated recognition methods and system of video-aware |
CN108875836A (en) * | 2018-06-27 | 2018-11-23 | 浙江大学 | A kind of simple-complicated activity collaboration recognition methods based on depth multi-task learning |
AU2018101751A4 (en) * | 2018-11-17 | 2018-12-20 | Feng Chen | A distributed attack detection and information security sharing algorithm over multitask networks |
CN109857843A (en) * | 2018-12-25 | 2019-06-07 | 科大讯飞股份有限公司 | Exchange method and system based on document |
-
2019
- 2019-12-25 CN CN201911355355.1A patent/CN111160443B/en active Active
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN105404877A (en) * | 2015-12-08 | 2016-03-16 | 商汤集团有限公司 | Human face attribute prediction method and apparatus based on deep study and multi-task study |
CN108875836A (en) * | 2018-06-27 | 2018-11-23 | 浙江大学 | A kind of simple-complicated activity collaboration recognition methods based on depth multi-task learning |
CN108846384A (en) * | 2018-07-09 | 2018-11-20 | 北京邮电大学 | Merge the multitask coordinated recognition methods and system of video-aware |
AU2018101751A4 (en) * | 2018-11-17 | 2018-12-20 | Feng Chen | A distributed attack detection and information security sharing algorithm over multitask networks |
CN109857843A (en) * | 2018-12-25 | 2019-06-07 | 科大讯飞股份有限公司 | Exchange method and system based on document |
Also Published As
Publication number | Publication date |
---|---|
CN111160443A (en) | 2020-05-15 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
Qin et al. | Imaging and fusing time series for wearable sensor-based human activity recognition | |
Chen et al. | Smartphone sensor-based human activity recognition using feature fusion and maximum full a posteriori | |
CN111160443B (en) | Activity and user identification method based on deep multitasking learning | |
Yger et al. | Riemannian approaches in brain-computer interfaces: a review | |
CN110515456B (en) | Electroencephalogram signal emotion distinguishing method and device based on attention mechanism | |
Raissi et al. | On parameter estimation approaches for predicting disease transmission through optimization, deep learning and statistical inference methods | |
US8935195B2 (en) | Method of identification and devices thereof | |
CN110415022B (en) | Method and device for processing user behavior sequence | |
Osama et al. | Forecasting Global Monkeypox Infections Using LSTM: A Non-Stationary Time Series Analysis | |
Zhang et al. | Diffusion kernel attention network for brain disorder classification | |
CN112766355B (en) | Electroencephalogram signal emotion recognition method under label noise | |
CN105608478B (en) | image feature extraction and classification combined method and system | |
CN106778708A (en) | A kind of expression shape change recognition methods of the active appearance models based on tensor | |
Chao et al. | Supervised nonnegative matrix factorization to predict ICU mortality risk | |
EP3660750A1 (en) | Method and system for classification of data | |
Li et al. | Smartphone-sensors based activity recognition using IndRNN | |
CN114708637A (en) | Face action unit detection method based on meta-learning | |
CN113705645B (en) | Self-adaptive joint model semi-supervised learning classification method for electroencephalogram signals | |
Akmal et al. | Classification analysis of tensor-based recovered missing EEG data | |
CN115392474B (en) | Local perception graph representation learning method based on iterative optimization | |
Jordao et al. | Latent hypernet: Exploring the layers of convolutional neural networks | |
Nia et al. | The Power of ANN-Random Forest Algorithm in Human Activities Recognition Using IMU Data | |
CN110613445A (en) | DWNN framework-based electrocardiosignal identification method | |
CN115664970A (en) | Network abnormal point detection method based on hyperbolic space | |
CN112560712B (en) | Behavior recognition method, device and medium based on time enhancement graph convolutional network |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |