CN111160443B - Activity and user identification method based on deep multitasking learning - Google Patents

Activity and user identification method based on deep multitasking learning Download PDF

Info

Publication number
CN111160443B
CN111160443B CN201911355355.1A CN201911355355A CN111160443B CN 111160443 B CN111160443 B CN 111160443B CN 201911355355 A CN201911355355 A CN 201911355355A CN 111160443 B CN111160443 B CN 111160443B
Authority
CN
China
Prior art keywords
user
activity
network
recognition
representing
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201911355355.1A
Other languages
Chinese (zh)
Other versions
CN111160443A (en
Inventor
陈岭
张毅
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Zhejiang University ZJU
Original Assignee
Zhejiang University ZJU
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Zhejiang University ZJU filed Critical Zhejiang University ZJU
Priority to CN201911355355.1A priority Critical patent/CN111160443B/en
Publication of CN111160443A publication Critical patent/CN111160443A/en
Application granted granted Critical
Publication of CN111160443B publication Critical patent/CN111160443B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • G06F18/2415Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches based on parametric or probabilistic models, e.g. based on likelihood ratio or false acceptance rate versus a false rejection rate
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/044Recurrent networks, e.g. Hopfield networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02DCLIMATE CHANGE MITIGATION TECHNOLOGIES IN INFORMATION AND COMMUNICATION TECHNOLOGIES [ICT], I.E. INFORMATION AND COMMUNICATION TECHNOLOGIES AIMING AT THE REDUCTION OF THEIR OWN ENERGY USE
    • Y02D10/00Energy efficient computing, e.g. low power processors, power management or thermal management

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Computation (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Artificial Intelligence (AREA)
  • General Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • Computing Systems (AREA)
  • Software Systems (AREA)
  • Molecular Biology (AREA)
  • Computational Linguistics (AREA)
  • Biophysics (AREA)
  • Biomedical Technology (AREA)
  • Mathematical Physics (AREA)
  • General Health & Medical Sciences (AREA)
  • Health & Medical Sciences (AREA)
  • Probability & Statistics with Applications (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Evolutionary Biology (AREA)
  • Image Analysis (AREA)

Abstract

The invention discloses an activity and user identification method based on deep multitask learning, which comprises the following steps: (1) Collecting sensor data of each user during activities by using a wearable sensor, and preprocessing; (2) Jointly constructing an activity recognition network and a user recognition network for predicting activities and users, wherein the activity recognition network and the user recognition network share part hidden parameters and introduce a mutual attention mechanism, and weighting each part of the characteristics by utilizing knowledge learned by the other party; (3) Constructing joint loss to perform collaborative optimization on the activity recognition network and the user recognition network to obtain an activity recognition model and a user recognition model with determined parameters; (4) And inputting the preprocessed sensor data into the activity recognition model and the user recognition model, obtaining an activity recognition result by using the activity recognition model, and obtaining a user recognition result by using the user recognition model. The method improves activity and user recognition capability.

Description

Activity and user identification method based on deep multitasking learning
Technical Field
The invention relates to the field of activity recognition and user recognition, in particular to an activity and user recognition method based on deep multitask learning.
Background
The method for deducing the activity and the user information based on the activity of the wearable sensor and the user identification by utilizing the data acquired by the sensor worn by the user is two key tasks in the fields of pervasive and mobile computing, and has important significance for realizing the computation with human centers. The wearable sensor-based activity and user identification can provide support for health support, skill assessment, biometric identification, and other applications.
Most wearable sensor-based activity recognition methods mix labeling samples of all training users, build an activity classifier using supervised learning methods, and use it directly for new users, ignoring differences in different user behavior patterns. Related studies have shown that different users have different patterns of behavior, which means that there is a distribution difference between the sensor data acquired from different users, and therefore, when an activity recognition model that performs well on a training user is directly applied to a new user, a significant performance degradation tends to occur. How to guarantee the recognition performance of new users becomes an important challenge for activity recognition based on wearable sensors. To cope with this challenge, several methods have been proposed at home and abroad, which can be roughly divided into two categories: the first type of method builds user-independent features for modeling, so that the generalization capability of the model on new users is guaranteed. Such methods have difficulty in fully utilizing the user information in the training data, and impair their activity characterization capabilities while enhancing feature user independence. The second class of methods builds a custom model for each user for its behavior pattern. Such methods often require data acquisition and model adaptation for each new user, which is limited in use at a high cost.
Most of the user identification methods based on the wearable sensors only support the identification of users in the walking activity scene, and although the methods can achieve good identification effect, the limited applicable scene limits the use of the methods in daily life. Extending the activity scenarios supported by wearable sensor-based user identification from walking to other daily activities is valuable, but also challenging, because there are significant differences in sensor data from activity scenario to activity scenario.
Existing approaches typically model activity recognition or user recognition tasks alone, ignoring the correlation between the two tasks. The activity related information acquired by the activity recognition task is helpful for the user to recognize task perception and adapt to different activity scenes; the user-related information learned by the user identification task enables the activity identification model to take into account the current user's behavior pattern when identifying human activity.
Disclosure of Invention
The invention aims to solve the technical problem of improving the generalization capability of an identification model by utilizing the correlation between two tasks of activity identification and user identification.
In order to solve the above problems, the present invention provides an activity and user identification method based on deep multitasking learning, comprising the following steps:
(1) Collecting sensor data of each user during activities by using a wearable sensor, and preprocessing;
(2) Jointly constructing an activity recognition network and a user recognition network for predicting activities and users, wherein the activity recognition network and the user recognition network share part hidden parameters and introduce a mutual attention mechanism, and weighting each part of the characteristics by utilizing knowledge learned by the other party;
(3) Constructing joint loss to perform collaborative optimization on the activity recognition network and the user recognition network to obtain an activity recognition model and a user recognition model with determined parameters;
(4) And inputting the preprocessed sensor data into the activity recognition model and the user recognition model, obtaining an activity recognition result by using the activity recognition model, and obtaining a user recognition result by using the user recognition model.
Compared with the prior method, the method has the advantages that:
1) And jointly constructing an activity recognition model and a user recognition model, and utilizing hidden parameters to carry out cross-task knowledge sharing so as to realize interaction promotion by utilizing the commonality and the difference between the two tasks.
2) A mutual attention mechanism is introduced between the activity recognition model and the user recognition model, so that the two models can mutually utilize knowledge learned by each other to give weight to each part of the feature, thereby adapting to the change of the user and the activity scene.
Drawings
In order to more clearly illustrate the embodiments of the present invention or the technical solutions in the prior art, the drawings that are required in the embodiments or the description of the prior art will be briefly described, and it is obvious that the drawings in the following description are only some embodiments of the present invention, and other drawings may be obtained according to these drawings without inventive effort for a person skilled in the art.
FIG. 1 is a general flow diagram of an activity and user identification method based on deep multitasking learning provided by an embodiment;
FIG. 2 is a general architecture diagram of an activity and user identification method based on deep multitasking learning provided by an embodiment;
FIG. 3 is a schematic diagram of a convolutional neural network architecture provided by an embodiment;
fig. 4 is a schematic diagram of a bidirectional long and short time memory network according to an embodiment.
Detailed Description
The present invention will be described in further detail with reference to the drawings and examples, in order to make the objects, technical solutions and advantages of the present invention more apparent. It should be understood that the detailed description is presented by way of example only and is not intended to limit the scope of the invention.
FIG. 1 is a general flow diagram of an activity and user identification method based on deep multitasking learning provided by an embodiment; fig. 2 is an overall architecture diagram of an activity and user recognition method based on deep multitasking learning provided in an embodiment, in fig. 2, conv1D and Pool1D represent a one-dimensional convolutional network layer and a one-dimensional max-pooling layer, respectively, U represents a bidirectional long-short-time memory network element, attNet represents an attention network, which outputs a weight vector w to distinguish the importance of a feature r,
Figure BDA0002335756710000041
representing Hadamard sum, ">
Figure BDA0002335756710000042
Representing the hadamard product.
Referring to fig. 1 and 2, the activity and user recognition method based on deep multitasking learning provided in the embodiment includes a data set construction stage, a hidden parameter initialization stage, and a model training stage.
Data set construction phase
The data set construction stage mainly collects sensor data, preprocesses the collected sensor data, and constructs a training data set, and the specific process is as follows:
step 1-1, acquiring sensor data of each user when different activities are performed by using a wearable sensor;
in this step, wearable sensor data is recorded as the user performs different activities. Common wearable sensors include accelerometers, gyroscopes, magnetometers, and the like. In order to ensure the recognition accuracy, the number, the type and the placement orientation of the sensors are required to be consistent when the training data set is constructed and when the training data set is actually used.
And step 1-2, carrying out outlier elimination and normalization processing on the acquired data, and dividing the processed data by utilizing a sliding window to obtain a data sample.
In this step, the processing of the acquired data mainly includes:
a) Abnormal value detection is carried out on the original sensor data, and invalid values (such as values exceeding a normal range and missing values) in the abnormal value detection are removed or mean filling is carried out.
b) Carrying out dispersion standardization processing on sensor data according to channels, so that the processed data are standardized to be within an [ -1,1] interval, and the conversion formula is as follows:
Figure BDA0002335756710000051
wherein X is the original value, X min To be the minimum value of the channel where the value is located, X max For the maximum value of the channel in which the value is located, x' is the value after normalization.
c) The processed data was divided using a fixed length sliding window, the sliding window length was set manually according to experience, and the overlap was set to 50%.
And step 1-3, taking each data sample, the corresponding active label and the user label as a training sample, and constructing a training data set. One training sample may be denoted (x, a, u), where x represents the data sample, a represents the active tag, and u represents the user tag.
Step 1-4, the training data set is batched according to a fixed size, and the total number of batches is N.
In this step, the batch size is manually set according to experience, and the total number of batches N is calculated as follows:
Figure BDA0002335756710000052
wherein M is total To train the total number of samples in the dataset, M batch Is of a batch size.
Hidden parameter initialization stage
The specific process is as follows:
step 2-1, sequentially selecting a batch of training samples from the training dataset, and repeating steps 2-2 to 2-5 for each training sample (x, a, u) in the batch.
Step 2-2, respectively processing the data sample x by using two independent convolutional neural networks in the active recognition network and the user recognition network to obtain an intermediate layer characteristic representation r of the sample a And r u
In this step, the two convolutional neural networks have the same network architecture and are respectively composed of three convolutional network layers and one max-pooling layer, as shown in fig. 3, wherein the convolutional operation and the max-pooling operation are performed along the time dimension. In fig. 3, conv1D and Pool1D represent a one-dimensional convolutional network layer and a one-dimensional max-pooling layer, respectively, the numbers before @ represent the size of the convolutional kernel or pooling region, the numbers after @ represent the number of generated feature maps, and ReLU represents the activation function.
Step 2-3, representing the intermediate layer characteristics as r a And r u The vector sequence with the length of l is expanded and is respectively input into two independent bidirectional long-short-time memory networks in the activity recognition network and the user recognition network to extract the time dependency relationship to obtain the characteristic representation vector sequence
Figure BDA0002335756710000061
And->
Figure BDA0002335756710000062
In this step, the intermediate layer features represent r a And r u The time dimension of (c) remains unchanged and the other dimensions are expanded. The two bidirectional long-short-time memory networks have the same network architecture, and each bidirectional long-short-time memory network layer comprises a bidirectional long-short-time memory network layer which is shown in a time expansion schematic diagram in fig. 4. In FIG. 4, L f And L b Respectively representing a forward long-short time memory network unit and a backward long-short time memory network unit, wherein the two long-short time memory network layers are opposite in directionThe short-time memory network layer is connected to the same output, one long-time memory network layer processes the input sequence in the forward direction, the other one processes in the backward direction, and the output can acquire the time dependence relationship between the forward direction and the backward direction at the same time.
Step 2-4, using two independent attention networks in the activity recognition network and the user recognition network to represent the feature vector sequence e a Mapping to weight vectors
Figure BDA0002335756710000063
Representing the vector sequence e for the feature u Each part is given weight; representing the feature by a vector sequence e u Mapping to weight vector->
Figure BDA0002335756710000064
Representing the vector sequence e for the feature a Each portion is given a weight.
In this step, the two attention networks have the same network architecture, and each includes a fully connected network layer, and the output of the fully connected network layer is normalized by a softmax function to obtain a weight vector, and the process is as shown in formulas (3) - (8):
ω a =AttNet a (e u ) (3)
Figure BDA0002335756710000065
Figure BDA0002335756710000066
ω u =AttNet u (e a ) (6)
Figure BDA0002335756710000067
Figure BDA0002335756710000071
wherein the method comprises the steps of,AttNet u (. Cndot.) and AttNet a (. Cndot.) represents the attention network, ω u And omega a Representing non-normalized weight vector, alpha a And alpha u Is a weighted sum of the feature representations.
Step 2-5, utilizing an activity classification output layer formed from fully-connected network layer and softmax function in the activity recognition network, according to the weighted sum alpha of characteristic representation a Obtaining an activity prediction result a'; using a user classification output layer in the user identification network, which consists of a fully connected network layer and a softmax function, a weighted sum alpha of feature representations is used u Obtaining a user prediction result u'.
In this step, the activity and user classification layer outputs probability distributions of different activity labels and different user labels, and the calculation formula is as follows:
Figure BDA0002335756710000072
Figure BDA0002335756710000073
wherein alpha is a And alpha u Representing the input of the last step, θ a And theta u Parameters representing the fully connected network layer, |a| represents the number of active labels, |u| represents the number of user labels, a i Representing the probability of predicting an active tag as i, u i Representing the probability of predicting the user tag as i.
And 2-6, calculating activity classification loss according to the activity prediction and the activity label for all training samples of the batch, calculating user classification loss according to the user prediction and the user label, and updating parameters of each part of the network according to the loss.
In this step, the loss is calculated using a cross entropy loss function, the calculation formula is as follows:
Figure BDA0002335756710000074
Figure BDA0002335756710000075
Figure BDA0002335756710000076
wherein L is a And L u All representing loss of a single sample, M batch Indicating the batch size, loss indicates the joint Loss of a batch of samples. Adam algorithm is used for network parameter update.
Step 2-7, if the specified iteration times are reached, entering step 2-8, otherwise returning to step 2-1.
And 2-8, stacking convolution kernels of the updated two convolution neural networks corresponding to the convolution network layers to form a high-order tensor, and decomposing by a Tucker to obtain an initial hidden parameter.
In this step, the convolution kernel of each convolution network layer may be represented as a d 1 ×d 2 ×d 3 The convolution kernels of the corresponding convolution network layers of the two convolution neural networks are stacked to form d by the dimension third-order tensor 1 ×d 2 ×d 3 ×d 4 (d 4 =2) a four-dimensional tensor T, which is subjected to a Tucker decomposition to obtain the initial hidden parameters, as shown in equation (14):
T=S· (1,2) R (1) · (1,2) R (2) · (1,2) R (3) · (1,2) R (4) (14)
wherein S is a v 1 ×v 2 ×v 3 ×v 4 Dimension fourth-order tensor, R (i) Is d i ×v i The dimension matrix, S and R are initial hidden parameters, R (4) Hidden parameters, S, R, specific to the identification network of the activity and the identification network of the user (1) ,R (2) And R is (3) Hidden parameters representing the sharing of an active identification network and a user identification network, subscript · (i,j) Representing coordinate axes for performing tensor dot product operations。
Model training stage
Step 3-1, sequentially selecting a batch of training samples from the training dataset, and repeating steps 3-2 to 3-3 for each training sample (x, a, u) in the batch.
Step 3-2, respectively processing the data sample x by using two convolutional neural networks sharing part of hidden parameters in the active recognition network and the user recognition network to obtain an intermediate layer characteristic representation r of the sample a And r u . The hidden parameters are stored and updated instead of the convolution kernels, specifically, in forward propagation, the hidden parameters are firstly used for generating the convolution kernels, and then the generated convolution kernels are used for carrying out convolution operation; in back propagation, hidden parameters are updated.
In this step, a corresponding convolution kernel is generated from the hidden parameter by inverse operation of the Tucker decomposition, and a convolution operation is performed using the generated convolution kernel. Wherein, part of hidden parameters are shared by the activity recognition network and the user recognition network, and when one network is trained, the updating of the shared hidden parameters also affects the convolution kernel generated by the other network, thereby realizing the cross-task sharing of knowledge.
Step 3-3, obtaining an activity prediction result a 'and a user prediction result u' according to the steps 2-3-2-5; namely, specifically comprises:
step 3-3-1, r a And r u Expanded into a vector sequence with the length of l, and respectively input into two independent bidirectional long-short-time memory networks in the activity recognition network and the user recognition network to obtain a characteristic representation vector sequence
Figure BDA0002335756710000091
And
Figure BDA0002335756710000092
the implementation details of the step are consistent with those of the step 2-3 in the hidden parameter initialization stage.
Step 3-3-2, using two independent attention networks of the activity recognition network and the user recognition network, e a Mapping to weight vectors
Figure BDA0002335756710000093
E is u Each part is given weight; will e u Mapping to weight vector->
Figure BDA0002335756710000094
E is a Each portion is given a weight.
The implementation details of the step are consistent with those of the step 2-4 in the hidden parameter initialization stage.
Step 3-3-3, obtaining an activity prediction result a' by using an activity classification output layer which is formed by a fully connected network layer and a softmax function in an activity recognition network; and obtaining a user prediction result u' by using a user classification output layer in the user identification network, wherein the user classification output layer consists of a fully connected network layer and a softmax function.
The implementation details of the step are consistent with those of the step 2-5 in the hidden parameter initialization stage.
And 3-4, updating parameters of each part of the network according to the steps 2-6.
In this step, for all training samples of the batch, activity classification losses are calculated according to the activity predictions and the activity labels, user classification losses are calculated according to the user predictions and the user labels, and parameters of each part of the network are updated according to the losses.
The implementation details of the step are consistent with those of the step 2-6 in the hidden parameter initialization stage.
And 3-5, if the specified training iteration times are reached, finishing training, obtaining an active recognition model and a user recognition model with determined parameters, and otherwise, returning to the step 3-1.
After the network parameters of the activity recognition model and the user recognition model are determined through the data set construction stage, the hidden parameter initialization stage and the model training stage, the activity recognition model and the user recognition model can be utilized for carrying out activity and user recognition, and the specific process is as follows:
for the acquired sensor data, performing outlier elimination and normalization processing on the data according to the steps 1-2 in the data set construction stage, and dividing the processed data to obtain a data sample;
processing the data sample by using an activity recognition model determined by the network parameters to obtain an activity recognition result;
and processing the data sample by using the user identification model determined by the network parameters to obtain a user identification result.
The invention jointly builds the activity recognition and user recognition models, utilizes hidden parameters to share information between the two models, further utilizes the commonality and the difference between the two tasks to realize mutual promotion, introduces a mutual attention mechanism, enables the two models to mutually utilize the knowledge learned by the other party to give weight to each part of the feature, thereby adapting to the change of the user and the activity scene, and has wide application prospect in the fields of health support, skill evaluation, biological recognition and the like.
The foregoing detailed description of the preferred embodiments and advantages of the invention will be appreciated that the foregoing description is merely illustrative of the presently preferred embodiments of the invention, and that no changes, additions, substitutions and equivalents of those embodiments are intended to be included within the scope of the invention.

Claims (5)

1. An activity and user identification method based on deep multitasking learning, comprising the following steps:
(1) Acquiring sensor data of each user during activities by using a wearable sensor, preprocessing to obtain a data sample x, taking the data sample x, a corresponding activity label a and a corresponding user label u as training samples, and constructing a training sample set;
(2) Selecting a portion of the training samples from the training sample set to jointly construct an activity recognition network and a user recognition network for predicting activities and users, wherein the activity recognition network and the user recognition network share a portion of hidden parameters and introduce a mutual attention mechanism, and weighting each portion of the features by utilizing knowledge learned by the other party, comprising:
(2-1) selecting a portion of the training samples from the training sample set;
(2-2) Using two independent volumes in an Activity recognition network and a subscriber recognition networkThe data samples x are processed by the neural network respectively to obtain an intermediate layer characteristic representation r of the samples a And r u
(2-3) representing the intermediate layer characteristics r a And r u The vector sequence with the length of l is expanded and is respectively input into two independent bidirectional long-short-time memory networks in the activity recognition network and the user recognition network to extract the time dependency relationship to obtain the characteristic representation vector sequence
Figure FDA0004127316950000011
And->
Figure FDA0004127316950000012
Wherein the interlayer features represent r a And r u The time dimension of the memory is kept unchanged, other dimensions are expanded, the two bidirectional long-short-time memory networks have the same network architecture and both comprise a bidirectional long-short-time memory network layer, the bidirectional long-short-time memory network layer connects the two long-short-time memory network layers with opposite directions to the same output, one long-short-time memory network layer processes an input sequence in a forward direction, the other long-short-time memory network layer processes in a backward direction, and the output can acquire the time dependence relationship of the forward direction and the backward direction at the same time;
(2-4) representing the feature vector sequence e with two independent attention networks of the activity recognition network and the user recognition network a Mapping to weight vectors
Figure FDA0004127316950000013
Representing the vector sequence e for the feature u Each part is given weight; representing the feature by a vector sequence e u Mapping to weight vector->
Figure FDA0004127316950000014
Representing the vector sequence e for the feature a Each part is given weight;
(2-5) utilizing an activity classification output layer in the activity recognition network, which consists of a fully connected network layer and a softmax function, according to a weighted sum alpha of feature representations a Obtaining an activity prediction result a'; utilization ofUser classification output layer in user identification network composed of fully connected network layer and softmax function according to weighted sum alpha of characteristic representation u Obtaining a user prediction result u';
(2-6) calculating an activity classification loss according to the activity prediction and the activity label, calculating a user classification loss according to the user prediction and the user label, and updating parameters of each part of the network according to the loss;
(2-7) stacking convolution kernels of the updated two convolution neural networks corresponding to the convolution network layers to form a higher-order tensor, and obtaining an initial hidden parameter through a Tucker decomposition, wherein the method comprises the following steps:
the convolution kernel of each convolution network layer may be represented as a d 1 ×d 2 ×d 3 The convolution kernels of the corresponding convolution network layers of the two convolution neural networks are stacked to form d by the dimension third-order tensor 1 ×d 2 ×d 3 ×d 4 A dimension fourth-order tensor T, where d 4 =2, performing a Tucker decomposition on the tensor T to obtain an initial hidden parameter, as shown in equation (1):
T=S· (1,2) R (1) · (1,2) R (2) · (1,2) R (3) · (1,2) R (4) (1)
wherein S is a v 1 ×v 2 ×v 3 ×v 4 Dimension fourth-order tensor, R (i) Is d i ×v i The dimension matrix, S and R are initial hidden parameters, R (4) Hidden parameters, S, R, specific to the identification network of the activity and the identification network of the user (1) ,R (2) And R is (3) Hidden parameters representing the sharing of an active identification network and a user identification network, subscript · (i,j) Representing coordinate axes for performing tensor dot product operations;
(3) Selecting part of training samples from the training sample set to construct a joint loss to carry out collaborative optimization on the active recognition network and the user recognition network, so as to obtain an active recognition model and a user recognition model with determined parameters;
(4) And inputting the preprocessed sensor data into the activity recognition model and the user recognition model, obtaining an activity recognition result by using the activity recognition model, and obtaining a user recognition result by using the user recognition model.
2. The deep multitasking learning-based activity and user recognition method of claim 1 in which in step (2-4) both attention networks have the same network architecture and each comprise a fully connected network layer whose outputs are normalized by a softmax function to obtain weight vectors as shown in equations (2) - (7):
ω a =AttNet a (e u ) (2)
Figure FDA0004127316950000031
Figure FDA0004127316950000032
ω u =AttNet u (e a ) (5)
Figure FDA0004127316950000033
Figure FDA0004127316950000034
wherein, attNet u (. Cndot.) and AttNet a (. Cndot.) represents the attention network, ω u And omega a Representing non-normalized weight vector, alpha a And alpha u Is a weighted sum of the feature representations.
3. The activity and user identification method based on deep multitasking learning of claim 1, characterized in that in step (2-5), the activity and user classification layer outputs probability distributions of different activity labels and different user labels, the calculation formula is as follows:
Figure FDA0004127316950000035
Figure FDA0004127316950000036
wherein alpha is a And alpha u Representing the input of the last step, θ a And theta u Parameters representing the fully connected network layer, |a| represents the number of active labels, |u| represents the number of user labels, a' i Representing the probability of predicting an active tag as i, u' i Representing the probability of predicting the user tag as i.
4. The deep multitasking learning based activity and user identification method of claim 1 in which step (3) specifically comprises:
(3-1) selecting a portion of the training samples from the training sample set;
(3-2) processing the data samples x respectively using two convolutional neural networks sharing part of the hidden parameters to obtain an intermediate layer characteristic representation r of the samples a And r u The hidden parameters are stored and updated instead of the convolution kernels, specifically, in forward propagation, the hidden parameters are firstly used for generating the convolution kernels, and then the generated convolution kernels are used for carrying out convolution operation; in back propagation, hidden parameters are updated;
(3-3) obtaining an activity prediction result a 'and a user prediction result u' according to the steps (2-3) to (2-5);
(3-4) updating the parameters of the network parts according to the step (2-6) to determine the network parameters, and obtaining the active recognition model and the user recognition model determined by the parameters.
5. The activity and user identification method based on deep multitasking learning of claim 1, characterized in that in step (2-6) the loss is calculated using a cross entropy loss function, the calculation formula is as follows:
Figure FDA0004127316950000041
Figure FDA0004127316950000042
Figure FDA0004127316950000043
wherein L is a And L u All representing loss of a single sample, M batch Representing the batch size, loss represents the joint Loss of a batch of samples, adam algorithm is used for network parameter update, |a| represents the number of active labels, |u| represents the number of user labels, a' i Representing the probability of predicting an active tag as i, u' i Representing the probability of predicting the user tag as i.
CN201911355355.1A 2019-12-25 2019-12-25 Activity and user identification method based on deep multitasking learning Active CN111160443B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201911355355.1A CN111160443B (en) 2019-12-25 2019-12-25 Activity and user identification method based on deep multitasking learning

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201911355355.1A CN111160443B (en) 2019-12-25 2019-12-25 Activity and user identification method based on deep multitasking learning

Publications (2)

Publication Number Publication Date
CN111160443A CN111160443A (en) 2020-05-15
CN111160443B true CN111160443B (en) 2023-05-23

Family

ID=70558071

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201911355355.1A Active CN111160443B (en) 2019-12-25 2019-12-25 Activity and user identification method based on deep multitasking learning

Country Status (1)

Country Link
CN (1) CN111160443B (en)

Families Citing this family (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111709243B (en) * 2020-06-19 2023-07-07 南京优慧信安科技有限公司 Knowledge extraction method and device based on deep learning
CN113472484B (en) * 2021-06-29 2022-08-05 哈尔滨工业大学 Internet of things equipment user feature code identification method based on cross entropy iterative learning
CN114462557B (en) * 2022-04-13 2022-07-01 北京大学 Physiological state identification and analysis method based on multi-source information fusion

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105404877A (en) * 2015-12-08 2016-03-16 商汤集团有限公司 Human face attribute prediction method and apparatus based on deep study and multi-task study
CN108846384A (en) * 2018-07-09 2018-11-20 北京邮电大学 Merge the multitask coordinated recognition methods and system of video-aware
CN108875836A (en) * 2018-06-27 2018-11-23 浙江大学 A kind of simple-complicated activity collaboration recognition methods based on depth multi-task learning
AU2018101751A4 (en) * 2018-11-17 2018-12-20 Feng Chen A distributed attack detection and information security sharing algorithm over multitask networks
CN109857843A (en) * 2018-12-25 2019-06-07 科大讯飞股份有限公司 Exchange method and system based on document

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105404877A (en) * 2015-12-08 2016-03-16 商汤集团有限公司 Human face attribute prediction method and apparatus based on deep study and multi-task study
CN108875836A (en) * 2018-06-27 2018-11-23 浙江大学 A kind of simple-complicated activity collaboration recognition methods based on depth multi-task learning
CN108846384A (en) * 2018-07-09 2018-11-20 北京邮电大学 Merge the multitask coordinated recognition methods and system of video-aware
AU2018101751A4 (en) * 2018-11-17 2018-12-20 Feng Chen A distributed attack detection and information security sharing algorithm over multitask networks
CN109857843A (en) * 2018-12-25 2019-06-07 科大讯飞股份有限公司 Exchange method and system based on document

Also Published As

Publication number Publication date
CN111160443A (en) 2020-05-15

Similar Documents

Publication Publication Date Title
Qin et al. Imaging and fusing time series for wearable sensor-based human activity recognition
Chen et al. Smartphone sensor-based human activity recognition using feature fusion and maximum full a posteriori
CN111160443B (en) Activity and user identification method based on deep multitasking learning
Yger et al. Riemannian approaches in brain-computer interfaces: a review
CN110515456B (en) Electroencephalogram signal emotion distinguishing method and device based on attention mechanism
Raissi et al. On parameter estimation approaches for predicting disease transmission through optimization, deep learning and statistical inference methods
US8935195B2 (en) Method of identification and devices thereof
CN110415022B (en) Method and device for processing user behavior sequence
Osama et al. Forecasting Global Monkeypox Infections Using LSTM: A Non-Stationary Time Series Analysis
Zhang et al. Diffusion kernel attention network for brain disorder classification
CN112766355B (en) Electroencephalogram signal emotion recognition method under label noise
CN105608478B (en) image feature extraction and classification combined method and system
CN106778708A (en) A kind of expression shape change recognition methods of the active appearance models based on tensor
Chao et al. Supervised nonnegative matrix factorization to predict ICU mortality risk
EP3660750A1 (en) Method and system for classification of data
Li et al. Smartphone-sensors based activity recognition using IndRNN
CN114708637A (en) Face action unit detection method based on meta-learning
CN113705645B (en) Self-adaptive joint model semi-supervised learning classification method for electroencephalogram signals
Akmal et al. Classification analysis of tensor-based recovered missing EEG data
CN115392474B (en) Local perception graph representation learning method based on iterative optimization
Jordao et al. Latent hypernet: Exploring the layers of convolutional neural networks
Nia et al. The Power of ANN-Random Forest Algorithm in Human Activities Recognition Using IMU Data
CN110613445A (en) DWNN framework-based electrocardiosignal identification method
CN115664970A (en) Network abnormal point detection method based on hyperbolic space
CN112560712B (en) Behavior recognition method, device and medium based on time enhancement graph convolutional network

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant