CN105678222B - A kind of mobile device-based Human bodys' response method - Google Patents

A kind of mobile device-based Human bodys' response method Download PDF

Info

Publication number
CN105678222B
CN105678222B CN201511009411.8A CN201511009411A CN105678222B CN 105678222 B CN105678222 B CN 105678222B CN 201511009411 A CN201511009411 A CN 201511009411A CN 105678222 B CN105678222 B CN 105678222B
Authority
CN
China
Prior art keywords
data
feature
disaggregated model
behavior
vector
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201511009411.8A
Other languages
Chinese (zh)
Other versions
CN105678222A (en
Inventor
潘赟
茹晨光
朱永光
朱怀宇
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Zhejiang University ZJU
Original Assignee
Zhejiang University ZJU
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Zhejiang University ZJU filed Critical Zhejiang University ZJU
Priority to CN201511009411.8A priority Critical patent/CN105678222B/en
Publication of CN105678222A publication Critical patent/CN105678222A/en
Application granted granted Critical
Publication of CN105678222B publication Critical patent/CN105678222B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/20Movements or behaviour, e.g. gesture recognition
    • G06V40/23Recognition of whole body movements, e.g. for sport training
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • G06F18/2411Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches based on the proximity to a decision surface, e.g. support vector machines
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/243Classification techniques relating to the number of classes
    • G06F18/24323Tree-organised classifiers
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/245Classification techniques relating to the decision surface
    • G06F18/2451Classification techniques relating to the decision surface linear, e.g. hyperplane
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2218/00Aspects of pattern recognition specially adapted for signal processing

Abstract

A kind of mobile device-based Human bodys' response method, utilizes the multiple sensors real-time data collection built in mobile device;The data that sensor obtains are corrected, are filtered, calculate a series of data preprocessing operation such as generation data and data segmentation;Feature extraction is carried out to pretreated data segment, by the character pair vector input equipment position disaggregated model of extraction, obtains device location classification;Select corresponding behavior disaggregated model that the character pair vector input behavior disaggregated model of extraction is obtained final Activity recognition result according to the device location classification of acquisition.The present invention provides that a kind of versatility is good, the higher mobile device-based Human bodys' response method of accuracy.

Description

A kind of mobile device-based Human bodys' response method
Technical field
The present invention relates to Activity recognition technical fields, relate in particular to a kind of mobile device-based Human bodys' response Method.
Background technique
Human bodys' response is a kind of by obtaining and analyzing human body behavior related data, judges the skill of human body behavior state Art.By knowing human body basis behavioral activity, which can be motion tracking, health monitoring, fall detection, the elderly's prison Shield, patient resumes training, complex behavior identifies, support industry manufactures, human-computer interaction, augmented reality, indoor positioning and navigation, a The research and application of the various fields such as the identification of people's feature, urbanization calculating provide human body relevant information, therefore have important answer With value and research significance.
Traditional Human bodys' response technology is based primarily upon image information progress.This method captures human body by video equipment The image information of movement analyzes the relevant image sequence of behavior, judges behavior classification.Behavior based on image information is known Qi Bu not be early, theoretical relative maturity, recognition accuracy is higher, but since method is more complex, influences vulnerable to background light, and image Acquire that equipment volume is big, power consumption is high, using being restricted.
Due to MEMS (Micro-Electro-Mechanical System, MEMS) manufacturing technology it is rapid into Step, various types of sensors continue to develop, such as accelerometer, gyroscope, magnetometer and barometer etc..These sensor energy The relevant information for enough collecting human motion, while having had both the advantages such as good portability, low in energy consumption, and to user without bothering, resist Environmental disturbances ability is strong, becomes the preference data source for realizing duration Human bodys' response instead of video equipment.
In addition, with the rapid progress of mobile device the relevant technologies, especially using iOS and Android platform as the intelligence of representative The equipment such as energy mobile phone, Intelligent flat have become the important carrier of all kinds of MEMS sensors.It is compared to wearing of specially designing Formula sensor device is worn, this kind of mobile device has more powerful processing, storage and network-connectivity, while built-in a variety of biographies Sensor carries or wears again extras without using person, and more convenient for user, friendly, development platform also has There is the feature of universality.Therefore mobile device-based Activity recognition becomes the focus of researcher.
Existing sensor-based Human bodys' response method is directed to the design of dedicated wearable device mostly, to being based on The Human bodys' response of common mobile devices is simultaneously not suitable for.Caused by due to the characteristics of wearable device, sensor is placed on use On the fixation position of family body, generally also there is fixed direction.And all kinds of mobile devices, such as smart phone, due in day That often lives carries and uses, and according to personal habits difference has different placement locations and different placement directions.It is related MEMS sensor is placed on the data characteristics generated when the different parts of human body and has bigger difference by studies have shown that.Therefore will Equipment is in the data that different location is collected into and mutually obscures the decline that will lead to whole recognition accuracy;And due to placement direction Indefinite, the method based on certain specific direction data also can not be used directly;In addition, the operational capability of smart phone is flat from desktop There are also larger gaps for platform, in order to realize real-time Activity recognition, need to choose and the suitable feature extracting method of design complexities And disaggregated model.Therefore actual conditions and operational capability for smart phone is needed to design suitable method.
Summary of the invention
Versatility in order to overcome the shortcomings of existing Human bodys' response method is poor, accuracy is poor, and the present invention provides A kind of versatility is good, the higher mobile device-based Human bodys' response method of accuracy.
The technical solution adopted by the present invention to solve the technical problems is:
A kind of mobile device-based Human bodys' response method, comprising the following steps:
Step (1), training device location disaggregated model cpWith the behavior disaggregated model ca based on distinct device positioni, cai ∈ C, C are the set being made of behavior disaggregated model, C={ ca1, ca2..., caI, behavior disaggregated model caiWith device location piWith one-to-one relationship, device location pi∈ P, P are the set being made of device location predetermined, P={ p1, p2..., pI, I is device location class number predetermined;
Step (2) acquires sensor raw data by mobile device built-in sensors in real time;
Step (3) carries out data prediction to the data that mobile device built-in sensors acquire in real time, obtains current time The corresponding data segment of window;
Step (4), according to position identification feature collection Fp, feature extraction is carried out to the data segment of acquisition, obtains the data segment Corresponding feature vectorNpIt is characterized collection FpIn number of features;By feature vector Vp Input equipment position disaggregated model cp, obtain device location classification pi
Step (5), according to Activity recognition feature set Fa, feature extraction is carried out to the data segment of acquisition, obtains the data segment Corresponding feature vectorNaIt is characterized collection FaIn number of features;According to step (4) The position classification p of acquisitioniSelect corresponding behavior disaggregated model cai, by feature vector VaInput behavior disaggregated model cai, output Behavior classification aj, aj∈ A, A are the set being made of behavior predetermined, A={ a1, a2..., aJ, J is predetermined Behavior class number, ajAs final recognition result.
Further, training device location disaggregated model c in the step (1)pWith row of the training based on distinct device position For disaggregated model caiThe step of it is as follows:
Step (1-1) acquires distinct device position and different human body behavior feelings by mobile device built-in sensors offline Sensor raw data under condition;
Step (1-2) carries out data prediction to the sensor raw data of acquisition, obtains corresponding different time window Data segment;
Step (1-3) carries out feature extraction to each data segment, obtains the corresponding feature of the data segment according to feature set F Vector Vk=[v1, v2..., vN]∈RN, N is characterized the number of features in collection F, k ∈ { 1,2 ..., K }, and all data segments extract Feature vector collectively form sample set S, S=[V1;V2;...;VK], feature vector number K is data segment number;
Step (1-4), to each feature vector V in sample set Sk, flag data acquire when device location classification pi, pi∈ P is obtained position training vector k ∈ { 1,2 ..., K }, and all training vectors after label constitute position Set training set Tp,
Step (1-5), utilizes training set TpTraining device location disaggregated model, obtains device location disaggregated model cp
Step (1-6), according to disaggregated model cp, the Partial Feature composition position in feature set F for the model is selected to know Other feature set Fp
Step (1-7) carries out feature selecting to feature set F, and selection is wherein suitable for the Partial Feature of Activity recognition model Form Activity recognition feature set Fa
Step (1-8), Screening Samples, which integrate, corresponds to device location category label as p in SiFeature inwardsForm sample Subset Si, meet
Step (1-9), to sample set SiIn each feature vectorBehavior classification a when flag data acquiresj, aj∈ A obtains Behavioral training vectorki∈ { 1,2 ..., Ki, all training vectors after label are constituted Behavioral training collection KiFor sample set SiThe number of middle feature vector;
Step (1-10), utilizes Behavioral training collectionTrain classification models obtain corresponding position piBehavior disaggregated model cai
To step (1-8)-(1-10), i=1,2 ..., I is taken to be repeated in respectively, I indicates device location predetermined Class number, training obtain the behavior disaggregated model ca under distinct device position1, ca2..., caI
Further, the sensor raw data in the step (1-1) includes different mobile device built-in sensors And the data on same sensor different directions;Sensor raw data, that is, X that each sampled point obtainsl=[x1, x2..., xP]∈RP, P is sensor raw data dimension, and l ∈ { 1,2 ..., L }, L are the total sampling length of sensing data.
Preferably, the data prediction in the step (3) includes the following steps:
Step (3-1) is corrected and filters to sensor raw data;
Step (3-2) calculates according to correction and filtered data and generates data Yl, Yl=[y1, y2..., yQ]∈RQ, Q To generate data dimension;By sensor raw data and data merging is generated, obtains Zl, Zl=[Xl, Yl]∈RP+Q, l ∈ 1, 2 ..., L };Data group on all sampled points is combined into sensor data set, i.e. [Z1;Z2;...;ZL];
Continuous sensor data set is divided into data segment according to time window by the segmentation of step (3-3) data.
Pretreatment in step (1-1) also uses the above process.
It is by feature vector V that device location classification is obtained in the step (4)pInput position disaggregated model cp, cpFor training Obtained decision-tree model;The feature vector of input selects qualified branch according to the splitting condition of decision tree step by step, It is being classified at leaf node as a result, i.e. position pi, pi∈P。
Acquisition behavior classification is by feature vector V in the step (5)aInput behavior disaggregated model cai, caiIt is trained The support vector cassification model arrived is made of g (J) a support vector machines, function g (x) by model multi-class classification policy It determines, J is class object, that is, behavior classification number;The corresponding Optimal Separating Hyperplane expression formula of each support vector machines is f (x), Functional value f (V) is calculated according to the feature vector V of input, is obtained classification results (f (V) >=0 or f (V) < 0);To it is all support to The classification results of amount machine are weighted processing, and peak is taken to export as behavior classification results.
Feature set F in step (1-3) is made of one-dimensional corresponding time domain every in data segment, frequency domain or time and frequency domain characteristics.
The device location disaggregated model of training is decision-tree model in step (1-5), and the classifier is by one group " if yes " The classifying rules of form constitutes tree structure;In training, the highest feature of each split vertexes selection evaluation function value is made For splitting condition.
Feature selecting in step (1-6) selects the feature set F for the input of position disaggregated model from feature set Fp,FpIt is screened using embedded mode, i.e. Decision-Tree Classifier Model c of the training for position identificationpAfterwards, by decision tree classification Model feature composition characteristic collection F selected to usep
Feature selecting in step (1-7) selects the feature set F for the input of behavior disaggregated model from feature set Fa,FaUsing filtering-encapsulation two-stage process screening;Filter process selection evaluation function successively comments all features Valence and sequence, screening and assessment function result are highest M first;Encapsulation process passes through the output of combination supporting vector machine disaggregated model As a result, search meets the minimal characteristic set of output threshold condition, as the feature set F inputted for behavior disaggregated modela
The behavior disaggregated model of training is supporting vector machine model in step (1-10), and the model is by g (J) a supporting vector Machine composition;Each support vector machines according to the training data of input construct largest interval Optimal Separating Hyperplane, the Optimal Separating Hyperplane by FormulaDefinition, wherein x is the feature vector of input, xiIt is supporting vector, yiIt is branch Hold the corresponding result label of vector, yiValue meets yi∈ { -1,1 }, n are supporting vector number, aiIt is the branch being calculated with b Hold vector parameter, K (x, xi) it is kernel function for input vector to be mapped to high dimension linear space.
Beneficial effects of the present invention are mainly manifested in: the present invention provides one kind to be based on mobile device built-in sensors data Human bodys' response method, this method can identify carrying of the mobile device with respect to human body or using position, and then select needle To the Activity recognition model of the position.Compared with prior art, method provided by the invention passes through above-mentioned layering identification model Building and suitable feature selection approach can adapt to different in daily life under the premise of guaranteeing that computation complexity is moderate Position of mobile equipment and direction, realize the Human bodys' response of high-accuracy.
Detailed description of the invention
Fig. 1 is the flow chart of mobile device-based Human bodys' response method of the invention;
Fig. 2 is device location disaggregated model of the invention and behavior disaggregated model training flow chart;
Fig. 3 is feature selecting flow chart of the invention.
Specific embodiment
The invention will be further described below in conjunction with the accompanying drawings.
Referring to Fig.1~Fig. 3, a kind of mobile device-based Human bodys' response method, comprising the following steps:
Step (1), training device location disaggregated model cpWith the behavior disaggregated model ca based on distinct device positioni, cai ∈ C, C are the set being made of behavior disaggregated model, C={ ca1, ca2..., caI, behavior disaggregated model caiWith device location piWith one-to-one relationship, device location pi∈ P, P are the set being made of device location predetermined, P={ p1, p2..., pI, I is device location class number predetermined;
Step (2) acquires sensor raw data by mobile device built-in sensors in real time;
Step (3) carries out data prediction to the data that mobile device built-in sensors acquire in real time, obtains current time The corresponding data segment of window;
Step (4), according to position identification feature collection Fp, feature extraction is carried out to the data segment of acquisition, obtains the data segment Corresponding feature vectorNpIt is characterized collection FpIn number of features;By feature vector Vp Input equipment position disaggregated model cp, obtain device location classification pi
Step (5), according to Activity recognition feature set Fa, feature extraction is carried out to the data segment of acquisition, obtains the data segment Corresponding feature vectorNaIt is characterized collection FaIn number of features;According to step (4) The position classification p of acquisitioniSelect corresponding behavior disaggregated model cai, by feature vector VaInput behavior disaggregated model cai, output Behavior classification aj, aj∈ A, A are the set being made of behavior predetermined, A={ a1, a2..., aJ, J is predetermined Behavior class number, ajAs final recognition result;
Training device location disaggregated model c in the step (1)pBehavior classification mould with training based on distinct device position Type caiSpecific step is as follows:
Step (1-1) acquires distinct device position and different human body behavior feelings by mobile device built-in sensors offline Sensor raw data under condition;
Step (1-2) carries out data prediction to the sensor raw data of acquisition, obtains corresponding different time window Data segment;
Step (1-3) carries out feature extraction to each data segment, obtains the corresponding feature of the data segment according to feature set F Vector Vk=[v1, v2..., vN]∈RN, N is characterized the number of features in collection F, k ∈ { 1,2 ..., K }, and all data segments extract Feature vector collectively form sample set S, S=[V1;V2;...;VK], feature vector number K is data segment number;
Step (1-4), to each feature vector V in sample set Sk, flag data acquire when device location classification pi, pi∈ P is obtained position training vector k ∈ { 1,2 ..., K }, and all training vectors after label constitute position Set training set Tp,
Step (1-5), utilizes training set TpTraining device location disaggregated model, obtains device location disaggregated model cp
Step (1-6), according to disaggregated model cp, the Partial Feature composition position in feature set F for the model is selected to know Other feature set Fp
Step (1-7) carries out feature selecting to feature set F, and selection is wherein suitable for the Partial Feature of Activity recognition model Form Activity recognition feature set Fa
Step (1-8), Screening Samples, which integrate, corresponds to device location category label as p in SiFeature inwardsForm sample Subset Si, meet
Step (1-9), to sample set SiIn each feature vectorBehavior classification a when flag data acquiresj, aj∈ A obtains Behavioral training vectorki∈ { 1,2 ..., Ki, all training vectors after label are constituted Behavioral training collection KiFor sample set SiThe number of middle feature vector;
Step (1-10), utilizes Behavioral training collectionTrain classification models obtain corresponding position piBehavior disaggregated model cai
To step (1-8)-(1-10), i=1,2 ..., I is taken to be repeated in respectively, training obtains under distinct device position Behavior disaggregated model ca1, ca2..., caI
Further, the sensor raw data in the step (2) and step (1-1) includes built in different mobile devices Data on sensor and same sensor different directions;Sensor raw data, that is, X that each sampled point obtainsl=[x1, x2..., xP]∈RP, P is sensor raw data dimension, and l ∈ { 1,2 ..., L }, L are the total sampling length of sensing data.
Data prediction in step (3) and step (1-2) includes the following steps:
Step (3-1) is corrected and filters to sensor raw data;
Step (3-2) calculates according to correction and filtered data and generates data Yl, Yl=[y1, y2..., yQ]∈RQ, Q To generate data dimension;By sensor raw data and data merging is generated, obtains Zl, Zl=[Xl, Yl]∈RP+Q, l ∈ 1, 2 ..., L };Data group on all sampled points is combined into sensor data set, i.e. [Z1;Z2;...;ZL];
Continuous sensor data set is divided into data segment according to time window by the segmentation of step (3-3) data.
It is by feature vector V that device location classification is obtained in step (4)pInput position disaggregated model cp, cpIt is obtained for training Decision-tree model;The feature vector of input selects qualified branch according to the splitting condition of decision tree step by step, in leaf It is being classified at node as a result, i.e. position pi, pi∈P。
Acquisition behavior classification is by feature vector V in step (5)aInput behavior disaggregated model cai, caiTraining obtains Support vector cassification model is made of g (J) a support vector machines, and function g (x) is determined by the multi-class classification policy of model, J is class object, that is, behavior classification number;The corresponding Optimal Separating Hyperplane expression formula of each support vector machines is f (x), according to defeated The feature vector V entered calculates functional value f (V), obtains classification results (f (V) >=0 or f (V) < 0);To all support vector machines Classification results are weighted processing, and peak is taken to export as behavior classification results.
Feature set F in step (1-3) is made of one-dimensional corresponding time domain every in data segment, frequency domain or time and frequency domain characteristics.
The device location disaggregated model of training is decision-tree model in step (1-5), and the classifier is by one group " if yes " The classifying rules of form constitutes tree structure;In training, the highest feature of each split vertexes selection evaluation function value is made For splitting condition.
Feature selecting in step (1-6) selects the feature set F for the input of position disaggregated model from feature set Fp,FpIt is screened using embedded mode, i.e. Decision-Tree Classifier Model c of the training for position identificationpAfterwards, by decision tree point The feature composition characteristic collection F used selected by class modelp
Feature selecting in step (1-7) selects the feature set F for the input of behavior disaggregated model from feature set Fa,FaUsing filtering-encapsulation two-stage process screening;Filter process selection evaluation function successively comments all features Valence and sequence, screening and assessment function result are highest M first;Encapsulation process passes through the output of combination supporting vector machine disaggregated model As a result, search meets the minimal characteristic set of output threshold condition, as the feature set F inputted for behavior disaggregated modela
The behavior disaggregated model of training is supporting vector machine model in step (1-10), and the model is by g (J) a supporting vector Machine composition;Each support vector machines according to the training data of input construct largest interval Optimal Separating Hyperplane, the Optimal Separating Hyperplane by FormulaDefinition, wherein x is the feature vector of input, xiIt is supporting vector, yiIt is branch Hold the corresponding result label of vector, yiValue meets yi∈ { -1,1 }, n are supporting vector number, aiIt is the branch being calculated with b Hold vector parameter, K (x, xi) it is kernel function for input vector to be mapped to high dimension linear space.
Mobile device of the present invention mainly includes smart phone, Intelligent flat computer etc., the tool that the present embodiment uses Body mobile device is 4 mobile phone of millet for running Android platform.Select acceleration transducer, the gyro built in 4 mobile phone of millet Instrument, baroceptor, light sensor and range sensor are acquired for data, and sample frequency is set as 50Hz.Data acquisition When, the factors such as the specific implementation of equipment placement direction and behavior are not required.The collected data of each sampled point Format is Xl=[x1, x2..., xP]∈RP, P=9, l ∈ { 1,2 ..., L }, L is the total sampling length of sensing data, L= 485050.The nine dimensions sensor raw data includes three-dimensional acceleration information, three-dimensional gyro data, one-dimensional air pressure transmission Sensor data, one-dimensional light sensor data and one-dimensional range-sensor data.
The present embodiment acquires sensor raw data using 4 cell phone apparatus of millet, for device location disaggregated model and The training and assessment of behavior disaggregated model.The number for participating in data sampling is 10 people, including 3 women and 7 males;It adopts The behavior of collection include sit, stand, walking, running, upstairs, go downstairs 6 kinds of classifications.Above-mentioned 6 kinds of behavior classifications are denoted as a respectively1, a2..., a6, Constituting action set A={ a1, a2..., a6}.The position of mobile equipment of acquisition includes in pocket of trousers, in coat pocket, both shoulders packet In, in shoulder bag, hold in side, be placed in one's ear, hold in front, place 8 kinds of classifications on the table.Above-mentioned 8 kinds of equipment Position is denoted as p respectively1, p2..., p8, constitution equipment location sets P={ p1, p2..., p8}。
As shown in Fig. 2, in the training process of disaggregated model, by the collected multiple sensors initial data of mobile device It needs by further pretreatment operation.These operations include the correction to acceleration information, to acceleration transducer, gyro The filtering and noise reduction of instrument and baroceptor data calculates according to correction and filtered data and generates data, to continuous data Data segmentation.
Wherein, generate data include three-dimensional acceleration of gravity data, three-dimensional linear acceleration data and it is one-dimensional plus Velocity amplitude data.Specific implementation includes that low-pass filtering is carried out to three-dimensional acceleration data, and low-pass filtering is calculated such as formula:
afiltered(l)=afiltered(l-1)+α×(aunfiltered(l)-afiltered(l-1)) (1)
Wherein l indicates current sampling point, afilteredIt (l) is filtered acceleration value, aunfilteredIt (l) is before filtering Acceleration value, factor alpha are cutoff frequency, are set as 0.05Hz.
It is three-dimensional gravity acceleration information, remaining high frequency section, i.e. a by the data obtained after low-pass filterhigh (l)=a (l)-afilteredIt (l) is linear acceleration data.
The calculating that data further include one-dimensional acceleration amplitude data is generated, i.e., ax、ay、 azFor the 3-axis acceleration data at current sampling point.Therefore, the form for generating data is Yl=[y1, y2..., yQ]∈RQ, Q =7.By correction, filtered sensor acquisition data and data merging is generated, obtains the data Z on each sampled pointl, Zl= [Xl, Yl]∈RP+Q, P+Q=16, l ∈ { 1,2 ..., L }.
Data segmentation is handled using slip window sampling the continuous data in each dimension of sensor data set, is adopted It is 2 seconds with length, i.e., the sliding window that is 100 comprising data sampling points is set as 50% between two adjacent windows and covers Lid.Collected continuous data is divided into the data segment of floor (T/50-1) number by the slip window sampling, wherein floor () For downward bracket function.Segmentation obtains summary data field number K=9551.
Feature set F is by mean value, standard deviation, variance, the degree of bias, kurtosis, maximum value, minimum value, mean square deviation, interquartile range, flat Equal absolute deviation, 11 category feature of coverage area are constituted, and every category feature generates 16 spies to P+Q=16 data axis of data segment Sign, therefore symbiosis is at N=11 × 16=176 features.
Enter feature extraction process by pretreated data.Feature extraction is the feature meter to data contained by data segment It calculates.To each data segment of above-mentioned acquisition, calculate separately the N item feature in feature set F, obtain each data segment calculate feature to Measure Vk=[v1, v2..., vN]∈RN, k ∈ { 1,2 ..., K }.The feature vector of all acquisitions forms sample set S=[V1; V2;...;VK]。
By each feature vector V in above-mentioned sample set SkWith position mark piForm training vectorpi∈P.By training vectorComposing training collection For The training of device location disaggregated model.
Training device location disaggregated model is using C4.5 Method And Principle training Decision-Tree Classifier Model.In training, each Split vertexes select the highest feature of information gain-ratio as splitting condition.The calculation of information gain-ratio is as follows:
In above formula, molecule indicates to divide bring information gain, the i.e. reduction of comentropy, Entropy (T) table by training set Show the comentropy gathered before division,Gather after expression division.Wherein Entropy is collection The comentropy of conjunction, piIt is the ratio for belonging to classification i in training set T, m is the number of target category;V (f) is the possible institute of feature f There is value set, v is the value set for wherein meeting splitting condition;TvIt is the son that characteristic value meets splitting condition in training set T Collection,It is TvThe ratio denominator term of middle feature vector number and the feature vector number in former set T indicates division Information Meter, Wherein;TiIt is the corresponding feature vector set of target category i,It is the corresponding feature vector number of target category i and former set The ratio of feature vector number in T.
According to above method principle, training forms Decision-Tree Classifier Model, and the size of the decision tree is 137, uses feature Number is 69, number of features in about feature set F69 above-mentioned feature composition characteristic collection, this feature collection are used for The feature set F of device location identificationp, as shown in Figure 3.
It is assessed using the device class model that training set obtains training.Assessment has used ten folding cross-validation methods, The average recognition accuracy of obtained device location is 97.2%.
By each feature vector V and behavior category label a in above-mentioned sample set SjComposition training is inwardsaj∈ A, 6 kinds of human body behavior classifications when corresponding data acquires.By training vectorIt constitutes Training setFor feature selecting.Feature selecting includes filtering and encapsulating two step process.Filtering To each single item feature calculation information gain-ratio in feature set F, information gain-ratio such as formula (2) is described, big according to information gain-ratio Small to be ranked up, the forward M item feature of selected and sorted takes M=100.Then, combination supporting vector machine model was packaged Journey further screens filtered feature set.Encapsulation process comprehensively considers the accuracy rate and recall rate of classification, using branch The F-Measure of vector machine classification results is held as evaluation criterion, calculation is as follows:
Wherein the accurate rate of P presentation class model output, R presentation class model export recall rate.It is filtered all special The F-Measure obtained when sign input support vector machines is 93.12%.After the preceding filtering to sequential search strategy search one by one Feature set, when disaggregated model output meet setting threshold value, i.e. when F-Measure is more than 88.46%, output signature search As a result, the result includes 49 features.These features constitute the feature set F for being used for Human bodys' responsea
To sample set S, screening position classification is piFeature vector, form sample set Si, meetI ∈ 1, 2 ..., 8 }.To sample set SiIn each feature vectorBehavior classification a when flag data acquiresj, aj∈ A is obtained Behavioral training vectorki∈ { 1,2 ..., Li, all training vector constituting action training sets after label LiFor feature vector number in behavior training set.Utilize Behavioral training collectionInstruction Practice disaggregated model, obtains behavior disaggregated model cai.Training behavior disaggregated model caiIt is realized using support vector machine method.
The supporting vector machine model target category is human body behavior, including 6 kinds of target categories, using one-to-one multi-class Classification policy, respective functionTherefore 15 support vector machines of training is needed to constitute complete mould of classifying more Type.Each support vector machines selects the corresponding data of two kinds of behavior classifications, according to the training data of input calculate supporting vector and Parameter constructs largest interval Optimal Separating Hyperplane, and the Optimal Separating Hyperplane is by formulaIt is fixed Justice, wherein x is the feature vector of input, and wherein x is the feature vector of input, xiIt is supporting vector, yiIt is that supporting vector is corresponding As a result label, yiValue meets yi∈ { -1,1 }, n are supporting vector number, aiIt is the supporting vector parameter being calculated, K with b (x, xi) it is kernel function of the non-linear low-dimensional maps feature vectors for that will input to higher dimensional space realization linear separability;Core letter Number uses Radial basis kernel function, by formulaIt defines, wherein xcFor kernel function center, σ is the width of function Parameter is spent, σ=1 is taken.
I=1,2 ..., 8 is taken to above-mentioned supporting vector machine model respectively, is repeated in, and then training obtains 8 kinds of different positions Set the behavior disaggregated model ca in the case of classification1, ca2..., ca8.Using training set to training obtain based on different location The assessment of behavior disaggregated model, as a result such as table 1:
Behavior disaggregated model based on different location Behavior classification accuracy
Equipment is located at the behavior disaggregated model ca in pocket of trousers1 96.4%
Equipment is located at the behavior disaggregated model ca of coat pocket2 98.0%
Equipment is located at the behavior disaggregated model ca in knapsack3 98.0%
Equipment is located at the behavior disaggregated model ca in shoulder bag4 98.7%
Behavior disaggregated model ca when holding using equipment5 97.5%
Behavior disaggregated model ca when handheld device is conversed6 98.6%
Behavior disaggregated model ca of handheld device when side7 95.4%
Behavior disaggregated model ca when equipment is located on table8 -
Table 1 is assessed based on the behavior disaggregated model of different location
The behavior classification method of differentiation is not added to position using training set and behavior classification method provided by the invention carries out Assessment, such as table 2, it can be seen that layering behavior classification method provided by the invention has a distinct increment to Activity recognition accuracy rate.
Whether table 2 distinguishes two kinds of classification method accuracys rate comparison of device location
The mould as shown in Figure 1, above-mentioned trained device location disaggregated model and the behavior based on distinct device position are classified Type has built the Activity recognition model of layering.In real-time identification process, specifically locate in advance comprising real-time data acquisition, data Reason, feature extraction and device location identification and Activity recognition process.
Under at an arbitrary position, pass through acceleration transducer, gyroscope, barometer, the light sensor built in mobile device With range sensor with the frequency collection real time data of 50Hz.
The data of acquisition are calculated by Data correction, filtering, generation data, are preset when the data points of sampling meet Window considerations when, that is, when meeting new sampled point number and reaching 50, together with 50 groups of samples before the window at sampling The window that length is 100, forms the data segment of current window.
According to feature set Fp, feature extraction is carried out to the data segment of acquisition, obtains the corresponding feature vector of the data segmentNp=69;By feature vector VpInput equipment position disaggregated model cp。cpFor training The Decision-Tree Classifier Model finished, the feature vector V of inputpIt is selected step by step according to the splitting condition of decision tree qualified Branch, the final result classified at leaf node, i.e. position pi, pi∈P。
According to feature set Fa, feature extraction is carried out to the data segment of acquisition, obtains the corresponding feature vector of the data segmentNa=49;By feature vector VaInput and device location classification piCorresponding behavior point Class model cai。caiFor the support vector cassification model that training obtains, it is made of 15 support vector machines, each supporting vector Machine classifies to two kinds of behavior classifications.Optimal Separating Hyperplane expression formula corresponding to each support vector machines, according to the spy of input Levy vector VaCalculate functional value f (Va), obtain classification results (f (Va) >=0 or f (Va) < 0), then to all classification results It votes, number of votes obtained is highest after adding up exports as behavior classification results.
It, can be with it should be noted that the behavior classification in the embodiment of the present invention includes but are not limited to above-mentioned situation Other a variety of mobile devices are adapted to carry, using position, can also be to other a variety of behavior category classifications, it can be according to actual It is determined using needing.

Claims (9)

1. a kind of mobile device-based Human bodys' response method, it is characterised in that: the following steps are included:
Step (1), training device location disaggregated model cpWith the behavior disaggregated model ca based on distinct device positioni, cai∈ C, C It is the set being made of behavior disaggregated model, C={ ca1, ca2..., caI, behavior disaggregated model caiWith device location piHave One-to-one relationship, device location pi∈ P, P are the set being made of device location predetermined, P={ p1, p2..., pI, I is device location class number predetermined;
Step (2) acquires sensor raw data by mobile device built-in sensors in real time;
Step (3) carries out data prediction to the data that mobile device built-in sensors acquire in real time, obtains actual time window Corresponding data segment;
Step (4), according to position identification feature collection Fp, feature extraction is carried out to the data segment of acquisition, it is corresponding to obtain the data segment Feature vectorNpIt is characterized collection FpIn number of features;By feature vector VpInput is set Standby position disaggregated model cp, obtain device location classification pi
Step (5), according to Activity recognition feature set Fa, feature extraction is carried out to the data segment of acquisition, it is corresponding to obtain the data segment Feature vectorNaIt is characterized collection FaIn number of features;It is obtained according to step (4) Position classification piSelect corresponding behavior disaggregated model cai, by feature vector VaInput behavior disaggregated model cai, export behavior class Other aj, aj∈ A, A are the set being made of behavior predetermined, A={ a1, a2..., aJ, J is behavior class predetermined Other number, ajAs final recognition result;
Training device location disaggregated model c in the step (1)pWith behavior disaggregated model ca of the training based on distinct device positioni The step of it is as follows:
Step (1-1), in the case of acquiring distinct device position and different human body behavior offline by mobile device built-in sensors Sensor raw data;
Step (1-2) carries out data prediction to the sensor raw data of acquisition, obtains the data of corresponding different time window Section;
Step (1-3) carries out feature extraction to each data segment, obtains the corresponding feature vector of the data segment according to feature set F Vk=[v1, v2..., vN]∈RN, N is characterized the number of features in collection F, k ∈ { 1,2 ..., K }, the spy that all data segments extract Sign vector collectively forms sample set S, S=[V1;V2;...;VK], feature vector number K is data segment number;
Step (1-4), to each feature vector V in sample set Sk, flag data acquire when device location classification pi, pi∈ P obtains position training vectorK ∈ { 1,2 ..., K }, all training vectors after label constitute position instruction Practice collection Tp,
Step (1-5), utilizes training set TpTraining device location disaggregated model, obtains device location disaggregated model cp
Step (1-6), according to disaggregated model cp, select to form position identification feature for the Partial Feature of the model in feature set F Collect Fp
Step (1-7) carries out feature selecting to feature set F, and selection is wherein formed suitable for the Partial Feature of Activity recognition model Activity recognition feature set Fa
Step (1-8), Screening Samples, which integrate, corresponds to device location category label as p in SiFeature vectorForm sample set Si, meet
Step (1-9), to sample set SiIn each feature vectorBehavior classification a when flag data acquiresi, aj∈ A, Obtain Behavioral training vector, ki∈ { 1,2 ..., Ki, all training vector constituting actions instruction after label Practice collection Ti a, KiFor sample set SiThe number of middle feature vector;
Step (1-10) utilizes Behavioral training collection Ti aTrain classification models obtain corresponding position piBehavior disaggregated model cai
To step (1-8)-(1-10), i=1,2 ..., I is taken to be repeated in respectively, I indicates device location classification predetermined Number, training obtain the behavior disaggregated model ca under distinct device position1, ca2..., CaI
2. mobile device-based Human bodys' response method as described in claim 1, it is characterised in that: in step (1-1) Sensor raw data include data on different mobile device built-in sensors and same sensor different directions;Each Sensor raw data, that is, X that sampled point obtainsl=[x1, x2..., xP]∈RP, P be sensor raw data dimension, l ∈ 1, 2 ..., L }, L is the total sampling length of sensing data.
3. mobile device-based Human bodys' response method as claimed in claim 1 or 2, it is characterised in that: the step (3) data prediction in includes the following steps:
Step (3-1) is corrected and filters to sensor raw data;
Step (3-2) calculates according to correction and filtered data and generates data Yl,
Yl=[y1, y2..., yQ]∈RQ, Q is to generate data dimension;By sensor raw data and data merging is generated, is obtained Zl, Zl=[Xl, Yl]∈RP+Q, l ∈ { 1,2 ..., L };Data group on all sampled points is combined into sensor data set, i.e. [Z1; Z2;...;ZL];
Continuous sensor data set is divided into data segment according to time window by the segmentation of step (3-3) data.
4. mobile device-based Human bodys' response method as claimed in claim 1 or 2, it is characterised in that: the step (4) it is by feature vector V that device location classification is obtained inpInput position disaggregated model cp, cpThe decision tree mould obtained for training Type;The feature vector of input selects qualified branch according to the splitting condition of decision tree step by step, obtains at leaf node Classification as a result, i.e. position pi, pi∈P。
5. mobile device-based Human bodys' response method as claimed in claim 1 or 2, it is characterised in that: the step (5) acquisition behavior classification is by feature vector V inaInput behavior disaggregated model cai, caiThe support vector machines point obtained for training Class model is made of g (J) a support vector machines, and function g (x) determines that J is class object by the multi-class classification policy of model That is the number of behavior classification;The corresponding Optimal Separating Hyperplane expression formula of each support vector machines is f (x), according to the feature of input to It measures V and calculates functional value f (V), obtain classification results, f (V) >=0 or f (V) < 0;To the classification results of all support vector machines into Row weighting processing, takes peak to export as behavior classification results.
6. mobile device-based Human bodys' response method as described in claim 1, it is characterised in that: in step (1-3) Feature set F by data segment per one-dimensional corresponding time domain, frequency domain or time and frequency domain characteristics form.
7. mobile device-based Human bodys' response method as described in claim 1, it is characterised in that: in step (1-5) Trained device location disaggregated model is decision-tree model, which is made of the classifying rules of one group of " if yes " form Tree structure;In training, each split vertexes select the highest feature of evaluation function value as splitting condition.
8. mobile device-based Human bodys' response method as described in claim 1, it is characterised in that: in step (1-6) Feature selecting selected from feature set F for position disaggregated model input feature set Fp,FpUsing embedded mode Screening, i.e. Decision-Tree Classifier Model c of the training for position identificationpAfterwards, the feature used as selected by Decision-Tree Classifier Model Composition characteristic collection Fp
Feature selecting in step (1-7) selects the feature set F for the input of behavior disaggregated model from feature set Fa,FaUsing filtering-encapsulation two-stage process screening;Filter process selection evaluation function successively evaluates all features And sequence, screening and assessment function result are highest M first;Encapsulation process passes through the output knot of combination supporting vector machine disaggregated model Fruit, search meets the minimal characteristic set of output threshold condition, as the feature set F inputted for behavior disaggregated modela
9. mobile device-based Human bodys' response method as claimed in claim 2, it is characterised in that: in step (1-10) Trained behavior disaggregated model is supporting vector machine model, which is made of g (J) a support vector machines;Each supporting vector Machine constructs largest interval Optimal Separating Hyperplane according to the training data of input, and the Optimal Separating Hyperplane is by formulaDefinition, wherein x is the feature vector of input, xiIt is supporting vector, yiIt is to support The corresponding result label of vector, yiValue meets yi∈ { -1,1 }, n are supporting vector number, aiIt is the support being calculated with b Vector parameter, K (x, xi) it is kernel function for input vector to be mapped to high dimension linear space.
CN201511009411.8A 2015-12-29 2015-12-29 A kind of mobile device-based Human bodys' response method Active CN105678222B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201511009411.8A CN105678222B (en) 2015-12-29 2015-12-29 A kind of mobile device-based Human bodys' response method

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201511009411.8A CN105678222B (en) 2015-12-29 2015-12-29 A kind of mobile device-based Human bodys' response method

Publications (2)

Publication Number Publication Date
CN105678222A CN105678222A (en) 2016-06-15
CN105678222B true CN105678222B (en) 2019-05-31

Family

ID=56297832

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201511009411.8A Active CN105678222B (en) 2015-12-29 2015-12-29 A kind of mobile device-based Human bodys' response method

Country Status (1)

Country Link
CN (1) CN105678222B (en)

Families Citing this family (22)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107643908A (en) * 2016-07-20 2018-01-30 中兴通讯股份有限公司 Voice application trigger control method, device and terminal
CN106295675B (en) * 2016-07-25 2019-09-27 南京邮电大学 A kind of Activity recognition method based on smart phone of high accuracy
CN106123911A (en) * 2016-08-06 2016-11-16 深圳市爱康伟达智能医疗科技有限公司 A kind of based on acceleration sensor with the step recording method of angular-rate sensor
CN107918688B (en) * 2016-10-10 2020-02-28 深圳云天励飞技术有限公司 Scene model dynamic estimation method, data analysis method and device and electronic equipment
CN106709471A (en) * 2017-01-05 2017-05-24 宇龙计算机通信科技(深圳)有限公司 Fall detection method and device
CN107092894A (en) * 2017-04-28 2017-08-25 孙恩泽 A kind of motor behavior recognition methods based on LSTM models
CN107495972A (en) * 2017-08-14 2017-12-22 哈尔滨工业大学(威海) One kind falls down detection algorithm
CN108196998A (en) * 2018-01-02 2018-06-22 联想(北京)有限公司 A kind of state identification method, mobile equipment and server
CN108245172B (en) * 2018-01-10 2020-04-07 山东大学 Human body posture recognition method free of position constraint
CN108392208B (en) * 2018-02-23 2020-12-15 大同公元三九八智慧养老服务有限公司 Falling judgment method based on multi-stage model
CN108921047B (en) * 2018-06-12 2021-11-26 江西理工大学 Multi-model voting mean value action identification method based on cross-layer fusion
CN108985195A (en) * 2018-06-29 2018-12-11 平安科技(深圳)有限公司 Activity recognition method, apparatus, computer equipment and storage medium
CN109784418B (en) * 2019-01-28 2020-11-17 东莞理工学院 Human behavior recognition method and system based on feature recombination
CN109889672B (en) * 2019-04-01 2020-12-25 中国科学技术大学 Friend recommendation method based on mobile phone sensor
CN111814523A (en) * 2019-04-12 2020-10-23 北京京东尚科信息技术有限公司 Human body activity recognition method and device
CN110427969B (en) * 2019-07-01 2020-11-27 创新先进技术有限公司 Data processing method and device and electronic equipment
CN110458033B (en) * 2019-07-17 2023-01-03 哈尔滨工程大学 Human body behavior sequence identification method based on wearable position sensor
CN110598744A (en) * 2019-08-12 2019-12-20 浙江大学 Real-time human body behavior recognition system and method based on inertial sensor and Edge TPU
CN111582361A (en) * 2020-05-06 2020-08-25 浙江大学 Human behavior recognition method based on inertial sensor
CN112471680A (en) * 2020-12-21 2021-03-12 浙江工业大学 Intelligent safety helmet capable of identifying working state of electric power worker and working method of intelligent safety helmet
CN112782365B (en) * 2020-12-30 2022-08-23 江苏智闻智能传感科技有限公司 Mixed gas identification method based on multistage classifier
CN115731602A (en) * 2021-08-24 2023-03-03 中国科学院深圳先进技术研究院 Human body activity recognition method, device, equipment and storage medium based on topological representation

Family Cites Families (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8165405B2 (en) * 2006-12-18 2012-04-24 Honda Motor Co., Ltd. Leveraging temporal, contextual and ordering constraints for recognizing complex activities in video
CN103268495B (en) * 2013-05-31 2016-08-17 公安部第三研究所 Human body behavior modeling recognition methods based on priori knowledge cluster in computer system
CN103886323B (en) * 2013-09-24 2017-02-15 清华大学 Behavior identification method based on mobile terminal and mobile terminal
CN105046215B (en) * 2015-07-07 2018-11-16 中国科学院上海高等研究院 The posture Activity recognition method not influenced by individual wearing position and wearing mode

Also Published As

Publication number Publication date
CN105678222A (en) 2016-06-15

Similar Documents

Publication Publication Date Title
CN105678222B (en) A kind of mobile device-based Human bodys&#39; response method
Yen et al. Human daily activity recognition performed using wearable inertial sensors combined with deep learning algorithms
Henpraserttae et al. Accurate activity recognition using a mobile phone regardless of device orientation and location
CN105310696B (en) A kind of fall detection model building method and corresponding fall detection method and device
CN106096662B (en) Human motion state identification based on acceleration transducer
Uddin et al. Random forests based recognition of human activities and postural transitions on smartphone
Mekruksavanich et al. Exercise activity recognition with surface electromyography sensor using machine learning approach
CN106228200A (en) A kind of action identification method not relying on action message collecting device
CN108171278B (en) Motion pattern recognition method and system based on motion training data
CN103400123A (en) Gait type identification method based on three-axis acceleration sensor and neural network
Ahmed et al. An approach to classify human activities in real-time from smartphone sensor data
CN105023022A (en) Tumble detection method and system
CN112464738A (en) Improved naive Bayes algorithm user behavior identification method based on mobile phone sensor
CN110532898A (en) A kind of physical activity recognition methods based on smart phone Multi-sensor Fusion
WO2023035093A1 (en) Inertial sensor-based human body behaviour recognition method
Al-Ghannam et al. Prayer activity monitoring and recognition using acceleration features with mobile phone
CN110674875A (en) Pedestrian motion mode identification method based on deep hybrid model
CN106295675A (en) A kind of Activity recognition method of based on smart mobile phone of high accuracy
Hung et al. Activity recognition with sensors on mobile devices
CN109770913A (en) A kind of abnormal gait recognition methods based on reverse transmittance nerve network
Wang et al. Energy expenditure estimation during normal ambulation using triaxial accelerometry and barometric pressure
Harasimowicz et al. Accelerometer-based human activity recognition and the impact of the sample size
Zhang et al. An efficient feature selection method for activity classification
CN111772639B (en) Motion pattern recognition method and device for wearable equipment
CN105760819A (en) Daily activity recognition method based on acceleration signal

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant