CN111354458A - Touch interactive motion user feature extraction method based on general drawing task and auxiliary disease detection system - Google Patents

Touch interactive motion user feature extraction method based on general drawing task and auxiliary disease detection system Download PDF

Info

Publication number
CN111354458A
CN111354458A CN201811591897.4A CN201811591897A CN111354458A CN 111354458 A CN111354458 A CN 111354458A CN 201811591897 A CN201811591897 A CN 201811591897A CN 111354458 A CN111354458 A CN 111354458A
Authority
CN
China
Prior art keywords
over
track
points
task
parameters
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201811591897.4A
Other languages
Chinese (zh)
Other versions
CN111354458B (en
Inventor
田丰
黄进
朱以诚
翟菲菲
王宏安
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Institute of Software of CAS
Original Assignee
Institute of Software of CAS
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Institute of Software of CAS filed Critical Institute of Software of CAS
Priority to CN201811591897.4A priority Critical patent/CN111354458B/en
Publication of CN111354458A publication Critical patent/CN111354458A/en
Application granted granted Critical
Publication of CN111354458B publication Critical patent/CN111354458B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G16INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
    • G16HHEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
    • G16H50/00ICT specially adapted for medical diagnosis, medical simulation or medical data mining; ICT specially adapted for detecting, monitoring or modelling epidemics or pandemics
    • G16H50/20ICT specially adapted for medical diagnosis, medical simulation or medical data mining; ICT specially adapted for detecting, monitoring or modelling epidemics or pandemics for computer-aided diagnosis, e.g. based on medical expert systems
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/40Detecting, measuring or recording for evaluating the nervous system
    • A61B5/4076Diagnosing or monitoring particular conditions of the nervous system
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/03Arrangements for converting the position or the displacement of a member into a coded form
    • G06F3/033Pointing devices displaced or positioned by the user, e.g. mice, trackballs, pens or joysticks; Accessories therefor
    • G06F3/0354Pointing devices displaced or positioned by the user, e.g. mice, trackballs, pens or joysticks; Accessories therefor with detection of 2D relative movements between the device, or an operating part thereof, and a plane or surface, e.g. 2D mice, trackballs, pens or pucks
    • G06F3/03545Pens or stylus
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0487Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser
    • G06F3/0488Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures
    • GPHYSICS
    • G16INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
    • G16HHEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
    • G16H40/00ICT specially adapted for the management or administration of healthcare resources or facilities; ICT specially adapted for the management or operation of medical equipment or devices
    • G16H40/60ICT specially adapted for the management or administration of healthcare resources or facilities; ICT specially adapted for the management or operation of medical equipment or devices for the operation of medical equipment or devices
    • G16H40/63ICT specially adapted for the management or administration of healthcare resources or facilities; ICT specially adapted for the management or operation of medical equipment or devices for the operation of medical equipment or devices for local operation

Landscapes

  • Engineering & Computer Science (AREA)
  • Health & Medical Sciences (AREA)
  • Biomedical Technology (AREA)
  • General Engineering & Computer Science (AREA)
  • Medical Informatics (AREA)
  • Theoretical Computer Science (AREA)
  • Public Health (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Physics & Mathematics (AREA)
  • General Health & Medical Sciences (AREA)
  • Primary Health Care (AREA)
  • Epidemiology (AREA)
  • Pathology (AREA)
  • Neurology (AREA)
  • General Physics & Mathematics (AREA)
  • Human Computer Interaction (AREA)
  • Heart & Thoracic Surgery (AREA)
  • Surgery (AREA)
  • Animal Behavior & Ethology (AREA)
  • Molecular Biology (AREA)
  • Biophysics (AREA)
  • Veterinary Medicine (AREA)
  • Business, Economics & Management (AREA)
  • General Business, Economics & Management (AREA)
  • Physiology (AREA)
  • Neurosurgery (AREA)
  • Data Mining & Analysis (AREA)
  • Databases & Information Systems (AREA)
  • Measurement Of The Respiration, Hearing Ability, Form, And Blood Characteristics Of Living Organisms (AREA)

Abstract

The invention belongs to the field of digital medical treatment, and particularly relates to a touch interactive motion user feature extraction method and an auxiliary disease detection system based on a general drawing task. The user characteristics which are irrelevant to the task are extracted by using the stylus and the drawing board, the hand motion function of a patient with the nervous system disease is evaluated, a general drawing task-based nervous system disease daily monitoring and early warning system is realized, a classification model of the system is analyzed and verified, and the result shows that the system can accurately predict the nervous system disease. The invention discloses a system for automatically identifying nervous system diseases in a general drawing task by using a pen interaction technology, an adopted feature extraction method can be used for constructing a robust and accurate identification model, and the drawing system constructed according to the system can be used for automatic diagnosis in the field including but not limited to nervous system disease detection under the condition of no supervision and no task relevance.

Description

Touch interactive motion user feature extraction method based on general drawing task and auxiliary disease detection system
Technical Field
The invention belongs to the field of digital medical treatment, and particularly relates to a diagnostic interactive software system for nervous system diseases, which utilizes a touch interactive motion user characteristic extraction method irrelevant to a specific drawing task and utilizes a touch pen, a drawing screen and a common desktop computer as input equipment.
Background
The nervous system diseases are caused by various causes such as bacterial infection or heredity in the central nervous system, peripheral nervous system, and autonomic nervous system, and the symptoms thereof are usually manifested as various problems such as dyskinesia and cognitive dysfunction, for example, decreased motor ability, decreased language ability, impaired memory, and memory loss. Common nervous system diseases such as Parkinson's disease, Alzheimer's disease and cerebral small vessel disease are clinically manifested as motor dysfunction such as resting tremor, muscular rigidity and balance disorder, and cognitive dysfunction such as memory loss, language and directional dysfunction.
Due to the change of environment and the aging of population, the number of patients suffering from nervous system diseases is increasing, and the diseases largely affect the quality of life of the patients and their families. Due to the complex pathogenesis, complex examination steps and high medical cost of the diseases, an economical and effective daily monitoring scheme is not obtained yet. With the continuous development of computer science and technology and the gradual deepening of human cognition on the relationship between physiological information and diseases, the method enables the physiological information to be obtained and analyzed through a computer, and the detection of the functional diseases of the nervous system becomes feasible.
The pen interaction is used as a mature touch interaction technology and has the characteristics of high interaction efficiency, low cognitive load and the like. The pen-moving state and the graph drawing result can effectively reflect the abnormal information of the nervous system disease patients on the motor and cognitive functions. For example, Atilla et al extract multiple features from handwriting, pressure and inclination of digital pen, and combine them to model train for diagnosing Parkinson's disease (reference: Atilla)
Figure BDA0001920465220000011
R ü diger Brause, Karsten Krakow.2006.hand writing Analysis for diagnosing and diagnosing Parkinson's disease. in Proceedings of the 7th international relationship on the parkinsons disease using digital pen-extracted information, finding features for quantitative assessment of bradykinesia, font undersize and hand tremor symptoms (reference: heat J Smits, Antti Tonen, Luc clinical, AIkvarls, Bernard A Constrain, Rutger C sizing, U.S. Pat. No. 5. see the relevant neural system of the diagnosis of neuropathy, U.S. Pat. No. 5. the important parameters of the diagnosis of neuropathy, I.S. Pat. No. 5. 7. the important parameters of the diagnosis of the disease, I.S. Pat. No. 5. 7. the important parameters of the diagnosis of the nerve system of the heart, I.S. 5. and 7. the important parameters of the diagnosis of the disease using digital pen-reading test system of the heart, I.S. 5. the diagnosis of the disease using the heart test of the heart test system of the heart and the heart test of the heart disease (reference: the important parameters of the diagnosis of the nerve of the disease, I.S. 7. the diagnosis of the heart, the diagnosis of the disease using heart test of the heart test system of the heart, the diagnosis of the heart, the heart test of the diagnosis of the heart, the diagnosis of the disease using heart, the heart test of the heart disease using the heart test system of the heart, the heart test of the heart, the heart test of the diagnosis of the heart of the diagnosis of the heart of the diagnosis of the heart of the diagnosis of the heart of the disease of the diagnosis of the heart of the diagnosis of the heart of the diagnosis of the heart of the diagnosis of the disease of the heart of the diagnosis of the disease of the diagnosis of the: stuart Hagler, Holly Jimison, Misha Pavel.2014. insulating functional a computer gate. effective journal of biological and Health information, 18(4),1442 and 1452.) the diagnosis effect of the method on different nervous system diseases was evaluated.
However, the above detection methods are specific tasks, the feature extraction method must be modified according to the designated mapping task, and the execution of the above tests must be performed under the supervision of medical staff, and the long-term real-time monitoring of patients in daily life cannot be realized.
Disclosure of Invention
The invention aims to provide a touch interactive motion user feature extraction method and an auxiliary disease detection system based on a general drawing task.
The invention considers the user characteristics irrelevant to the task to evaluate the hand motion function of the patient with nervous system disease, and comprises data acquisition, track segmentation, parameter calculation, characteristic extraction and classification diagnosis.
The technical scheme adopted by the invention is as follows:
a touch interactive motion user feature extraction method based on a general drawing task comprises the following steps:
1) collecting drawing tracks of different drawing tasks of a user;
2) carrying out track segmentation on the collected drawing tracks to form a task-independent sub-track set;
3) calculating motion parameters of all points of any section of sub-track in the sub-track set;
4) track-level features are extracted from the motion parameters and statistical analysis is performed on all sub-tracks to extract graph-level features.
Further, step 1) collecting a drawing track of a user by using a touch pen and a drawing screen; the touch control pen is provided with a pressure sensor and a gyroscope and can acquire X and Y direction positions, touch control point pressure and touch control equipment space postures, and the touch control equipment space postures comprise azimuth angles, elevation angles and self-rotation angles.
Further, the drawing trace of the user collected in step 1) includes:
1.1) time equidistant trajectory, requiring the time interval between two sampling points to be consistent, for calculating time-domain correlation features
1.2) space equidistant trajectory, the space distance between two sampling points is required to be consistent, and the spatial equidistant trajectory is used for calculating the position correlation characteristics.
Further, the track segmentation of step 2) includes:
2.1) segmenting long and complex tracks by using local geometric extreme values;
and 2.2) clustering the dense over-segmentation points by utilizing time and space information, and selecting the optimal segmentation point as a final segmentation point.
Further, step 2.1) performing SER resampling on the track to obtain curvatures of all points and putting the curvatures into a linked list, then obtaining a local extreme value of the curvatures from the linked list by using a Gaussian filter, and screening out the over-cut points through a curvature threshold value and putting the over-cut points into an over-cut point list; and 2.2) traversing all the over-cutting points, recording other over-cutting points in a space-time domain range R of all the over-cutting points in a data structure named nearIndexes, then sorting the over-cutting points in a descending order according to the quantity of the nearIndexes of the over-cutting points, finally traversing the sorted over-cutting points, removing other over-cutting points recorded in the nearIndexes of each over-cutting point from a list of the over-cutting points, and taking the remaining list of the over-cutting points as the final over-cutting points.
Further, the motion parameters in the step 3) include behavior parameters and kinematic parameters; the behavior parameters are parameters directly reflecting the drawing behaviors and decisions of the user; the kinematic parameters are first and second order kinematic parameters obtained by deriving a part of the behavior parameters with respect to time.
Further, extracting track level characteristics by adopting a statistical method, a signal processing method and an information entropy method in the step 4); the statistical method comprises summation, mean, maximum value, minimum value, quartile and standard deviation; the signal processing method comprises a time domain method and a frequency domain method, wherein the time domain method utilizes zero crossing times, zero crossing frequency and root mean square to extract features, and the frequency domain method utilizes dominant frequency, dominant frequency energy and dominant frequency energy ratio to extract features; the information entropy method utilizes the fragrance concentration entropy and the Rayleigh entropy to extract features.
Further, the extracting of the graphics-level features in step 4) is to extract two types of graphics-level features of the whole drawing task: statistical attributes of all plot motion parameters, and statistical attributes of all sub-trajectory features.
A touch interactive motion user feature extraction device based on a general drawing task comprises:
the data acquisition module is responsible for acquiring drawing tracks of different drawing tasks of a user;
the track segmentation module is responsible for carrying out track segmentation on the collected drawing tracks to form a task-independent sub-track set;
the parameter calculation module is responsible for calculating motion parameters of all points of any section of sub-track in the sub-track set;
and the characteristic extraction module is responsible for extracting the track level characteristics from the motion parameters and performing statistical analysis on all the sub-tracks to extract the graph level characteristics.
A general drawing task-based auxiliary disease detection system comprises a general drawing task-based touch interactive motion user feature extraction device, a classification model training module and a disease diagnosis module; the classification model training module screens the characteristics related to the diseases from the characteristics extracted by the characteristic extraction module by using a characteristic selection algorithm to train a classification model; and the disease diagnosis module performs automatic diagnosis of diseases by using the trained classification model.
According to the invention, the touch pen and the drawing board are used for extracting the user characteristics irrelevant to the task, the hand motion function of the patient with the nervous system disease is evaluated, a general drawing task-based nervous system disease daily monitoring and early warning system is realized, the classification model of the system is analyzed and verified, the result shows that the scheme can accurately predict the nervous system disease, and a plurality of results capable of providing inspiration for the diagnosis of the nervous system disease are found.
The system constructed by the invention is the first system for automatically identifying nervous system diseases in a general drawing task by utilizing a pen interaction technology. The given feature extraction method can be used for constructing a robust and accurate recognition model, and various application programs can be constructed according to the method. The mapping system constructed in accordance with the present system can be used in unsupervised, task-independent situations for automated diagnosis including, but not limited to, the field of neurological disease detection.
Drawings
FIG. 1 is a system framework diagram of the present invention.
FIG. 2 is a diagram of hardware used in the present invention.
Fig. 3 is a flowchart of the algorithm workflow (universal trajectory segmentation algorithm workflow).
FIG. 4 is a diagram illustrating a graph level feature.
FIG. 5 is a flowchart of the steps for extracting user features from a drawing task.
Detailed Description
In order to make the aforementioned objects, features and advantages of the present invention comprehensible, the present invention shall be described in further detail with reference to the following detailed description and accompanying drawings.
The system framework of the invention is shown in fig. 1, and the hardware schematic diagram used is shown in fig. 2. And obtaining a feature set of the user irrelevant to a specific drawing task by using the user drawing information collected by the hardware equipment and using a universal track segmentation algorithm, and automatically performing classification diagnosis on diseases by using a machine learning algorithm. The details of the system involved are as follows:
1. data acquisition: the interactive software system for diagnosing the nervous system diseases integrates different drawing tasks, and acquires multi-dimensional drawing and motion information when a tested person performs a test by using a stylus (a digital pen in figure 1) and a drawing screen (a liquid crystal digital screen in figure 1) as input devices. The system is also responsible for the storage, management and preprocessing of these data.
Specifically, the user drawing trace data is acquired based on a stylus equipped with a pressure sensor and a gyroscope. Besides the X, Y directional position that can be obtained by a general touch device, it can also provide touch point pressure (pressure), and spatial attitude of the touch device, including azimuth (azimuth), elevation (elevation), and rotation angle (rotation).
The user drawing trace used includes the following two:
1) time equidistant trace, requiring the time interval between two sampling points to be consistent, for calculating time-domain correlation characteristics
2) And the spatial equidistant trajectory requires that the spatial distance between two sampling points is consistent for calculating the position correlation characteristics.
2. Track segmentation: and carrying out unified track segmentation on data obtained from different drawing tasks by using a universal track segmentation algorithm to form a task-independent sub-track set.
The general track segmentation algorithm: the algorithm mainly comprises two steps: 1) and segmenting the long and complex track by utilizing the local geometric extreme value, and intentionally not smoothing the track. 2) And clustering the dense over-segmentation points by using time and space information, and selecting the optimal segmentation point as a final segmentation point. The following describes the track segmentation algorithm employed in the present invention.
Firstly, an algorithm carries out SER resampling on a track, on the basis, curvatures of all points are obtained and are put into a linked list points _ cur, then, a Gaussian filter is used for obtaining a local extreme value of the curvatures from the points _ cur, and an over-cut point is screened out through a curvature threshold value A and is put into an over-cut point list.
The local extremum is calculated in the same way as for UMC. The related introduction to the UMC method is as follows:
in the track task, any attribute can be regarded as a signal s (n), and it is assumed that the signal includes two components: intentional Motion Component (IMC) i (n) and Unintentional Motion Component (UMC) u (n):
s(n)=i(n)+u(n)
to obtain the involuntary moving parts, the above equation is transformed:
u(n)=s(n)-i(n)
where s (n) can be obtained directly from the touch device, and i (n) is unknown. Estimating i (n) using a gaussian filter:
Figure BDA0001920465220000051
wherein the content of the first and second substances,
Figure BDA0001920465220000052
representing the estimated IMC at time n, s (n + k) is the signal obtained directly from the device, offset by k from time n, w is the length of half the processing time window, and σ represents the standard deviation in a gaussian filter. Because the original signal is smoothed by the Gaussian filter, the unintentional motion of the user is excluded
Figure BDA0001920465220000053
Furthermore, the UMC can thus be calculated by subtracting the IMC from the original signal:
Figure BDA0001920465220000054
secondly, traversing all the over-segmentation points by the algorithm, and recording other over-segmentation points in a space-time domain range R of all the over-segmentation points in a data structure named as nearIndexes of the over-segmentation points;
thirdly, sorting the over-cut points in a descending order according to the quantity of the nearIndexes of the over-cut points by an algorithm;
and finally, traversing the sorted over-cut points, and removing other over-cut points recorded in the nearIndexes of each over-cut point from the over-cut point list. Thus, the remaining list of over-cut points is the final cut point.
The workflow of the algorithm is shown in fig. 3. Wherein (a) the graph is an original stroke, and (b) the graph is an over-segmentation point screened by a UMC local curvature extreme value. The overcut points 1 and 7 are generated directly from the start and end points of the trajectory, and with UMC, some inappropriately large curvature points have been removed, but are still too dense at the several cut points 2, 3, 4, 5. The algorithm then calculates all of these overcut points' neighbors within the domain R (only the spatial neighborhood is considered in the illustration for simplicity), the neighborhood of overcut point 3 containing 2, 4, 5, which is the point with the most neighboring overcut points of all overcut points, and thus its neighborhood list containing 2, 4, 5 is removed, resulting in the final cutpoints 1, 3, 6, 7 as labeled in fig. 3 (c). The numbers 1-7 in the figure represent the cut points of the track, and the dotted circle in the figure represents the neighborhood range of the over-cut point.
3. And (3) parameter calculation: on the segmented track, the motion parameters of each point are calculated by taking track points as units, so that the user characteristics are extracted from the motion parameters by using a method of statistics, signal processing and information entropy in a later step 4. The motion parameters are mainly classified into behavior parameters and kinematics parameters. The behavior parameters are parameters which directly reflect the drawing behaviors and decisions of the user; kinematic parameters refer to first and second order motion parameters obtained by deriving a portion of the behavior parameters with respect to time.
4. Feature extraction: and extracting the drawing features from all the motion parameters by using a drawing feature extraction method irrelevant to the task, wherein the feature extraction comprises a feature extraction method and a feature extraction strategy.
1) The feature extraction method comprises the following steps:
the feature extraction method used in the algorithm comprises a statistical method, a signal processing method and an information entropy method.
a) And (5) a statistical method. Including Sum (Sum), Mean (Mean), maximum (Max), minimum (Min), quartile (Q1, Q2, and Q3), standard deviation STD.
b) Provided is a signal processing method. Including time domain methods and frequency domain methods. In the time domain, three signal processing methods are utilized: number of zero crossing times (NTZ), zero crossing frequency (RTZ), Root Mean Square (RMS). The three processing methods are explained as follows:
NTZ: for a signal which can take a positive value or a negative value, within a period of sampling time, if a value obtained by the t-th sampling is opposite to that obtained by the next sampling, the signal is regarded as a zero crossing point, namely, the NTZ value is increased by 1.
RTZ: the inverse of the value of NTZ is the zero-crossing frequency.
RMS: the calculation formula is as follows:
Figure BDA0001920465220000061
where N is the length of the signal, siIs the ith signal value in the signal sequence. Generally, RMS reflects the average intensity of a signal, which is related to the amplitude of the signal.
In the frequency domain, a Main Frequency (MF), a main frequency Energy (EMF), and a main frequency Energy Ratio (ER) are utilized. The computation of these three frequency domain features relies on a Fast Fourier Transform (FFT) algorithm. Assuming that the sampling frequency is Fs and the number of samples is N, the result of the FFT is N complex numbers, any one of which is associated with a frequency. More specifically, the relationship between the ith complex number and the frequency f is:
f=(i-1)×Fs/N
then, the amplitude at frequency f can be calculated by:
Figure BDA0001920465220000071
wherein z isf=af+bfi is the complex number associated with frequency f. Then we can obtain the energy at frequency f by squaring the amplitude:
Figure BDA0001920465220000075
then, EMF can be found from the maximum of the energy in all frequency domains, and correspondingly, MF is the frequency of the frequency domain with the maximum energy, and ER is the ratio of MF to the energy of all frequency domains.
c) An information entropy method. The user features are extracted by using two information entropies, including a fragrance entropy (Shannenentrypy) and a Rayleigh entropy (Rinyi entrypy). The entropy of the aroma of a random signal X is defined as follows:
Figure BDA0001920465220000072
where p (x) is the probability density function of the value x, which is estimated using MLE (Maximum Likelihood estimation):
Figure BDA0001920465220000073
where I is an indicator function, and therefore, p (x) reflects the probability that sample x appears in the population.
The Rayleigh entropy is an expansion form of the fragrance entropy, and is defined as follows:
Figure BDA0001920465220000074
where r ≧ 0 is the order of rayleigh entropy, and the aroma entropy is actually the case where r is 1, and the case where r is 2 is what is commonly called rayleigh entropy.
2) A feature extraction strategy:
the motion parameters are behavior and motion descriptions based on the user drawing task, which are motion information of "point" level, and for each sub-track, motion information of "track" level, i.e. track level feature (track level feature), is extracted by using a feature extraction method of statistics, information processing and information entropy, and finally, motion information of "graph" level, i.e. graph level feature, is extracted by performing statistical analysis on all sub-tracks of the whole graph, as shown in fig. 4. Where pi at a point level feature represents the ith point in the trajectory, di represents the deviation of the point from the point after filtering, ai represents the acceleration of the point, vi represents the velocity of the point, and ri represents the curvature of the point.
The feature extraction methods were divided into three "method groups" (MGs) that made them suitable for processing different types of data, as shown in table 1 below. MG1 includes statistical methods for extracting overall statistical attributes from the data. The MG2 processes vector data with timing information. Scalars with timing information can then be processed simultaneously using two sets of methods. MG3 is used to extract statistical attributes from the overall map data and features.
TABLE 1 all feature extraction methods and method group partitioning
Figure BDA0001920465220000081
The steps for extracting user features from a drawing task are as follows, and FIG. 5 illustrates these four steps:
(1) and processing the original track of the drawing task by using a general track segmentation algorithm to obtain a sub-track set.
(2) And for any section of sub-track, calculating the behavior parameters and the motion parameters of all points on the track.
(3) And extracting track level characteristics from the behavior parameters and the motion parameters of the track by using the MG1 and the MG 2.
(4) Extracting two types of graphic level features (graphic level features) of the whole drawing task: statistical attributes of all plot motion parameters, and statistical attributes of all sub-trajectory features. For the first type, MG1 is used to extract features from the motion parameters of all drawing points, referred to as SPM (static properties on all sampling points); for the second type, MG3 is used to extract features from all sub-track features, that is, "features of features", referred to as ssf (static properties on all track level features for short).
In step (3) of the feature extraction strategy, the MG1 and the MG2 are used, the types of the motion parameters are considered, and for different motion parameters, a matching method group is established, and all the matching method groups and the calculated feature numbers are shown in table 2 below.
TABLE 2 method set and calculated feature number
Figure BDA0001920465220000091
The correspondence of motion parameters to symbols in the above table is shown in table 3 below, from different motion parameters, trace-level features are extracted from sub-traces, then graphics-level features 32 × +36 × -940 SSF are extracted from these trace-level features, then statistical attributes are directly calculated from all point sets, graphics-level features 8 × -64 are extracted, and 64+ 940-1004 dimensional features are composed.
TABLE 3 motion parameters and symbols
Figure BDA0001920465220000101
5. And (3) classified diagnosis: on the obtained total data feature set, a feature selection algorithm is used to screen out features most relevant to diseases, and the features are used as input data (such as but not limited to decision tree (C4.5), Naive Bayes (NB) and linear Support Vector Machine (SVM)) of a classification model to train so as to obtain a classification model for automatic diagnosis of diseases.
1. The system hardware is composed of a touch pen, a displayable drawing board and a common desktop computer, for example, the touch pen is a Wacom drawing pen with the length of 14.4cm, the diameter of 1cm, the weight of 19g and the sampling rate of 100Hz, the model of the displayable drawing board is Wacom Cintiq DTK-1300, the screen size is 13.3 inches, and the resolution is 1280 × 800 pixels, the main reason for selecting the equipment is that besides basic handwriting, the equipment also supports the acquisition of data such as pressure sense and pen body posture, so that more comprehensive hand motion characteristics can be acquired, the software system is installed on the matched computer, is developed by a C # programming language, operates under a NET Framework 4.0, and performs the unified user data storage and management by using SQLite and a file system, the handwriting data structure is adopted by the system, the sampling point data set is an ordered sampling point data structure:
data={timestamp,x,y,pressure,azimuth,altitude,rotation,pen_event}
wherein timestamp is a timestamp, and PEN _ event is an enumeration type for marking PEN status, and is used for marking events such as PEN DOWN (PEN DOWN), drawing (PEN MOVE), PEN UP (PEN UP), and the like.
2. After the test is started, the system starts to record user data, and after the user test is completed, the system combines all sample data obtained by 5 tests to obtain a total data set containing 1070 samples (214 samples per test), which is called ODS (orthogonal dataset). Thereafter, the 6 data sets were analyzed separately.
3. All 5 test data are processed by using a task-independent drawing task feature extraction method, 1004-dimensional features are obtained in each sample of each test, feature selection is carried out on the feature set, the most important feature set in the evaluation of the nervous system disease-oriented hand motion function is found out, and the features are analyzed. The feature analysis is carried out on the data sets comprising the ODS and the 5 tests, the feature set which is excellent in different tasks can be obtained through the feature analysis of the ODS, and the feature set which is suitable for different degrees of freedom and different scenes can be obtained through the feature analysis of the 5 test data sets.
4. And (3) carrying out data cleaning and normalization, and specifically comprising the following steps:
1) removing samples lacking more than 30% of feature data; 2) calculating the mean value and the standard deviation of each dimension characteristic in the non-empty data; 3) replacing null value features of corresponding dimensions of the sample data by using the feature mean value of each dimension; 4) replacing feature data of each dimension beyond three times of standard deviation by using the feature mean value of each dimension; 5) the feature vectors extracted for the pathological features were normalized to [0,1] using the maximum minimum.
5. The optimal subset is searched by using a characteristic selection method based on correlation so as to increase the operation speed without losing the most main characteristic information, and the algorithm utilizes the following heuristic indexes to guide the searching process of the optimal subset:
Figure BDA0001920465220000111
wherein M issIs the goodness of the subset S containing k features,
Figure BDA0001920465220000112
is the average feature-class correlation coefficient,
Figure BDA0001920465220000113
is the average feature-feature internal correlation coefficient.
6. The feature selection algorithm picks 27 optimal feature subsets in the ODS, picking the corresponding features. The features are classified in three different ways to analyze the performance of different classes of features in the mapping task. (1) The relevant features are classified into 9 categories of position, UMO, curvature, velocity, acceleration, pressure, azimuth, altitude and rotation angle according to the motion parameter category. (2) The related features are divided into two categories of MG1 (statistics) and MG2 (signal and entropy) according to the track level feature extraction method group. (3) The related features are divided into two categories of SPM (statistical properties of point parameters) and SSF (statistical properties of trace-level features) according to the feature category of the graph level.
7. Training a diagnosis model:
in the stage of diagnosing model training, we can use different classifiers to train, and the implementation process is described in detail below by taking a support vector machine as an example.
Support Vector Machines (SVMs) are a class of supervised learning in machine learning, and can construct a hyperplane, which is used for classification or regression. For training of Support Vector Machines (SVMs), the following scheme is used:
1) and taking the features extracted in the steps as input, and labeling the sample.
2) And performing feature extraction on the collected user drawing tasks, and taking the extracted user drawing tasks as the input of a support vector machine to obtain a final classification result.
The method and system of the present invention can be used not only for the related hardware devices mentioned in the above specific trial solution, but also for deployment and use in, but not limited to, tablet, mobile phone, and interactive device with touch function.
The above embodiments are only intended to illustrate the technical solution of the present invention and not to limit the same, and a person skilled in the art can modify the technical solution of the present invention or substitute the same without departing from the spirit and scope of the present invention, and the scope of the present invention should be determined by the claims.

Claims (10)

1. A touch interactive motion user feature extraction method based on a general drawing task is characterized by comprising the following steps:
1) collecting drawing tracks of different drawing tasks of a user;
2) carrying out track segmentation on the collected drawing tracks to form a task-independent sub-track set;
3) calculating motion parameters of all points of any section of sub-track in the sub-track set;
4) track-level features are extracted from the motion parameters and statistical analysis is performed on all sub-tracks to extract graph-level features.
2. The method of claim 1, wherein step 1) collects a drawing trace of the user using a stylus and a drawing screen; the touch control pen is provided with a pressure sensor and a gyroscope and can acquire X and Y direction positions, touch control point pressure and touch control equipment space postures, and the touch control equipment space postures comprise azimuth angles, elevation angles and self-rotation angles.
3. The method of claim 1, wherein the step 1) of collecting the user's drawing trace comprises:
1.1) time equidistant trajectory, requiring the time interval between two sampling points to be consistent, for calculating time-domain correlation features
1.2) space equidistant trajectory, the space distance between two sampling points is required to be consistent, and the spatial equidistant trajectory is used for calculating the position correlation characteristics.
4. The method of claim 1, wherein step 2) the track slicing comprises:
2.1) segmenting long and complex tracks by using local geometric extreme values;
and 2.2) clustering the dense over-segmentation points by utilizing time and space information, and selecting the optimal segmentation point as a final segmentation point.
5. The method of claim 4, wherein step 2.1) SER resampling the trajectory to obtain curvatures for all points and placing them in a linked list, then using gaussian filter to obtain local extrema of curvature from the linked list, and screening out the over-cut points by a curvature threshold to place them in the over-cut point list; and 2.2) traversing all the over-cutting points, recording other over-cutting points in a space-time domain range R of all the over-cutting points in a data structure named nearIndexes, then sorting the over-cutting points in a descending order according to the quantity of the nearIndexes of the over-cutting points, finally traversing the sorted over-cutting points, removing other over-cutting points recorded in the nearIndexes of each over-cutting point from a list of the over-cutting points, and taking the remaining list of the over-cutting points as the final over-cutting points.
6. The method of claim 1, wherein the motion parameters of step 3) include both behavioral parameters and kinematic parameters; the behavior parameters are parameters directly reflecting the drawing behaviors and decisions of the user; the kinematic parameters are first and second order kinematic parameters obtained by deriving a part of the behavior parameters with respect to time.
7. The method of claim 1, wherein step 4) adopts a statistical method, a signal processing method and an information entropy method to extract track-level features; the statistical method comprises summation, mean, maximum value, minimum value, quartile and standard deviation; the signal processing method comprises a time domain method and a frequency domain method, wherein the time domain method utilizes zero crossing times, zero crossing frequency and root mean square to extract features, and the frequency domain method utilizes dominant frequency, dominant frequency energy and dominant frequency energy ratio to extract features; the information entropy method utilizes the fragrance concentration entropy and the Rayleigh entropy to extract features.
8. The method of claim 1, wherein the extracting of the graphics-level features of step 4) is extracting two types of graphics-level features of the whole drawing task: statistical attributes of all plot motion parameters, and statistical attributes of all sub-trajectory features.
9. A touch interactive motion user feature extraction device based on a general drawing task is characterized by comprising the following steps:
the data acquisition module is responsible for acquiring drawing tracks of different drawing tasks of a user;
the track segmentation module is responsible for carrying out track segmentation on the collected drawing tracks to form a task-independent sub-track set;
the parameter calculation module is responsible for calculating motion parameters of all points of any section of sub-track in the sub-track set;
and the characteristic extraction module is responsible for extracting the track level characteristics from the motion parameters and performing statistical analysis on all the sub-tracks to extract the graph level characteristics.
10. An auxiliary disease detection system based on a general drawing task, which is characterized by comprising the touch interactive motion user feature extraction device based on the general drawing task as claimed in claim 9, and a classification model training module and a disease diagnosis module; the classification model training module screens the characteristics related to the diseases from the characteristics extracted by the characteristic extraction module by using a characteristic selection algorithm to train a classification model; and the disease diagnosis module performs automatic diagnosis of diseases by using the trained classification model.
CN201811591897.4A 2018-12-20 2018-12-20 Touch interactive motion user feature extraction method and auxiliary disease detection system based on universal drawing task Active CN111354458B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201811591897.4A CN111354458B (en) 2018-12-20 2018-12-20 Touch interactive motion user feature extraction method and auxiliary disease detection system based on universal drawing task

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201811591897.4A CN111354458B (en) 2018-12-20 2018-12-20 Touch interactive motion user feature extraction method and auxiliary disease detection system based on universal drawing task

Publications (2)

Publication Number Publication Date
CN111354458A true CN111354458A (en) 2020-06-30
CN111354458B CN111354458B (en) 2023-11-14

Family

ID=71198111

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201811591897.4A Active CN111354458B (en) 2018-12-20 2018-12-20 Touch interactive motion user feature extraction method and auxiliary disease detection system based on universal drawing task

Country Status (1)

Country Link
CN (1) CN111354458B (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP4350489A1 (en) * 2022-10-07 2024-04-10 BIC Violex Single Member S.A. Writing instrument and method therefor

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20050234309A1 (en) * 2004-01-07 2005-10-20 David Klapper Method and apparatus for classification of movement states in Parkinson's disease
CN105068743A (en) * 2015-06-12 2015-11-18 西安交通大学 Mobile terminal user identity authentication method based on multi-finger touch behavior characteristics
CN106095104A (en) * 2016-06-20 2016-11-09 电子科技大学 Continuous gesture path dividing method based on target model information and system
CN107045393A (en) * 2016-02-05 2017-08-15 中国科学院软件研究所 Line test multi-channel data collecting method and system based on digital pen
CN107273677A (en) * 2017-06-08 2017-10-20 中国科学院软件研究所 A kind of multi-channel nerve function quantitative evaluation system
CN108962397A (en) * 2018-06-06 2018-12-07 中国科学院软件研究所 A kind of multichannel multitask the nervous system disease assistant diagnosis system based on pen and voice

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20050234309A1 (en) * 2004-01-07 2005-10-20 David Klapper Method and apparatus for classification of movement states in Parkinson's disease
CN105068743A (en) * 2015-06-12 2015-11-18 西安交通大学 Mobile terminal user identity authentication method based on multi-finger touch behavior characteristics
CN107045393A (en) * 2016-02-05 2017-08-15 中国科学院软件研究所 Line test multi-channel data collecting method and system based on digital pen
CN106095104A (en) * 2016-06-20 2016-11-09 电子科技大学 Continuous gesture path dividing method based on target model information and system
CN107273677A (en) * 2017-06-08 2017-10-20 中国科学院软件研究所 A kind of multi-channel nerve function quantitative evaluation system
CN108962397A (en) * 2018-06-06 2018-12-07 中国科学院软件研究所 A kind of multichannel multitask the nervous system disease assistant diagnosis system based on pen and voice

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP4350489A1 (en) * 2022-10-07 2024-04-10 BIC Violex Single Member S.A. Writing instrument and method therefor

Also Published As

Publication number Publication date
CN111354458B (en) 2023-11-14

Similar Documents

Publication Publication Date Title
Tonekaboni et al. Unsupervised representation learning for time series with temporal neighborhood coding
CN111046731B (en) Transfer learning method and recognition method for gesture recognition based on surface electromyographic signals
US20180144182A1 (en) Analyzing digital holographic microscopy data for hematology applications
Exarchos et al. EEG transient event detection and classification using association rules
Jadhav et al. Modular neural network based arrhythmia classification system using ECG signal data
CN108511055B (en) Ventricular premature beat recognition system and method based on classifier fusion and diagnosis rules
Fahimi et al. On metrics for measuring scanpath similarity
CN112818883B (en) Deep learning detection and positioning method for interested target based on eye movement signal
CN109817339A (en) Patient's group technology and device based on big data
Liu et al. Analyzing outliers cautiously
CN111783887B (en) Classified lie detection identification method based on fMRI (magnetic resonance imaging) small-world brain network computer
Jasm et al. Deep image mining for convolution neural network
CN107045624B (en) Electroencephalogram signal preprocessing and classifying method based on maximum weighted cluster
Raykov et al. Probabilistic modelling of gait for robust passive monitoring in daily life
Ingle et al. Lung Cancer Types Prediction Using Machine Learning Approach
CN111053552A (en) QRS wave detection method based on deep learning
CN108962379B (en) Mobile phone auxiliary detection system for cranial nerve system diseases
CN111354458A (en) Touch interactive motion user feature extraction method based on general drawing task and auxiliary disease detection system
Ng et al. Automated identification of epileptic and alcoholic EEG signals using recurrence quantification analysis
CN116484290A (en) Depression recognition model construction method based on Stacking integration
CN113171102B (en) ECG data classification method based on continuous deep learning
CN115358260A (en) Electroencephalogram sleep staging method and device, electronic equipment and storage medium
CN112022149B (en) Atrial fibrillation detection method based on electrocardiosignals
Toranjsimin et al. Robust Low Complexity Framework for Early Diagnosis of Autism Spectrum Disorder Based on Cross Wavelet Transform and Deep Transfer Learning
CN113133742B (en) Knee joint signal feature extraction and classification method and device based on time domain multidimensional feature fusion

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant