CN117828281B - Behavior intention recognition method, system and terminal based on cross-mode hypergraph - Google Patents

Behavior intention recognition method, system and terminal based on cross-mode hypergraph Download PDF

Info

Publication number
CN117828281B
CN117828281B CN202410247777.1A CN202410247777A CN117828281B CN 117828281 B CN117828281 B CN 117828281B CN 202410247777 A CN202410247777 A CN 202410247777A CN 117828281 B CN117828281 B CN 117828281B
Authority
CN
China
Prior art keywords
modal
cross
segment
feature
erasure
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202410247777.1A
Other languages
Chinese (zh)
Other versions
CN117828281A (en
Inventor
任卫红
董潜
高宇
王志永
王焦乐
刘洪海
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Shenzhen Graduate School Harbin Institute of Technology
Original Assignee
Shenzhen Graduate School Harbin Institute of Technology
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Shenzhen Graduate School Harbin Institute of Technology filed Critical Shenzhen Graduate School Harbin Institute of Technology
Priority to CN202410247777.1A priority Critical patent/CN117828281B/en
Publication of CN117828281A publication Critical patent/CN117828281A/en
Application granted granted Critical
Publication of CN117828281B publication Critical patent/CN117828281B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02DCLIMATE CHANGE MITIGATION TECHNOLOGIES IN INFORMATION AND COMMUNICATION TECHNOLOGIES [ICT], I.E. INFORMATION AND COMMUNICATION TECHNOLOGIES AIMING AT THE REDUCTION OF THEIR OWN ENERGY USE
    • Y02D10/00Energy efficient computing, e.g. low power processors, power management or thermal management

Landscapes

  • Image Analysis (AREA)

Abstract

The invention discloses a behavior intention recognition method, a system and a terminal based on cross-mode hypergraph, wherein the method comprises the following steps: obtaining a plurality of different segment modal characteristics of a target object in a target time period, performing time domain information enhancement processing and cross-modal enhancement processing on each segment modal characteristic to obtain a single-mode enhancement characteristic and a cross-mode enhancement characteristic corresponding to each segment modal characteristic, obtaining a pre-erasure time domain characteristic, a pre-erasure airspace characteristic, a post-erasure time domain characteristic and a post-erasure airspace characteristic corresponding to each node in a cross-mode hypergraph, fusing to obtain a final fused characteristic, and finally obtaining a behavior prediction result of the target object according to the final fused characteristic. The invention combines physical signals and physiological signals, fully utilizes the complementarity of information among different modes, realizes cross-mode interaction and enhancement in time dimension and space dimension, can effectively eliminate uncertainty among modes, and realizes cognition and behavior detection of patients.

Description

Behavior intention recognition method, system and terminal based on cross-mode hypergraph
Technical Field
The invention relates to the technical field of computer vision, in particular to a behavior intention recognition method, a system, a terminal and a computer readable storage medium based on cross-mode hypergraph.
Background
The behavior of ICU-acquired muscle weakness patients (ICU-AW) is intended to understand that the rehabilitation stages of patients are different from each other due to the complexity of ICU environments, and the problems of deep and hidden are two. The behavioral intention recognition system for ICU-AW patients uses computer vision techniques and machine learning techniques to extract information from the patient's physiological and physical signals to recognize the patient's current behaviors (shaking head, opening mouth, opening eyes, etc.) and intentions (seeking analgesic assistance, skin dryness, hunger, etc.).
At present, the behavior intention recognition methods aiming at patients are mainly divided into two main categories: the recognition method based on the single-mode signals and the recognition method based on the multi-mode signals are difficult to comprehensively understand the behavior intention of the patient by using the single-mode signals, and generally depend on a complex machine learning model, so that the result has poor interpretability. The complementarity of different modal signals can be utilized by using the multi-modal signals, additional clues are provided for semantic understanding and behavior analysis, the multi-modal signals can be divided into physical signals and physiological signals, but the problems of different time lags and multiple dimensions exist between the physical signals and the physiological signals, so that uncertainty among the multi-modal signals is caused, and effective alignment and fusion cannot be carried out.
Accordingly, the prior art is still in need of improvement and development.
Disclosure of Invention
The invention mainly aims to provide a behavior intention recognition method, a system, a terminal and a computer readable storage medium based on a cross-mode hypergraph, and aims to solve the problem that in the prior art, when behavior intention of a patient is recognized, physical signals and physiological signals are difficult to align and fuse effectively, so that the accuracy of a generated behavior intention prediction result is low.
In order to achieve the above object, the present invention provides a behavior intention recognition method based on a cross-modal hypergraph, the behavior intention recognition method based on the cross-modal hypergraph includes the following steps:
Acquiring a plurality of different modal signals of a target object in a target time period, and preprocessing the different modal signals according to the target time period and each modal signal to obtain a plurality of fragment modal characteristics corresponding to each modal signal;
Constructing a cross-modal hypergraph according to all the segment modal characteristics, and performing time domain information enhancement processing on each segment modal characteristic in the cross-modal hypergraph to obtain a single-mode enhancement characteristic corresponding to each segment modal characteristic;
Calculating cross-modal enhancement features corresponding to each segment modal feature according to a cross-modal attention mechanism, obtaining pre-erasure time domain features corresponding to each node in the cross-modal hypergraph according to the cross-modal enhancement features corresponding to each segment modal feature, and obtaining pre-erasure airspace features corresponding to each node according to the segment modal features of each node;
Performing time erasure processing on the cross-modal hypergraph according to the single-mode enhancement feature and the cross-mode enhancement feature corresponding to each segment mode feature, and acquiring an erased time domain feature and an erased space domain feature corresponding to each node in the cross-modal hypergraph after the time erasure processing;
And carrying out fusion processing on all the pre-erasure time domain features, all the pre-erasure space domain features, all the post-erasure time domain features and all the post-erasure space domain features to obtain final fusion features, and obtaining a behavior prediction result of the target object in a target time period according to the final fusion features.
Optionally, in the behavior intention recognition method based on cross-modal hypergraph, the obtaining a plurality of different modal signals of the target object in a target time period, and preprocessing according to the target time period and each modal signal to obtain a plurality of segment modal features corresponding to each modal signal specifically includes:
Acquiring a plurality of different modal signals of a target object in a target time period, wherein the different modal signals are acquired by different preset sensors;
Preprocessing each modal signal to obtain a modal characteristic corresponding to each modal signal;
Uniformly dividing each modal feature according to the preset time segment length to obtain a plurality of segment modal features corresponding to each modal feature
Optionally, in the behavior intention recognition method based on a cross-modal hypergraph, constructing a cross-modal hypergraph according to all the segment modal features, and performing time domain information enhancement processing on each segment modal feature in the cross-modal hypergraph to obtain a single-mode enhancement feature corresponding to each segment modal feature, which specifically includes:
Respectively integrating the segment modal characteristics corresponding to each modal signal to obtain nodes corresponding to each modal signal, and constructing a cross-modal hypergraph according to all the nodes;
performing linear mapping operation on segment modal characteristics of each node in the cross-modal hypergraph respectively to obtain a key vector, a value vector and a query vector corresponding to each segment modal characteristic;
constructing a similarity matrix according to the key vector, the value vector and the query vector corresponding to each segment modal feature;
and respectively carrying out time domain information enhancement processing on all the segment modal characteristics according to the similarity matrix to obtain single-mode enhancement characteristics corresponding to each segment modal characteristic.
Optionally, in the behavior intention recognition method based on a cross-modal hypergraph, calculating a cross-modal enhancement feature corresponding to each segment modal feature according to a cross-modal attention mechanism, and obtaining a pre-erasure time domain feature corresponding to each node in the cross-modal hypergraph according to the cross-modal enhancement feature corresponding to each segment modal feature, and obtaining a pre-erasure airspace feature corresponding to each node according to the segment modal feature of each node, including:
Acquiring a time segment corresponding to each segment modal feature, and classifying all the segment modal features according to the time segments to divide a plurality of segment modal features belonging to the same time into the same group to obtain a plurality of modal feature groups;
Taking any segment modal feature in any one modal feature group as a feature to be processed according to a cross-modal attention mechanism, and performing cross-modal enhancement processing on the feature to be processed according to all segment modal features in the same modal feature group as the feature to be processed to obtain a cross-modal enhancement feature corresponding to the feature to be processed;
after the cross-modal enhancement processing of all the segment modal features is completed, obtaining cross-modal enhancement features corresponding to each segment modal feature;
carrying out fusion processing on the single-mode enhancement features and the cross-mode enhancement features belonging to the same node to obtain pre-erasure time domain features corresponding to each node;
And integrating the segment modal characteristics of each node to obtain node characteristics corresponding to each node, and performing coding processing according to each node characteristic to obtain the pre-erasure airspace characteristics corresponding to each node.
Optionally, the behavior intention recognition method based on cross-mode hypergraph, wherein the encoding processing is performed according to each node feature to obtain a pre-erasure airspace feature corresponding to each node, specifically includes:
Respectively carrying out coding processing on each node characteristic through a public coder and a private coder to obtain a public characteristic and a private characteristic corresponding to each node characteristic;
Respectively carrying out graph distillation treatment on the public characteristic and the private characteristic of each node characteristic to obtain public distillation loss corresponding to each public characteristic and private distillation loss corresponding to each private characteristic;
Optimizing the corresponding public characteristics and private characteristics according to each public distillation loss and each private distillation loss to obtain final public characteristics and final private characteristics corresponding to each node;
and respectively carrying out fusion processing and linear mapping on the final public features and the final private features of each node to obtain the pre-erasure airspace features corresponding to each node.
Optionally, in the behavior intention recognition method based on a cross-modal hypergraph, the performing time erasure processing on the cross-modal hypergraph according to a single-mode enhancement feature and a cross-modal enhancement feature corresponding to each segment mode feature, and acquiring an post-erasure time domain feature and an post-erasure space domain feature corresponding to each node in the cross-modal hypergraph after the time erasure processing specifically includes:
respectively carrying out addition processing on the single-mode enhancement feature and the cross-mode enhancement feature corresponding to each segment mode feature to obtain an addition feature corresponding to each segment mode feature, and obtaining a time attention diagram of each segment mode feature according to each addition feature;
Performing time erasure processing on all the time attention force diagrams according to a preset time erasure template, and taking the time attention force diagrams meeting preset conditions as post-erasure time attention force diagrams;
Acquiring all the segment modal characteristics corresponding to the time attention map after erasure and taking the segment modal characteristics as the segment modal characteristics after erasure corresponding to each node;
performing time domain information enhancement processing on the erased fragment modal characteristics, and then calculating the erased fragment modal characteristics according to a cross-modal attention mechanism to obtain erased cross-modal enhancement characteristics corresponding to the erased fragment modal characteristics;
and obtaining the erased time domain characteristic corresponding to each node according to the erased single-mode characteristic and the erased cross-mode enhancement characteristic corresponding to each node, and obtaining the erased airspace characteristic corresponding to each node according to each erased segment mode characteristic.
Optionally, in the behavior intention recognition method based on cross-modal hypergraph, the fusing processing is performed on all the pre-erasure time domain features, all the pre-erasure space domain features, all the post-erasure time domain features and all the post-erasure space domain features to obtain a final fused feature, and a behavior prediction result of the target object in a target time period is obtained according to the final fused feature, which specifically includes:
Respectively carrying out fusion processing on the pre-erasure time domain features, the pre-erasure space domain features, the post-erasure time domain features and the post-erasure space domain features corresponding to each node to obtain space-time features corresponding to each node;
the space-time characteristics of all the nodes are connected in series to obtain the final fusion characteristics;
And analyzing and processing the final fusion characteristics by using a pre-trained graph neural network to obtain a behavior prediction result of the target object in a target time period.
In addition, in order to achieve the above object, the present invention further provides a behavior intention recognition system based on a cross-modal hypergraph, wherein the behavior intention recognition system based on the cross-modal hypergraph includes:
The segment characteristic acquisition module is used for acquiring a plurality of different modal signals of a target object in a target time period, and preprocessing the different modal signals according to the target time period and each modal signal to obtain a plurality of segment modal characteristics corresponding to each modal signal;
the single-mode enhancement module is used for constructing a cross-mode hypergraph according to all the segment mode characteristics, and carrying out time domain information enhancement processing on each segment mode characteristic in the cross-mode hypergraph to obtain single-mode enhancement characteristics corresponding to each segment mode characteristic;
The pre-erasure feature acquisition module is used for calculating cross-modal enhancement features corresponding to each segment modal feature according to a cross-modal attention mechanism, obtaining pre-erasure time domain features corresponding to each node in the cross-modal hypergraph according to the cross-modal enhancement features corresponding to each segment modal feature, and obtaining pre-erasure airspace features corresponding to each node according to the segment modal features of each node;
The post-erasure feature acquisition module is used for performing time erasure processing on the cross-modal hypergraph according to the single-mode enhancement feature and the cross-mode enhancement feature corresponding to each segment mode feature, and acquiring post-erasure time domain features and post-erasure space domain features corresponding to each node in the cross-modal hypergraph after the time erasure processing;
And the prediction result generation module is used for carrying out fusion processing on all the pre-erasure time domain features, all the pre-erasure space domain features, all the post-erasure time domain features and all the post-erasure space domain features to obtain final fusion features, and obtaining a behavior prediction result of the target object in a target time period according to the final fusion features.
In addition, to achieve the above object, the present invention also provides a terminal, wherein the terminal includes: a memory, a processor, and a cross-modal hypergraph-based behavior intent recognition program stored on the memory and executable on the processor, which when executed by the processor, implements the steps of the cross-modal hypergraph-based behavior intent recognition method as described above.
In addition, in order to achieve the above object, the present invention also provides a computer-readable storage medium storing a behavior intention recognition program based on a cross-modal hypergraph, which when executed by a processor, implements the steps of the behavior intention recognition method based on a cross-modal hypergraph as described above.
According to the method, a plurality of different modal signals of a target object in a target time period are obtained, and preprocessing is carried out according to the target time period and each modal signal to obtain a plurality of fragment modal characteristics corresponding to each modal signal; constructing a cross-modal hypergraph according to all the segment modal characteristics, and performing time domain information enhancement processing on each segment modal characteristic in the cross-modal hypergraph to obtain a single-mode enhancement characteristic corresponding to each segment modal characteristic; calculating cross-modal enhancement features corresponding to each segment modal feature according to a cross-modal attention mechanism, obtaining pre-erasure time domain features corresponding to each node in the cross-modal hypergraph according to the cross-modal enhancement features corresponding to each segment modal feature, and obtaining pre-erasure airspace features corresponding to each node according to the segment modal features of each node; performing time erasure processing on the cross-modal hypergraph according to the single-mode enhancement feature and the cross-mode enhancement feature corresponding to each segment mode feature, and acquiring an erased time domain feature and an erased space domain feature corresponding to each node in the cross-modal hypergraph after the time erasure processing; and carrying out fusion processing on all the pre-erasure time domain features, all the pre-erasure space domain features, all the post-erasure time domain features and all the post-erasure space domain features to obtain final fusion features, and obtaining a behavior prediction result of the target object in a target time period according to the final fusion features. The invention uses physical signals and physiological signals, fully utilizes the complementarity of information among different modes, realizes cross-mode interaction and enhancement in time dimension and space dimension, can effectively eliminate uncertainty among modes, and realizes cognition and behavior detection of patients.
Drawings
FIG. 1 is a flow chart of a preferred embodiment of the behavior intent recognition method based on cross-modal hypergraph of the present invention;
FIG. 2 is a schematic diagram of a hardware module of the behavior intent recognition method based on cross-modal hypergraph of the present invention;
FIG. 3 is a schematic diagram of signal presentation of a behavior intent recognition method based on cross-modal hypergraph of the present invention;
FIG. 4 is a flow chart of multi-modal signal acquisition of the behavior intent recognition method based on cross-modal hypergraph of the present invention;
FIG. 5 is a flow chart of time domain feature extraction of the behavior intention recognition method based on cross-modal hypergraph of the invention;
FIG. 6 is a flow chart of airspace feature extraction of the behavior intention recognition method based on cross-modal hypergraph;
FIG. 7 is a complete flow chart of behavior prediction for a behavior intent recognition method based on cross-modal hypergraph of the present invention;
FIG. 8 is a schematic diagram of a preferred embodiment of the behavior intent recognition system based on cross-modal hypergraphs of the present invention;
FIG. 9 is a schematic diagram of the operating environment of a preferred embodiment of the terminal of the present invention.
Detailed Description
In order to make the objects, technical solutions and advantages of the present invention more clear and clear, the present invention will be further described in detail below with reference to the accompanying drawings and examples. It should be understood that the specific embodiments described herein are for purposes of illustration only and are not intended to limit the scope of the invention.
According to the behavior intention recognition method based on the cross-modal hypergraph, as shown in fig. 1, the behavior intention recognition method based on the cross-modal hypergraph comprises the following steps:
Step S10, acquiring a plurality of different modal signals of a target object in a target time period, and preprocessing the different modal signals according to the target time period and each modal signal to obtain a plurality of fragment modal characteristics corresponding to each modal signal.
Specifically, in a preferred embodiment of the present invention, a behavioral intention recognition system for an ICU-AW patient is designed, where the system is a somatosensory network system suitable for ICU scenes, and includes a hardware module for acquiring modal signals of the patient, such as a wearable physiological and physical signal sensing module and a multi-view visual sensing module, so as to collect a plurality of different modal signals of the patient in a target time period, and then pre-process the acquired modal signals in different manners, so as to obtain a plurality of segment modal characteristics corresponding to each modal signal.
Further, the acquiring a plurality of different modal signals of the target object in a target time period, and preprocessing according to the target time period and each modal signal to obtain a plurality of segment modal characteristics corresponding to each modal signal, specifically includes:
Acquiring a plurality of different modal signals of a target object in a target time period, wherein the different modal signals are acquired by different preset sensors; preprocessing each modal signal to obtain a modal characteristic corresponding to each modal signal; and uniformly dividing each modal feature according to the preset time segment length to obtain a plurality of segment modal features corresponding to each modal feature.
Specifically, the hardware module of the invention is shown in fig. 2, which comprises a wearable physiological physical signal sensing module and a multi-view visual sensing module, wherein the wearable physiological physical signal sensing module comprises a scalp electroencephalogram module, an myoelectricity module, an ultrasonic module, an IMU (Inertial Measurement Unit ) module and a force touch sensing module, the multi-view sensing module mainly comprises a plurality of RGBD cameras (three-dimensional color depth cameras) and can capture facial expressions of a patient (target object) in real time, the behavior intention recognition system is matched with upper computer software and can display signals of all modes simultaneously so as to observe the state of the patient in real time, a left signal column can select to observe signals of a certain mode or signals of a plurality of modes simultaneously, the right signal column can display original signals in real time, and a menu column can control interface display and the like.
The system can realize independent work of each mode or synchronous acquisition of all modes, and when each mode sensor is arranged, repeated positions are subjected to integrated design so as to synchronously detect myoelectricity, ultrasound and IMU of limb movement. Because the acquisition of each mode adopts an independent module, the synchronous operation of each mode needs to be realized by using a synchronous center during synchronous acquisition so as to obtain time-aligned multi-mode signals. The method for synchronously aligning the multi-mode data is shown in fig. 4, the synchronous center synchronously transmits the signals for starting acquisition to each module to enable the modules to synchronously start working, the visual module positions and shares the time tag with the synchronous center, and the synchronous center transmits the time tag to each module to enable the visual module to be synchronized, so that a plurality of different mode signals of a patient in a target time period are obtained.
After a plurality of different modal signals are acquired, the system needs to synchronously preprocess each modal signal so as to obtain the corresponding modal characteristics of each modal signal, and the preprocessing modes of the different modal signals are different, specifically as follows:
Electroencephalogram signal: firstly, converting an electroencephalogram signal into a time-frequency domain feature map by using a wavelet transformation method, then constructing a space transformation matrix according to the space position distribution of an electroencephalogram acquisition electrode, converting the time-frequency domain feature map into a 3D feature map containing a time-frequency-space domain, and extracting features by using a convolutional neural network and a cyclic neural network; firstly, a convolutional neural network is used for extracting features of a frequency domain and a spatial domain, the features are extracted at the tail end of a structure, the features are converted into feature vectors, the feature vectors are used as input of a cyclic neural network, and the cyclic neural network is used for extracting time features to obtain preprocessed electroencephalogram features.
Myoelectric signal: denoising the electromyographic signals by using a low-pass filter, then learning the spatial characteristics of the high-density electromyographic signals by using a graph neural network, specifically, regarding different channel signals as independent nodes, modeling the inter-channel relationship of the signals by using an adjacency matrix, calculating the dependency relationship of each node by using a graph convolution module, and then extracting the electromyographic characteristics by using a depth separable convolution network.
Ultrasonic signal: the ultrasonic sensor is placed at the muscles such as the rectus leg muscle, the extensor carpi forearm, the deep flexor digitorum of finger, the longus hallucination muscle and the like to collect multi-channel ultrasonic signals, then a Gaussian band-pass filter is used for filtering ultrasonic echo signals to obtain echo data corresponding to each channel, the echo data of each channel are uniformly cut into a plurality of sections, the characteristics are extracted from the cut data by wavelet packet transformation, the characteristics extracted from the echo data of the same channel jointly form the characteristic vector of each channel, and finally the characteristic vectors of different ultrasonic channels are connected in series to obtain ultrasonic characteristics.
Haptic signal: according to the piezoelectric signals and the piezoresistive signals acquired by the haptic signal acquisition module, the piezoelectric signals and the piezoresistive signals are aggregated by using a convolution layer of a convolution neural network, then feature vectors of the piezoelectric signals and the piezoresistive signals are extracted by using a pooling layer to obtain piezoelectric feature vectors and piezoresistive feature vectors, and the piezoelectric feature vectors and the piezoresistive feature vectors are connected in series and then input into a multi-layer perceptron to obtain the haptic characteristics.
IMU signal (inertial signal): the IMU signal is low-pass filtered in time to eliminate high frequency interference, then a convolutional neural network is used to extract local features with the convolutional layer, and then a pre-trained transducer (a neural network model) is used to extract global features of the IMU as IMU features (inertial features).
Visual signal: RGBD cameras with different visual angles can shoot different human body parts. For a face, firstly, a face detection network is used for extracting the face in an image and then standardization is carried out according to the RGB (color mode) data of the image; for the body, different parts of the human body such as shoulders, waist, knees, feet, elbows and the like are abstracted into characteristic points, and each part corresponds to one characteristic point. And then extracting facial expression characteristics, key point characteristics of the human face, line of sight characteristics and skeleton point characteristics of the human body of a patient through a main characteristic extraction module of a pre-trained expression recognition network, a line of sight estimation network and a skeleton point monitoring network according to the human face and a plurality of characteristic points, and finally, performing serial splicing on the four characteristics and outputting the four characteristics through a multi-layer perceptron to obtain visual characteristics of the external behavior state of the patient.
After the electroencephalogram feature, myoelectricity feature, ultrasonic feature, IMU feature, force touch feature and visual feature of the patient are obtained, each modal feature is uniformly divided according to the preset time segment length, and a plurality of segment modal features corresponding to each modal feature are obtained; the brain electrical characteristic, myoelectrical characteristic, ultrasonic characteristic, IMU characteristic, force touch characteristic and visual characteristic of the patient are respectively provided as,/>,/>,/>,/>,/>Taking the example of recording and acquiring a plurality of modal characteristics within a target period of time of a patient, the target period of time is divided into/>Each time segment, each time segment randomly extracts k frames of continuous frames, and then the 3D-ResNet (a neural network) characteristic network is used for mapping the characteristic of each mode into the same dimension to obtain/>,/>Representing segment modality characteristics,/>Representing the number of time slices,/>Representing the dimension of the feature,/>The method is characterized by comprising a set of all segment modal characteristics of the same time period, namely the segment modal characteristics comprise segment brain electrical characteristics, segment myoelectrical characteristics, segment ultrasonic characteristics, segment IMU characteristics, segment force touch characteristics and segment visual characteristics.
And step S20, constructing a cross-modal hypergraph according to all the segment modal characteristics, and carrying out time domain information enhancement processing on each segment modal characteristic in the cross-modal hypergraph to obtain a single-mode enhancement characteristic corresponding to each segment modal characteristic.
Specifically, after the electroencephalogram feature, myoelectric feature, ultrasonic feature, IMU feature, force touch feature and visual feature of a patient are obtained, because the problems of different time lags and multiple dimensions exist among the multi-modal features, a cross-modal hypergraph information fusion algorithm is adopted, a graph substructure is utilized to integrate modal information, the graph is defined as a directed graph containing nodes and edges, space construction of the cross-modal hypergraph is carried out, wherein different nodes represent the features of different modalities, and edges represent the relationship among the nodes.
And then carrying out time domain information enhancement processing on all the segment modal characteristics contained in the cross-modal hypergraph through a self-attention mechanism to obtain single-mode enhancement characteristics corresponding to each segment modal characteristic.
Further, constructing a cross-modal hypergraph according to all the segment modal features, and performing time domain information enhancement processing on each segment modal feature in the cross-modal hypergraph to obtain a single-modal enhancement feature corresponding to each segment modal feature, which specifically comprises:
Respectively integrating the segment modal characteristics corresponding to each modal signal to obtain nodes corresponding to each modal signal, and constructing a cross-modal hypergraph according to all the nodes; performing linear mapping operation on segment modal characteristics of each node in the cross-modal hypergraph respectively to obtain a key vector, a value vector and a query vector corresponding to each segment modal characteristic; constructing a similarity matrix according to the key vector, the value vector and the query vector corresponding to each segment modal feature; and respectively carrying out time domain information enhancement processing on all the segment modal characteristics according to the similarity matrix to obtain single-mode enhancement characteristics corresponding to each segment modal characteristic.
Specifically, segment modal features corresponding to the same modal feature are integrated to obtain nodes corresponding to each modal feature, and then construction of a cross-modal hypergraph is completed according to all the nodes; for each node in the cross-mode hypergraph, the time domain information of each node needs to be enhanced, and considering that different segments in the time domain have correlation, the intra-mode correlation modeling is firstly carried out on the mode characteristics corresponding to each node, namely, the time domain information enhancement processing is respectively carried out on the segment mode characteristics belonging to the same node.
Taking the long-distance dependency relationship between different frames of a single modality as an example, taking a visual feature as an example, based on a self-attention mechanism, performing linear mapping on all segment visual features in nodes corresponding to the visual feature to obtain a corresponding Key (Key) vector, a Value (Value) vector and a Query (Query) vector, and using a similarity matrix for time correlation between segment visual features of different time segmentsTo express, specifically: /(I); Wherein/>For the scale-up and scale-down coefficients, Q is the query vector of all segment visual features, K is the key vector of all segment visual features, and the similarity matrix formula represents that normalization processing is performed on each line of data in brackets, that is, the time correlation between segment visual features corresponding to all different time segments in the visual feature nodes is encoded in the similarity matrix.
And then respectively carrying out time domain information enhancement processing on the visual characteristics of each segment according to the similarity matrix, wherein the specific formula is as follows: ; wherein G, F and V represent the segment visual features after enhancement (i.e., the single-mode enhancement features), the segment visual features before enhancement, and the value vectors corresponding to the segment visual features, respectively, and G and F have the same dimensions.
And respectively carrying out time domain information enhancement processing on all the segment modal characteristics of the rest nodes, thereby obtaining single-mode enhancement characteristics corresponding to all the segment modal characteristics in the cross-mode hypergraph.
Step S30, calculating cross-modal enhancement features corresponding to each segment modal feature according to a cross-modal attention mechanism, obtaining pre-erasure time domain features corresponding to each node in the cross-modal hypergraph according to the cross-modal enhancement features corresponding to each segment modal feature, and obtaining pre-erasure airspace features corresponding to each node according to the segment modal features of each node.
Specifically, after obtaining the unimodal enhancement feature corresponding to each segment modal feature in each node, performing cross-modal enhancement on a plurality of different unimodal enhancement features (in the preferred embodiment of the present invention, each time segment has six different unimodal enhancement features) of the same time segment, to obtain the cross-modal enhancement feature corresponding to each unimodal enhancement feature, and obtaining the pre-erasure time domain feature corresponding to each node according to all the cross-modal enhancement features corresponding to each node, and then obtaining the pre-erasure airspace feature corresponding to each node according to the segment modal feature of each node.
Further, the calculating a cross-modal enhancement feature corresponding to each segment modal feature according to a cross-modal attention mechanism, and obtaining a pre-erasure time domain feature corresponding to each node in the cross-modal hypergraph according to the cross-modal enhancement feature corresponding to each segment modal feature, and obtaining a pre-erasure airspace feature corresponding to each node according to the segment modal feature of each node specifically includes:
Acquiring a time segment corresponding to each segment modal feature, and classifying all the segment modal features according to the time segments to divide a plurality of segment modal features belonging to the same time into the same group to obtain a plurality of modal feature groups; taking any segment modal feature in any one modal feature group as a feature to be processed according to a cross-modal attention mechanism, and performing cross-modal enhancement processing on the feature to be processed according to all segment modal features in the same modal feature group as the feature to be processed to obtain a cross-modal enhancement feature corresponding to the feature to be processed; after the cross-modal enhancement processing of all the segment modal features is completed, obtaining cross-modal enhancement features corresponding to each segment modal feature; carrying out fusion processing on the single-mode enhancement features and the cross-mode enhancement features belonging to the same node to obtain pre-erasure time domain features corresponding to each node; and integrating the segment modal characteristics of each node to obtain node characteristics corresponding to each node, and performing coding processing according to each node characteristic to obtain the pre-erasure airspace characteristics corresponding to each node.
Specifically, firstly, a time segment corresponding to each segment modal feature is obtained, then, a plurality of segment modal features belonging to the same time segment are divided into the same group, a plurality of modal feature groups (actually, each modal feature group comprises six segment modal features) are obtained, then, information complementarity among different modalities is utilized, and a cross-modal attention is used for learning modal representation, as shown in fig. 5, by taking a visual feature as an example, the cross-modal attention is represented as: ; wherein/> AndRepresenting segment visual features in a set of modal features that are not cross-modal enhanced and then re-using the set of modal features according to the formula; Wherein/>For the number of cross-modal enhancements,/>Is a weight parameter,/>Representing remaining segment modal features other than segment visual features,/>Representing that the rest segment modal characteristics are used for enhancing segment visual characteristics, and in order to obtain cross-modal enhancement characteristics corresponding to the segment visual characteristics, segment electroencephalogram characteristics, segment myoelectric characteristics, segment ultrasonic characteristics, segment IMU characteristics and segment force touch characteristics in the modal characteristic group are sequentially used for cross-modal enhancement processing, so that cross-modal enhancement characteristics/>, corresponding to the segment visual characteristics, are obtainedIn practice, 5 cross-modal enhancement processes are performed on the remaining 1 segment modal features by using 5 segment modal features in the same modal feature group; and similarly, carrying out the same processing on all the segment modal characteristics in all the modal characteristic groups, and obtaining cross-modal enhancement characteristics corresponding to each segment modal characteristic after the processing is completed.
Then, carrying out fusion processing on all single-mode enhancement features and all cross-mode enhancement features of each node (each mode), and obtaining the time domain features before erasure corresponding to each node; and integrating the segment modal characteristics of each node to obtain node characteristics corresponding to each node, and simultaneously carrying out coding processing on the node characteristics to obtain the pre-erasure airspace characteristics corresponding to each node.
Further, the encoding processing is performed according to each node characteristic to obtain a pre-erasure airspace characteristic corresponding to each node, which specifically includes:
Respectively carrying out coding processing on each node characteristic through a public coder and a private coder to obtain a public characteristic and a private characteristic corresponding to each node characteristic; respectively carrying out graph distillation treatment on the public characteristic and the private characteristic of each node characteristic to obtain public distillation loss corresponding to each public characteristic and private distillation loss corresponding to each private characteristic; optimizing the corresponding public characteristics and private characteristics according to each public distillation loss and each private distillation loss to obtain final public characteristics and final private characteristics corresponding to each node; and respectively carrying out fusion processing and linear mapping on the final public features and the final private features of each node to obtain the pre-erasure airspace features corresponding to each node.
Specifically, as shown in fig. 6, after defining the pre-erasure time domain features of each node, the pre-erasure spatial domain features of each node need to be calculated, and a common encoder is used firstAnd private encoder/>The node characteristics of each node are respectively encoded, and the specific expression is as follows: /(I); Wherein/>And/>The node characteristics are represented by public characteristics existing in each node and private characteristics which belong to modes independently.
Then using knowledge distillation algorithm based on graph, for convenience of formula description, the method is now unifiedRepresenting common featuresAnd/>. Active/>Nodes (modalities) representing diagrams,/>Representing modality/>To/>Distillation intensity/>The method comprises the following steps: /(I); Wherein/>Representing a concatenation of features,/>And/>Representing a fully connected layer,/>Expressed/>Public or private features,/>Representation/>Public or private features of (a).
From the modalityTo/>Is defined as the difference logits (output vector) between them, is defined as/>. For modality/>Distillation loss/>The method comprises the following steps: /(I); Wherein/>Representing injection modality/>Is defined by a node set.
From distillation intensityCan form square matrix as element,/>As element composition matrix/>. Graph distillation loss/>The method comprises the following steps: /(I); Wherein/>Representing element-by-element multiplication,/>Representing the L1 norm.
According to the description of the formula, graph distillation is carried out on the public features and the private features corresponding to each node, so that the public distillation loss corresponding to each public feature and the private distillation loss corresponding to each private feature can be obtained respectively, then optimization processing is carried out by using each public distillation loss as the corresponding public feature, optimization processing is carried out by using each private distillation loss as the corresponding private feature, the final public feature and the final private feature corresponding to each node are obtained, and then the final public feature and the final private feature of each node are connected in series and then are subjected to linear mapping to obtain the space domain feature before erasure corresponding to each node.
And S40, performing time erasure processing on the cross-modal hypergraph according to the single-mode enhancement feature and the cross-mode enhancement feature corresponding to each segment mode feature, and acquiring the time domain feature and the space domain feature after erasure corresponding to each node in the cross-modal hypergraph after the time erasure processing.
Specifically, in order to learn more time slices more significant for behavior intention recognition, time erasure processing needs to be performed on all time slices, and then the post-erasure time domain features and the post-erasure space domain features of each node after the time erasure processing are acquired.
Further, the time erasure processing is performed on the cross-modal hypergraph according to the single-mode enhancement feature and the cross-mode enhancement feature corresponding to each segment mode feature, and the time domain feature and the space domain feature after erasure corresponding to each node in the cross-modal hypergraph after the time erasure processing are obtained, which specifically comprises:
Respectively carrying out addition processing on the single-mode enhancement feature and the cross-mode enhancement feature corresponding to each segment mode feature to obtain an addition feature corresponding to each segment mode feature, and obtaining a time attention diagram of each segment mode feature according to each addition feature; performing time erasure processing on all the time attention force diagrams according to a preset time erasure template, and taking the time attention force diagrams meeting preset conditions as post-erasure time attention force diagrams; acquiring all the segment modal characteristics corresponding to the time attention map after erasure and taking the segment modal characteristics as the segment modal characteristics after erasure corresponding to each node; performing time domain information enhancement processing on the erased fragment modal characteristics, and then calculating the erased fragment modal characteristics according to a cross-modal attention mechanism to obtain erased cross-modal enhancement characteristics corresponding to the erased fragment modal characteristics; and obtaining the erased time domain characteristic corresponding to each node according to the erased single-mode characteristic and the erased cross-mode enhancement characteristic corresponding to each node, and obtaining the erased airspace characteristic corresponding to each node according to each erased segment mode characteristic.
Specifically, taking a visual feature as an example, acquiring a single-mode enhancement feature G and a cross-mode enhancement feature corresponding to a certain segment visual featureThe unimodal enhancement features and cross-modal enhancement features of this segment visual feature are feature added and then a temporal attention attempt/>, using a multi-layer perceptron (MLP)The specific formula is as follows: /(I)Then, the template E is erased by using the preset time, which is specifically expressed as: /(I); Wherein/>Is a manually set superparameter,/>Represents a more meaningful time segment for behavioral intent recognition, in order to learn more useful features further, will/>Temporary discard, retention/>Time slices of (i.e./>)The time segment of the time frame is a time segment meeting the preset condition.
Similarly, after the time erasure processing is completed on all the segment modal characteristics, all the segment modal characteristics are matchedThe time attention map of (2) is used as a post-erasure time attention map, the post-erasure segment modal characteristics corresponding to each of the erasure time attention maps are obtained, the time domain information enhancement processing is repeated according to all the post-erasure segment modal characteristics to obtain post-erasure single-mode enhancement characteristics corresponding to each of the post-erasure segment modal characteristics, and the post-erasure cross-mode enhancement processing is performed on each of the post-erasure segment modal characteristics to obtain post-erasure cross-mode enhancement characteristics corresponding to each of the post-erasure segment modal characteristics; then obtaining erased time domain features corresponding to each node according to erased single-mode enhancement features and erased cross-mode enhancement features corresponding to each node, and obtaining erased airspace features corresponding to each node according to all erased segment mode features corresponding to each node; the steps of acquiring the time domain features after erasure and the space domain features after erasure are identical to the above, and are not described in detail herein.
And S50, carrying out fusion processing on all the pre-erasure time domain features, all the pre-erasure airspace features, all the post-erasure time domain features and all the post-erasure airspace features to obtain final fusion features, and obtaining a behavior prediction result of the target object in a target time period according to the final fusion features.
Specifically, after the pre-erasure time domain feature, the pre-erasure airspace feature, the post-erasure time domain feature and the post-erasure airspace feature corresponding to each node are obtained, corresponding fusion processing is needed to obtain final fusion features, the final fusion features reflect the behavior features of the patient in the target time period, so that corresponding behavior intention prediction can be performed on the patient in the target time period according to the final fusion features, and a behavior prediction result of the patient in the target time period is obtained.
Further, the fusing processing is performed on all the pre-erasure time domain features, all the pre-erasure airspace features, all the post-erasure time domain features and all the post-erasure airspace features to obtain a final fused feature, and a behavior prediction result of the target object in a target time period is obtained according to the final fused feature, which specifically includes:
Respectively carrying out fusion processing on the pre-erasure time domain features, the pre-erasure space domain features, the post-erasure time domain features and the post-erasure space domain features corresponding to each node to obtain space-time features corresponding to each node; the space-time characteristics of all the nodes are connected in series to obtain the final fusion characteristics; and analyzing and processing the final fusion characteristics by using a pre-trained graph neural network to obtain a behavior prediction result of the target object in a target time period.
Specifically, as shown in fig. 7, after the pre-erasure time domain feature, the pre-erasure space domain feature, the post-erasure time domain feature and the post-erasure space domain feature corresponding to each node are obtained, the four features of each node need to be fused to obtain a space-time feature corresponding to each node, and the specific formula is as follows: ; wherein/> For space-time characteristics,/>For pre-erasure time domain features,/>For post-erasure temporal feature,/>To erase the pre-spatial features,/>To erase the post-spatial features.
After the space-time characteristics corresponding to each node are obtained, the space-time characteristics of all the nodes are connected in series to be used as final fusion characteristics, the final fusion characteristics are input into a pre-trained graph neural network, the pre-trained graph neural network analyzes and processes the final fusion characteristics, so that the score of each behavior category of a patient in a target time period is obtained, and the behavior category with the highest score is selected to be used as a behavior prediction result of the patient in the target time period.
Further, as shown in fig. 8, based on the above behavior intention recognition method based on the cross-modal hypergraph, the present invention further correspondingly provides a behavior intention recognition system based on the cross-modal hypergraph, where the behavior intention recognition system based on the cross-modal hypergraph includes:
The segment characteristic obtaining module 51 is configured to obtain a plurality of different modal signals of a target object in a target time period, and perform preprocessing according to the target time period and each modal signal to obtain a plurality of segment modal characteristics corresponding to each modal signal;
The unimodal enhancement module 52 is configured to construct a cross-modal hypergraph according to all the segment modal features, and perform time domain information enhancement processing on each segment modal feature in the cross-modal hypergraph to obtain a unimodal enhancement feature corresponding to each segment modal feature;
the pre-erasure feature obtaining module 53 is configured to calculate a cross-modal enhancement feature corresponding to each segment modal feature according to a cross-modal attention mechanism, obtain a pre-erasure time domain feature corresponding to each node in the cross-modal hypergraph according to the cross-modal enhancement feature corresponding to each segment modal feature, and obtain a pre-erasure airspace feature corresponding to each node according to the segment modal feature of each node;
The post-erasure feature obtaining module 54 is configured to perform a time erasure process on the cross-modal hypergraph according to the single-mode enhancement feature and the cross-mode enhancement feature corresponding to each of the segment mode features, and obtain post-erasure time domain features and post-erasure space domain features corresponding to each node in the cross-modal hypergraph after the time erasure process;
and the prediction result generating module 55 is configured to perform fusion processing on all the pre-erasure time domain features, all the pre-erasure space domain features, all the post-erasure time domain features and all the post-erasure space domain features to obtain a final fusion feature, and obtain a behavior prediction result of the target object in a target time period according to the final fusion feature.
Further, as shown in fig. 9, based on the above behavior intention recognition method and system based on cross-modal hypergraph, the present invention further provides a terminal correspondingly, where the terminal includes a processor 10, a memory 20 and a display 30. Fig. 9 shows only some of the components of the terminal, but it should be understood that not all of the illustrated components are required to be implemented and that more or fewer components may alternatively be implemented.
The memory 20 may in some embodiments be an internal storage unit of the terminal, such as a hard disk or a memory of the terminal. The memory 20 may in other embodiments also be an external storage device of the terminal, such as a plug-in hard disk provided on the terminal, a smart memory card (SMART MEDIA CARD, SMC), a Secure Digital (SD) card, a flash memory card (FLASH CARD), etc. Further, the memory 20 may also include both an internal storage unit and an external storage device of the terminal. The memory 20 is used for storing application software installed in the terminal and various data, such as program codes of the installation terminal. The memory 20 may also be used to temporarily store data that has been output or is to be output. In one embodiment, the memory 20 has stored thereon a cross-modal hypergraph based behavior intent recognition program 40, the cross-modal hypergraph based behavior intent recognition program 40 being executable by the processor 10 to implement the cross-modal hypergraph based behavior intent recognition method of the present application.
The processor 10 may in some embodiments be a central processing unit (Central Processing Unit, CPU), microprocessor or other data processing chip for executing program code or processing data stored in the memory 20, such as performing the cross-modal hypergraph-based behavior intent recognition method or the like.
The display 30 may be an LED display, a liquid crystal display, a touch-sensitive liquid crystal display, an OLED (Organic Light-Emitting Diode) touch, or the like in some embodiments. The display 30 is used for displaying information at the terminal and for displaying a visual user interface. The components 10-30 of the terminal communicate with each other via a system bus.
The invention also provides a computer readable storage medium, wherein the computer readable storage medium stores a cross-modal hypergraph-based behavior intention recognition program which, when executed by a processor, implements the steps of the cross-modal hypergraph-based behavior intention recognition method described above.
It should be noted that, in this document, the terms "comprises," "comprising," or any other variation thereof, are intended to cover a non-exclusive inclusion, such that a process, method, article, or terminal that comprises a list of elements does not include only those elements but may include other elements not expressly listed or inherent to such process, method, article, or terminal. Without further limitation, an element defined by the phrase "comprising one … …" does not exclude the presence of other like elements in a process, method, article, or terminal that comprises the element.
Of course, those skilled in the art will appreciate that implementing all or part of the above described methods may be accomplished by a computer program for instructing relevant hardware (e.g., processor, controller, etc.), the program may be stored on a computer readable storage medium, and the program may include the above described methods when executed. The computer readable storage medium may be a memory, a magnetic disk, an optical disk, etc.
It is to be understood that the invention is not limited in its application to the examples described above, but is capable of modification and variation in light of the above teachings by those skilled in the art, and that all such modifications and variations are intended to be included within the scope of the appended claims.

Claims (10)

1. The behavior intention recognition method based on the cross-modal hypergraph is characterized by comprising the following steps of:
Acquiring a plurality of different modal signals of a target object in a target time period, and preprocessing the different modal signals according to the target time period and each modal signal to obtain a plurality of fragment modal characteristics corresponding to each modal signal;
Constructing a cross-modal hypergraph according to all the segment modal characteristics, and performing time domain information enhancement processing on each segment modal characteristic in the cross-modal hypergraph to obtain a single-mode enhancement characteristic corresponding to each segment modal characteristic;
Calculating cross-modal enhancement features corresponding to each segment modal feature according to a cross-modal attention mechanism, obtaining pre-erasure time domain features corresponding to each node in the cross-modal hypergraph according to the cross-modal enhancement features corresponding to each segment modal feature, and obtaining pre-erasure airspace features corresponding to each node according to the segment modal features of each node;
Performing time erasure processing on the cross-modal hypergraph according to the single-mode enhancement feature and the cross-mode enhancement feature corresponding to each segment mode feature, and acquiring an erased time domain feature and an erased space domain feature corresponding to each node in the cross-modal hypergraph after the time erasure processing;
And carrying out fusion processing on all the pre-erasure time domain features, all the pre-erasure space domain features, all the post-erasure time domain features and all the post-erasure space domain features to obtain final fusion features, and obtaining a behavior prediction result of the target object in a target time period according to the final fusion features.
2. The behavior intention recognition method based on cross-modal hypergraph according to claim 1, wherein the obtaining a plurality of different modal signals of a target object in a target time period, and preprocessing according to the target time period and each modal signal to obtain a plurality of segment modal features corresponding to each modal signal specifically comprises:
Acquiring a plurality of different modal signals of a target object in a target time period, wherein the different modal signals are acquired by different preset sensors;
Preprocessing each modal signal to obtain a modal characteristic corresponding to each modal signal;
and uniformly dividing each modal feature according to the preset time segment length to obtain a plurality of segment modal features corresponding to each modal feature.
3. The behavior intention recognition method based on cross-modal hypergraph according to claim 1, wherein the steps of constructing a cross-modal hypergraph according to all the segment modal features, and performing time domain information enhancement processing on each segment modal feature in the cross-modal hypergraph to obtain a single-modal enhancement feature corresponding to each segment modal feature comprise the following steps:
Respectively integrating the segment modal characteristics corresponding to each modal signal to obtain nodes corresponding to each modal signal, and constructing a cross-modal hypergraph according to all the nodes;
performing linear mapping operation on segment modal characteristics of each node in the cross-modal hypergraph respectively to obtain a key vector, a value vector and a query vector corresponding to each segment modal characteristic;
constructing a similarity matrix according to the key vector, the value vector and the query vector corresponding to each segment modal feature;
and respectively carrying out time domain information enhancement processing on all the segment modal characteristics according to the similarity matrix to obtain single-mode enhancement characteristics corresponding to each segment modal characteristic.
4. The behavior intention recognition method based on a cross-modal hypergraph according to claim 1, wherein the calculating a cross-modal enhancement feature corresponding to each segment modal feature according to a cross-modal attention mechanism, and obtaining a pre-erasure time domain feature corresponding to each node in the cross-modal hypergraph according to the cross-modal enhancement feature corresponding to each segment modal feature, and obtaining a pre-erasure airspace feature corresponding to each node according to the segment modal feature of each node specifically includes:
Acquiring a time segment corresponding to each segment modal feature, and classifying all the segment modal features according to the time segments to divide a plurality of segment modal features belonging to the same time into the same group to obtain a plurality of modal feature groups;
Taking any segment modal feature in any one modal feature group as a feature to be processed according to a cross-modal attention mechanism, and performing cross-modal enhancement processing on the feature to be processed according to all segment modal features in the same modal feature group as the feature to be processed to obtain a cross-modal enhancement feature corresponding to the feature to be processed;
after the cross-modal enhancement processing of all the segment modal features is completed, obtaining cross-modal enhancement features corresponding to each segment modal feature;
carrying out fusion processing on the single-mode enhancement features and the cross-mode enhancement features belonging to the same node to obtain pre-erasure time domain features corresponding to each node;
And integrating the segment modal characteristics of each node to obtain node characteristics corresponding to each node, and performing coding processing according to each node characteristic to obtain the pre-erasure airspace characteristics corresponding to each node.
5. The behavior intention recognition method based on cross-modal hypergraph according to claim 4, wherein the encoding processing is performed according to each node characteristic to obtain a pre-erasure airspace characteristic corresponding to each node, and the method specifically comprises:
Respectively carrying out coding processing on each node characteristic through a public coder and a private coder to obtain a public characteristic and a private characteristic corresponding to each node characteristic;
Respectively carrying out graph distillation treatment on the public characteristic and the private characteristic of each node characteristic to obtain public distillation loss corresponding to each public characteristic and private distillation loss corresponding to each private characteristic;
Optimizing the corresponding public characteristics and private characteristics according to each public distillation loss and each private distillation loss to obtain final public characteristics and final private characteristics corresponding to each node;
and respectively carrying out fusion processing and linear mapping on the final public features and the final private features of each node to obtain the pre-erasure airspace features corresponding to each node.
6. The behavior intention recognition method based on a cross-modal hypergraph according to claim 1, wherein the performing time erasure processing on the cross-modal hypergraph according to a single-mode enhancement feature and a cross-modal enhancement feature corresponding to each segment mode feature, and acquiring an post-erasure time domain feature and an post-erasure space domain feature corresponding to each node in the cross-modal hypergraph after the time erasure processing specifically comprises:
respectively carrying out addition processing on the single-mode enhancement feature and the cross-mode enhancement feature corresponding to each segment mode feature to obtain an addition feature corresponding to each segment mode feature, and obtaining a time attention diagram of each segment mode feature according to each addition feature;
Performing time erasure processing on all the time attention force diagrams according to a preset time erasure template, and taking the time attention force diagrams meeting preset conditions as post-erasure time attention force diagrams;
Acquiring all the segment modal characteristics corresponding to the time attention map after erasure and taking the segment modal characteristics as the segment modal characteristics after erasure corresponding to each node;
performing time domain information enhancement processing on the erased fragment modal characteristics, and then calculating the erased fragment modal characteristics according to a cross-modal attention mechanism to obtain erased cross-modal enhancement characteristics corresponding to the erased fragment modal characteristics;
and obtaining the erased time domain characteristic corresponding to each node according to the erased single-mode characteristic and the erased cross-mode enhancement characteristic corresponding to each node, and obtaining the erased airspace characteristic corresponding to each node according to each erased segment mode characteristic.
7. The behavior intention recognition method based on cross-modal hypergraph according to claim 1, wherein the fusing processing is performed on all the pre-erasure time domain features, all the pre-erasure space domain features, all the post-erasure time domain features and all the post-erasure space domain features to obtain a final fused feature, and the behavior prediction result of the target object in the target time period is obtained according to the final fused feature, and specifically includes:
Respectively carrying out fusion processing on the pre-erasure time domain features, the pre-erasure space domain features, the post-erasure time domain features and the post-erasure space domain features corresponding to each node to obtain space-time features corresponding to each node;
the space-time characteristics of all the nodes are connected in series to obtain the final fusion characteristics;
And analyzing and processing the final fusion characteristics by using a pre-trained graph neural network to obtain a behavior prediction result of the target object in a target time period.
8. A cross-modal hypergraph-based behavior intent recognition system, the cross-modal hypergraph-based behavior intent recognition system comprising:
The segment characteristic acquisition module is used for acquiring a plurality of different modal signals of a target object in a target time period, and preprocessing the different modal signals according to the target time period and each modal signal to obtain a plurality of segment modal characteristics corresponding to each modal signal;
the single-mode enhancement module is used for constructing a cross-mode hypergraph according to all the segment mode characteristics, and carrying out time domain information enhancement processing on each segment mode characteristic in the cross-mode hypergraph to obtain single-mode enhancement characteristics corresponding to each segment mode characteristic;
The pre-erasure feature acquisition module is used for calculating cross-modal enhancement features corresponding to each segment modal feature according to a cross-modal attention mechanism, obtaining pre-erasure time domain features corresponding to each node in the cross-modal hypergraph according to the cross-modal enhancement features corresponding to each segment modal feature, and obtaining pre-erasure airspace features corresponding to each node according to the segment modal features of each node;
The post-erasure feature acquisition module is used for performing time erasure processing on the cross-modal hypergraph according to the single-mode enhancement feature and the cross-mode enhancement feature corresponding to each segment mode feature, and acquiring post-erasure time domain features and post-erasure space domain features corresponding to each node in the cross-modal hypergraph after the time erasure processing;
And the prediction result generation module is used for carrying out fusion processing on all the pre-erasure time domain features, all the pre-erasure space domain features, all the post-erasure time domain features and all the post-erasure space domain features to obtain final fusion features, and obtaining a behavior prediction result of the target object in a target time period according to the final fusion features.
9. A terminal, the terminal comprising: memory, a processor and a cross-modal hypergraph based behavior intent recognition program stored on the memory and executable on the processor, which when executed by the processor implements the steps of the cross-modal hypergraph based behavior intent recognition method as claimed in any one of claims 1-7.
10. A computer readable storage medium, characterized in that it stores a cross-modal hypergraph based behavior intention recognition program, which when executed by a processor implements the steps of the cross-modal hypergraph based behavior intention recognition method according to any of claims 1-7.
CN202410247777.1A 2024-03-05 2024-03-05 Behavior intention recognition method, system and terminal based on cross-mode hypergraph Active CN117828281B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202410247777.1A CN117828281B (en) 2024-03-05 2024-03-05 Behavior intention recognition method, system and terminal based on cross-mode hypergraph

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202410247777.1A CN117828281B (en) 2024-03-05 2024-03-05 Behavior intention recognition method, system and terminal based on cross-mode hypergraph

Publications (2)

Publication Number Publication Date
CN117828281A CN117828281A (en) 2024-04-05
CN117828281B true CN117828281B (en) 2024-05-07

Family

ID=90513794

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202410247777.1A Active CN117828281B (en) 2024-03-05 2024-03-05 Behavior intention recognition method, system and terminal based on cross-mode hypergraph

Country Status (1)

Country Link
CN (1) CN117828281B (en)

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113792768A (en) * 2021-08-27 2021-12-14 清华大学 Hypergraph neural network classification method and device
WO2023024017A1 (en) * 2021-08-26 2023-03-02 Ebay Inc. Multi-modal hypergraph-based click prediction
CN116432053A (en) * 2023-03-21 2023-07-14 浙江师范大学 Multi-mode data representation method based on modal interaction deep hypergraph neural network
CN116548985A (en) * 2023-05-06 2023-08-08 平安科技(深圳)有限公司 Electroencephalogram signal feature extraction method and device, computer equipment and storage medium
CN117037017A (en) * 2023-05-23 2023-11-10 南开大学 Video emotion detection method based on key frame erasure
CN117173394A (en) * 2023-08-07 2023-12-05 山东大学 Weak supervision salient object detection method and system for unmanned aerial vehicle video data

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2023024017A1 (en) * 2021-08-26 2023-03-02 Ebay Inc. Multi-modal hypergraph-based click prediction
CN113792768A (en) * 2021-08-27 2021-12-14 清华大学 Hypergraph neural network classification method and device
CN116432053A (en) * 2023-03-21 2023-07-14 浙江师范大学 Multi-mode data representation method based on modal interaction deep hypergraph neural network
CN116548985A (en) * 2023-05-06 2023-08-08 平安科技(深圳)有限公司 Electroencephalogram signal feature extraction method and device, computer equipment and storage medium
CN117037017A (en) * 2023-05-23 2023-11-10 南开大学 Video emotion detection method based on key frame erasure
CN117173394A (en) * 2023-08-07 2023-12-05 山东大学 Weak supervision salient object detection method and system for unmanned aerial vehicle video data

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
甲状腺、乳腺超声影像自动分析技术综述;龚勋;杨菲;杜章锦;师恩;赵绪;杨子奇;邹海鹏;罗俊;;软件学报;20200715(07);第319-356页 *

Also Published As

Publication number Publication date
CN117828281A (en) 2024-04-05

Similar Documents

Publication Publication Date Title
Du et al. Representation learning of temporal dynamics for skeleton-based action recognition
Kanade et al. First-person vision
EP3811337A1 (en) System for predicting articulated object feature location
US9239956B2 (en) Method and apparatus for coding of eye and eye movement data
CN111222486B (en) Training method, device and equipment for hand gesture recognition model and storage medium
CN109685037B (en) Real-time action recognition method and device and electronic equipment
CN109993131B (en) Design intention distinguishing system and method based on multi-mode signal fusion
CN109765991A (en) Social interaction system is used to help system and non-transitory computer-readable storage media that user carries out social interaction
CN108960171B (en) Method for converting gesture recognition into identity recognition based on feature transfer learning
CN113435236A (en) Home old man posture detection method, system, storage medium, equipment and application
CN111915618B (en) Peak response enhancement-based instance segmentation algorithm and computing device
Li et al. Global co-occurrence feature learning and active coordinate system conversion for skeleton-based action recognition
CN115205933A (en) Facial expression recognition method, device, equipment and readable storage medium
CN117828281B (en) Behavior intention recognition method, system and terminal based on cross-mode hypergraph
CN110956599A (en) Picture processing method and device, storage medium and electronic device
CN113627349B (en) Dynamic facial expression recognition method based on self-attention transformation network
CN115359265A (en) Key point extraction method, device, equipment and storage medium
CN116700471A (en) Method and system for enhancing user experience of virtual reality system
Weibel et al. Hands that speak: An integrated approach to studying complex human communicative body movements
JP2021170247A (en) Information processing device, information processing system, information processing method and program
Ghosh et al. AV-Gaze: A study on the effectiveness of audio guided visual attention estimation for non-profilic faces
Wang et al. UMSNet: An Universal Multi-sensor Network for Human Activity Recognition
CN114723963B (en) Task action and object physical attribute identification method based on visual touch signal
CN112884076B (en) Sensor data generation model and method for generating confrontation network based on conditional expression
Kalra et al. Fixed do solfège based object detection and positional analysis for the visually impaired

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant