CN117747064A - Depression clinical decision method and system based on AI - Google Patents

Depression clinical decision method and system based on AI Download PDF

Info

Publication number
CN117747064A
CN117747064A CN202311664055.8A CN202311664055A CN117747064A CN 117747064 A CN117747064 A CN 117747064A CN 202311664055 A CN202311664055 A CN 202311664055A CN 117747064 A CN117747064 A CN 117747064A
Authority
CN
China
Prior art keywords
data
module
strategy
feature
depression
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202311664055.8A
Other languages
Chinese (zh)
Inventor
沈敏
曹音
董贯忠
王成贵
和法亮
郑万强
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Jiangsu Fumin Artificial Intelligence Technology Co ltd
Changzhou Second Peoples Hospital
Original Assignee
Jiangsu Fumin Artificial Intelligence Technology Co ltd
Changzhou Second Peoples Hospital
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Jiangsu Fumin Artificial Intelligence Technology Co ltd, Changzhou Second Peoples Hospital filed Critical Jiangsu Fumin Artificial Intelligence Technology Co ltd
Priority to CN202311664055.8A priority Critical patent/CN117747064A/en
Publication of CN117747064A publication Critical patent/CN117747064A/en
Pending legal-status Critical Current

Links

Landscapes

  • Measurement Of The Respiration, Hearing Ability, Form, And Blood Characteristics Of Living Organisms (AREA)

Abstract

The invention relates to the technical field of artificial intelligence, in particular to an AI-based depression clinical decision method and system, comprising the following steps: based on the depression patients, the data preprocessing technology is adopted to clean and synchronize voice, video, text and physiological signals, so as to generate multi-mode patient data. According to the invention, personalized depression treatment prediction and assessment are realized through multi-mode data analysis and feature fusion, the emotion recognition accuracy of voice and video data is improved by combining a long-term memory network and a three-dimensional convolutional neural network, the robustness of a prediction model is enhanced by a random forest and gradient lifting decision tree, the treatment strategy is optimized by a depth strategy gradient algorithm, the success rate is continuously improved by matching with a simulation platform, accurate interpretation of a patient is realized by real-time facial expression tracking and emotion analysis, the dynamic adjustment of clinical decisions is realized by an online learning algorithm, the state change of the patient is responded timely, and the individuation and the accuracy of treatment are improved.

Description

Depression clinical decision method and system based on AI
Technical Field
The invention relates to the technical field of artificial intelligence, in particular to an AI-based depression clinical decision method and system.
Background
Artificial intelligence is a field of research on how computer systems exhibit intelligent behavior. It encompasses a number of sub-fields of machine learning, deep learning, natural language processing, computer vision, and the like. AI technology is widely used in the medical field, particularly in clinical decision support systems. AI is capable of processing and analyzing large amounts of medical data, identifying patterns, assisting doctors in diagnosing, predicting disease progression, providing personalized treatment regimens, and improving patient care.
The AI-based depression clinical decision method is a method for diagnosis, prediction and treatment support of depression by combining medical data and psychological knowledge by utilizing an artificial intelligence technology. The method aims at assisting doctors in diagnosing depression by analyzing multidimensional data such as physiology, psychology and behavior of patients, predicting the disease development trend of the patients, making personalized treatment plans, providing timely intervention and support, improving the life quality of the patients and reducing the psychological health risks of the patients suffering from depression. The AI-based depression clinical decision method provides more accurate, efficient and personalized depression diagnosis and treatment schemes through the support of artificial intelligence technology, and provides better medical service and mental health support for patients.
The existing depression treatment decision-making method mostly depends on subjective experience of doctors and linear statistical models, and lacks deep and dynamic analysis on emotion and behavior patterns of patients. The methods have low efficiency when processing complex multi-mode data, and often cannot realize deep fusion of data features, so that the recognition accuracy and the prediction accuracy are limited. In predicting the response of the treatment, the traditional method fails to effectively utilize the machine learning algorithm to mine the potential nonlinear relationship of the data, so that the treatment scheme often lacks individualization and dynamic adaptation capability. Furthermore, the lack of integration of an immediate feedback mechanism makes it difficult to adjust the treatment regimen in real time, which affects the maximization of the therapeutic effect to some extent. Therefore, the existing method has limitation in the aspect of multidimensional complexity of treating depression, and the potential of clinical big data and advanced algorithms in treatment individuation and treatment effect improvement cannot be fully exerted.
Disclosure of Invention
The invention aims to solve the defects in the prior art, and provides an AI-based depression clinical decision method and an AI-based depression clinical decision system.
In order to achieve the above purpose, the present invention adopts the following technical scheme: an AI-based depression clinical decision method comprising the steps of:
S1: based on the depression patient, adopting a data preprocessing technology to clean and synchronize voice, video, text and physiological signals to generate multi-mode patient data;
s2: based on the multi-modal patient data, adopting a long-short-term memory network and a three-dimensional convolutional neural network to analyze voice and video data, and carrying out feature fusion to generate a psychological behavior pattern recognition result;
s3: based on the psychological behavior pattern recognition result, adopting a random forest and gradient lifting decision tree to predict the treatment response of the patient, evaluating the treatment effect, and generating a depression prediction result and a treatment effect evaluation report;
s4: based on the treatment effect evaluation report, adopting a depth deterministic strategy gradient algorithm to optimize a strategy, and testing on a simulation platform to generate an optimized depression treatment strategy;
s5: based on video data in the multi-mode patient data, adopting a convolutional neural network and cyclic neural network mixed model to track facial expressions and analyze emotional states, and generating facial expression recognition and emotion analysis reports;
s6: based on the optimized depression treatment strategy, facial expression recognition and emotion analysis report, adopting an online learning algorithm to adjust clinical decisions in real time, and generating optimized clinical decisions and dynamic adjustment strategies;
The multi-modal patient data is specifically a patient information set subjected to noise reduction, outlier processing and time stamp alignment, the psychological behavior pattern recognition result comprises an emotion state label, language expression habit and limb language characteristics, the depression prediction result is specifically probability distribution of future health states of the patient, the treatment effect evaluation report comprises benefit comparison of multiple treatment schemes, and the facial expression recognition and emotion analysis report comprises facial muscle activity data and emotion change-based time sequence analysis.
As a further scheme of the invention, based on the depression patient, the data preprocessing technology is adopted to clean and synchronize voice, video, text and physiological signals, and the steps for generating the multi-mode patient data are specifically as follows:
s101: performing frequency domain analysis by adopting Fourier transform based on original multi-mode data of a depression patient, and processing missing values by mean filling and nearest neighbor interpolation to generate preprocessed multi-mode data;
s102: resampling and time alignment are carried out by adopting a time sequence analysis method based on the preprocessed multi-mode data, so as to generate time-synchronous multi-mode data;
S103: based on the time-synchronous multi-mode data, extracting key features by adopting a Mel frequency cepstrum coefficient and a deep learning feature automatic encoder to generate a multi-mode feature data set;
s104: and based on the multi-modal feature data set, adopting a weighted fusion strategy to fuse multi-modal data and generating fused multi-modal feature data.
As a further scheme of the invention, based on the multi-modal patient data, a long-term memory network and a three-dimensional convolutional neural network are adopted to analyze voice and video data, and feature fusion is carried out, so that the step of generating a psychological behavior pattern recognition result is specifically as follows:
s201: based on the fused multi-mode characteristic data, extracting dynamic time characteristics of voice by adopting a long-short-period memory network, and generating voice time sequence characteristics;
s202: based on the fused multi-mode characteristic data, extracting the spatial characteristics and dynamic changes of the video by adopting a three-dimensional convolutional neural network, and generating the space-time characteristics of the video;
s203: based on the voice time sequence features and the video space-time features, performing depth feature fusion by using feature cascading and fusion strategies to generate voice video depth feature fusion data;
S204: and carrying out pattern recognition by adopting a support vector machine based on the voice video depth feature fusion data to generate a psychological behavior pattern recognition result.
As a further scheme of the present invention, based on the psychological behavior pattern recognition result, a random forest and gradient lifting decision tree is adopted to predict the therapeutic response of the patient, and evaluate the therapeutic effect, and the steps of generating the predicted result of depression and the evaluation report of therapeutic effect are specifically as follows:
s301: based on the psychological behavior pattern recognition result, ranking the feature importance by adopting a random forest algorithm, analyzing the change before and after treatment, and generating baseline feature analysis data;
s302: constructing a prediction model by adopting a gradient lifting decision tree based on the baseline characteristic analysis data, and performing treatment response prediction to generate treatment response prediction data;
s303: based on the treatment response prediction data, performing model verification and tuning, and using k-fold cross verification to ensure generalization capability of the model to generate a treatment effect evaluation model;
s304: and analyzing the model performance by adopting a confusion matrix and a receiver operation characteristic curve based on the treatment effect evaluation model, and generating a depression predicted result and a treatment effect evaluation report.
As a further scheme of the invention, based on the treatment effect evaluation report, a depth deterministic strategy gradient algorithm optimization strategy is adopted, and a simulation platform is used for testing, and the steps for generating the optimized depression treatment strategy are specifically as follows:
s401: based on the treatment effect evaluation report, performing scale processing on the data by adopting a Z-score standardization method, and generating normalized evaluation data;
s402: based on the normalized evaluation data, training a strategy network by adopting a depth deterministic strategy gradient algorithm to generate a primarily optimized treatment strategy model;
s403: based on the primarily optimized treatment strategy model, performing Monte Carlo simulation in a simulation environment, collecting decision performance, and generating simulation test data;
s404: and based on the simulation test data, the DDPG algorithm refinement strategy is applied again, and the optimized depression treatment strategy is generated.
As a further scheme of the present invention, based on the video data in the multi-modal patient data, a convolutional neural network and cyclic neural network hybrid model is adopted to track facial expressions and analyze emotional states, and the steps of generating facial expression recognition and emotion analysis reports specifically include:
S501: based on video data in the multi-mode patient data, performing frame extraction and gray scale processing through an OpenCV tool to generate preprocessed video frame data;
s502: based on the preprocessed video frame data, extracting spatial features by adopting a convolutional neural network to generate facial feature data;
s503: based on the facial feature data, analyzing the expression sequence by applying a long-term memory network to generate facial expression dynamic data;
s504: and based on the facial expression dynamic data, integrating facial feature data and facial expression dynamic data, and generating facial expression recognition and emotion analysis reports by adopting an emotion recognition algorithm.
As a further scheme of the invention, based on the optimized depression treatment strategy and facial expression recognition and emotion analysis report, the clinical decision is adjusted in real time by adopting an online learning algorithm, and the steps of generating the optimized clinical decision and dynamic adjustment strategy are specifically as follows:
s601: based on the optimized depression treatment strategy, parameters are adjusted by using an online learning method, and initially adjusted clinical decision parameters are generated;
s602: optimizing decision parameters based on the facial expression recognition and emotion analysis report in combination with reinforcement learning technology to generate emotion-adjusted clinical decision parameters;
S603: based on the emotion-adjusted clinical decision parameters, real-time monitoring and dynamic adjustment are carried out, and self-adaptive control strategies are used for generating real-time monitoring adjustment data;
s604: based on the real-time monitoring and adjusting data, final online learning optimization is executed, real-time performance and individuation of decisions are ensured, and optimized clinical decisions and dynamic adjusting strategies are generated.
An AI-based depression clinical decision system is used for executing the AI-based depression clinical decision method, and comprises a data preprocessing module, a multi-mode feature extraction module, a feature fusion and mode identification module, a treatment strategy optimization module, an emotion analysis module, a decision parameter optimization module and a system integration and optimization module.
As a further scheme of the invention, the data preprocessing module adopts a Fourier transform algorithm to perform frequency domain analysis based on original multi-mode data of a patient suffering from depression, and utilizes a mean value filling method and a nearest neighbor interpolation method to process a missing value so as to generate preprocessed multi-mode data;
the multi-modal feature extraction module performs complex sampling and time alignment by adopting a time sequence analysis method based on the preprocessed multi-modal data, automatically extracts key features by utilizing a Mel frequency cepstrum coefficient and a deep learning feature automatic encoder, and generates a time-synchronous multi-modal feature data set;
The feature fusion and pattern recognition module utilizes a time-synchronous multi-modal feature data set, utilizes a weighted fusion strategy to fuse data, analyzes patterns through a support vector machine algorithm, and generates fused multi-modal feature data and a psychological behavior pattern recognition result;
the treatment strategy optimization module sorts the importance of the features by adopting a random forest algorithm based on the psychological behavior pattern recognition result, and establishes a prediction model by utilizing a gradient lifting decision tree algorithm to generate baseline feature analysis data and treatment response prediction data;
the emotion analysis module extracts voice and video features by using a long-term memory network algorithm and a three-dimensional convolutional neural network according to the fused multi-modal feature data, analyzes the voice and video features by applying an emotion recognition algorithm, and generates voice video depth feature fusion data and facial expression recognition and emotion analysis reports;
the decision parameter tuning module is used for adjusting parameters based on treatment response prediction data and facial expression recognition and emotion analysis reports by utilizing an online learning method and a reinforcement learning technology to generate preliminarily adjusted clinical decision parameters and emotion adjusted clinical decision parameters;
The system integration and optimization module analyzes model performance by adopting a confusion matrix and a receiver operation characteristic curve, performs large-scale processing on data by a Z score standardization method, trains a strategy network by utilizing a depth deterministic strategy gradient algorithm, performs Monte Carlo simulation in a simulation environment, collects decision expression, applies a depth deterministic strategy gradient algorithm refinement strategy again, and generates an optimized treatment strategy and an optimized depression treatment strategy.
As a further scheme of the invention, the data preprocessing module comprises a frequency domain analysis sub-module, a missing value processing sub-module and a data normalization sub-module;
the multi-mode feature extraction module comprises a time alignment sub-module, a key feature extraction sub-module and a feature data set construction sub-module;
the feature fusion and pattern recognition module comprises a data fusion sub-module, a pattern recognition sub-module and a feature cascade sub-module;
the treatment strategy optimization module comprises a feature importance ordering sub-module, a prediction model construction sub-module and a model verification and tuning sub-module;
the emotion analysis module comprises a voice feature extraction sub-module, a video feature extraction sub-module and a feature fusion sub-module;
The decision parameter tuning module comprises a feature selection sub-module, a model selection sub-module, a super parameter tuning sub-module and a model evaluation sub-module;
the system integration and optimization module comprises a performance analysis sub-module, a data scale sub-module, a strategy network training sub-module, a Monte Carlo simulation sub-module and a strategy refinement sub-module.
Compared with the prior art, the invention has the advantages and positive effects that:
according to the invention, through multi-mode data comprehensive analysis and multi-level feature fusion, a more accurate and personalized depression treatment prediction and assessment system is provided. The accuracy of voice and video data in the psychological behavior pattern recognition is improved by using the long-short-term memory network and the three-dimensional convolutional neural network, so that emotion recognition and analysis are more detailed and comprehensive. The combination of random forests and gradient boosting decision trees is used to predict treatment response, enhancing the robustness and interpretation of the predictive model. The depth deterministic strategy gradient algorithm optimizes the treatment strategy, provides a simulation platform for testing and verification, and can continuously improve the treatment success rate. Real-time tracking of facial expressions and emotional state analysis enable more accurate interpretation of patient feedback. The method realizes the dynamic adjustment of clinical decisions through an online learning algorithm, ensures that the treatment scheme can timely respond to the tiny change of the state of a patient, and improves the individuation and the accuracy of treatment.
Drawings
FIG. 1 is a schematic workflow diagram of the present invention;
FIG. 2 is a S1 refinement flowchart of the present invention;
FIG. 3 is a S2 refinement flowchart of the present invention;
FIG. 4 is a S3 refinement flowchart of the present invention;
FIG. 5 is a S4 refinement flowchart of the present invention;
FIG. 6 is a S5 refinement flowchart of the present invention;
FIG. 7 is a S6 refinement flowchart of the present invention;
FIG. 8 is a system flow diagram of the present invention;
FIG. 9 is a schematic diagram of a system framework of the present invention.
Detailed Description
The present invention will be described in further detail with reference to the drawings and examples, in order to make the objects, technical solutions and advantages of the present invention more apparent. It should be understood that the specific embodiments described herein are for purposes of illustration only and are not intended to limit the scope of the invention.
In the description of the present invention, it should be understood that the terms "length," "width," "upper," "lower," "front," "rear," "left," "right," "vertical," "horizontal," "top," "bottom," "inner," "outer," and the like indicate orientations or positional relationships based on the orientation or positional relationships shown in the drawings, merely to facilitate describing the present invention and simplify the description, and do not indicate or imply that the devices or elements referred to must have a specific orientation, be configured and operated in a specific orientation, and therefore should not be construed as limiting the present invention. Furthermore, in the description of the present invention, the meaning of "a plurality" is two or more, unless explicitly defined otherwise.
Example 1
Referring to fig. 1, the present invention provides a technical solution: an AI-based depression clinical decision method comprising the steps of:
s1: based on the depression patient, adopting a data preprocessing technology to clean and synchronize voice, video, text and physiological signals to generate multi-mode patient data;
s2: based on multi-mode patient data, adopting a long-term memory network and a three-dimensional convolutional neural network to analyze voice and video data, and carrying out feature fusion to generate a psychological behavior mode recognition result;
s3: based on the psychological behavior pattern recognition result, adopting a random forest and gradient lifting decision tree to predict the treatment response of the patient, evaluating the treatment effect, and generating a depression prediction result and a treatment effect evaluation report;
s4: based on the treatment effect evaluation report, adopting a depth deterministic strategy gradient algorithm to optimize the strategy, and testing on a simulation platform to generate an optimized depression treatment strategy;
s5: based on video data in the multi-mode patient data, tracking facial expressions and analyzing the emotional states by adopting a convolutional neural network and cyclic neural network mixed model to generate facial expression recognition and emotion analysis reports;
S6: based on the optimized depression treatment strategy, facial expression recognition and emotion analysis report, adopting an online learning algorithm to adjust clinical decisions in real time, and generating an optimized clinical decision and a dynamic adjustment strategy;
the multi-modal patient data is specifically a patient information set subjected to noise reduction, outlier processing and time stamp alignment, the psychological behavior pattern recognition result comprises an emotion state label, language expression habit and limb language characteristics, the depression prediction result is specifically probability distribution of future health states of the patient, the treatment effect evaluation report comprises benefit comparison of a multi-treatment scheme, and the facial expression recognition and emotion analysis report comprises facial muscle activity data and emotion change-based time series analysis.
Firstly, the voice, video, text and physiological signals are cleaned and synchronized through a data preprocessing technology to generate multi-mode patient data, so that the quality and usability of the data are greatly improved, and a solid foundation is provided for subsequent analysis and prediction. And secondly, the method adopts a long-term memory network and a three-dimensional convolutional neural network to analyze voice and video data, performs feature fusion, and generates a psychological behavior pattern recognition result.
In addition, the method also adopts a random forest and gradient lifting decision tree to predict the treatment response of the patient, evaluates the treatment effect and generates a depression prediction result and a treatment effect evaluation report. Meanwhile, the method also adopts a depth deterministic strategy gradient algorithm optimization strategy, and tests are carried out on a simulation platform to generate an optimized depression treatment strategy.
Finally, the method also adopts a convolutional neural network and cyclic neural network mixed model to track facial expressions and analyze emotional states, and generates facial expression recognition and emotion analysis reports. In general, the clinical decision method for the depression based on the AI can improve the diagnosis accuracy, optimize the treatment strategy, improve the treatment effect, provide more comprehensive patient information for doctors, and have important significance for improving the treatment effect of depression patients.
Referring to fig. 2, based on the depression patient, the steps of cleaning and synchronizing the voice, video, text and physiological signals by adopting the data preprocessing technology to generate the multi-modal patient data are specifically as follows:
s101: performing frequency domain analysis by adopting Fourier transform based on original multi-mode data of a depression patient, and processing missing values by mean filling and nearest neighbor interpolation to generate preprocessed multi-mode data;
s102: resampling and time alignment are carried out by adopting a time sequence analysis method based on the preprocessed multi-modal data, so as to generate time-synchronous multi-modal data;
s103: based on time-synchronous multi-modal data, extracting key features by adopting a Mel frequency cepstrum coefficient and a deep learning feature automatic encoder to generate a multi-modal feature data set;
s104: based on the multi-modal feature data set, adopting a weighted fusion strategy to fuse multi-modal data and generating fused multi-modal feature data.
The original multi-modal data is frequency domain analyzed using fourier transforms. By converting the time domain signal into a frequency domain signal, the frequency characteristics of the data can be better understood. Next, the missing values are processed using mean padding and nearest neighbor interpolation. For missing data points, the missing values can be filled in by calculating the average value of the adjacent data points so as to ensure the integrity of the data.
Resampling and time alignment of the preprocessed multimodal data using time series analysis. Since data of different modalities may have different sampling rates and time scales, it is necessary to unify the data on a common time axis. By resampling and time alignment, it can be ensured that the data of the various modalities are synchronized in time.
Key features are extracted using mel-frequency cepstral coefficients (MFCCs) and a deep learning feature auto-encoder. MFCC is a feature extraction method commonly used for speech signals, and captures the spectral characteristics of the speech signals. While deep learning feature auto-encoders learn higher level abstract features. By combining the MFCC and the deep learning feature auto-encoder, more representative key features can be extracted from the multi-modal data.
And fusing the multi-modal feature data set by using a weighted fusion strategy. Because the data of different modes has different information amounts and importance, a weight can be allocated to each mode to reflect the contribution degree of the data in the final fusion result. Through a weighted fusion strategy, the characteristics of all modes can be effectively integrated together to generate the fused multi-mode characteristic data.
Referring to fig. 3, based on multi-modal patient data, the steps of performing voice and video data analysis by using a long-short-term memory network and a three-dimensional convolutional neural network, and performing feature fusion to generate a psycho-behavioral pattern recognition result are specifically as follows:
s201: based on the fused multi-mode characteristic data, extracting the dynamic time characteristic of the voice by adopting a long-period memory network to generate a voice time sequence characteristic;
s202: based on the fused multi-mode characteristic data, extracting the spatial characteristics and dynamic changes of the video by adopting a three-dimensional convolutional neural network, and generating the space-time characteristics of the video;
s203: based on the voice time sequence features and the video space-time features, performing depth feature fusion by using feature cascading and fusion strategies to generate voice video depth feature fusion data;
s204: and carrying out pattern recognition by adopting a support vector machine based on the depth feature fusion data of the voice and video to generate a psychological behavior pattern recognition result.
And extracting the voice time sequence characteristics of the fused multi-mode characteristic data by using a long-short-term memory network (LSTM). LSTM is a recurrent neural network structure capable of capturing long-term dependencies, which can efficiently process voice data with a time dimension. By inputting the fused features into the LSTM network, the dynamic time characteristics of the speech data can be learned and represented as continuous time series feature vectors.
And extracting video space-time characteristics of the fused multi-mode characteristic data by using a three-dimensional convolutional neural network (3D-CNN). The 3D-CNN may take into account both the spatial and temporal dimensions of the video data, thereby better capturing motion information and spatial variations in the video. By inputting the fused features into the 3D-CNN network, the spatio-temporal features of the video data can be learned and represented as continuous video spatio-temporal feature vectors.
And carrying out depth feature fusion on the voice time sequence features and the video space-time features by using feature cascading and fusion strategies. The feature cascade connection refers to connecting features of different modes according to a certain sequence to form a new feature vector. The feature fusion refers to performing operations such as weighted average or maximum pooling on features of different modes to obtain a comprehensive feature representation. Through feature cascading and fusion strategies, features of voice and video can be effectively integrated together to generate voice video depth feature fusion data.
And performing pattern recognition on the audio and video depth feature fusion data by using a Support Vector Machine (SVM) to generate a psychological behavior pattern recognition result. SVM is a commonly used classification algorithm that can separate different classes of data by finding an optimal hyperplane. By training an SVM classifier, the voice video depth feature fusion data can be mapped to a specific psychological behavior pattern, so that the psychological behavior pattern of a depression patient can be identified.
Referring to fig. 4, based on the psychological behavior pattern recognition result, the treatment response prediction of the patient is performed by adopting a random forest and gradient lifting decision tree, and the treatment effect is evaluated, and the steps for generating the depression prediction result and the treatment effect evaluation report are specifically as follows:
s301: based on the psychological behavior pattern recognition result, ranking the feature importance by adopting a random forest algorithm, analyzing the change before and after treatment, and generating baseline feature analysis data;
s302: constructing a prediction model by adopting a gradient lifting decision tree based on the baseline characteristic analysis data, and predicting treatment response to generate treatment response prediction data;
s303: based on the treatment response prediction data, performing model verification and optimization, and using k-fold cross verification to ensure the generalization capability of the model to generate a treatment effect evaluation model;
s304: and analyzing the model performance by adopting the confusion matrix and the receiver operation characteristic curve based on the treatment effect evaluation model, and generating a depression predicted result and a treatment effect evaluation report.
The feature importance was ranked using a random forest algorithm and analyzed for changes before and after treatment. Random forests are an ensemble learning algorithm that can classify or regression predictions by voting or averaging across multiple decision trees. By calculating the relative importance of each feature in the random forest, it can be determined which features have a greater impact on the treatment response. From these baseline signature analysis data, the relationship between different signatures and treatment responses can be understood.
And constructing a prediction model by using a gradient lifting decision tree to predict treatment response. The gradient lifting decision tree is an iterative optimized decision tree algorithm, which can improve the performance of the model by continuously adding new decision trees. By inputting the baseline characteristic analysis data into the gradient-lifting decision tree model, the response of the patient to different treatment methods can be predicted, and treatment response prediction data can be generated.
Model verification and optimization are carried out on the treatment response prediction data. To ensure generalization ability of the model, the performance of the model on different subsets can be evaluated using a k-fold cross-validation method. The accuracy and stability of the model can be further improved by adjusting the hyper-parameters of the model and selecting the appropriate feature subset. Finally, a validated and tuned treatment effect assessment model is generated.
Model performance is analyzed using the confusion matrix and receiver operating characteristic curve (ROC curve) and depression prediction results and treatment effect assessment reports are generated. The confusion matrix may be used to evaluate the accuracy, recall, and precision of the classification model. The ROC curve can be used to evaluate the balance between true and false positive rates of the classification model at different thresholds. By comprehensively analyzing the indexes, a predicted depression result and a treatment effect evaluation report can be obtained, and a reference basis is provided for clinical decision.
Referring to fig. 5, based on the treatment effect evaluation report, a depth deterministic strategy gradient algorithm is adopted to optimize the strategy, and a simulation platform is used for testing, so that the optimized treatment strategy for the depression is specifically generated by the following steps:
s401: based on the treatment effect evaluation report, adopting a Z-score standardization method to carry out scale processing on the data and generating normalized evaluation data;
s402: based on the normalized evaluation data, training a strategy network by adopting a depth deterministic strategy gradient algorithm to generate a primarily optimized treatment strategy model;
s403: based on the primarily optimized treatment strategy model, performing Monte Carlo simulation in a simulation environment, collecting decision performance, and generating simulation test data;
s404: based on the simulation test data, the DDPG algorithm refinement strategy is applied again, and the optimized depression treatment strategy is generated.
In order to eliminate dimensional differences among different features, a Z-score standardization method can be adopted to conduct large-scale processing on data. And converting the value of each characteristic into a Z score corresponding to the value by calculating the mean value and the standard deviation of each sample, thereby realizing the normalization processing of the data.
Next, a depth deterministic strategy gradient algorithm (DDPG) is used to train the strategy network, generating a primarily optimized treatment strategy model. DDPG is an algorithm based on reinforcement learning, which can be used to solve the problem of continuous motion space. In the training process, the normalized evaluation data is used as input, and parameters of the strategy network are updated continuously and iteratively, so that the strategy network can select the optimal action according to the current environment state.
Then, monte Carlo simulation is carried out in a simulation environment, decision expressions are collected, and simulation test data are generated. Monte Carlo simulation is a commonly used random sampling method that can be used to estimate the performance of a system or process. The primarily optimized treatment strategy model can be applied to a simulation environment, and decision-making performance data are collected through multiple times of running simulation experiments. These data can be used to evaluate the efficacy and stability of a treatment strategy.
And finally, based on the simulation test data, the DDPG algorithm refinement strategy is applied again, and the optimized depression treatment strategy is generated. By analyzing the simulation test data, it can be found that the policy network has shortcomings in certain situations. Therefore, the hyper-parameters of the DDPG algorithm can be further adjusted, or the structure of the strategy network is improved, so that the performance and the robustness of the treatment strategy are further improved.
Referring to fig. 6, based on video data in multi-modal patient data, a convolutional neural network and cyclic neural network hybrid model is adopted to track facial expressions and analyze emotional states, and the steps of generating facial expression recognition and emotion analysis reports are specifically as follows:
s501: based on video data in the multi-mode patient data, performing frame extraction and gray scale processing through an OpenCV tool to generate preprocessed video frame data;
S502: based on the preprocessed video frame data, extracting spatial features by adopting a convolutional neural network to generate facial feature data;
s503: based on facial feature data, applying a long-term and short-term memory network to analyze the expression sequence and generating facial expression dynamic data;
s504: based on the facial expression dynamic data, integrating facial feature data and facial expression dynamic data, and generating facial expression recognition and emotion analysis reports by adopting an emotion recognition algorithm.
Frame extraction and gray scale processing are performed on video data in multi-modal patient data using an OpenCV tool. By decomposing the video into successive image frames, each frame may be pre-processed, including graying, scaling, etc., to facilitate subsequent feature extraction and analysis.
Convolutional Neural Networks (CNNs) are employed to extract spatial features. CNN is a deep learning model that automatically learns spatial dependency information in images. By training a CNN model on the preprocessed video frame data, the correlation between each pixel and surrounding pixels can be learned, thereby extracting facial feature data. These feature data can be used to further analyze the facial expression.
Long and short term memory networks (LSTM) are applied to analyze expression sequences. LSTM is a Recurrent Neural Network (RNN) that can capture long-term dependencies in time series data. By inputting the facial feature data into one LSTM model, the law of change of facial expression can be learned, and facial expression dynamic data can be generated. These data can help to better understand the trend of facial expression changes.
Facial feature data and facial expression dynamic data are integrated together, and facial expression recognition and emotion analysis reports are generated by adopting an emotion recognition algorithm. The emotion recognition algorithm can judge the emotion state expressed by the current facial expression according to the facial feature data and the facial expression dynamic data. By analyzing the emotional states of the plurality of samples, a detailed facial expression recognition and emotion analysis report can be generated, and a reference basis is provided for doctors.
Referring to fig. 7, based on the optimized depression treatment strategy and facial expression recognition and emotion analysis report, the clinical decision is adjusted in real time by adopting an online learning algorithm, and the steps of generating the optimized clinical decision and dynamic adjustment strategy are specifically as follows:
s601: based on the optimized depression treatment strategy, parameters are adjusted by using an online learning method, and initially adjusted clinical decision parameters are generated;
S602: optimizing decision parameters based on facial expression recognition and emotion analysis reports by combining reinforcement learning technology, and generating emotion-adjusted clinical decision parameters;
s603: based on the clinical decision parameters of emotion adjustment, real-time monitoring and dynamic adjustment are carried out, and self-adaptive control strategies are used for generating real-time monitoring adjustment data;
s604: based on the real-time monitoring adjustment data, final online learning optimization is executed, real-time performance and individuation of decisions are ensured, and optimized clinical decisions and dynamic adjustment strategies are generated.
And carrying out parameter adjustment on the optimized depression treatment strategy by using an online learning method. Online learning is a learning method that can continuously update model parameters in a data stream. By using facial expression recognition and emotion analysis reports as feedback signals, the therapeutic effect and emotion state of a patient can be monitored in real time, and parameters of a therapeutic strategy can be adjusted according to the information. This may make the treatment strategy more personalized and targeted.
Decision parameters are further optimized in conjunction with reinforcement learning techniques. Reinforcement learning is a method of learning optimal behavior through interactions with an environment. By reporting facial expression recognition and emotion analysis as reward signals, a reinforcement learning model can be trained to select the best clinical decision parameters. Thus, the treatment strategy can be more flexible and adaptive.
And carrying out real-time monitoring and dynamic adjustment, and generating real-time monitoring adjustment data by using an adaptive control strategy. Adaptive control is a method that can automatically adjust the control strategy according to the state change of the system. By monitoring the therapeutic effect and emotional state of the patient in real time, the parameters of the therapeutic strategy can be continuously adjusted to adapt to the changing requirements of the patient. This can improve the effect and satisfaction of the treatment.
And executing final online learning optimization, ensuring the real-time performance and individuation of the decision, and generating an optimized clinical decision and a dynamic adjustment strategy. By continuously and iteratively updating model parameters and using an adaptive control strategy, real-time optimization and personalized adjustment of clinical decisions can be realized. This can improve the effectiveness of the treatment and patient satisfaction.
Referring to fig. 8, an AI-based depression clinical decision system is used for executing the above-mentioned AI-based depression clinical decision method, and the system includes a data preprocessing module, a multi-mode feature extraction module, a feature fusion and mode recognition module, a treatment strategy optimization module, an emotion analysis module, a decision parameter optimization module, and a system integration and optimization module.
The data preprocessing module is used for carrying out frequency domain analysis by adopting a Fourier transform algorithm based on original multi-mode data of a depression patient, and processing a missing value by utilizing a mean value filling method and a nearest neighbor interpolation method to generate preprocessed multi-mode data;
the multi-modal feature extraction module performs complex sampling and time alignment by adopting a time sequence analysis method based on the preprocessed multi-modal data, automatically extracts key features by utilizing a Mel frequency cepstrum coefficient and a deep learning feature automatic encoder, and generates a time-synchronous multi-modal feature data set;
the feature fusion and pattern recognition module utilizes a time-synchronous multi-mode feature data set, utilizes a weighted fusion strategy to fuse the data, analyzes the patterns through a support vector machine algorithm, and generates fused multi-mode feature data and a psychological behavior pattern recognition result;
the treatment strategy optimization module sorts the importance of the features by adopting a random forest algorithm based on the psychological behavior pattern recognition result, and establishes a prediction model by utilizing a gradient lifting decision tree algorithm to generate baseline feature analysis data and treatment response prediction data;
the emotion analysis module extracts voice and video features by using a long-short-term memory network algorithm and a three-dimensional convolutional neural network according to the fused multi-modal feature data, analyzes the voice and video features by applying an emotion recognition algorithm, and generates voice video depth feature fusion data and facial expression recognition and emotion analysis reports;
The decision parameter tuning module is used for adjusting parameters based on treatment response prediction data and facial expression recognition and emotion analysis reports by utilizing an online learning method and a reinforcement learning technology to generate preliminarily adjusted clinical decision parameters and emotion adjusted clinical decision parameters;
the system integration and optimization module analyzes model performance by adopting a confusion matrix and a receiver operation characteristic curve, performs large-scale processing on data by a Z-score standardization method, trains a strategy network by utilizing a depth deterministic strategy gradient algorithm, performs Monte Carlo simulation in a simulation environment, collects decision expression, applies a depth deterministic strategy gradient algorithm refinement strategy again, and generates an optimized treatment strategy and an optimized depression treatment strategy.
The system can rapidly and accurately identify the psychological behavior mode and the emotional state of the patient through the analysis of the multimodal data and the automatic extraction of the key characteristics, thereby improving the diagnosis and treatment efficiency. Meanwhile, according to the feature importance ordering of the patient and the establishment of the prediction model, personalized treatment strategy optimization is realized. In addition, various information such as voice, video and facial expression are comprehensively considered, so that a more comprehensive diagnosis result is provided. The parameters are adjusted in real time through online learning and reinforcement learning technologies, and an optimized treatment strategy is generated by utilizing a depth deterministic strategy gradient algorithm, so that the treatment effect is further improved.
Referring to fig. 9, the data preprocessing module includes a frequency domain analysis sub-module, a missing value processing sub-module, and a data normalization sub-module;
the multi-mode feature extraction module comprises a time alignment sub-module, a key feature extraction sub-module and a feature data set construction sub-module;
the feature fusion and pattern recognition module comprises a data fusion sub-module, a pattern recognition sub-module and a feature cascade sub-module;
the treatment strategy optimization module comprises a feature importance sequencing sub-module, a prediction model construction sub-module and a model verification and tuning sub-module;
the emotion analysis module comprises a voice feature extraction sub-module, a video feature extraction sub-module and a feature fusion sub-module;
the decision parameter tuning module comprises a feature selection sub-module, a model selection sub-module, a super parameter tuning sub-module and a model evaluation sub-module;
the system integration and optimization module comprises a performance analysis sub-module, a data scale sub-module, a strategy network training sub-module, a Monte Carlo simulation sub-module and a strategy refinement sub-module.
In the data preprocessing module, a frequency domain analysis submodule adopts a Fourier transform algorithm to carry out frequency domain analysis on original multi-mode data of a patient suffering from depression; the missing value processing submodule processes missing values in the preprocessed multi-mode data by using a mean value filling method and a nearest neighbor interpolation method; and the data normalization sub-module performs normalization processing on the processed data.
In the multi-mode feature extraction module, a time alignment submodule carries out repeated sampling and time alignment on the preprocessed multi-mode data by adopting a time sequence analysis method; the key feature extraction submodule automatically extracts key features by utilizing the mel frequency cepstrum coefficient and the deep learning feature automatic encoder; the feature dataset construction submodule generates a time-synchronized multimodal feature dataset.
In the feature fusion and mode identification module, a data fusion submodule fuses the multi-mode feature data sets synchronized in time by using a weighted fusion strategy; the pattern recognition sub-module analyzes the fused characteristic data through a support vector machine algorithm to generate fused multi-pattern characteristic data; and the feature cascading sub-module cascades the fused multi-mode feature data with the psychological behavior mode recognition result.
In the treatment strategy optimization module, a feature importance ranking submodule adopts a random forest algorithm to rank the importance of features of the fused multi-mode feature data; the prediction model construction submodule establishes a prediction model by utilizing a gradient lifting decision tree algorithm; the model verification and tuning sub-module verifies and tunes the prediction model to generate baseline characteristic analysis data and treatment response prediction data.
In the emotion analysis module, a voice feature extraction sub-module extracts voice features by using a long-term memory network algorithm and a three-dimensional convolutional neural network; the video feature extraction submodule extracts video features by using a long-term and short-term memory network algorithm and a three-dimensional convolutional neural network; and the feature fusion sub-module fuses the voice features and the video features to generate voice video depth feature fusion data. And the facial expression recognition and emotion analysis report is generated by analyzing the fused multi-modal feature data and the voice video depth feature fusion data.
In the decision parameter tuning module, a feature selection submodule selects important features according to treatment response prediction data and facial expression recognition and emotion analysis reports; the model selection submodule selects a proper model for parameter adjustment; the super parameter tuning sub-module adjusts the super parameters of the model; the model evaluation sub-module evaluates the adjusted model performance.
In the system integration and optimization module, a performance analysis sub-module adopts a confusion matrix and a receiver operation characteristic curve to analyze the performance of the model; the data scale submodule carries out scale processing on the data by utilizing a Z score standardization method; the strategy network training submodule trains a strategy network by using a depth deterministic strategy gradient algorithm; the Monte Carlo simulation sub-module performs Monte Carlo simulation in a simulation environment and collects decision expressions; the strategy refinement sub-module applies the depth deterministic strategy gradient algorithm refinement strategy again, generating an optimized treatment strategy and an optimized depression treatment strategy.
The present invention is not limited to the above embodiments, and any equivalent embodiments which can be changed or modified by the technical disclosure described above can be applied to other fields, but any simple modification, equivalent changes and modification made to the above embodiments according to the technical matter of the present invention will still fall within the scope of the technical disclosure.

Claims (10)

1. An AI-based depression clinical decision method, comprising the steps of:
based on the depression patient, adopting a data preprocessing technology to clean and synchronize voice, video, text and physiological signals to generate multi-mode patient data;
based on the multi-modal patient data, adopting a long-short-term memory network and a three-dimensional convolutional neural network to analyze voice and video data, and carrying out feature fusion to generate a psychological behavior pattern recognition result;
based on the psychological behavior pattern recognition result, adopting a random forest and gradient lifting decision tree to predict the treatment response of the patient, evaluating the treatment effect, and generating a depression prediction result and a treatment effect evaluation report;
Based on the treatment effect evaluation report, adopting a depth deterministic strategy gradient algorithm to optimize a strategy, and testing on a simulation platform to generate an optimized depression treatment strategy;
based on video data in the multi-mode patient data, adopting a convolutional neural network and cyclic neural network mixed model to track facial expressions and analyze emotional states, and generating facial expression recognition and emotion analysis reports;
based on the optimized depression treatment strategy, facial expression recognition and emotion analysis report, adopting an online learning algorithm to adjust clinical decisions in real time, and generating optimized clinical decisions and dynamic adjustment strategies;
the multi-modal patient data is specifically a patient information set subjected to noise reduction, outlier processing and time stamp alignment, the psychological behavior pattern recognition result comprises an emotion state label, language expression habit and limb language characteristics, the depression prediction result is specifically probability distribution of future health states of the patient, the treatment effect evaluation report comprises benefit comparison of multiple treatment schemes, and the facial expression recognition and emotion analysis report comprises facial muscle activity data and emotion change-based time sequence analysis.
2. The AI-based depression clinical decision method of claim 1, wherein the step of generating multimodal patient data based on depression patient data by cleaning and synchronizing voice, video, text and physiological signals using data preprocessing techniques comprises:
performing frequency domain analysis by adopting Fourier transform based on original multi-mode data of a depression patient, and processing missing values by mean filling and nearest neighbor interpolation to generate preprocessed multi-mode data;
resampling and time alignment are carried out by adopting a time sequence analysis method based on the preprocessed multi-mode data, so as to generate time-synchronous multi-mode data;
based on the time-synchronous multi-mode data, extracting key features by adopting a Mel frequency cepstrum coefficient and a deep learning feature automatic encoder to generate a multi-mode feature data set;
and based on the multi-modal feature data set, adopting a weighted fusion strategy to fuse multi-modal data and generating fused multi-modal feature data.
3. The AI-based depression clinical decision method of claim 1, wherein based on the multimodal patient data, voice and video data analysis is performed using a long-term memory network and a three-dimensional convolutional neural network, and feature fusion is performed, and the step of generating a psychobehavioral pattern recognition result specifically comprises:
Based on the fused multi-mode characteristic data, extracting dynamic time characteristics of voice by adopting a long-short-period memory network, and generating voice time sequence characteristics;
based on the fused multi-mode characteristic data, extracting the spatial characteristics and dynamic changes of the video by adopting a three-dimensional convolutional neural network, and generating the space-time characteristics of the video;
based on the voice time sequence features and the video space-time features, performing depth feature fusion by using feature cascading and fusion strategies to generate voice video depth feature fusion data;
and carrying out pattern recognition by adopting a support vector machine based on the voice video depth feature fusion data to generate a psychological behavior pattern recognition result.
4. The AI-based depression clinical decision method of claim 1, wherein based on the psychographic pattern recognition result, the treatment response prediction of the patient is performed using a random forest and gradient boost decision tree, and the treatment effect is evaluated, and the step of generating a depression prediction result and a treatment effect evaluation report is specifically:
based on the psychological behavior pattern recognition result, ranking the feature importance by adopting a random forest algorithm, analyzing the change before and after treatment, and generating baseline feature analysis data;
Constructing a prediction model by adopting a gradient lifting decision tree based on the baseline characteristic analysis data, and performing treatment response prediction to generate treatment response prediction data;
based on the treatment response prediction data, performing model verification and tuning, and using k-fold cross verification to ensure generalization capability of the model to generate a treatment effect evaluation model;
and analyzing the model performance by adopting a confusion matrix and a receiver operation characteristic curve based on the treatment effect evaluation model, and generating a depression predicted result and a treatment effect evaluation report.
5. The AI-based depression clinical decision method of claim 1, wherein based on the treatment effect evaluation report, a depth deterministic strategy gradient algorithm optimization strategy is adopted, and the simulation platform is tested, and the step of generating an optimized depression treatment strategy specifically comprises:
based on the treatment effect evaluation report, performing scale processing on the data by adopting a Z-score standardization method, and generating normalized evaluation data;
based on the normalized evaluation data, training a strategy network by adopting a depth deterministic strategy gradient algorithm to generate a primarily optimized treatment strategy model;
Based on the primarily optimized treatment strategy model, performing Monte Carlo simulation in a simulation environment, collecting decision performance, and generating simulation test data;
and based on the simulation test data, the DDPG algorithm refinement strategy is applied again, and the optimized depression treatment strategy is generated.
6. The AI-based depression clinical decision method of claim 1, wherein tracking facial expressions and analyzing emotional states based on video data in the multimodal patient data using a convolutional neural network and cyclic neural network hybrid model, the step of generating facial expression recognition and emotional analysis reports is specifically:
based on video data in the multi-mode patient data, performing frame extraction and gray scale processing through an OpenCV tool to generate preprocessed video frame data;
based on the preprocessed video frame data, extracting spatial features by adopting a convolutional neural network to generate facial feature data;
based on the facial feature data, analyzing the expression sequence by applying a long-term memory network to generate facial expression dynamic data;
and based on the facial expression dynamic data, integrating facial feature data and facial expression dynamic data, and generating facial expression recognition and emotion analysis reports by adopting an emotion recognition algorithm.
7. The AI-based depression clinical decision method of claim 1, wherein based on the optimized depression treatment strategy and facial expression recognition and emotion analysis report, the clinical decision is adjusted in real time using an online learning algorithm, and the steps of generating the optimized clinical decision and dynamic adjustment strategy are specifically as follows:
based on the optimized depression treatment strategy, parameters are adjusted by using an online learning method, and initially adjusted clinical decision parameters are generated;
optimizing decision parameters based on the facial expression recognition and emotion analysis report in combination with reinforcement learning technology to generate emotion-adjusted clinical decision parameters;
based on the emotion-adjusted clinical decision parameters, real-time monitoring and dynamic adjustment are carried out, and self-adaptive control strategies are used for generating real-time monitoring adjustment data;
based on the real-time monitoring and adjusting data, final online learning optimization is executed, real-time performance and individuation of decisions are ensured, and optimized clinical decisions and dynamic adjusting strategies are generated.
8. An AI-based depression clinical decision system, characterized in that it comprises a data preprocessing module, a multi-modal feature extraction module, a feature fusion and mode recognition module, a treatment strategy optimization module, a mood analysis module, a decision parameter optimization module, and a system integration and optimization module according to the AI-based depression clinical decision method of any one of claims 1-7.
9. The AI-based depression clinical decision system of claim 8, wherein the data preprocessing module performs frequency domain analysis based on raw multi-modal data of a depression patient using a fourier transform algorithm and processes missing values using a mean-fill method and a nearest neighbor interpolation method to generate preprocessed multi-modal data;
the multi-modal feature extraction module performs complex sampling and time alignment by adopting a time sequence analysis method based on the preprocessed multi-modal data, automatically extracts key features by utilizing a Mel frequency cepstrum coefficient and a deep learning feature automatic encoder, and generates a time-synchronous multi-modal feature data set;
the feature fusion and pattern recognition module utilizes a time-synchronous multi-modal feature data set, utilizes a weighted fusion strategy to fuse data, analyzes patterns through a support vector machine algorithm, and generates fused multi-modal feature data and a psychological behavior pattern recognition result;
the treatment strategy optimization module sorts the importance of the features by adopting a random forest algorithm based on the psychological behavior pattern recognition result, and establishes a prediction model by utilizing a gradient lifting decision tree algorithm to generate baseline feature analysis data and treatment response prediction data;
The emotion analysis module extracts voice and video features by using a long-term memory network algorithm and a three-dimensional convolutional neural network according to the fused multi-modal feature data, analyzes the voice and video features by applying an emotion recognition algorithm, and generates voice video depth feature fusion data and facial expression recognition and emotion analysis reports;
the decision parameter tuning module is used for adjusting parameters based on treatment response prediction data and facial expression recognition and emotion analysis reports by utilizing an online learning method and a reinforcement learning technology to generate preliminarily adjusted clinical decision parameters and emotion adjusted clinical decision parameters;
the system integration and optimization module analyzes model performance by adopting a confusion matrix and a receiver operation characteristic curve, performs large-scale processing on data by a Z score standardization method, trains a strategy network by utilizing a depth deterministic strategy gradient algorithm, performs Monte Carlo simulation in a simulation environment, collects decision expression, applies a depth deterministic strategy gradient algorithm refinement strategy again, and generates an optimized treatment strategy and an optimized depression treatment strategy.
10. The AI-based depression clinical decision making system of claim 8, wherein the data preprocessing module comprises a frequency domain analysis sub-module, a missing value processing sub-module, a data normalization sub-module;
The multi-mode feature extraction module comprises a time alignment sub-module, a key feature extraction sub-module and a feature data set construction sub-module;
the feature fusion and pattern recognition module comprises a data fusion sub-module, a pattern recognition sub-module and a feature cascade sub-module;
the treatment strategy optimization module comprises a feature importance ordering sub-module, a prediction model construction sub-module and a model verification and tuning sub-module;
the emotion analysis module comprises a voice feature extraction sub-module, a video feature extraction sub-module and a feature fusion sub-module;
the decision parameter tuning module comprises a feature selection sub-module, a model selection sub-module, a super parameter tuning sub-module and a model evaluation sub-module;
the system integration and optimization module comprises a performance analysis sub-module, a data scale sub-module, a strategy network training sub-module, a Monte Carlo simulation sub-module and a strategy refinement sub-module.
CN202311664055.8A 2023-12-06 2023-12-06 Depression clinical decision method and system based on AI Pending CN117747064A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202311664055.8A CN117747064A (en) 2023-12-06 2023-12-06 Depression clinical decision method and system based on AI

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202311664055.8A CN117747064A (en) 2023-12-06 2023-12-06 Depression clinical decision method and system based on AI

Publications (1)

Publication Number Publication Date
CN117747064A true CN117747064A (en) 2024-03-22

Family

ID=90260063

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202311664055.8A Pending CN117747064A (en) 2023-12-06 2023-12-06 Depression clinical decision method and system based on AI

Country Status (1)

Country Link
CN (1) CN117747064A (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN117976148A (en) * 2024-04-01 2024-05-03 成都市双流区妇幼保健院 Machine learning-based prediction method and system for mental health problems of children

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN117976148A (en) * 2024-04-01 2024-05-03 成都市双流区妇幼保健院 Machine learning-based prediction method and system for mental health problems of children

Similar Documents

Publication Publication Date Title
US20210201003A1 (en) Synthetic data for neural network training using vectors
CN108351862B (en) Method and apparatus for determining developmental progress using artificial intelligence and user input
JP2022009339A (en) Platform and system for digital personalized medical treatment
CN110188836B (en) Brain function network classification method based on variational self-encoder
CN117457147B (en) Personalized nursing planning method and system for rehabilitation patient
US11670421B2 (en) Method and system enabling digital biomarker data integration and analysis for clinical treatment impact
Vairachilai et al. Body sensor 5 G networks utilising deep learning architectures for emotion detection based on EEG signal processing
CN117438029A (en) Intelligent evaluation system for wound severity of orthopedic wound patient
CN117198468A (en) Intervention scheme intelligent management system based on behavior recognition and data analysis
Meshram et al. Diagnosis of depression level using multimodal approaches using deep learning techniques with multiple selective features
Bakariya et al. Facial emotion recognition and music recommendation system using CNN-based deep learning techniques
CN117009668A (en) Self-adaptive learning content recommendation method and system based on deep reinforcement learning
CN117747064A (en) Depression clinical decision method and system based on AI
CN113360772B (en) Interpretable recommendation model training method and device
Ahmadieh et al. Visual image reconstruction based on EEG signals using a generative adversarial and deep fuzzy neural network
CN106845386A (en) A kind of action identification method based on dynamic time warping Yu Multiple Kernel Learning
Kukkar et al. A novel methodology using RNN+ LSTM+ ML for predicting student’s academic performance
Wang et al. Epileptic seizures prediction based on unsupervised learning for feature extraction
Shaffi et al. Multi-Planar MRI-Based Classification of Alzheimer's Disease Using Tree-Based Machine Learning Algorithms
Eeswar et al. Better you: Automated tool that evaluates mental health and provides guidance for university students
CN117850601B (en) System and method for automatically detecting vital signs of handheld PDA
Moosavi et al. Early mental stress detection using q-learning embedded starling murmuration optimiser-based deep learning model
Fofanah et al. Adaptive Neuro-Fuzzy Inference System with Non-Linear Regression Model for Online Learning Framework
Bekenova et al. Emotion recognition and classification based on audio data using AI
CN117158904B (en) Old people cognitive disorder detection system and method based on behavior analysis

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination