CN116503841A - A mental health intelligent emotion recognition method - Google Patents

A mental health intelligent emotion recognition method Download PDF

Info

Publication number
CN116503841A
CN116503841A CN202310478600.8A CN202310478600A CN116503841A CN 116503841 A CN116503841 A CN 116503841A CN 202310478600 A CN202310478600 A CN 202310478600A CN 116503841 A CN116503841 A CN 116503841A
Authority
CN
China
Prior art keywords
model
emotion recognition
recognition
algorithm
face
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202310478600.8A
Other languages
Chinese (zh)
Inventor
姚尧
袁礼承
徐锋
陈冠伟
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Good Feeling Health Industry Group Co ltd
Original Assignee
Good Feeling Health Industry Group Co ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Good Feeling Health Industry Group Co ltd filed Critical Good Feeling Health Industry Group Co ltd
Priority to CN202310478600.8A priority Critical patent/CN116503841A/en
Publication of CN116503841A publication Critical patent/CN116503841A/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L25/00Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
    • G10L25/48Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use
    • G10L25/51Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use for comparison or discrimination
    • G10L25/63Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use for comparison or discrimination for estimating an emotional state
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/40Scenes; Scene-specific elements in video content
    • G06V20/46Extracting features or characteristics from the video content, e.g. video fingerprints, representative shots or key frames
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/50Context or environment of the image
    • G06V20/59Context or environment of the image inside of a vehicle, e.g. relating to seat occupancy, driver state or inner lighting conditions
    • G06V20/597Recognising the driver's state or behaviour, e.g. attention or drowsiness
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/172Classification, e.g. identification
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/174Facial expression recognition
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02DCLIMATE CHANGE MITIGATION TECHNOLOGIES IN INFORMATION AND COMMUNICATION TECHNOLOGIES [ICT], I.E. INFORMATION AND COMMUNICATION TECHNOLOGIES AIMING AT THE REDUCTION OF THEIR OWN ENERGY USE
    • Y02D10/00Energy efficient computing, e.g. low power processors, power management or thermal management

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Health & Medical Sciences (AREA)
  • General Physics & Mathematics (AREA)
  • Human Computer Interaction (AREA)
  • General Health & Medical Sciences (AREA)
  • Oral & Maxillofacial Surgery (AREA)
  • Psychiatry (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Acoustics & Sound (AREA)
  • Signal Processing (AREA)
  • Computational Linguistics (AREA)
  • Hospice & Palliative Care (AREA)
  • Child & Adolescent Psychology (AREA)
  • Measurement Of The Respiration, Hearing Ability, Form, And Blood Characteristics Of Living Organisms (AREA)

Abstract

The invention provides a mental health intelligent emotion recognition method, which comprises the following steps: a user terminal and a service cloud; the service cloud creates a model, trains the model, adjusts model parameters and identifies the subsequent service after emotion; the user terminal is responsible for shooting, storing and expression recognition of voice and pictures and uploading the voice and pictures to the service remote end. And the server side continuously trains a model according to the collected characteristic expression data set, optimizes model parameters, forms an effective model algorithm and continuously improves the accuracy and the effectiveness of emotion recognition. The face photo of the driver can be acquired in real time, the facial expression is recognized, and a preliminary emotion recognition report is provided.

Description

一种心理健康智能情绪识别方法A mental health intelligent emotion recognition method

技术领域technical field

本发明涉及情绪识别领域,尤其涉及一种心理健康智能情绪识别方法。The invention relates to the field of emotion recognition, in particular to a mental health intelligent emotion recognition method.

背景技术Background technique

心理情绪是人们对客观事物的一种反馈方式,更是人们情感的呈现方式。情绪作为人类心理活动的重要组成部分,在组织和指导人们的行为、交流与预测他人意愿方面,都起着举足重轻的作用。由于生理变化、表情变化和声音的变化等因素,使人们表达出自己的情绪时产生对应信号,并且引起他人推断,这一过程就是情绪的识别过程。Psychological emotion is a way of people's feedback to objective things, and it is also a way of presenting people's emotions. As an important part of human psychological activities, emotions play a pivotal role in organizing and guiding people's behavior, communicating and predicting the wishes of others. Due to factors such as physiological changes, facial expression changes, and voice changes, people express their emotions and generate corresponding signals, and cause others to infer. This process is the process of emotion recognition.

好心情是一家聚焦于CNS(中枢神经)领域精神心理健康服务的互联网医疗平台,通过大量数据以及文献分析,探讨了情绪的定义与分类,并对情绪识别研究的未来方向及其应用价值进行了展望。Good mood is an Internet medical platform focusing on mental and mental health services in the CNS (central nervous system). Through a large amount of data and literature analysis, it discusses the definition and classification of emotions, and discusses the future direction of emotion recognition research and its application value. outlook.

长期以来,长途货车司机的出险率较高,造成较大的人身和财产损失,通过后续跟进分析,大多事故均是由于货车司机有较大的情绪波动引起的不正当驾驶操作引起的,保险公司为了提供更好的服务,降低出险率,提出通过实时监控货车司机人脸表情,分析货车司机的情绪,当司机产生危险情绪时,及时上报到服务端,通过服务端语音干预的服务,平复司机情绪,规范司机安全驾驶行为,从而降低事故发生的概率。For a long time, the accident rate of long-distance truck drivers has been relatively high, resulting in relatively large personal and property losses. Through follow-up analysis, most of the accidents were caused by improper driving operations caused by truck drivers’ large emotional fluctuations. Insurance In order to provide better services and reduce the risk of accidents, the company proposes to monitor truck drivers’ facial expressions in real time, analyze truck drivers’ emotions, and report to the server in time when the driver has a dangerous emotion. The driver's emotions can regulate the driver's safe driving behavior, thereby reducing the probability of accidents.

基于以上问题分析,综合目前市面上的产品解决方案。提出产品解决方案为,通过对货车司机的人脸录入,人脸建模,创建好人脸识别模型。终端设备实时的检测驾驶位置的人脸特征照片,识别货车司机身份,通过人脸表情照片,识别货车司机驾驶中的人脸表情,综合时间维度,分析出货车司机的情绪,当货车司机情绪出现异常,实时上报分析结果,照片信息到服务端,产生告警。服务端接收到告警信息,人工分析危险行为,通过服务端指令下发,下发语音包,语音温馨提示行为的产品服务模式,有效终止驾驶员的危险驾驶情绪和行为,达到降低事故的服务方式。Based on the analysis of the above problems, the product solutions currently on the market are synthesized. The proposed product solution is to create a good face recognition model by entering and modeling the face of the truck driver. The terminal device detects the facial feature photos of the driving position in real time, identifies the identity of the truck driver, and recognizes the facial expressions of the truck driver while driving through the facial expression photos, and analyzes the truck driver's emotions comprehensively in the time dimension. When the truck driver's emotions appear Abnormal, report analysis results in real time, photo information to the server, and generate an alarm. The server receives the alarm information, manually analyzes the dangerous behavior, sends out the command through the server, sends out the voice package, and the product service mode of warm voice reminder behavior, effectively terminates the driver's dangerous driving emotions and behaviors, and achieves the service mode of reducing accidents .

发明内容Contents of the invention

鉴于上述问题,提出了本发明以便提供克服上述问题或者至少部分地解决上述问题的一种心理健康智能情绪识别方法。In view of the above problems, the present invention is proposed to provide a mental health intelligent emotion recognition method that overcomes the above problems or at least partially solves the above problems.

根据本发明的一个方面,提供了一种心理健康智能情绪识别方法,所述情绪识别方法包括:用户终端和服务云端;According to one aspect of the present invention, a mental health intelligent emotion recognition method is provided, and the emotion recognition method includes: a user terminal and a service cloud;

所述服务云端创建模型,模型训练,模型参数调整及情绪识别后的后续服务;The service cloud creates a model, model training, model parameter adjustment and follow-up services after emotion recognition;

所述用户终端负责语音和照片的拍摄、存储、表情识别,并上传至所述服务远端。The user terminal is responsible for taking, storing, and facial expression recognition of voice and photos, and uploading them to the remote service end.

可选的,所述情绪识别的方法具体包括:Optionally, the method for emotion recognition specifically includes:

通过所述用户终端将人脸照片上传到所述服务云端;Upload the face photo to the service cloud through the user terminal;

所述服务云端将所述人脸照片通过人脸分析模型,创建好人脸识别模型参数,再将参数下发到所述用户终端的模型用于终端的人脸识别;The service cloud passes the face photo through the face analysis model, creates face recognition model parameters, and then sends the parameters to the model of the user terminal for face recognition of the terminal;

所述用户终端通过定时拍摄照片组,识别人脸,分析人脸表情结果,分析出当前人的情绪,并将分析结果上传到所述服务云端。The user terminal regularly takes photos, recognizes faces, analyzes the results of facial expressions, analyzes the current emotions of people, and uploads the analysis results to the service cloud.

可选的,所述将分析结果上传到所述服务云端之后还包括:Optionally, after uploading the analysis results to the service cloud, it also includes:

所述服务云端根据不同的情绪结果,有差别的提供后续服务。The service cloud provides follow-up services differently according to different emotional results.

可选的,所述人脸识别算法具体包括:人脸识别算法,情绪分析算法会随着业务量的增加,业务场景的变化,数据量的增大,动态的调整算法模型和算法参数。Optionally, the face recognition algorithm specifically includes: a face recognition algorithm, and a sentiment analysis algorithm that dynamically adjusts the algorithm model and algorithm parameters as business volume increases, business scenarios change, and data volume increases.

可选的,所述表情识别的结果会和业务线进行深度融合,为业务线的决策提供辅助支持。Optionally, the expression recognition result will be deeply integrated with the business line to provide auxiliary support for the business line's decision-making.

本发明提供的一种心理健康智能情绪识别方法,所述情绪识别方法包括:用户终端和服务云端;所述服务云端创建模型,模型训练,模型参数调整及情绪识别后的后续服务;所述用户终端负责语音和照片的拍摄、存储、表情识别,并上传至所述服务远端。服务端根据收集到的特征表情数据集,不断的训练模型,优化模型参数,形成有效的模型算法,不断提升情绪识别的准确性和有效性。能够实时的采集驾驶员的人脸照片,识别人脸表情,出具初步的情绪识别报告。A mental health intelligent emotion recognition method provided by the present invention, the emotion recognition method includes: a user terminal and a service cloud; the service cloud creates a model, model training, model parameter adjustment and follow-up services after emotion recognition; the user The terminal is responsible for taking, storing, and facial expression recognition of voice and photos, and uploading them to the remote end of the service. Based on the collected characteristic expression data sets, the server continuously trains the model, optimizes the model parameters, forms an effective model algorithm, and continuously improves the accuracy and effectiveness of emotion recognition. It can collect the driver's face photos in real time, recognize facial expressions, and issue a preliminary emotion recognition report.

上述说明仅是本发明技术方案的概述,为了能够更清楚了解本发明的技术手段,而可依照说明书的内容予以实施,并且为了让本发明的上述和其它目的、特征和优点能够更明显易懂,以下特举本发明的具体实施方式。The above description is only an overview of the technical solution of the present invention. In order to better understand the technical means of the present invention, it can be implemented according to the contents of the description, and in order to make the above and other purposes, features and advantages of the present invention more obvious and understandable , the specific embodiments of the present invention are enumerated below.

附图说明Description of drawings

为了更清楚地说明本发明实施例的技术方案,下面将对实施例描述中所需要使用的附图作简单地介绍,显而易见地,下面描述中的附图仅仅是本发明的一些实施例,对于本领域普通技术人员来讲,在不付出创造性劳动的前提下,还可以根据这些附图获得其它的附图。In order to more clearly illustrate the technical solutions of the embodiments of the present invention, the following will briefly introduce the accompanying drawings that need to be used in the description of the embodiments. Obviously, the accompanying drawings in the following description are only some embodiments of the present invention. For Those of ordinary skill in the art can also obtain other drawings based on these drawings without making creative efforts.

图1为本发明实施例提供的人脸情绪识别的流程图;Fig. 1 is the flowchart of facial emotion recognition provided by the embodiment of the present invention;

图2为本发明实施例提供的面部情绪识别+语音情绪识别的方法流程框图。Fig. 2 is a flowchart of a method for facial emotion recognition + voice emotion recognition provided by an embodiment of the present invention.

具体实施方式Detailed ways

下面将参照附图更详细地描述本公开的示例性实施例。虽然附图中显示了本公开的示例性实施例,然而应当理解,可以以各种形式实现本公开而不应被这里阐述的实施例所限制。相反,提供这些实施例是为了能够更透彻地理解本公开,并且能够将本公开的范围完整的传达给本领域的技术人员。Exemplary embodiments of the present disclosure will be described in more detail below with reference to the accompanying drawings. Although exemplary embodiments of the present disclosure are shown in the drawings, it should be understood that the present disclosure may be embodied in various forms and should not be limited by the embodiments set forth herein. Rather, these embodiments are provided for more thorough understanding of the present disclosure and to fully convey the scope of the present disclosure to those skilled in the art.

本发明的说明书实施例和权利要求书及附图中的术语“包括”和“具有”以及他们的任何变形,意图在于覆盖不排他的包含,例如,包含了一系列步骤或单元。The terms "comprising" and "having" and any variations thereof in the description, embodiments, claims and drawings of the present invention are intended to cover non-exclusive inclusion, for example, including a series of steps or units.

下面结合附图和实施例,对本发明的技术方案做进一步的详细描述。The technical solutions of the present invention will be described in further detail below in conjunction with the accompanying drawings and embodiments.

通过对货车司机的人脸录入,人脸建模,创建好人脸识别模型。终端设备实时的检测驾驶位置的人脸特征照片,识别货车司机身份,通过人脸表情照片,识别货车司机驾驶中的人脸表情,综合时间维度,分析出货车司机的情绪,当货车司机情绪出现异常,实时上报分析结果,照片信息到服务端,产生告警。服务端接收到告警信息,人工分析危险行为,通过服务端指令下发,下发语音包,语音温馨提示行为的产品服务模式,有效终止驾驶员的危险驾驶情绪和行为,达到降低事故的服务方式。Create a good face recognition model by entering and modeling the face of the truck driver. The terminal device detects the facial feature photos of the driving position in real time, identifies the identity of the truck driver, and recognizes the facial expressions of the truck driver while driving through the facial expression photos, and analyzes the truck driver's emotions comprehensively in the time dimension. When the truck driver's emotions appear Abnormal, report analysis results in real time, photo information to the server, and generate an alarm. The server receives the alarm information, manually analyzes the dangerous behavior, sends out the command through the server, sends out the voice package, and the product service mode of warm voice reminder behavior, effectively terminates the driver's dangerous driving emotions and behaviors, and achieves the service mode of reducing accidents .

目前的产品解决方案为,服务端远程录入人脸信息,生成人脸建模基础信息,下发人脸模型到终端设备,终端设备植入人脸识别算法,表情识别算法,终端设备控制照片的录入频次,表情识别的频次,保存关键特征表情数据,上传到服务端。终端设备可接收服务端的语音播放指令,服务端的算法参数调整和算法升级。The current product solution is to remotely input face information on the server side, generate basic face modeling information, and deliver the face model to the terminal device. The terminal device is embedded with face recognition algorithms and expression recognition algorithms, and the terminal device controls the photos. Input frequency, frequency of expression recognition, save key feature expression data, and upload to the server. The terminal device can receive voice playback instructions from the server, and algorithm parameter adjustment and algorithm upgrade from the server.

服务端根据收集到的特征表情数据集,不断的训练模型,优化模型参数,形成有效的模型算法,不断提升情绪识别的准确性和有效性。通过不断丰富语音服务包的内容和表现形式,提供更多更好更优质的后续服务。Based on the collected characteristic expression data sets, the server continuously trains the model, optimizes the model parameters, forms an effective model algorithm, and continuously improves the accuracy and effectiveness of emotion recognition. Provide more, better and higher-quality follow-up services by continuously enriching the content and presentation of the voice service package.

目前已经可以实时的采集驾驶员的人脸照片,识别人脸表情,出具初步的情绪识别报告。At present, it is possible to collect the driver's face photos in real time, recognize facial expressions, and issue a preliminary emotion recognition report.

情绪识别包含用户终端和服务端云端两大部分。云端负责模型的创建,模型训练,模型参数调整,以及情绪识别后的后续服务提供。终端负责语音和照片的拍摄,存储,表情识别,上传云端等功能。Emotion recognition includes two parts: user terminal and server cloud. The cloud is responsible for model creation, model training, model parameter adjustment, and subsequent service provision after emotion recognition. The terminal is responsible for voice and photo shooting, storage, expression recognition, uploading to the cloud and other functions.

情绪识别基本技术路线是:通过终端将人脸照片上传到云端,云端将照片通过人脸分析模型,创建好人脸识别模型参数,再将参数下发到终端模型用于终端的人脸识别。终端通过定时拍摄照片组,识别人脸,分析人脸表情结果,分析出当前人的情绪,并将分析结果上传到云端。云端根据不同的情绪结果,有差别的提供后续服务。The basic technical route of emotion recognition is: upload face photos to the cloud through the terminal, and the cloud passes the photos through the face analysis model to create face recognition model parameters, and then send the parameters to the terminal model for face recognition on the terminal. The terminal takes photo groups at regular intervals, recognizes faces, analyzes the results of facial expressions, analyzes the current emotions of people, and uploads the analysis results to the cloud. According to different emotional results, the cloud provides follow-up services differently.

人脸识别算法,情绪分析算法会随着业务量的增加,业务场景的变化,数据量的增大,动态的调整算法模型和算法参数。不断的提升算法的准确度和算法的业务场景覆盖范围。The face recognition algorithm and emotion analysis algorithm will dynamically adjust the algorithm model and algorithm parameters as the business volume increases, the business scene changes, and the data volume increases. Continuously improve the accuracy of the algorithm and the business scenario coverage of the algorithm.

表情识别目前只在linux系统上植入,未来会兼容其他终端设备,覆盖多场景的应用。随着算法的深入,未来对硬件算力会有较大需求,势必会和硬件厂商深度绑定,推动硬件的发展。Expression recognition is currently only implanted in the Linux system, and will be compatible with other terminal devices in the future, covering applications in multiple scenarios. With the deepening of the algorithm, there will be a greater demand for hardware computing power in the future, and it is bound to be deeply bound with hardware manufacturers to promote the development of hardware.

表情识别的结果会和其他业务线进行深度融合,为其他业务线的决策提供辅助支持。The results of expression recognition will be deeply integrated with other business lines to provide auxiliary support for the decision-making of other business lines.

人脸情绪识别+语音情绪识别的核心技术分两个层面。一个是算法模型,一个是多场景设备的兼容。The core technology of face emotion recognition + speech emotion recognition is divided into two levels. One is the algorithm model, and the other is the compatibility of multi-scenario devices.

算法模型方面,目前市面上的人脸识别算法和表情识别算法,存在识别度不高,识别结果简单等特点,特别是语音情绪识别的数据集,算法识别度特别低。我司的算法会根据业务场景,大量数据训练,不断的提升算法模型的识别准确度,表情识别的准确度,情绪分析的多场景覆盖,单个人的表情变化趋势。不同地域,不同职业的人群表情基本特点模型搭建,抑郁症患者的情绪特点模型。In terms of algorithm models, the current face recognition algorithms and expression recognition algorithms on the market have the characteristics of low recognition degree and simple recognition results, especially for the voice emotion recognition data set, the algorithm recognition degree is particularly low. Our algorithm will continuously improve the recognition accuracy of the algorithm model, the accuracy of expression recognition, the multi-scenario coverage of emotion analysis, and the expression change trend of a single person based on business scenarios and a large amount of data training. Build a model of the basic characteristics of facial expressions of people in different regions and occupations, and model the emotional characteristics of patients with depression.

目前人脸识别+表情识别算法模型大多在固定终端设定。对于各种移动终端,工业终端的覆盖很少。市面上的硬件设备也大多不支持算法模型的植入,特别是算法模型的动态升级都不具备可操作性。工业级的硬件也存在价格昂贵,算力有限的特点,我司可根据业务场景,不断覆盖业务场景,减低成本,使得这个技术可以惠及到更多人群。Currently, face recognition + expression recognition algorithm models are mostly set on fixed terminals. For various mobile terminals, industrial terminals have little coverage. Most of the hardware devices on the market do not support the implantation of algorithm models, especially the dynamic upgrade of algorithm models is not operable. Industrial-grade hardware is also expensive and has limited computing power. Our company can continuously cover business scenarios and reduce costs according to business scenarios, so that this technology can benefit more people.

人脸算法在工业级的应用目前基本都是基础的C语言植入,但是大量的算法模型都是python语言。At present, the application of face algorithm at the industrial level is basically implanted in the basic C language, but a large number of algorithm models are in python language.

如图1所示,面部情绪识别是基于人工智能技术,用于分析来源于不同图片,视频中的情绪技的技术。一般从摄像头、社交媒体页面、视频库等获取信号,进行静态与动态的面部表情检测,再将情绪状态归类;面部情绪识别通常基于深度学习算法,经过人脸信息预处理、特征学习、情绪识别三阶段;在精神心理疾病领域,面部情绪识别技术可用于预测患病几率、辅助诊断,以及辅助提升医疗人员的照护质量。目前面部情绪识别存在数据准确性、算法公平、数据隐私以及用户反应性等内生风险。As shown in Figure 1, facial emotion recognition is based on artificial intelligence technology and is used to analyze emotions from different pictures and videos. Generally, signals are obtained from cameras, social media pages, video libraries, etc., static and dynamic facial expression detection is performed, and then emotional states are classified; facial emotion recognition is usually based on deep learning algorithms, after face information preprocessing, feature learning, emotion There are three stages of identification; in the field of mental illness, facial emotion recognition technology can be used to predict the probability of illness, assist in diagnosis, and assist in improving the quality of care for medical personnel. At present, facial emotion recognition has inherent risks such as data accuracy, algorithm fairness, data privacy, and user reactivity.

如图2所示,语音情绪识别,即给定一段语音信号,计算机自动判断出说话人的多维度信息。人的语音产生包含大脑认知活动和身体肌肉运动的复杂的多系统协调过程。语音信号包含声学、语言】情感三层信息,其运动纤维高度机械化,所生成的语音信号具有客观的、可重复的特征。语音识别流程经过语音信号处理、特征提取、情绪建模三个阶段,其中涉及的识别算法有传统算法、基于深度学习的算法和端到端算法。语音生物标志物是临床结果相关的语音音频信号的特征或特征组合,可用于精神心理疾病的筛查、诊断、病情检测,以及AI+CBT、数字疗法的干预手段。目前也有受情绪定义模糊、数据稀缺且标注困难的影响,语音情绪识别技术难度大。As shown in Figure 2, speech emotion recognition, that is, given a speech signal, the computer automatically judges the multi-dimensional information of the speaker. Human speech production is a complex multi-system coordination process involving brain cognitive activities and body muscle movements. Speech signals contain three layers of information: acoustics, language, and emotion. The motor fibers are highly mechanized, and the generated speech signals have objective and repeatable characteristics. The speech recognition process goes through three stages: speech signal processing, feature extraction, and emotion modeling. The recognition algorithms involved include traditional algorithms, algorithms based on deep learning, and end-to-end algorithms. Speech biomarkers are features or feature combinations of speech and audio signals related to clinical outcomes, which can be used for screening, diagnosis, and disease detection of mental and psychological diseases, as well as intervention methods for AI+CBT and digital therapy. At present, due to the vague definition of emotion, the scarcity of data and the difficulty of labeling, speech emotion recognition technology is difficult.

面部情绪识别+语音情绪识别可有效提升识别精度,可相互验证识别过程中产生的数据干扰因素,用户情绪影响干扰因素影响结果的准确性。Facial emotion recognition + voice emotion recognition can effectively improve the recognition accuracy, and can mutually verify the data interference factors generated in the recognition process, and the user's emotions affect the interference factors and affect the accuracy of the results.

有益效果:服务端远程录入人脸信息,生成人脸建模基础信息,下发人脸模型到终端设备,终端设备植入人脸识别算法,表情识别算法,终端设备控制照片的录入频次,表情识别的频次,保存关键特征表情数据,上传到服务端。终端设备可接收服务端的语音播放指令,服务端的算法参数调整和算法升级。Beneficial effects: the server remotely enters face information, generates basic face modeling information, and sends the face model to the terminal device. The terminal device is embedded with a face recognition algorithm, an expression recognition algorithm, and the terminal device controls the frequency of photo entry, expression The frequency of recognition, save key feature expression data, and upload to the server. The terminal device can receive voice playback instructions from the server, and algorithm parameter adjustment and algorithm upgrade from the server.

服务端根据收集到的特征表情数据集,不断的训练模型,优化模型参数,形成有效的模型算法,不断提升情绪识别的准确性和有效性。通过不断丰富语音服务包的内容和表现形式,提供更多更好更优质的后续服务。Based on the collected characteristic expression data sets, the server continuously trains the model, optimizes the model parameters, forms an effective model algorithm, and continuously improves the accuracy and effectiveness of emotion recognition. Provide more, better and higher-quality follow-up services by continuously enriching the content and presentation of the voice service package.

目前已经能够实时的采集驾驶员的人脸照片,识别人脸表情,出具初步的情绪识别报告。At present, it has been able to collect the driver's face photos in real time, recognize facial expressions, and issue a preliminary emotion recognition report.

以上的具体实施方式,对本发明的目的、技术方案和有益效果进行了进一步详细说明,所应理解的是,以上仅为本发明的具体实施方式而已,并不用于限定本发明的保护范围,凡在本发明的精神和原则之内,所做的任何修改、等同替换、改进等,均应包含在本发明的保护范围之内。The above specific implementation manners have further described the purpose, technical solutions and beneficial effects of the present invention in detail. It should be understood that the above are only specific implementation modes of the present invention, and are not used to limit the protection scope of the present invention. Within the spirit and principles of the present invention, any modifications, equivalent replacements, improvements, etc., shall be included in the protection scope of the present invention.

Claims (5)

1.一种心理健康智能情绪识别方法,其特征在于,所述情绪识别方法包括:用户终端和服务云端;1. A mental health intelligent emotion recognition method is characterized in that, the emotion recognition method comprises: a user terminal and a service cloud; 所述服务云端创建模型,模型训练,模型参数调整及情绪识别后的后续服务;The service cloud creates a model, model training, model parameter adjustment and follow-up services after emotion recognition; 所述用户终端负责语音和照片的拍摄、存储、表情识别,并上传至所述服务远端。The user terminal is responsible for taking, storing, and facial expression recognition of voice and photos, and uploading them to the remote service end. 2.根据权利要求1所述的一种心理健康智能情绪识别方法,其特征在于,所述情绪识别的方法具体包括:2. a kind of mental health intelligent emotion recognition method according to claim 1, is characterized in that, the method for described emotion recognition specifically comprises: 通过所述用户终端将人脸照片上传到所述服务云端;Upload the face photo to the service cloud through the user terminal; 所述服务云端将所述人脸照片通过人脸分析模型,创建好人脸识别模型参数,再将参数下发到所述用户终端的模型用于终端的人脸识别;The service cloud passes the face photo through the face analysis model, creates face recognition model parameters, and then sends the parameters to the model of the user terminal for face recognition of the terminal; 所述用户终端通过定时拍摄照片组,识别人脸,分析人脸表情结果,分析出当前人的情绪,并将分析结果上传到所述服务云端。The user terminal regularly takes photos, recognizes faces, analyzes the results of facial expressions, analyzes the current emotions of people, and uploads the analysis results to the service cloud. 3.根据权利要求2所述的一种心理健康智能情绪识别方法,其特征在于,所述将分析结果上传到所述服务云端之后还包括:3. A kind of mental health intelligent emotion recognition method according to claim 2, is characterized in that, after described analysis result is uploaded to described service cloud, also comprises: 所述服务云端根据不同的情绪结果,有差别的提供后续服务。The service cloud provides follow-up services differently according to different emotional results. 4.根据权利要求2所述的一种心理健康智能情绪识别方法,其特征在于,所述人脸识别算法具体包括:人脸识别算法,情绪分析算法会随着业务量的增加,业务场景的变化,数据量的增大,动态的调整算法模型和算法参数。4. A mental health intelligent emotion recognition method according to claim 2, characterized in that, the face recognition algorithm specifically comprises: a face recognition algorithm, and the emotion analysis algorithm will increase with the increase in business volume, Changes, increase in the amount of data, and dynamically adjust the algorithm model and algorithm parameters. 5.根据权利要求1所述的一种心理健康智能情绪识别方法,其特征在于,所述表情识别的结果会和业务线进行深度融合,为业务线的决策提供辅助支持。5. A mental health intelligent emotion recognition method according to claim 1, characterized in that the result of the facial expression recognition will be deeply integrated with the business line to provide auxiliary support for the decision-making of the business line.
CN202310478600.8A 2023-04-28 2023-04-28 A mental health intelligent emotion recognition method Pending CN116503841A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202310478600.8A CN116503841A (en) 2023-04-28 2023-04-28 A mental health intelligent emotion recognition method

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202310478600.8A CN116503841A (en) 2023-04-28 2023-04-28 A mental health intelligent emotion recognition method

Publications (1)

Publication Number Publication Date
CN116503841A true CN116503841A (en) 2023-07-28

Family

ID=87324499

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202310478600.8A Pending CN116503841A (en) 2023-04-28 2023-04-28 A mental health intelligent emotion recognition method

Country Status (1)

Country Link
CN (1) CN116503841A (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN117496580A (en) * 2023-11-23 2024-02-02 好心情健康产业集团有限公司 Facial expression intelligent recognition robot terminal based on multi-person synchronous interaction

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108960023A (en) * 2017-09-19 2018-12-07 炬大科技有限公司 A kind of portable Emotion identification device
CN113221821A (en) * 2021-05-28 2021-08-06 中国工商银行股份有限公司 Business data pushing method and device and server
CN113741863A (en) * 2021-07-29 2021-12-03 南方电网深圳数字电网研究院有限公司 Application program generation method based on algorithm model, electronic device and storage medium
CN115543089A (en) * 2022-10-20 2022-12-30 昆明奥智科技有限公司 Virtual human emotion interaction system and method based on five-dimensional emotion model

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108960023A (en) * 2017-09-19 2018-12-07 炬大科技有限公司 A kind of portable Emotion identification device
CN113221821A (en) * 2021-05-28 2021-08-06 中国工商银行股份有限公司 Business data pushing method and device and server
CN113741863A (en) * 2021-07-29 2021-12-03 南方电网深圳数字电网研究院有限公司 Application program generation method based on algorithm model, electronic device and storage medium
CN115543089A (en) * 2022-10-20 2022-12-30 昆明奥智科技有限公司 Virtual human emotion interaction system and method based on five-dimensional emotion model

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN117496580A (en) * 2023-11-23 2024-02-02 好心情健康产业集团有限公司 Facial expression intelligent recognition robot terminal based on multi-person synchronous interaction
CN117496580B (en) * 2023-11-23 2024-05-07 好心情健康产业集团有限公司 Facial expression intelligent recognition robot terminal based on multi-person synchronous interaction

Similar Documents

Publication Publication Date Title
US10878818B2 (en) Methods and apparatus for silent speech interface
US12205595B2 (en) Wearable for suppressing sound other than a wearer's voice
US12204627B2 (en) Using a wearable to interpret facial skin micromovements
US20170187880A1 (en) Coordinating voice calls between representatives and customers to influence an outcome of the call
US12105876B2 (en) System and method for using gestures and expressions for controlling speech applications
CN110262665A (en) Method and apparatus for output information
US12230407B2 (en) Medical intelligence system and method
CN110175526A (en) Dog Emotion identification model training method, device, computer equipment and storage medium
CN113468983A (en) Emotion analysis method, device, equipment and storage medium
US11699043B2 (en) Determination of transcription accuracy
CN118098587A (en) An AI suicide risk analysis method and system based on digital doctor
CN116503841A (en) A mental health intelligent emotion recognition method
Bunton Effects of nasal port area on perception of nasality and measures of nasalance based on computational modeling
US11556720B2 (en) Context information reformation and transfer mechanism at inflection point
CN206892866U (en) Intelligent dialogue device with scenario analysis function
CN115471890A (en) Vehicle interaction method and device, vehicle and storage medium
CN119182748B (en) Instant messaging method, system and related device
CN119539711A (en) A personalized remote trial system
CN118587757A (en) AR-based emotional data processing method, device and electronic device
JP2025051679A (en) system
CN117095573A (en) PBL teaching analysis system and method based on audio and video
JP2025048829A (en) system
JP2025048852A (en) system
JP2025048984A (en) system
JP2025049250A (en) system

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination