WO2021217973A1 - Emotion information recognition method and apparatus, and storage medium and computer device - Google Patents

Emotion information recognition method and apparatus, and storage medium and computer device Download PDF

Info

Publication number
WO2021217973A1
WO2021217973A1 PCT/CN2020/111036 CN2020111036W WO2021217973A1 WO 2021217973 A1 WO2021217973 A1 WO 2021217973A1 CN 2020111036 W CN2020111036 W CN 2020111036W WO 2021217973 A1 WO2021217973 A1 WO 2021217973A1
Authority
WO
WIPO (PCT)
Prior art keywords
posture
emotion
information
matrix
emotion intensity
Prior art date
Application number
PCT/CN2020/111036
Other languages
French (fr)
Chinese (zh)
Inventor
喻凌威
周宸
周宝
陈远旭
Original Assignee
平安科技(深圳)有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 平安科技(深圳)有限公司 filed Critical 平安科技(深圳)有限公司
Publication of WO2021217973A1 publication Critical patent/WO2021217973A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/174Facial expression recognition
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/048Activation functions
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/049Temporal neural networks, e.g. delay elements, oscillating neurons or pulsed inputs
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/40Scenes; Scene-specific elements in video content
    • G06V20/41Higher-level, semantic clustering, classification or understanding of video scenes, e.g. detection, labelling or Markovian modelling of sport events or news items
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/20Movements or behaviour, e.g. gesture recognition

Definitions

  • This application relates to the field of artificial intelligence technology, in particular to an emotional information recognition method, device, storage medium, and computer equipment.
  • the inventor realizes that at present, the traditional emotion information recognition technology is only to infer people’s emotions through facial expressions.
  • this emotion information recognition method ignores the emotional intensity of the body language that people naturally make in social situations.
  • the role played has led to low accuracy of emotion information recognition and low efficiency of emotion information recognition.
  • the present application provides an emotional information recognition method, device, storage medium, and computer equipment, the main purpose of which is to improve the accuracy and efficiency of emotional information recognition.
  • a method for identifying emotional information including:
  • the corresponding emotion type is retrieved and fed back.
  • an emotional information recognition device including:
  • a receiving unit configured to receive an emotional information recognition request, where the emotional information recognition request carries human posture information
  • a conversion unit configured to convert the human body posture information into a posture matrix containing posture feature points by using a preset posture conversion algorithm
  • a processing unit configured to process the posture matrix according to a preset emotion intensity algorithm to obtain emotion intensity data
  • the feedback unit is used to retrieve and feed back the corresponding emotion type according to the emotion intensity data.
  • a storage medium in which at least one executable instruction is stored, and the executable instruction causes a processor to perform the following steps: receiving an emotional information recognition request, and the emotional information recognizing The request carries human posture information; the human posture information is converted into a posture matrix containing posture feature points using a preset posture conversion algorithm; the posture matrix is processed according to a preset emotion intensity algorithm to obtain emotion intensity data ; According to the emotional intensity data, retrieve and feed back the corresponding emotional type.
  • a computer device which includes a processor, a memory, a communication interface, and a communication bus.
  • the processor, the memory, and the communication interface communicate with each other through the communication bus, and
  • the memory is used to store at least one executable instruction, the executable instruction causes the processor to perform the following steps: receiving an emotional information recognition request, the emotional information recognition request carries human posture information; using a preset posture conversion
  • the algorithm converts the human body posture information into a posture matrix containing posture feature points; processes the posture matrix according to a preset emotion intensity algorithm to obtain emotion intensity data; according to the emotion intensity data, retrieves and feeds back the corresponding emotion type.
  • This application can improve the accuracy and efficiency of emotional information recognition.
  • Fig. 1 shows a flow chart of an emotional information recognition method provided by an embodiment of the present application
  • FIG. 2 shows a schematic diagram of human body feature points based on Euler angles according to an embodiment of the present application
  • FIG. 3 shows a schematic structural diagram of an emotional information recognition device provided by an embodiment of the present application
  • Fig. 4 shows a schematic diagram of the physical structure of a computer device provided by an embodiment of the present application.
  • the technical solution of this application can be applied to the fields of artificial intelligence, block chain and/or big data technology, and the data involved can be stored in a database or can be stored through a block chain, which is not limited by this application.
  • the traditional emotion information recognition technology only stays at inferring human emotions through facial expressions.
  • this emotion information recognition method ignores the emotional intensity of the body language that people naturally make in social situations. As a result, the accuracy of emotion information recognition is not high, and the efficiency of emotion information recognition is relatively low.
  • an embodiment of the present application provides a method for identifying emotional information. As shown in FIG. 1, the method includes:
  • the emotional information recognition request may specifically be sent by the server.
  • images or images with human body posture information can be acquired through a camera placed inside the robot, and the human body posture information can be used for emotion analysis, so that the robot can take different types of emotions to the user according to the obtained emotion types. Measures, such as handling business in advance, or changing windows, etc., to improve business handling efficiency.
  • the posture conversion algorithm may specifically express the rotation degrees of freedom of each joint point of the human body through Euler angles.
  • the obtained human body posture information it can be processed by a preset algorithm to obtain a matrix containing 13 feature points. Since the human body joints have angular rotation motion, the embodiment of this application uses Euler angles to determine the position of each feature point.
  • the rotation degree of freedom is expressed, so that the coordinate position and the rotation degree of freedom of each feature point of the human body in each frame of image can be obtained through the human body posture information.
  • the posture of the human body can be abstracted into a posture matrix represented by 13 feature points, and the matrix can be subsequently analyzed to obtain emotion types corresponding to different posture matrices.
  • the number of feature points corresponding to the embodiment of the present application can be set according to the requirements of the service type. For example, if the accuracy requirement is high, the finger joints can also be abstracted as feature points.
  • the emotion intensity algorithm may specifically include processing the posture matrix through a pre-trained emotion intensity model.
  • the emotion intensity model may specifically be a two-layer LSTM-RNN structure.
  • the prior art only uses facial recognition as a dimension to recognize emotional information, it is easy to cause recognition errors and result in low recognition accuracy.
  • the embodiment of this application is innovative on this basis and adopts a two-layer LSTM-RNN structure, which can Combining the two dimensions of facial recognition and human posture to recognize emotion types greatly improves the accuracy of emotion information recognition.
  • corresponding emotion intensity data can be obtained, and the emotion intensity data can be used to correspondingly search for emotion types, so as to adopt a corresponding processing method.
  • the emotion intensity data retrieve and feed back the corresponding emotion type.
  • the corresponding emotion type is retrieved locally, and the emotion type is used to respond to the emotion information recognition request. For example, if the emotion intensity is 1, the emotion type corresponding to the emotion intensity 1 can be found locally as anger. After the emotion type of anger is fed back, the robot can be controlled to take measures to handle the business in advance for the user.
  • the corresponding relationship between the emotion intensity data and the emotion type can be established in advance, and the emotion intensity data, the emotion type, and the corresponding relationship between the emotion intensity data and the emotion type can be saved At the local level, different measures can be taken to appease. In practical application scenarios, such as when handling business at the front desk of a bank, users often wait and queue. During this period, users may be impatient, confused, and need help in self-service handling. In the process of estimating the emotional strength of the human body posture, it can comfort the user and assist the user in handling business according to the user's expression and action posture.
  • the step 103 may specifically include: using a pre-trained emotion intensity model to process the posture matrix and the obtained emotion vector information to obtain the emotion intensity.
  • the pre-trained emotional intensity model may be an LSTM-RNN structure, which can capture millimeter-level differential streams, so this structure is very suitable for handling complex, variable length, and highly intrinsically related human postures sequence.
  • LSTM-RNN structure a two-layer LSTM-RNN structure can be adopted, and the structure input variable can be a variable-length action series ⁇ X 1 , X 2 , X 3 ,...X n-1 , X n ⁇
  • the variable-length action sequence may specifically be a human body posture image
  • the X n may be any frame of the human body posture image in the human body posture image.
  • the emotion vector corresponding to the facial expression can also be input as another parameter.
  • the emotion vector can be used to represent the emotion type of the actual person, specifically, each human face in the human body posture image
  • An expression can correspond to a different emotion type.
  • the emotion type can be represented by a value between 0 and 1, for example, 1 is used to represent anger, 2 is used to represent surprise, etc., and the human body posture image and the emotion vector
  • it is input into the pre-trained emotional intensity model for processing, and emotional intensity data can be obtained.
  • emotional intensity data can be obtained for example, for the human posture of slightly open arms, if the recognized facial expression is normal, then the emotional strength of calmness can be obtained; but if the recognized facial expression is staring, then the emotional strength of surprise can be obtained.
  • the accuracy of recognition can be improved and correct feedback can be made.
  • a person if a person’s emotions are dissatisfaction, he may frown, and if irritated by improper behavior, he may express strong dissatisfaction by shrugging his shoulders. But if you can detect this emotional fluctuation in time and make adaptive adjustments based on people's reactions, people's dissatisfaction will be alleviated and users will be more accepted.
  • the movement of slightly opening the arms is just a normal posture when a person expresses his thoughts calmly; but if his expression is surprised, it will be much more surprised than the naturally hanging arms. Therefore, in order to enable the robot to obtain a more comprehensive analysis of human body posture actions, the emotion type can be included in the process of emotion intensity estimation.
  • using the pre-trained emotion strength model to process the posture matrix and the obtained emotion vector information to obtain the emotion strength may specifically include: using a sigmoid function to simultaneously perform the input posture matrix and emotion vector information. Process and output the obtained emotional intensity data, which is stored in the blockchain.
  • the posture matrix extracted from the human posture image and the emotion vector information are simultaneously input into the pre-trained emotion intensity model, and the sigmoid function in the model can be called for processing.
  • the sigmoid function can be used to convert any real number Converted to a certain number between 0-1 as the probability. For example, after the sigmoid function processes the posture matrix and emotion vector information, the probability of different emotion intensity data can be obtained, such as anger 93%, less than 5% , Happy 1%, excited 1%, the emotional intensity data with the highest probability can be output.
  • the above-mentioned emotional intensity data may also be stored in a node of a blockchain.
  • the emotional intensity data may also be stored in a node of the blockchain.
  • a blockchain network can be established in advance, and the recording nodes in the blockchain network can be used to record emotional intensity data, the emotional intensity data can be packaged and stored in a new block, and the generated management key Save it in the record node for retrieval and feedback when needed.
  • the emotional intensity data is stored through the blockchain technology, which can greatly ensure the security of the emotional intensity data, and it is easy to retrieve the data, and can improve the efficiency of emotion recognition.
  • the step 102 may specifically include: obtaining Euler angle parameters of each feature point; and determining the posture of each feature point based on the human body static model coordinate system according to the Euler angle parameters matrix.
  • Euler angles indicate that the Euler angles can be used to determine a set of three independent angular parameters of the position of a fixed-point rotating rigid body.
  • the rotational freedom of a joint point can be represented by a set of three angular parameters of rpy.
  • a standardization of the body feature data of different people is realized by inputting a pre-defined human skeleton model.
  • Each feature point in the pre-defined model has a pre-defined coordinate system.
  • the key point extraction technology in OpenPose can be used to determine that each joint is in the human body.
  • the OpenPose human body gesture recognition is an open source library developed by Carnegie Mellon University (CMU) based on convolutional neural networks and supervised learning and using caffe as the framework. It can realize posture estimation of human body movements, facial expressions, and finger movements. It is suitable for single and multiple people, and has excellent robustness.
  • the input is an image
  • the basic model can be VGG19
  • the model output is a matrix that can be represented by the above posture.
  • r, p, y in the matrix can respectively represent the Euler angle parameters of each feature point.
  • the method may further include: performing recognition processing on the acquired facial expression information by using a preset facial recognition algorithm to obtain corresponding emotion vector information.
  • the specific process of the facial recognition algorithm may include: reading facial expression images, using the top of the head as a reference point to estimate the approximate position of facial features, and evenly setting mark points on the contours of each feature part of the face; passing through the center of the brow and the pupils The central axis fitted by the midpoint of the connection line and the three points at the center of the mouth divides the face into two symmetrical parts. Without scaling, translation, or rotation, the image will be adjusted to be symmetrical with respect to the central axis.
  • the facial expression shape model divides the left eye/right eye, left eyebrow/right eyebrow and mouth into different areas, and define these areas as feature candidate areas; For each feature candidate area, the feature vector is extracted by the difference image method. By performing the difference operation between all the image sequences in the image processed in the previous step and the neutral expression image in the database, the average value of the difference value in each feature candidate area is the largest The facial expression feature vector is extracted from the image sequence.
  • the emotion vector data corresponding to the facial expression feature vector is retrieved locally.
  • the emotion vector data may indicate the type of emotion expressed by facial expressions. For example, through the facial expression of frowning, the emotion vector dissatisfaction can be correspondingly obtained.
  • the method may further include: according to the human body posture information, establishing a homogeneous transformation matrix of the human body joint points based on the human body static model coordinate system; The coordinates of the nodes are related, and the joint points are determined as the characteristic points of the posture of the human body.
  • the G coordinate system can be a human body static model coordinate system, and the representation method of the coordinate system can be specifically described as follows: "Skeleton tree", the relative relationship between feature points and feature points in the skeleton tree is statically stored as a predefined model; secondly, since everyone's body structure is the same but the skeleton length is different, homogeneous transformation can be introduced
  • the matrix T represents the rigid transformation of different individuals with respect to the corresponding points on the static model. The position of any point can be obtained by matrix multiplication.
  • the method may further include: training an emotional strength model according to the RNN-LSTM model, sample pose data, and preset emotional pose tags.
  • model training pre-training needs to be performed by making a data set or using a data set on the network. For example, for a certain posture, the actual emotion of the observed person needs to be determined and data labeling is completed.
  • the specific process of the model training may include: RNN-based improved long-term and short-term memory modeling. Since many data in real life have both temporal and spatial features, such as human body motion trajectory, continuous frames of video, etc., human body The same is true for posture, each time point corresponds to a set of data, and the data often has certain spatial characteristics. Therefore, in order to carry out classification and prediction work on such a time series, it is necessary to model and extract features in time and space.
  • the commonly used time modeling tool is the Recurrent Neural Network (RNN) correlation model (LSTM). Because of its unique gate structure design and powerful extraction of time series features, it is widely used in forecasting problems and has achieved good results. .
  • the traditional LSTM structure includes three structures: input gate, output gate, forget gate, and a neural node (cell), where the input can be the body's posture representation in the current frame at time t, The output can be a posture descriptor, which is used to describe the type of current posture.
  • a double-layer LSTM-RNN structure is formed by connecting n LSTM structures horizontally, because a continuous image stream is often required when determining the human body posture, so through ⁇ X 1 , X 2 , X 3 ,. . X n-1 ,X n ⁇ to represent such a video stream and serve as the input of the model.
  • the training of the model requires a data set for pre-training, and then a second training is performed by creating a data set to achieve a relatively robust effect.
  • the current posture image stream of a person is recorded through the camera, the posture description matrix is obtained through the key point extraction method of OpenPose, and the current person's emotion type strength is asked to complete the data labeling, and then the training is carried out.
  • the structure of the LSTM unit is still adopted, but the structure of a double-layer LSTM and a fully connected layer are adopted.
  • the double-layer structure can increase the correlation detection of the timing.
  • This application provides an emotional information recognition method. Compared with the prior art inferring human emotions only through facial expressions, this application receives an emotional information recognition request that carries human posture information; using predictions The posture conversion algorithm is assumed to convert the human posture information into a posture matrix containing posture feature points; the posture matrix is processed according to a preset emotion intensity algorithm to obtain emotion intensity data; according to the emotion intensity data, retrieve and merge Feedback corresponding emotion type. Therefore, the accuracy and efficiency of emotion information recognition can be improved through the dual dimensions of human body posture and facial expression. In addition, this application also uses blockchain technology to store data, which can improve the security of emotional information.
  • an embodiment of the present application provides an emotional information recognition device.
  • the device includes: a receiving unit 21, a conversion unit 22, a processing unit 23 and a feedback unit 24.
  • the receiving unit 21 may be configured to receive an emotion information recognition request, where the emotion information recognition request carries human posture information;
  • the conversion unit 22 may be used to convert the human body posture information into a posture matrix containing posture feature points by using a preset posture conversion algorithm;
  • the processing unit 23 may be configured to process the posture matrix according to a preset emotion intensity algorithm to obtain emotion intensity data;
  • the feedback unit 24 may be used to retrieve and feed back the corresponding emotion type according to the emotion intensity data.
  • processing unit 23 includes:
  • the processing module 231 may be used to process the posture matrix and the obtained emotion vector information using a pre-trained emotion intensity model to obtain the emotion intensity.
  • processing module 231 may be specifically configured to simultaneously process the input posture matrix and emotion vector information by using the sigmoid function, and output the obtained emotion intensity data.
  • processing module 231 may be specifically used to store the emotional intensity data using blockchain technology.
  • the conversion unit 22 includes:
  • the obtaining module 221 may be used to obtain Euler angle parameters of each feature point
  • the determining module 222 may be used to determine the posture matrix of each feature point in the human body static model coordinate system according to the Euler angle parameters.
  • the device further includes:
  • the recognition unit 25 may be used to perform recognition processing on the acquired facial expression information by using a preset facial recognition algorithm to obtain corresponding emotion vector information.
  • the device further includes:
  • the establishing unit 26 may be used to establish, according to the human body posture information, a homogeneous transformation matrix of the human body joint points based on the human body static model coordinate system;
  • the determining unit 27 may be used to determine the coordinates of each joint point by matrix multiplication, and determine the joint point as a feature point of the posture of the human body.
  • the device further includes:
  • the training unit 28 may be used to train the emotional intensity model according to the RNN-LSTM model, sample pose data, and preset emotional pose labels.
  • an embodiment of the present application further provides a storage medium in which at least one executable instruction is stored, and the executable instruction causes the processor to perform the following steps: receiving Emotion information recognition request, said emotion information recognition request carries human body posture information; using a preset posture conversion algorithm to convert the human body posture information into a posture matrix containing posture feature points; according to the preset emotion intensity algorithm The posture matrix is processed to obtain emotion intensity data; according to the emotion intensity data, the corresponding emotion type is retrieved and fed back.
  • the executable instruction can also implement other steps of the method in the foregoing embodiment when executed by the processor, which will not be repeated here.
  • the storage medium involved in the present application may be a computer-readable storage medium, and the storage medium, such as a computer-readable storage medium, may be non-volatile or volatile.
  • an embodiment of the present application also provides a computer device.
  • the processor 31, the communication interface 32, and the memory 33 communicate with each other through the communication bus 34.
  • the communication interface 34 is used to communicate with other devices, such as network elements such as user terminals or other servers.
  • the processor 31 is configured to execute a program, and specifically can execute relevant steps in the foregoing embodiment of the emotion information recognition method.
  • the program may include program code, and the program code includes computer operation instructions.
  • the processor 31 may be a central processing unit CPU, or an Application Specific Integrated Circuit (ASIC), or one or more integrated circuits configured to implement the embodiments of the present application.
  • ASIC Application Specific Integrated Circuit
  • the one or more processors included in the terminal may be processors of the same type, such as one or more CPUs; or processors of different types, such as one or more CPUs and one or more ASICs.
  • the memory 33 is used to store programs.
  • the memory 33 may include a high-speed RAM memory, and may also include a non-volatile memory (non-volatile memory), for example, at least one disk memory.
  • the program can specifically be used to make the processor 31 perform the following operations: receive an emotion information recognition request, the emotion information recognition request carries human posture information; use a preset posture conversion algorithm to convert the human posture information to include posture features Point posture matrix; process the posture matrix according to a preset emotion intensity algorithm to obtain emotion intensity data; retrieve and feed back the corresponding emotion type according to the emotion intensity data.
  • the emotion information recognition request carries human body posture information; use a preset posture conversion algorithm to convert the human body posture information into a posture matrix containing posture feature points;
  • the posture matrix is processed according to a preset emotion intensity algorithm to obtain emotion intensity data; according to the emotion intensity data, the corresponding emotion type is retrieved and fed back. Therefore, the accuracy and efficiency of emotion information recognition can be improved through the dual dimensions of human body posture and facial expression.
  • this application also uses blockchain technology to store data, which can improve the security of emotional information.
  • the blockchain referred to in this application is a new application mode of computer technology such as distributed data storage, point-to-point transmission, consensus mechanism, and encryption algorithm.
  • Blockchain essentially a decentralized database, is a series of data blocks associated with cryptographic methods. Each data block contains a batch of network transaction information for verification. The validity of the information (anti-counterfeiting) and the generation of the next block.
  • the blockchain can include the underlying platform of the blockchain, the platform product service layer, and the application service layer.
  • modules or units or components in the embodiments can be combined into one module or unit or component, and in addition, they can be divided into multiple sub-modules or sub-units or sub-components. Except that at least some of such features and/or processes or units are mutually exclusive, any combination can be used to compare all the features disclosed in this specification (including the accompanying claims, abstract and drawings) and any method or methods disclosed in this manner or All the processes or units of the equipment are combined. Unless expressly stated otherwise, each feature disclosed in this specification (including the accompanying claims, abstract and drawings) may be replaced by an alternative feature providing the same, equivalent or similar purpose.
  • the various component embodiments of the present application may be implemented by hardware, or by software modules running on one or more processors, or by a combination of them.
  • a microprocessor or a digital signal processor (DSP) may be used in practice to implement some or all of the functions of some or all of the components in the embodiments of the present application.
  • This application can also be implemented as a device or device program (for example, a computer program and a computer program product) for executing part or all of the methods described herein.
  • Such a program for implementing the present application may be stored on a computer-readable medium, or may have the form of one or more signals.
  • Such a signal can be downloaded from an Internet website, or provided on a carrier signal, or provided in any other form.

Abstract

Disclosed are an emotion information recognition method and apparatus, and a storage medium and a computer device, which relate to the technical field of artificial intelligence, and aim to respectively process a human body posture and facial expression, which have been obtained, into a posture matrix and emotion vector information, and to simultaneously process, by means of a pre-trained emotion intensity recognition model, the posture matrix and the vector information to obtain emotion intensity data and correspondingly recognize an emotion type, thereby improving the accuracy and efficiency of emotion information recognition. The method comprises: receiving an emotion information recognition request, wherein the emotion information recognition request carries human body posture information; transforming the human body posture information into a posture matrix including posture feature points by using a preset posture transformation algorithm; processing the posture matrix according to a preset emotion intensity algorithm, so as to obtain emotion intensity data; and retrieving and feeding back a corresponding emotion type according to the emotion intensity data. In addition, the present application further relates to blockchain technology, and the emotion intensity data may be stored in a blockchain.

Description

情感信息识别方法、装置、存储介质及计算机设备Emotional information recognition method, device, storage medium and computer equipment
本申请要求于2020年4月28日提交中国专利局、申请号为202010349534.0,发明名称为“情感信息识别方法、装置、存储介质及计算机设备”的中国专利申请的优先权,其全部内容通过引用结合在本申请中。This application claims the priority of a Chinese patent application filed with the Chinese Patent Office on April 28, 2020, the application number is 202010349534.0, and the invention title is "emotional information recognition method, device, storage medium and computer equipment", the entire content of which is incorporated by reference Incorporated in this application.
技术领域Technical field
本申请涉及人工智能技术领域,特别是涉及一种情感信息识别方法、装置、存储介质及计算机设备。This application relates to the field of artificial intelligence technology, in particular to an emotional information recognition method, device, storage medium, and computer equipment.
背景技术Background technique
随着大数据的发展,让机器人拥有社交和服务的能力并在人机交互的过程中具备实时读取人情感强度和波动的能力越来越成为人们的愿望和需求。在实际业务办理过程中,如果机器人能及时察觉到人的情感波动并能根据人的反应做适应性调整,人的不满就会缓解,机器人的表现也会得到用户更高的接受度。With the development of big data, allowing robots to have social and service capabilities and the ability to read human emotional intensity and fluctuations in real time during human-computer interaction has increasingly become people's desires and needs. In the actual business processing process, if the robot can detect human emotion fluctuations in time and can make adaptive adjustments according to human response, human dissatisfaction will be alleviated, and the robot's performance will be more accepted by users.
发明人意识到,目前,传统的情感信息识别技术还仅仅停留在通过面部表情推断人的情绪,然而,这种情感信息识别方法忽略了人在社交场合中自然而然做出的肢体语言在情感强度中所扮演的作用,导致情感信息识别的准确率不高,且情感信息识别的效率比较低。The inventor realizes that at present, the traditional emotion information recognition technology is only to infer people’s emotions through facial expressions. However, this emotion information recognition method ignores the emotional intensity of the body language that people naturally make in social situations. The role played has led to low accuracy of emotion information recognition and low efficiency of emotion information recognition.
发明内容Summary of the invention
有鉴于此,本申请提供一种情感信息识别方法、装置、存储介质及计算机设备,主要目的在于提高情感信息识别的准确率和效率。In view of this, the present application provides an emotional information recognition method, device, storage medium, and computer equipment, the main purpose of which is to improve the accuracy and efficiency of emotional information recognition.
依据本申请一个方面,提供了一种情感信息识别方法,包括:According to one aspect of this application, a method for identifying emotional information is provided, including:
接收情感信息识别请求,所述情感信息识别请求中携带有人体姿态信息;Receiving an emotional information recognition request, where the emotional information recognition request carries human posture information;
利用预设的姿态转换算法将所述人体姿态信息转换为包含姿态特征点的姿态矩阵;Using a preset posture conversion algorithm to convert the human posture information into a posture matrix containing posture feature points;
根据预设的情感强度算法对所述姿态矩阵进行处理,得到情感强度数据;Processing the posture matrix according to a preset emotion intensity algorithm to obtain emotion intensity data;
根据所述情感强度数据,检索并反馈对应的情感类型。According to the emotion intensity data, the corresponding emotion type is retrieved and fed back.
依据本申请第二方面,提供了一种情感信息识别装置,包括:According to the second aspect of the present application, there is provided an emotional information recognition device, including:
接收单元,用于接收情感信息识别请求,所述情感信息识别请求中携带有人体姿态信息;A receiving unit, configured to receive an emotional information recognition request, where the emotional information recognition request carries human posture information;
转换单元,用于利用预设姿态转换算法将所述人体姿态信息转换为包含姿态特征点的姿态矩阵;A conversion unit, configured to convert the human body posture information into a posture matrix containing posture feature points by using a preset posture conversion algorithm;
处理单元,用于根据预设的情感强度算法对所述姿态矩阵进行处理,得到情感强度数据;A processing unit, configured to process the posture matrix according to a preset emotion intensity algorithm to obtain emotion intensity data;
反馈单元,用于根据所述情感强度数据,检索并反馈对应的情感类型。The feedback unit is used to retrieve and feed back the corresponding emotion type according to the emotion intensity data.
依据本申请第三方面,提供了一种存储介质,所述存储介质中存储有至少一可执行指令,所述可执行指令使处理器执行以下步骤:接收情感信息识别请求,所述情感信息识别请求中携带有人体姿态信息;利用预设的姿态转换算法将所述人体姿态信息转换为包含姿态特征点的姿态矩阵;根据预设的情感强度算法对所述姿态矩阵进行处理,得到情感强度数据;根据所述情感强度数据,检索并反馈对应的情感类型。According to the third aspect of the present application, there is provided a storage medium in which at least one executable instruction is stored, and the executable instruction causes a processor to perform the following steps: receiving an emotional information recognition request, and the emotional information recognizing The request carries human posture information; the human posture information is converted into a posture matrix containing posture feature points using a preset posture conversion algorithm; the posture matrix is processed according to a preset emotion intensity algorithm to obtain emotion intensity data ; According to the emotional intensity data, retrieve and feed back the corresponding emotional type.
依据本申请第四方面,提供了一种计算机设备,包括处理器、存储器、通信接口和通信总线所述处理器、所述存储器和所述通信接口通过所述通信总线完成相互间的通信,所述存储器用于存放至少一可执行指令,所述可执行指令使所述处理器执行以下步骤:接收情感信息识别请求,所述情感信息识别请求中携带有人体姿态信息;利用预设的姿态转换算法将所述人体姿态信息转换为包含姿态特征点的姿态矩阵;根据预设的情感强度算法对所述姿态矩阵进行处理,得到情感强度数据;根据所述情感强度数据,检索并反馈对应的情感类型。According to the fourth aspect of the present application, a computer device is provided, which includes a processor, a memory, a communication interface, and a communication bus. The processor, the memory, and the communication interface communicate with each other through the communication bus, and The memory is used to store at least one executable instruction, the executable instruction causes the processor to perform the following steps: receiving an emotional information recognition request, the emotional information recognition request carries human posture information; using a preset posture conversion The algorithm converts the human body posture information into a posture matrix containing posture feature points; processes the posture matrix according to a preset emotion intensity algorithm to obtain emotion intensity data; according to the emotion intensity data, retrieves and feeds back the corresponding emotion type.
本申请能够提高情感信息识别的准确率和效率。This application can improve the accuracy and efficiency of emotional information recognition.
附图说明Description of the drawings
通过阅读下文优选实施方式的详细描述,各种其他的优点和益处对于本领域普通技术人员将变得清楚明了。附图仅用于示出优选实施方式的目的,而并不认为是对本申请的限制。而且在整个附图中,用相同的参考符号表示相同的部件。在附图中:By reading the detailed description of the preferred embodiments below, various other advantages and benefits will become clear to those of ordinary skill in the art. The drawings are only used for the purpose of illustrating the preferred embodiments, and are not considered as a limitation to the application. Also, throughout the drawings, the same reference symbols are used to denote the same components. In the attached picture:
图1示出了本申请实施例提供的一种情感信息识别方法流程图;Fig. 1 shows a flow chart of an emotional information recognition method provided by an embodiment of the present application;
图2示出了本申请实施例提供的一种基于欧拉角的人体特征点示意图;FIG. 2 shows a schematic diagram of human body feature points based on Euler angles according to an embodiment of the present application;
图3示出了本申请实施例提供的一种情感信息识别装置的结构示意图;FIG. 3 shows a schematic structural diagram of an emotional information recognition device provided by an embodiment of the present application;
图4示出了本申请实施例提供的一种计算机设备的实体结构示意图。Fig. 4 shows a schematic diagram of the physical structure of a computer device provided by an embodiment of the present application.
具体实施方式Detailed ways
下面将参照附图更详细地描述本公开的示例性实施例。虽然附图中显示了本公开的示例性实施例,然而应当理解,可以以各种形式实现本公开而不应被这里阐述的实施例所限制。相反,提供这些实施例是为了能够更透彻地理解本公开,并且能够将本公开的范围完整的传达给本领域的技术人员。Hereinafter, exemplary embodiments of the present disclosure will be described in more detail with reference to the accompanying drawings. Although the drawings show exemplary embodiments of the present disclosure, it should be understood that the present disclosure can be implemented in various forms and should not be limited by the embodiments set forth herein. On the contrary, these embodiments are provided to enable a more thorough understanding of the present disclosure and to fully convey the scope of the present disclosure to those skilled in the art.
本申请的技术方案可应用于人工智能、区块链和/或大数据技术领域,涉及的数据可存储于数据库中,或者可以通过区块链存储,本申请不做限定。The technical solution of this application can be applied to the fields of artificial intelligence, block chain and/or big data technology, and the data involved can be stored in a database or can be stored through a block chain, which is not limited by this application.
如背景技术所述,目前,传统的情感信息识别技术还仅仅停留在通过面部表情推断人的情绪,然而,这种情感信息识别方法忽略了人在社交场合中自然而然做出的肢体语言在情感强度中所扮演的作用,导致情感信息识别的准确率不高,且情感信息识别的效率比较低。As mentioned in the background art, at present, the traditional emotion information recognition technology only stays at inferring human emotions through facial expressions. However, this emotion information recognition method ignores the emotional intensity of the body language that people naturally make in social situations. As a result, the accuracy of emotion information recognition is not high, and the efficiency of emotion information recognition is relatively low.
为了解决上述问题,本申请实施例提供了一种情感信息识别方法,如图1所示,所述方法包括:In order to solve the foregoing problem, an embodiment of the present application provides a method for identifying emotional information. As shown in FIG. 1, the method includes:
101、接收情感信息识别请求,所述情感信息识别请求中携带有人体姿态信息。101. Receive an emotional information recognition request, where the emotional information recognition request carries human posture information.
其中,所述情感信息识别请求具体可以为服务器发送的。在实际应用场景中,可以通过安置于机器人内部的摄像头获取带有人体姿态信息的影像或图像,所述人体姿态信息可以用于进行情感分析,从而使机器人根据得到的情感类型对用户采取不同的措施,例如业务提前办理,或者更换窗口等,以提高业务办理效率。Wherein, the emotional information recognition request may specifically be sent by the server. In actual application scenarios, images or images with human body posture information can be acquired through a camera placed inside the robot, and the human body posture information can be used for emotion analysis, so that the robot can take different types of emotions to the user according to the obtained emotion types. Measures, such as handling business in advance, or changing windows, etc., to improve business handling efficiency.
102、利用预设的姿态转换算法将所述人体姿态信息转换为包含姿态特征点的姿态矩阵。102. Use a preset posture conversion algorithm to convert the human posture information into a posture matrix containing posture feature points.
其中,所述姿态转换算法具体可以为通过欧拉角表示人体各个关节点的旋转自由度。对于获取的人体姿态信息,可以通过预设的算法进行处理,得到包含13个特征点的矩阵,由于人体关节存在角度旋转运动,因此,本申请实施例采用欧拉角对每个特征点处的旋转自由度进行表示,从而可以通过所述人体姿态信息,得到每一帧图像中人体各特征点处的坐标位置以及旋转自由度。通过步骤102可以将人体姿态抽象成一个由13个特征点表示的姿态矩阵,后续可以对所述矩阵进行分析,从而得到不同姿态矩阵对应的情感类型。Wherein, the posture conversion algorithm may specifically express the rotation degrees of freedom of each joint point of the human body through Euler angles. For the obtained human body posture information, it can be processed by a preset algorithm to obtain a matrix containing 13 feature points. Since the human body joints have angular rotation motion, the embodiment of this application uses Euler angles to determine the position of each feature point. The rotation degree of freedom is expressed, so that the coordinate position and the rotation degree of freedom of each feature point of the human body in each frame of image can be obtained through the human body posture information. Through step 102, the posture of the human body can be abstracted into a posture matrix represented by 13 feature points, and the matrix can be subsequently analyzed to obtain emotion types corresponding to different posture matrices.
需要说明的是,本申请实施例所对应的特征点数量可以根据业务类型的需求进行设置,例如,若精度要求较高,可以将手指关节也抽象为特征点。It should be noted that the number of feature points corresponding to the embodiment of the present application can be set according to the requirements of the service type. For example, if the accuracy requirement is high, the finger joints can also be abstracted as feature points.
103、根据预设的情感强度算法对所述姿态矩阵进行处理,得到情感强度数据。103. Process the posture matrix according to a preset emotion intensity algorithm to obtain emotion intensity data.
其中,所述情感强度算法具体可以包括通过预先训练的情感强度模型对所述姿态矩阵进行处理,对于本申请实施例,所述情感强度模型具体可以为双层LSTM-RNN结构。由于现有技术中仅通过面部识别这一个维度进行情感信息识别,容易造成识别误差,导致识别准确率低,而本申请实施例在此基础上进行创新,采用双层LSTM-RNN结构,从而可以将面部识别与人体姿态这两个维度结合起来进行情感类型的识别,极大的提高了情感信息识别的准确性。具体地,根据预设的情感强度算法,对所述姿态矩阵进行处理,可以得到对应的情感强度数据,所述情感强度数据可以用于对应查找情感类型,从而采取对应的处理 方法。Wherein, the emotion intensity algorithm may specifically include processing the posture matrix through a pre-trained emotion intensity model. For the embodiment of the present application, the emotion intensity model may specifically be a two-layer LSTM-RNN structure. As the prior art only uses facial recognition as a dimension to recognize emotional information, it is easy to cause recognition errors and result in low recognition accuracy. However, the embodiment of this application is innovative on this basis and adopts a two-layer LSTM-RNN structure, which can Combining the two dimensions of facial recognition and human posture to recognize emotion types greatly improves the accuracy of emotion information recognition. Specifically, by processing the posture matrix according to a preset emotion intensity algorithm, corresponding emotion intensity data can be obtained, and the emotion intensity data can be used to correspondingly search for emotion types, so as to adopt a corresponding processing method.
104、根据所述情感强度数据,检索并反馈对应的情感类型。104. According to the emotion intensity data, retrieve and feed back the corresponding emotion type.
其中,得到所述情感强度数据后,在本地检索对应的情感类型,并利用所述情感类型响应所述情感信息识别请求。例如,得到情感强度为1,则可以在本地查找情感强度1对应的情感类型为愤怒,反馈愤怒这一情感类型后,可以控制机器人对用户采取提前办理业务的措施。对于本申请实施例,可以预先建立所述情感强度数据与所述情感类型的对应关系,并将所述情感强度数据、所述情感类型以及所述情感强度数据和所述情感类型的对应关系保存在本地,从而可以采取不同的措施进行安抚。在实际应用场景中,如在银行前台办理业务时,经常会出现用户等待和排队的现象,在这期间,用户可能会出现等待不耐烦、排队无所适从、以及自助办理业务时需要帮助的情况,通过人体姿态估计情感强度的过程中可以根据用户的表情和动作姿态进行安抚用户、辅助用户办理业务。Wherein, after the emotion intensity data is obtained, the corresponding emotion type is retrieved locally, and the emotion type is used to respond to the emotion information recognition request. For example, if the emotion intensity is 1, the emotion type corresponding to the emotion intensity 1 can be found locally as anger. After the emotion type of anger is fed back, the robot can be controlled to take measures to handle the business in advance for the user. For the embodiment of the present application, the corresponding relationship between the emotion intensity data and the emotion type can be established in advance, and the emotion intensity data, the emotion type, and the corresponding relationship between the emotion intensity data and the emotion type can be saved At the local level, different measures can be taken to appease. In practical application scenarios, such as when handling business at the front desk of a bank, users often wait and queue. During this period, users may be impatient, confused, and need help in self-service handling. In the process of estimating the emotional strength of the human body posture, it can comfort the user and assist the user in handling business according to the user's expression and action posture.
进一步的,为了更好的说明上述情感信息识别方法的过程,作为对上述实施例的细化和扩展,本申请实施例提供了几种可选实施例,但不限于此,具体如下所示:Further, in order to better explain the process of the foregoing emotion information recognition method, as a refinement and extension of the foregoing embodiment, the embodiments of the present application provide several optional embodiments, but are not limited thereto, and the details are as follows:
在本申请的一个可选实施例,所述步骤103具体可以包括:利用预先训练的情感强度模型对所述姿态矩阵以及获取的情感矢量信息进行处理,得到情感强度。In an optional embodiment of the present application, the step 103 may specifically include: using a pre-trained emotion intensity model to process the posture matrix and the obtained emotion vector information to obtain the emotion intensity.
其中,所述预先训练的情感强度模型可以为LSTM-RNN结构,所述LSTM-RNN结构可以捕捉毫米级的差分流,因此该结构非常适合处理复杂、变长度以及内在关联程度较高的人体姿态序列。对于本申请实施例,可以采用双层LSTM-RNN结构,所述结构输入变量可以为变长度的动作系列{X 1,X 2,X 3,......X n-1,X n},所述变长度的动作序列具体可以为人体姿态影像,所述X n可以为所述人体姿态图像中任意一帧人体姿态图像。另外,对于本申请实施例,还可以将面部表情对应的情感矢量作为另一个参数进行输入,所述情感矢量可以用于表示实际人的情感类型,具体地,人体姿态影像中的每个人体面部表情可以对应一个不同的情感类型,所述情感类型可以用一个0~1之间的值来表示,例如,用1来表示愤怒,2表示惊讶等,将所述人体姿态影像和所述情感矢量同时输入到预先训练的情感强度模型中进行处理,可以得到情感强度数据。例如,对于微微张开双臂这一人体姿态,如果识别的面部表情是正常的,那么可以得到冷静这一情感强度;但是如果识别的面部表情是瞪眼张口的,那么可以得到惊讶这一情感强度。也就是说,通过将情感类型与人体姿态相结合,可以提高识别的准确性,从而做出正确的反馈。例如,若人的情感为不满,他可能会皱眉,如果被不当行为激怒,他也许会以耸肩的方式来表达强烈的不满。但如果能及时察觉到这种情感波动并能根据人的反应做适应性调整,人的不满就会缓解,也会得到用户更高的接受度。再如,微微张开双臂这个动作在一个人冷静表达想法的时候只是一个正常的姿势;但如果他的表情是惊讶的,这会比双臂自然垂下的惊讶程度要高得多。因此,为了能使机器人通过人的身体姿态动作得到更全面的分析,可以在情感强度估计的过程中将情感类型也包括在内。 Wherein, the pre-trained emotional intensity model may be an LSTM-RNN structure, which can capture millimeter-level differential streams, so this structure is very suitable for handling complex, variable length, and highly intrinsically related human postures sequence. For the embodiment of the present application, a two-layer LSTM-RNN structure can be adopted, and the structure input variable can be a variable-length action series {X 1 , X 2 , X 3 ,...X n-1 , X n }, the variable-length action sequence may specifically be a human body posture image, and the X n may be any frame of the human body posture image in the human body posture image. In addition, for the embodiment of the present application, the emotion vector corresponding to the facial expression can also be input as another parameter. The emotion vector can be used to represent the emotion type of the actual person, specifically, each human face in the human body posture image An expression can correspond to a different emotion type. The emotion type can be represented by a value between 0 and 1, for example, 1 is used to represent anger, 2 is used to represent surprise, etc., and the human body posture image and the emotion vector At the same time, it is input into the pre-trained emotional intensity model for processing, and emotional intensity data can be obtained. For example, for the human posture of slightly open arms, if the recognized facial expression is normal, then the emotional strength of calmness can be obtained; but if the recognized facial expression is staring, then the emotional strength of surprise can be obtained. That is to say, by combining the emotion type with the human body posture, the accuracy of recognition can be improved and correct feedback can be made. For example, if a person’s emotions are dissatisfaction, he may frown, and if irritated by improper behavior, he may express strong dissatisfaction by shrugging his shoulders. But if you can detect this emotional fluctuation in time and make adaptive adjustments based on people's reactions, people's dissatisfaction will be alleviated and users will be more accepted. For another example, the movement of slightly opening the arms is just a normal posture when a person expresses his thoughts calmly; but if his expression is surprised, it will be much more surprised than the naturally hanging arms. Therefore, in order to enable the robot to obtain a more comprehensive analysis of human body posture actions, the emotion type can be included in the process of emotion intensity estimation.
对于本申请实施例,所述利用预先训练的情感强度模型对所述姿态矩阵以及获取的情感矢量信息进行处理,得到情感强度具体还可以包括:利用sigmoid函数对输入的姿态矩阵以及情感矢量信息同时进行处理,并输出得到的情感强度数据,该情感强度数据存储于区块链中。For the embodiment of the present application, using the pre-trained emotion strength model to process the posture matrix and the obtained emotion vector information to obtain the emotion strength may specifically include: using a sigmoid function to simultaneously perform the input posture matrix and emotion vector information. Process and output the obtained emotional intensity data, which is stored in the blockchain.
具体地,将所述人体姿态影像提取的姿态矩阵和所述情感矢量信息同时输入到预先训练的情感强度模型中,可以调用模型中的sigmoid函数进行处理,所述sigmoid函数可以用于把任意实数转换为0-1之间的某个数作为概率,例如,所述sigmoid函数对所述姿态矩阵和情感矢量信息进行处理后,可以得到不同情感强度数据的概率,如愤怒93%,不满5%,开心1%,兴奋1%,则可以将所述概率最高的情感强度数据进行输出。Specifically, the posture matrix extracted from the human posture image and the emotion vector information are simultaneously input into the pre-trained emotion intensity model, and the sigmoid function in the model can be called for processing. The sigmoid function can be used to convert any real number Converted to a certain number between 0-1 as the probability. For example, after the sigmoid function processes the posture matrix and emotion vector information, the probability of different emotion intensity data can be obtained, such as anger 93%, less than 5% , Happy 1%, excited 1%, the emotional intensity data with the highest probability can be output.
需要强调的是,为进一步保证上述情感强度数据的私密和安全性,上述情感强度数据还可以存储于一区块链的节点中。It should be emphasized that, in order to further ensure the privacy and security of the above-mentioned emotional intensity data, the above-mentioned emotional intensity data may also be stored in a node of a blockchain.
其中,为了保证所述情感强度数据的私密和安全性,所述情感强度数据还可以存储于区块链的节点中。具体地,可以预先建立区块链网络,并利用所述区块链网络中的记录节点记录情感强度数据,将所述情感强度数据打包存储在新的区块中,并将生成的管理密钥保存在记录节点中,以便需要时进行调取和反馈。本申请实施例通过区块链技术存储情感强度数据,能够极大的保证情感强度数据的安全性,并且易于调取数据,能够提高情感识别的效率。Wherein, in order to ensure the privacy and security of the emotional intensity data, the emotional intensity data may also be stored in a node of the blockchain. Specifically, a blockchain network can be established in advance, and the recording nodes in the blockchain network can be used to record emotional intensity data, the emotional intensity data can be packaged and stored in a new block, and the generated management key Save it in the record node for retrieval and feedback when needed. In the embodiment of the present application, the emotional intensity data is stored through the blockchain technology, which can greatly ensure the security of the emotional intensity data, and it is easy to retrieve the data, and can improve the efficiency of emotion recognition.
在本申请的另一个可选实施例,所述步骤102具体可以包括:获取各特征点的欧拉角参数;根据所述欧拉角参数确定基于人体静态模型坐标系下每个特征点的姿态矩阵。In another optional embodiment of the present application, the step 102 may specifically include: obtaining Euler angle parameters of each feature point; and determining the posture of each feature point based on the human body static model coordinate system according to the Euler angle parameters matrix.
其中,由于人体骨架长短的不同,导致不同人体对应的姿态表示都不一样,因此为了简化数据的复杂度,将骨架长度这项特征剔除,只保留关节点的旋转自由度这一特征,可以通过欧拉角表示,所述欧拉角可以用来确定定点转动刚体位置的3个一组独立角参量,例如,一个关节点的旋转自由度可以通过r-p-y三个角参量为一组进行表示。Among them, due to the difference in the length of the human body skeleton, the corresponding posture representations of different human bodies are different. Therefore, in order to simplify the complexity of the data, the feature of the skeleton length is eliminated, and only the rotation freedom of the joint points is retained. Euler angles indicate that the Euler angles can be used to determine a set of three independent angular parameters of the position of a fixed-point rotating rigid body. For example, the rotational freedom of a joint point can be represented by a set of three angular parameters of rpy.
通过输入一个预定义的人体骨架模型实现不同人身体特征数据的一个标准化,预定义模型中每个特征点有一个预定义的坐标系,可以利用OpenPose中的关键点提取技术确定出各个关节在人体上的位置,所述OpenPose人体姿态识别是美国卡耐基梅隆大学(CMU)基于卷积神经网络和监督学习并以caffe为框架开发的开源库。可以实现人体动作、面部表情、手指运动等姿态估计。适用于单人和多人,具有极好的鲁棒性。具体地,输入为图像,基础模型可以为VGG19,模型输出为可以上述姿态表示的矩阵,对于本申请实施例,仅需要提取出关键点的信息作为模型输入即可送进双层LSTM-RNN的网络中。同时,OpenCV也提供了对于openpose开源框架的调用接口,也可以通过这种方式计算关键点信息。接下来对提取出的关键点附近的点云进行提取,估计出关键点相对于预定义模型相同关键点的朝向,通过各个关节相对于预定义的关节点的变换关系,并用欧拉角进行表示。通过这种方式,在i时刻(帧)人的姿态即可用以下矩阵X i进行表示: A standardization of the body feature data of different people is realized by inputting a pre-defined human skeleton model. Each feature point in the pre-defined model has a pre-defined coordinate system. The key point extraction technology in OpenPose can be used to determine that each joint is in the human body. At the position above, the OpenPose human body gesture recognition is an open source library developed by Carnegie Mellon University (CMU) based on convolutional neural networks and supervised learning and using caffe as the framework. It can realize posture estimation of human body movements, facial expressions, and finger movements. It is suitable for single and multiple people, and has excellent robustness. Specifically, the input is an image, the basic model can be VGG19, and the model output is a matrix that can be represented by the above posture. For the embodiment of this application, only the key point information needs to be extracted as the model input to send it to the two-layer LSTM-RNN Network. At the same time, OpenCV also provides a call interface for the openpose open source framework, and key point information can also be calculated in this way. Next, extract the point cloud near the extracted key points, estimate the orientation of the key points relative to the same key point of the predefined model, and express the transformation relationship of each joint relative to the predefined joint point by Euler angles. . In this way, the posture of the person at time i (frame) can be represented by the following matrix X i :
Figure PCTCN2020111036-appb-000001
Figure PCTCN2020111036-appb-000001
其中,矩阵中r、p、y分别可以表示每个特征点的欧拉角参数。Among them, r, p, y in the matrix can respectively represent the Euler angle parameters of each feature point.
在本申请的又一个可选实施例,所述方法还可以包括:利用预设的面部识别算法对获取的面部表情信息进行识别处理,得到对应的情感矢量信息。In yet another optional embodiment of the present application, the method may further include: performing recognition processing on the acquired facial expression information by using a preset facial recognition algorithm to obtain corresponding emotion vector information.
其中,所述面部识别算法的具体过程可以包括:读取面部表情图像,以头顶为基准点估计面部特征的大概位置,在面部的各特征部位轮廓上均匀的设置标记点;通过眉心、两瞳孔连线的中点和口中央三点拟合出的中轴线将人脸分为左右对称的两部分,在不缩放、不平移、不旋转的条件下,调整图像将,将相对于中轴线对称的标记点调整到同一水平线,并建立面部表情形状模型;在面部表情形状模型中按照左眼/右眼,左眉毛/右眉毛和嘴划分为不同区域,并且将这些区域定义为特征候选区域;针对每一个特征候选区域,采用差分图像法提取特征向量,通过将上一步处理后图像中的所有图像序列与数据库中的中性表情的图像进行差分运算,从各特征候选区域内差分值均值最大的图像序列中提取面部表情特征向量。Wherein, the specific process of the facial recognition algorithm may include: reading facial expression images, using the top of the head as a reference point to estimate the approximate position of facial features, and evenly setting mark points on the contours of each feature part of the face; passing through the center of the brow and the pupils The central axis fitted by the midpoint of the connection line and the three points at the center of the mouth divides the face into two symmetrical parts. Without scaling, translation, or rotation, the image will be adjusted to be symmetrical with respect to the central axis. Adjust the mark points of the to the same horizontal line, and establish a facial expression shape model; in the facial expression shape model, divide the left eye/right eye, left eyebrow/right eyebrow and mouth into different areas, and define these areas as feature candidate areas; For each feature candidate area, the feature vector is extracted by the difference image method. By performing the difference operation between all the image sequences in the image processed in the previous step and the neutral expression image in the database, the average value of the difference value in each feature candidate area is the largest The facial expression feature vector is extracted from the image sequence.
在得到所述面部表情特征向量后,在本地检索所述面部表情特征向量对应的情感矢量数据。所述情感矢量数据可以表示面部表情表达的情感类型,例如,通过皱眉这一面部表情,可以对应得到情感矢量不满。After the facial expression feature vector is obtained, the emotion vector data corresponding to the facial expression feature vector is retrieved locally. The emotion vector data may indicate the type of emotion expressed by facial expressions. For example, through the facial expression of frowning, the emotion vector dissatisfaction can be correspondingly obtained.
在本申请的再一个可选实施例,所述方法还可以包括:根据所述人体姿态信息,建立 人体关节点基于人体静态模型坐标系下的齐次变换矩阵;通过矩阵相乘的方式确定各关节点坐标,并将所述关节点确定为人体姿态的特征点。In still another optional embodiment of the present application, the method may further include: according to the human body posture information, establishing a homogeneous transformation matrix of the human body joint points based on the human body static model coordinate system; The coordinates of the nodes are related, and the joint points are determined as the characteristic points of the posture of the human body.
其中,如图2所示,所述G坐标系可以为人体静态模型坐标系,所述坐标系的表示方法可以具体阐述为:首先,在获取的人的姿态影像提取一个包含13个特征点的“骨架树”,所述骨架树中特征点与特征点之间的相对关系作为预定义的模型静态存储;其次,由于每个人的身体构造是相同的而骨架长短不同,因此可以引入齐次变换矩阵T来表示不同个体相对于静态模型上对应点的刚性变换,任意一点的位置可以通过矩阵相乘的方式得到。Wherein, as shown in Figure 2, the G coordinate system can be a human body static model coordinate system, and the representation method of the coordinate system can be specifically described as follows: "Skeleton tree", the relative relationship between feature points and feature points in the skeleton tree is statically stored as a predefined model; secondly, since everyone's body structure is the same but the skeleton length is different, homogeneous transformation can be introduced The matrix T represents the rigid transformation of different individuals with respect to the corresponding points on the static model. The position of any point can be obtained by matrix multiplication.
在本申请的再一个可选实施例,所述方法还可以包括:根据RNN-LSTM模型、样本姿态数据以及预设的情绪姿态标签,训练情感强度模型。In still another optional embodiment of the present application, the method may further include: training an emotional strength model according to the RNN-LSTM model, sample pose data, and preset emotional pose tags.
具体地,在模型训练前,需要通过制作数据集或者采用网络上的数据集先进行预训练,例如对于某个姿态,需要确定被观察人的实际情绪,完成数据标注。所述模型训练的具体过程可以包括:基于RNN的改进的长短期记忆建模,由于现实生活中的许多数据都同时具有时间特征和空间特征,例如人体的运动轨迹,连续帧的视频等,人体姿态也是这样,每个时间点对应一组数据,而数据往往又具有一定的空间特征。因此要在这样的时间序列上开展分类,预测等工作,就必须在时间和空间上对其进行建模和特征抽取。常用的时间建模工具是循环神经网络(RNN)相关模型(LSTM),由于其特有的门结构设计,对时间序列特征具有强大的抽取能力,因此被广泛应用于预测问题并取得了良好的成果。传统的LSTM结构,包含输入门(input gate),输出门(output gate),遗忘门(forget gate)三个结构以及一个神经节点(cell),其中输入可以为t时刻当前帧人体的姿态表示,输出可以为姿态描述子,用于描述当前姿态的种类。对于本申请实施例,通过n各LSTM结构横向连接便组成双层LSTM-RNN结构,因为在确定人体姿态时,往往需要一段连续的影像流,因此通过{X 1,X 2,X 3,......X n-1,X n}来表示这样的视频流并作为模型的输入。 Specifically, before model training, pre-training needs to be performed by making a data set or using a data set on the network. For example, for a certain posture, the actual emotion of the observed person needs to be determined and data labeling is completed. The specific process of the model training may include: RNN-based improved long-term and short-term memory modeling. Since many data in real life have both temporal and spatial features, such as human body motion trajectory, continuous frames of video, etc., human body The same is true for posture, each time point corresponds to a set of data, and the data often has certain spatial characteristics. Therefore, in order to carry out classification and prediction work on such a time series, it is necessary to model and extract features in time and space. The commonly used time modeling tool is the Recurrent Neural Network (RNN) correlation model (LSTM). Because of its unique gate structure design and powerful extraction of time series features, it is widely used in forecasting problems and has achieved good results. . The traditional LSTM structure includes three structures: input gate, output gate, forget gate, and a neural node (cell), where the input can be the body's posture representation in the current frame at time t, The output can be a posture descriptor, which is used to describe the type of current posture. For the embodiment of the present application, a double-layer LSTM-RNN structure is formed by connecting n LSTM structures horizontally, because a continuous image stream is often required when determining the human body posture, so through {X 1 , X 2 , X 3 ,. ..... X n-1 ,X n } to represent such a video stream and serve as the input of the model.
需要说明的是,对于模型的训练需要数据集进行预训练,再通过制作数据集的方式进行二次训练以达到比较鲁棒的效果。比如,通过摄像头记录一个人当前的姿态影像流,通过OpenPose的关键点提取方式得到姿态描述矩阵,并询问当前人的情感类型强度完成数据标注,进而进行训练。对于本申请实施例,依然采用LSTM单元的结构,但采用双层LSTM的结构以及一个全连接层,双层的结构可以增加时序的相关性检测。It should be noted that the training of the model requires a data set for pre-training, and then a second training is performed by creating a data set to achieve a relatively robust effect. For example, the current posture image stream of a person is recorded through the camera, the posture description matrix is obtained through the key point extraction method of OpenPose, and the current person's emotion type strength is asked to complete the data labeling, and then the training is carried out. For the embodiment of the present application, the structure of the LSTM unit is still adopted, but the structure of a double-layer LSTM and a fully connected layer are adopted. The double-layer structure can increase the correlation detection of the timing.
本申请提供一种情感信息识别方法,与现有技术仅通过面部表情推断人的情绪相比,本申请通过将接收情感信息识别请求,所述情感信息识别请求中携带有人体姿态信息;利用预设的姿态转换算法将所述人体姿态信息转换为包含姿态特征点的姿态矩阵;根据预设的情感强度算法对所述姿态矩阵进行处理,得到情感强度数据;根据所述情感强度数据,检索并反馈对应的情感类型。从而能够通过人体姿态和面部表情双重维度提高情感信息识别的准确率和效率。此外,本申请还利用区块链技术存储数据,能够提高情感信息的安全性。This application provides an emotional information recognition method. Compared with the prior art inferring human emotions only through facial expressions, this application receives an emotional information recognition request that carries human posture information; using predictions The posture conversion algorithm is assumed to convert the human posture information into a posture matrix containing posture feature points; the posture matrix is processed according to a preset emotion intensity algorithm to obtain emotion intensity data; according to the emotion intensity data, retrieve and merge Feedback corresponding emotion type. Therefore, the accuracy and efficiency of emotion information recognition can be improved through the dual dimensions of human body posture and facial expression. In addition, this application also uses blockchain technology to store data, which can improve the security of emotional information.
进一步地,作为图1的具体实现,本申请实施例提供了一种情感信息识别装置,如图3所示,所述装置包括:接收单元21、转换单元22、处理单元23和反馈单元24。Further, as a specific implementation of FIG. 1, an embodiment of the present application provides an emotional information recognition device. As shown in FIG. 3, the device includes: a receiving unit 21, a conversion unit 22, a processing unit 23 and a feedback unit 24.
接收单元21,可以用于接收情感信息识别请求,所述情感信息识别请求中携带有人体姿态信息;The receiving unit 21 may be configured to receive an emotion information recognition request, where the emotion information recognition request carries human posture information;
转换单元22,可以用于利用预设姿态转换算法将所述人体姿态信息转换为包含姿态特征点的姿态矩阵;The conversion unit 22 may be used to convert the human body posture information into a posture matrix containing posture feature points by using a preset posture conversion algorithm;
处理单元23,可以用于根据预设的情感强度算法对所述姿态矩阵进行处理,得到情感强度数据;The processing unit 23 may be configured to process the posture matrix according to a preset emotion intensity algorithm to obtain emotion intensity data;
反馈单元24,可以用于根据所述情感强度数据,检索并反馈对应的情感类型。The feedback unit 24 may be used to retrieve and feed back the corresponding emotion type according to the emotion intensity data.
进一步地,所述处理单元23,包括:Further, the processing unit 23 includes:
处理模块231,可以用于利用预先训练的情感强度模型对所述姿态矩阵以及获取的情感矢量信息进行处理,得到情感强度。The processing module 231 may be used to process the posture matrix and the obtained emotion vector information using a pre-trained emotion intensity model to obtain the emotion intensity.
进一步地,所述处理模块231具体可以用于利用sigmoid函数对输入的姿态矩阵以及情感矢量信息同时进行处理,并输出得到的情感强度数据。Further, the processing module 231 may be specifically configured to simultaneously process the input posture matrix and emotion vector information by using the sigmoid function, and output the obtained emotion intensity data.
进一步地,所述处理模块231具体还可以用于利用区块链技术存储所述情感强度数据。Further, the processing module 231 may be specifically used to store the emotional intensity data using blockchain technology.
进一步地,所述转换单元22,包括:Further, the conversion unit 22 includes:
获取模块221,可以用于获取各特征点的欧拉角参数;The obtaining module 221 may be used to obtain Euler angle parameters of each feature point;
确定模块222,可以用于根据所述欧拉角参数确定基于人体静态模型坐标系下每个特征点的姿态矩阵。The determining module 222 may be used to determine the posture matrix of each feature point in the human body static model coordinate system according to the Euler angle parameters.
进一步地,所述装置还包括:Further, the device further includes:
识别单元25,可以用于利用预设的面部识别算法对获取的面部表情信息进行识别处理,得到对应的情感矢量信息。The recognition unit 25 may be used to perform recognition processing on the acquired facial expression information by using a preset facial recognition algorithm to obtain corresponding emotion vector information.
进一步地,所述装置还包括:Further, the device further includes:
建立单元26,可以用于根据所述人体姿态信息,建立人体关节点基于人体静态模型坐标系下的齐次变换矩阵;The establishing unit 26 may be used to establish, according to the human body posture information, a homogeneous transformation matrix of the human body joint points based on the human body static model coordinate system;
确定单元27,可以用于通过矩阵相乘的方式确定各关节点坐标,并将所述关节点确定为人体姿态的特征点。The determining unit 27 may be used to determine the coordinates of each joint point by matrix multiplication, and determine the joint point as a feature point of the posture of the human body.
进一步地,所述装置还包括:Further, the device further includes:
训练单元28,可以用于根据RNN-LSTM模型、样本姿态数据以及预设的情绪姿态标签,训练情感强度模型。The training unit 28 may be used to train the emotional intensity model according to the RNN-LSTM model, sample pose data, and preset emotional pose labels.
基于上述如图1所示方法,相应的,本申请实施例还提供了一种存储介质,所述存储介质中存储有至少一可执行指令,所述可执行指令使处理器执行以下步骤:接收情感信息识别请求,所述情感信息识别请求中携带有人体姿态信息;利用预设的姿态转换算法将所述人体姿态信息转换为包含姿态特征点的姿态矩阵;根据预设的情感强度算法对所述姿态矩阵进行处理,得到情感强度数据;根据所述情感强度数据,检索并反馈对应的情感类型。可选的,该可执行指令使处理器执行时还可实现上述实施例中方法的其他步骤,这里不再赘述。进一步可选的,本申请涉及的存储介质可以是计算机可读存储介质,该存储介质如计算机可读存储介质可以是非易失性的,也可以是易失性的。Based on the above method shown in FIG. 1, correspondingly, an embodiment of the present application further provides a storage medium in which at least one executable instruction is stored, and the executable instruction causes the processor to perform the following steps: receiving Emotion information recognition request, said emotion information recognition request carries human body posture information; using a preset posture conversion algorithm to convert the human body posture information into a posture matrix containing posture feature points; according to the preset emotion intensity algorithm The posture matrix is processed to obtain emotion intensity data; according to the emotion intensity data, the corresponding emotion type is retrieved and fed back. Optionally, the executable instruction can also implement other steps of the method in the foregoing embodiment when executed by the processor, which will not be repeated here. Further optionally, the storage medium involved in the present application may be a computer-readable storage medium, and the storage medium, such as a computer-readable storage medium, may be non-volatile or volatile.
基于上述如图1所示方法和如图3所示装置的实施例,本申请实施例还提供了一种计算机设备,如图4所示,处理器(processor)31、通信接口(Communications Interface)32、存储器(memory)33、以及通信总线34。其中:处理器31、通信接口32、以及存储器33通过通信总线34完成相互间的通信。通信接口34,用于与其它设备比如用户端或其它服务器等的网元通信。处理器31,用于执行程序,具体可以执行上述情感信息识别方法实施例中的相关步骤。具体地,程序可以包括程序代码,该程序代码包括计算机操作指令。处理器31可能是中央处理器CPU,或者是特定集成电路ASIC(Application Specific Integrated Circuit),或者是被配置成实施本申请实施例的一个或多个集成电路。Based on the above-mentioned method shown in FIG. 1 and the embodiment of the apparatus shown in FIG. 3, an embodiment of the present application also provides a computer device. As shown in FIG. 4, a processor 31 and a communication interface (Communications Interface) 32. A memory (memory) 33, and a communication bus 34. Among them, the processor 31, the communication interface 32, and the memory 33 communicate with each other through the communication bus 34. The communication interface 34 is used to communicate with other devices, such as network elements such as user terminals or other servers. The processor 31 is configured to execute a program, and specifically can execute relevant steps in the foregoing embodiment of the emotion information recognition method. Specifically, the program may include program code, and the program code includes computer operation instructions. The processor 31 may be a central processing unit CPU, or an Application Specific Integrated Circuit (ASIC), or one or more integrated circuits configured to implement the embodiments of the present application.
终端包括的一个或多个处理器,可以是同一类型的处理器,如一个或多个CPU;也可以是不同类型的处理器,如一个或多个CPU以及一个或多个ASIC。存储器33,用于存放程序。存储器33可能包含高速RAM存储器,也可能还包括非易失性存储器(non-volatile memory),例如至少一个磁盘存储器。程序具体可以用于使得处理器31执行以下操作:接收情感信息识别请求,所述情感信息识别请求中携带有人体姿态信息;利用预设的姿态转换算法将所述人体姿态信息转换为包含姿态特征点的姿态矩阵;根据预设的情感强度算法对所述姿态矩阵进行处理,得到情感强度数据;根据所述情感强度数据,检索并反馈对 应的情感类型。The one or more processors included in the terminal may be processors of the same type, such as one or more CPUs; or processors of different types, such as one or more CPUs and one or more ASICs. The memory 33 is used to store programs. The memory 33 may include a high-speed RAM memory, and may also include a non-volatile memory (non-volatile memory), for example, at least one disk memory. The program can specifically be used to make the processor 31 perform the following operations: receive an emotion information recognition request, the emotion information recognition request carries human posture information; use a preset posture conversion algorithm to convert the human posture information to include posture features Point posture matrix; process the posture matrix according to a preset emotion intensity algorithm to obtain emotion intensity data; retrieve and feed back the corresponding emotion type according to the emotion intensity data.
通过本申请的技术方案,能够接收情感信息识别请求,所述情感信息识别请求中携带有人体姿态信息;利用预设的姿态转换算法将所述人体姿态信息转换为包含姿态特征点的姿态矩阵;根据预设的情感强度算法对所述姿态矩阵进行处理,得到情感强度数据;根据所述情感强度数据,检索并反馈对应的情感类型。从而能够从而能够通过人体姿态和面部表情双重维度提高情感信息识别的准确率和效率。此外,本申请还利用区块链技术存储数据,能够提高情感信息的安全性。Through the technical solution of the present application, it is possible to receive an emotion information recognition request, the emotion information recognition request carries human body posture information; use a preset posture conversion algorithm to convert the human body posture information into a posture matrix containing posture feature points; The posture matrix is processed according to a preset emotion intensity algorithm to obtain emotion intensity data; according to the emotion intensity data, the corresponding emotion type is retrieved and fed back. Therefore, the accuracy and efficiency of emotion information recognition can be improved through the dual dimensions of human body posture and facial expression. In addition, this application also uses blockchain technology to store data, which can improve the security of emotional information.
本申请所指区块链是分布式数据存储、点对点传输、共识机制、加密算法等计算机技术的新型应用模式。区块链(Blockchain),本质上是一个去中心化的数据库,是一串使用密码学方法相关联产生的数据块,每一个数据块中包含了一批次网络交易的信息,用于验证其信息的有效性(防伪)和生成下一个区块。区块链可以包括区块链底层平台、平台产品服务层以及应用服务层等。The blockchain referred to in this application is a new application mode of computer technology such as distributed data storage, point-to-point transmission, consensus mechanism, and encryption algorithm. Blockchain, essentially a decentralized database, is a series of data blocks associated with cryptographic methods. Each data block contains a batch of network transaction information for verification. The validity of the information (anti-counterfeiting) and the generation of the next block. The blockchain can include the underlying platform of the blockchain, the platform product service layer, and the application service layer.
需要说明的是,本申请实施例提供的一种情感信息识别装置所涉及各功能模块的其他相应描述,可以参考图1所示方法的对应描述,在此不再赘述。It should be noted that, for other corresponding descriptions of the functional modules involved in the emotional information recognition device provided in the embodiment of the present application, reference may be made to the corresponding description of the method shown in FIG. 1, and details are not repeated here.
在上述实施例中,对各个实施例的描述都各有侧重,某个实施例中没有详述的部分,可以参见其他实施例的相关描述。In the above-mentioned embodiments, the description of each embodiment has its own emphasis. For parts that are not described in detail in an embodiment, reference may be made to related descriptions of other embodiments.
可以理解的是,上述方法及装置中的相关特征可以相互参考。另外,上述实施例中的“第一”、“第二”等是用于区分各实施例,而并不代表各实施例的优劣。It can be understood that the relevant features in the above method and device can be referred to each other. In addition, the “first”, “second”, etc. in the foregoing embodiments are used to distinguish the embodiments, and do not represent the advantages and disadvantages of the embodiments.
所属领域的技术人员可以清楚地了解到,为描述的方便和简洁,上述描述的系统,装置和单元的具体工作过程,可以参考前述方法实施例中的对应过程,在此不再赘述。Those skilled in the art can clearly understand that, for the convenience and conciseness of the description, the specific working process of the above-described system, device, and unit can refer to the corresponding process in the foregoing method embodiment, which will not be repeated here.
在此提供的算法和显示不与任何特定计算机、虚拟系统或者其它设备固有相关。各种通用系统也可以与基于在此的示教一起使用。根据上面的描述,构造这类系统所要求的结构是显而易见的。此外,本申请也不针对任何特定编程语言。应当明白,可以利用各种编程语言实现在此描述的本申请的内容,并且上面对特定语言所做的描述是为了披露本申请的最佳实施方式。The algorithms and displays provided here are not inherently related to any particular computer, virtual system or other equipment. Various general-purpose systems can also be used with the teaching based on this. Based on the above description, the structure required to construct this type of system is obvious. In addition, this application is not aimed at any specific programming language. It should be understood that various programming languages can be used to implement the content of the application described herein, and the above description of a specific language is for the purpose of disclosing the best embodiment of the application.
在此处所提供的说明书中,说明了大量具体细节。然而,能够理解,本申请的实施例可以在没有这些具体细节的情况下实践。在一些实例中,并未详细示出公知的方法、结构和技术,以便不模糊对本说明书的理解。In the instructions provided here, a lot of specific details are explained. However, it can be understood that the embodiments of the present application can be practiced without these specific details. In some instances, well-known methods, structures, and technologies are not shown in detail, so as not to obscure the understanding of this specification.
类似地,应当理解,为了精简本公开并帮助理解各个发明方面中的一个或多个,在上面对本申请的示例性实施例的描述中,本申请的各个特征有时被一起分组到单个实施例、图、或者对其的描述中。然而,并不应将该公开的方法解释成反映如下意图:即所要求保护的本申请要求比在每个权利要求中所明确记载的特征更多的特征。更确切地说,如下面的权利要求书所反映的那样,发明方面在于少于前面公开的单个实施例的所有特征。因此,遵循具体实施方式的权利要求书由此明确地并入该具体实施方式,其中每个权利要求本身都作为本申请的单独实施例。Similarly, it should be understood that, in order to simplify the present disclosure and help understand one or more of the various inventive aspects, in the above description of the exemplary embodiments of the present application, the various features of the present application are sometimes grouped together into a single embodiment, Figure, or its description. However, the disclosed method should not be interpreted as reflecting the intention that the claimed application requires more features than the features explicitly recorded in each claim. More precisely, as reflected in the following claims, the inventive aspect lies in less than all the features of a single embodiment disclosed previously. Therefore, the claims following the specific embodiment are thus explicitly incorporated into the specific embodiment, wherein each claim itself serves as a separate embodiment of the application.
本领域那些技术人员可以理解,可以对实施例中的设备中的模块进行自适应性地改变并且把它们设置在与该实施例不同的一个或多个设备中。可以把实施例中的模块或单元或组件组合成一个模块或单元或组件,以及此外可以把它们分成多个子模块或子单元或子组件。除了这样的特征和/或过程或者单元中的至少一些是相互排斥之外,可以采用任何组合对本说明书(包括伴随的权利要求、摘要和附图)中公开的所有特征以及如此公开的任何方法或者设备的所有过程或单元进行组合。除非另外明确陈述,本说明书(包括伴随的权利要求、摘要和附图)中公开的每个特征可以由提供相同、等同或相似目的的替代特征来代替。Those skilled in the art can understand that it is possible to adaptively change the modules in the device in the embodiment and set them in one or more devices different from the embodiment. The modules or units or components in the embodiments can be combined into one module or unit or component, and in addition, they can be divided into multiple sub-modules or sub-units or sub-components. Except that at least some of such features and/or processes or units are mutually exclusive, any combination can be used to compare all the features disclosed in this specification (including the accompanying claims, abstract and drawings) and any method or methods disclosed in this manner or All the processes or units of the equipment are combined. Unless expressly stated otherwise, each feature disclosed in this specification (including the accompanying claims, abstract and drawings) may be replaced by an alternative feature providing the same, equivalent or similar purpose.
此外,本领域的技术人员能够理解,尽管在此所述的一些实施例包括其它实施例中所 包括的某些特征而不是其它特征,但是不同实施例的特征的组合意味着处于本申请的范围之内并且形成不同的实施例。例如,在下面的权利要求书中,所要求保护的实施例的任意之一都可以以任意的组合方式来使用。In addition, those skilled in the art can understand that although some embodiments described herein include certain features included in other embodiments but not other features, the combination of features of different embodiments means that they are within the scope of the present application. Within and form different embodiments. For example, in the following claims, any one of the claimed embodiments can be used in any combination.
本申请的各个部件实施例可以以硬件实现,或者以在一个或者多个处理器上运行的软件模块实现,或者以它们的组合实现。本领域的技术人员应当理解,可以在实践中使用微处理器或者数字信号处理器(DSP)来实现根据本申请实施例中的一些或者全部部件的一些或者全部功能。本申请还可以实现为用于执行这里所描述的方法的一部分或者全部的设备或者装置程序(例如,计算机程序和计算机程序产品)。这样的实现本申请的程序可以存储在计算机可读介质上,或者可以具有一个或者多个信号的形式。这样的信号可以从因特网网站上下载得到,或者在载体信号上提供,或者以任何其他形式提供。The various component embodiments of the present application may be implemented by hardware, or by software modules running on one or more processors, or by a combination of them. Those skilled in the art should understand that a microprocessor or a digital signal processor (DSP) may be used in practice to implement some or all of the functions of some or all of the components in the embodiments of the present application. This application can also be implemented as a device or device program (for example, a computer program and a computer program product) for executing part or all of the methods described herein. Such a program for implementing the present application may be stored on a computer-readable medium, or may have the form of one or more signals. Such a signal can be downloaded from an Internet website, or provided on a carrier signal, or provided in any other form.
应该注意的是上述实施例对本申请进行说明而不是对本申请进行限制,并且本领域技术人员在不脱离所附权利要求的范围的情况下可设计出替换实施例。在权利要求中,不应将位于括号之间的任何参考符号构造成对权利要求的限制。单词“包含”不排除存在未列在权利要求中的元件或步骤。位于元件之前的单词“一”或“一个”不排除存在多个这样的元件。本申请可以借助于包括有若干不同元件的硬件以及借助于适当编程的计算机来实现。在列举了若干装置的单元权利要求中,这些装置中的若干个可以是通过同一个硬件项来具体体现。单词第一、第二、以及第三等的使用不表示任何顺序。可将这些单词解释为名称。It should be noted that the above-mentioned embodiments illustrate rather than limit the application, and those skilled in the art can design alternative embodiments without departing from the scope of the appended claims. In the claims, any reference signs placed between parentheses should not be constructed as a limitation to the claims. The word "comprising" does not exclude the presence of elements or steps not listed in the claims. The word "a" or "an" preceding an element does not exclude the presence of multiple such elements. The application can be realized by means of hardware including several different elements and by means of a suitably programmed computer. In the unit claims listing several devices, several of these devices may be embodied in the same hardware item. The use of the words first, second, and third, etc. do not indicate any order. These words can be interpreted as names.

Claims (20)

  1. 一种情感信息识别方法,其中,包括:An emotional information recognition method, which includes:
    接收情感信息识别请求,所述情感信息识别请求中携带有人体姿态信息;Receiving an emotional information recognition request, where the emotional information recognition request carries human posture information;
    利用预设的姿态转换算法将所述人体姿态信息转换为包含姿态特征点的姿态矩阵;Using a preset posture conversion algorithm to convert the human posture information into a posture matrix containing posture feature points;
    根据预设的情感强度算法对所述姿态矩阵进行处理,得到情感强度数据;Processing the posture matrix according to a preset emotion intensity algorithm to obtain emotion intensity data;
    根据所述情感强度数据,检索并反馈对应的情感类型。According to the emotion intensity data, the corresponding emotion type is retrieved and fed back.
  2. 根据权利要求1所述的方法,其中,所述根据预设的情感强度算法对所述姿态矩阵进行处理,得到情感强度数据,包括:The method according to claim 1, wherein the processing the posture matrix according to a preset emotion intensity algorithm to obtain emotion intensity data comprises:
    利用预先训练的情感强度模型对所述姿态矩阵以及获取的情感矢量信息进行处理,得到情感强度数据。The pre-trained emotion intensity model is used to process the posture matrix and the obtained emotion vector information to obtain emotion intensity data.
  3. 根据权利要求2所述的方法,其中,所述利用预先训练的情感强度模型对所述姿态矩阵以及获取的情感矢量信息进行处理,得到情感强度数据,包括:The method according to claim 2, wherein said using a pre-trained emotion intensity model to process the posture matrix and the obtained emotion vector information to obtain emotion intensity data comprises:
    利用sigmoid函数对输入的姿态矩阵以及情感矢量信息同时进行处理,并输出得到的情感强度数据,所述情感强度数据存储于区块链中。The sigmoid function is used to process the input posture matrix and emotion vector information at the same time, and the obtained emotion intensity data is output, and the emotion intensity data is stored in the blockchain.
  4. 根据权利要求3所述的方法,其中,所述利用预设姿态转换算法将所述人体姿态信息转换为包含姿态特征点的姿态矩阵,包括:The method according to claim 3, wherein said using a preset posture conversion algorithm to convert said human posture information into a posture matrix containing posture feature points comprises:
    获取各特征点的欧拉角参数;Obtain the Euler angle parameters of each feature point;
    根据所述欧拉角参数确定基于人体静态模型坐标系下每个特征点的姿态矩阵。The posture matrix of each feature point in the coordinate system of the human body static model is determined according to the Euler angle parameters.
  5. 根据权利要求4所述的方法,其中,所述利用预先训练的情感强度模型对所述姿态矩阵以及获取的情感矢量信息进行处理,得到情感强度数据之前,所述方法还包括:The method according to claim 4, wherein the method further comprises: before the posture matrix and the obtained emotion vector information are processed by using a pre-trained emotion intensity model to obtain emotion intensity data, the method further comprises:
    利用预设的面部识别算法对获取的面部表情信息进行识别处理,得到对应的情感矢量信息。A preset facial recognition algorithm is used to recognize and process the acquired facial expression information to obtain the corresponding emotion vector information.
  6. 根据权利要求5所述的方法,其中,所述利用预设姿态转换算法将所述人体姿态信息转换为包含姿态特征点的姿态矩阵之前,所述方法还包括:The method according to claim 5, wherein before said converting said human body posture information into a posture matrix containing posture feature points by using a preset posture conversion algorithm, said method further comprises:
    根据所述人体姿态信息,建立人体关节点基于人体静态模型坐标系下的齐次变换矩阵;According to the human body posture information, establish a homogeneous transformation matrix of the human body joint points based on the human body static model coordinate system;
    通过矩阵相乘的方式确定各关节点坐标,并将所述关节点确定为人体姿态的特征点。The coordinates of each joint point are determined by matrix multiplication, and the joint points are determined as the characteristic points of the posture of the human body.
  7. 根据权利要求6所述的方法,其中,所述利用预先训练的情感强度模型对所述姿态矩阵以及获取的情感适量信息进行处理,得到情感强度数据之前,所述方法还包括:7. The method according to claim 6, wherein said using a pre-trained emotional intensity model to process said posture matrix and the acquired emotional appropriate amount of information, and before obtaining emotional intensity data, said method further comprises:
    根据RNN-LSTM模型、样本姿态数据以及预设的情绪姿态标签,训练情感强度模型。According to the RNN-LSTM model, sample pose data, and preset emotional pose labels, the emotional intensity model is trained.
  8. 一种情感信息识别装置,其中,包括:An emotional information recognition device, which includes:
    接收单元,用于接收情感信息识别请求,所述情感信息识别请求中携带有人体姿态信息;A receiving unit, configured to receive an emotional information recognition request, where the emotional information recognition request carries human posture information;
    转换单元,用于利用预设姿态转换算法将所述人体姿态信息转换为包含姿态特征点的姿态矩阵;A conversion unit, configured to convert the human body posture information into a posture matrix containing posture feature points by using a preset posture conversion algorithm;
    处理单元,用于根据预设的情感强度算法对所述姿态矩阵进行处理,得到情感强度数据;A processing unit, configured to process the posture matrix according to a preset emotion intensity algorithm to obtain emotion intensity data;
    反馈单元,用于根据所述情感强度数据,检索并反馈对应的情感类型。The feedback unit is used to retrieve and feed back the corresponding emotion type according to the emotion intensity data.
  9. 一种存储介质,其上存储有计算机程序,所述存储介质中存储有至少一可执行指令,所述可执行指令使处理器执行以下步骤:A storage medium on which a computer program is stored, and at least one executable instruction is stored in the storage medium, and the executable instruction causes a processor to perform the following steps:
    接收情感信息识别请求,所述情感信息识别请求中携带有人体姿态信息;Receiving an emotional information recognition request, where the emotional information recognition request carries human posture information;
    利用预设的姿态转换算法将所述人体姿态信息转换为包含姿态特征点的姿态矩阵;Using a preset posture conversion algorithm to convert the human posture information into a posture matrix containing posture feature points;
    根据预设的情感强度算法对所述姿态矩阵进行处理,得到情感强度数据;Processing the posture matrix according to a preset emotion intensity algorithm to obtain emotion intensity data;
    根据所述情感强度数据,检索并反馈对应的情感类型。According to the emotion intensity data, the corresponding emotion type is retrieved and fed back.
  10. 根据权利要求9所述的存储介质,其中,所述根据预设的情感强度算法对所述姿 态矩阵进行处理,得到情感强度数据时,具体执行以下步骤:The storage medium according to claim 9, wherein when the posture matrix is processed according to a preset emotion intensity algorithm to obtain emotion intensity data, the following steps are specifically executed:
    利用预先训练的情感强度模型对所述姿态矩阵以及获取的情感矢量信息进行处理,得到情感强度数据。The pre-trained emotion intensity model is used to process the posture matrix and the obtained emotion vector information to obtain emotion intensity data.
  11. 根据权利要求10所述的存储介质,其中,所述利用预先训练的情感强度模型对所述姿态矩阵以及获取的情感矢量信息进行处理,得到情感强度数据时,具体执行以下步骤:10. The storage medium according to claim 10, wherein the pre-trained emotion intensity model is used to process the posture matrix and the obtained emotion vector information to obtain emotion intensity data, specifically performing the following steps:
    利用sigmoid函数对输入的姿态矩阵以及情感矢量信息同时进行处理,并输出得到的情感强度数据,所述情感强度数据存储于区块链中。The sigmoid function is used to process the input posture matrix and emotion vector information at the same time, and the obtained emotion intensity data is output, and the emotion intensity data is stored in the blockchain.
  12. 根据权利要求11所述的存储介质,其中,所述利用预设姿态转换算法将所述人体姿态信息转换为包含姿态特征点的姿态矩阵时,具体执行以下步骤:The storage medium according to claim 11, wherein the following steps are specifically performed when the human body posture information is converted into a posture matrix containing posture feature points by using a preset posture conversion algorithm:
    获取各特征点的欧拉角参数;Obtain the Euler angle parameters of each feature point;
    根据所述欧拉角参数确定基于人体静态模型坐标系下每个特征点的姿态矩阵。The posture matrix of each feature point in the coordinate system of the human body static model is determined according to the Euler angle parameters.
  13. 根据权利要求12所述的存储介质,其中,所述利用预先训练的情感强度模型对所述姿态矩阵以及获取的情感矢量信息进行处理,得到情感强度数据之前,所述可执行指令还使处理器执行:The storage medium according to claim 12, wherein the pre-trained emotion intensity model is used to process the posture matrix and the obtained emotion vector information, and before the emotion intensity data is obtained, the executable instruction further causes the processor implement:
    利用预设的面部识别算法对获取的面部表情信息进行识别处理,得到对应的情感矢量信息。A preset facial recognition algorithm is used to recognize and process the acquired facial expression information to obtain the corresponding emotion vector information.
  14. 根据权利要求13所述的存储介质,其中,所述利用预设姿态转换算法将所述人体姿态信息转换为包含姿态特征点的姿态矩阵之前,所述可执行指令还使处理器执行:The storage medium according to claim 13, wherein, before the human body posture information is converted into a posture matrix containing posture feature points by using a preset posture conversion algorithm, the executable instruction further causes the processor to execute:
    根据所述人体姿态信息,建立人体关节点基于人体静态模型坐标系下的齐次变换矩阵;According to the human body posture information, establish a homogeneous transformation matrix of the human body joint points based on the human body static model coordinate system;
    通过矩阵相乘的方式确定各关节点坐标,并将所述关节点确定为人体姿态的特征点。The coordinates of each joint point are determined by matrix multiplication, and the joint points are determined as the characteristic points of the posture of the human body.
  15. 一种计算机设备,包括处理器、存储器、通信接口和通信总线所述处理器、所述存储器和所述通信接口通过所述通信总线完成相互间的通信,所述存储器用于存放至少一可执行指令,所述可执行指令使所述处理器执行以下步骤:A computer device includes a processor, a memory, a communication interface, and a communication bus. The processor, the memory, and the communication interface communicate with each other through the communication bus, and the memory is used to store at least one executable Instructions, the executable instructions cause the processor to perform the following steps:
    接收情感信息识别请求,所述情感信息识别请求中携带有人体姿态信息;Receiving an emotional information recognition request, where the emotional information recognition request carries human posture information;
    利用预设的姿态转换算法将所述人体姿态信息转换为包含姿态特征点的姿态矩阵;Using a preset posture conversion algorithm to convert the human posture information into a posture matrix containing posture feature points;
    根据预设的情感强度算法对所述姿态矩阵进行处理,得到情感强度数据;Processing the posture matrix according to a preset emotion intensity algorithm to obtain emotion intensity data;
    根据所述情感强度数据,检索并反馈对应的情感类型。According to the emotion intensity data, the corresponding emotion type is retrieved and fed back.
  16. 根据权利要求15所述的计算机设备,其中,所述根据预设的情感强度算法对所述姿态矩阵进行处理,得到情感强度数据时,具体执行以下步骤:15. The computer device according to claim 15, wherein when the posture matrix is processed according to a preset emotion intensity algorithm to obtain emotion intensity data, the following steps are specifically executed:
    利用预先训练的情感强度模型对所述姿态矩阵以及获取的情感矢量信息进行处理,得到情感强度数据。The pre-trained emotion intensity model is used to process the posture matrix and the obtained emotion vector information to obtain emotion intensity data.
  17. 根据权利要求16所述的计算机设备,其中,所述利用预先训练的情感强度模型对所述姿态矩阵以及获取的情感矢量信息进行处理,得到情感强度数据时,具体执行以下步骤:15. The computer device according to claim 16, wherein the pre-trained emotion intensity model is used to process the posture matrix and the obtained emotion vector information to obtain emotion intensity data, specifically performing the following steps:
    利用sigmoid函数对输入的姿态矩阵以及情感矢量信息同时进行处理,并输出得到的情感强度数据,所述情感强度数据存储于区块链中。The sigmoid function is used to process the input posture matrix and emotion vector information at the same time, and the obtained emotion intensity data is output, and the emotion intensity data is stored in the blockchain.
  18. 根据权利要求17所述的计算机设备,其中,所述利用预设姿态转换算法将所述人体姿态信息转换为包含姿态特征点的姿态矩阵时,具体执行以下步骤:18. The computer device according to claim 17, wherein the following steps are specifically performed when the human body posture information is converted into a posture matrix containing posture feature points by using a preset posture conversion algorithm:
    获取各特征点的欧拉角参数;Obtain the Euler angle parameters of each feature point;
    根据所述欧拉角参数确定基于人体静态模型坐标系下每个特征点的姿态矩阵。The posture matrix of each feature point in the coordinate system of the human body static model is determined according to the Euler angle parameters.
  19. 根据权利要求18所述的计算机设备,其中,所述利用预先训练的情感强度模型对所述姿态矩阵以及获取的情感矢量信息进行处理,得到情感强度数据之前,所述可执行指令还使处理器执行:18. The computer device according to claim 18, wherein the pre-trained emotion intensity model is used to process the posture matrix and the obtained emotion vector information, and before the emotion intensity data is obtained, the executable instruction further causes the processor implement:
    利用预设的面部识别算法对获取的面部表情信息进行识别处理,得到对应的情感矢量 信息。The preset facial recognition algorithm is used to recognize the acquired facial expression information to obtain the corresponding emotion vector information.
  20. 根据权利要求19所述的计算机设备,其中,所述利用预设姿态转换算法将所述人体姿态信息转换为包含姿态特征点的姿态矩阵之前,所述可执行指令还使处理器执行:20. The computer device according to claim 19, wherein, before the human body posture information is converted into a posture matrix containing posture feature points by using a preset posture conversion algorithm, the executable instruction further causes the processor to execute:
    根据所述人体姿态信息,建立人体关节点基于人体静态模型坐标系下的齐次变换矩阵;According to the human body posture information, establish a homogeneous transformation matrix of the human body joint points based on the human body static model coordinate system;
    通过矩阵相乘的方式确定各关节点坐标,并将所述关节点确定为人体姿态的特征点。The coordinates of each joint point are determined by matrix multiplication, and the joint points are determined as the characteristic points of the posture of the human body.
PCT/CN2020/111036 2020-04-28 2020-08-25 Emotion information recognition method and apparatus, and storage medium and computer device WO2021217973A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN202010349534.0 2020-04-28
CN202010349534.0A CN111680550A (en) 2020-04-28 2020-04-28 Emotion information identification method and device, storage medium and computer equipment

Publications (1)

Publication Number Publication Date
WO2021217973A1 true WO2021217973A1 (en) 2021-11-04

Family

ID=72452275

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2020/111036 WO2021217973A1 (en) 2020-04-28 2020-08-25 Emotion information recognition method and apparatus, and storage medium and computer device

Country Status (2)

Country Link
CN (1) CN111680550A (en)
WO (1) WO2021217973A1 (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114998834A (en) * 2022-06-06 2022-09-02 杭州中威电子股份有限公司 Medical warning system based on face image and emotion recognition

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2022141894A1 (en) * 2020-12-31 2022-07-07 苏州源想理念文化发展有限公司 Three-dimensional feature emotion analysis method capable of fusing expression and limb motion
CN113255557B (en) * 2021-06-08 2023-08-15 苏州优柿心理咨询技术有限公司 Deep learning-based video crowd emotion analysis method and system

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR20160095735A (en) * 2015-02-04 2016-08-12 단국대학교 천안캠퍼스 산학협력단 Method and system for complex and multiplex emotion recognition of user face
CN105868694A (en) * 2016-03-24 2016-08-17 中国地质大学(武汉) Dual-mode emotion identification method and system based on facial expression and eyeball movement
CN108805087A (en) * 2018-06-14 2018-11-13 南京云思创智信息科技有限公司 Semantic temporal fusion association based on multi-modal Emotion identification system judges subsystem
CN109145754A (en) * 2018-07-23 2019-01-04 上海电力学院 Merge the Emotion identification method of facial expression and limb action three-dimensional feature
CN110147729A (en) * 2019-04-16 2019-08-20 深圳壹账通智能科技有限公司 User emotion recognition methods, device, computer equipment and storage medium
CN111401116A (en) * 2019-08-13 2020-07-10 南京邮电大学 Bimodal emotion recognition method based on enhanced convolution and space-time L STM network

Family Cites Families (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106803098A (en) * 2016-12-28 2017-06-06 南京邮电大学 A kind of three mode emotion identification methods based on voice, expression and attitude
CN108596039B (en) * 2018-03-29 2020-05-05 南京邮电大学 Bimodal emotion recognition method and system based on 3D convolutional neural network
CN109684911B (en) * 2018-10-30 2021-05-11 百度在线网络技术(北京)有限公司 Expression recognition method and device, electronic equipment and storage medium
CN109815938A (en) * 2019-02-27 2019-05-28 南京邮电大学 Multi-modal affective characteristics recognition methods based on multiclass kernel canonical correlation analysis

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR20160095735A (en) * 2015-02-04 2016-08-12 단국대학교 천안캠퍼스 산학협력단 Method and system for complex and multiplex emotion recognition of user face
CN105868694A (en) * 2016-03-24 2016-08-17 中国地质大学(武汉) Dual-mode emotion identification method and system based on facial expression and eyeball movement
CN108805087A (en) * 2018-06-14 2018-11-13 南京云思创智信息科技有限公司 Semantic temporal fusion association based on multi-modal Emotion identification system judges subsystem
CN109145754A (en) * 2018-07-23 2019-01-04 上海电力学院 Merge the Emotion identification method of facial expression and limb action three-dimensional feature
CN110147729A (en) * 2019-04-16 2019-08-20 深圳壹账通智能科技有限公司 User emotion recognition methods, device, computer equipment and storage medium
CN111401116A (en) * 2019-08-13 2020-07-10 南京邮电大学 Bimodal emotion recognition method based on enhanced convolution and space-time L STM network

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114998834A (en) * 2022-06-06 2022-09-02 杭州中威电子股份有限公司 Medical warning system based on face image and emotion recognition

Also Published As

Publication number Publication date
CN111680550A (en) 2020-09-18

Similar Documents

Publication Publication Date Title
WO2019174439A1 (en) Image recognition method and apparatus, and terminal and storage medium
WO2021217973A1 (en) Emotion information recognition method and apparatus, and storage medium and computer device
CN109948475B (en) Human body action recognition method based on skeleton features and deep learning
Storey et al. 3DPalsyNet: A facial palsy grading and motion recognition framework using fully 3D convolutional neural networks
WO2020103700A1 (en) Image recognition method based on micro facial expressions, apparatus and related device
Deng et al. MVF-Net: A multi-view fusion network for event-based object classification
CN108363973B (en) Unconstrained 3D expression migration method
AU2014253687B2 (en) System and method of tracking an object
CN110210426B (en) Method for estimating hand posture from single color image based on attention mechanism
CN112329525A (en) Gesture recognition method and device based on space-time diagram convolutional neural network
Sabater et al. Domain and view-point agnostic hand action recognition
Rajalakshmi et al. Multi-semantic discriminative feature learning for sign gesture recognition using hybrid deep neural architecture
Neverova Deep learning for human motion analysis
Podder et al. Time efficient real time facial expression recognition with CNN and transfer learning
Wan et al. Midoriko chatbot: LSTM-based emotional 3D avatar
Pattar et al. Intention and engagement recognition for personalized human-robot interaction, an integrated and deep learning approach
Khatun et al. A systematic review on the chronological development of bangla sign language recognition systems
Wu et al. Limb pose aware networks for monocular 3D pose estimation
Rastgoo et al. Word separation in continuous sign language using isolated signs and post-processing
CN112906520A (en) Gesture coding-based action recognition method and device
JP2022531828A (en) Hand movement symmetry recognition method and system
Bayegizova et al. EFFECTIVENESS OF THE USE OF ALGORITHMS AND METHODS OF ARTIFICIAL TECHNOLOGIES FOR SIGN LANGUAGE RECOGNITION FOR PEOPLE WITH DISABILITIES.
Usman et al. Skeleton-based motion prediction: A survey
Senanayaka et al. Continuous American Sign Language Recognition Using Computer Vision And Deep Learning Technologies
CN111178141B (en) LSTM human body behavior identification method based on attention mechanism

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 20933591

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 20933591

Country of ref document: EP

Kind code of ref document: A1