WO2021114841A1 - User report generating method and terminal device - Google Patents

User report generating method and terminal device Download PDF

Info

Publication number
WO2021114841A1
WO2021114841A1 PCT/CN2020/119300 CN2020119300W WO2021114841A1 WO 2021114841 A1 WO2021114841 A1 WO 2021114841A1 CN 2020119300 W CN2020119300 W CN 2020119300W WO 2021114841 A1 WO2021114841 A1 WO 2021114841A1
Authority
WO
WIPO (PCT)
Prior art keywords
conversation
vector
sentence
text
emotional
Prior art date
Application number
PCT/CN2020/119300
Other languages
French (fr)
Chinese (zh)
Inventor
邓悦
金戈
徐亮
Original Assignee
平安科技(深圳)有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 平安科技(深圳)有限公司 filed Critical 平安科技(深圳)有限公司
Publication of WO2021114841A1 publication Critical patent/WO2021114841A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/30Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
    • G06F16/33Querying
    • G06F16/332Query formulation
    • G06F16/3329Natural language query formulation or dialogue systems
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/30Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
    • G06F16/35Clustering; Classification
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/30Semantic analysis
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q10/00Administration; Management
    • G06Q10/06Resources, workflows, human or project management; Enterprise or organisation planning; Enterprise or organisation modelling
    • G06Q10/063Operations research, analysis or management
    • G06Q10/0639Performance analysis of employees; Performance analysis of enterprise or organisation operations
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q10/00Administration; Management
    • G06Q10/10Office automation; Time management
    • G06Q10/105Human resources
    • G06Q10/1053Employment or hiring

Definitions

  • This application belongs to the field of artificial intelligence technology, and in particular relates to a method for generating a user report and a terminal device.
  • the inventor realized how to efficiently screen interviewers and determine the personality characteristics of the interviewers, which directly affects the efficiency of the interview and the speed of decision-making.
  • the inventor found that through the user analysis report, the situation of the interviewed user can be quickly understood, and the interview efficiency can be greatly improved.
  • the inventor found that it mainly relies on the interviewer to analyze the interviewer’s personality. By collecting the interviewer’s answers to preset questions, the interviewer’s personality characteristics are subjectively determined, and the user analysis report is generated. People realize that the existing user analysis reports are completed by manpower, and the generation efficiency is low, thereby reducing the efficiency of personnel management.
  • the embodiment of the present application provides a method and terminal device for generating user reports to solve the problem of existing user report generation technology, which relies on manpower to complete, and the report generation efficiency is low, thereby reducing the efficiency of personnel management. .
  • the first aspect of the embodiments of the present application provides a method for generating a user report, including:
  • a personality analysis report of the target user is generated.
  • the embodiment of the application collects the voice signal of the target user during the conversation with the target user, converts the voice signal into the corresponding conversation text, and performs semantic analysis on the conversation text to obtain the corresponding conversation content collection, which is based on the conversation
  • the conversation word vector of each conversation keyword in the content collection generates the emotional feature value corresponding to the voice signal, and based on the emotional feature value of all voice signals, determines the personality type of the target user, and generates a personality analysis report about the target user, so as to be able to
  • the target user’s language is used to determine the personality, which achieves the purpose of automatically outputting an analysis report.
  • this embodiment does not rely on the interviewer or the conversation object to manually fill in or subjectively judge, and does not require the user to spend extra time writing a personality analysis report on the target user, thereby greatly reducing user operations, and
  • the above process can determine the emotional characteristic value through the voice signals at different stages in the conversation process, instead of using a single utterance or sentence to judge the personality, thereby improving the accuracy of the personality analysis report.
  • FIG. 1 is an implementation flowchart of a method for generating a user report provided by the first embodiment of the present application
  • FIG. 2 is a specific implementation flow chart of a method S103 for generating a user report provided by the second embodiment of the present application;
  • FIG. 3 is a specific implementation flow chart of a method S1031 for generating a user report provided by the third embodiment of the present application;
  • FIG. 4 is a specific implementation flow chart of a method S301 for generating a user report provided by the fourth embodiment of the present application;
  • FIG. 5 is a specific implementation flowchart of a method S302 for generating a user report provided by the fifth embodiment of the present application;
  • FIG. 6 is a specific implementation flow chart of a method S1034 for generating a user report provided by the sixth embodiment of the present application;
  • FIG. 7 is a specific implementation flowchart of a method S104 for generating a user report provided by the seventh embodiment of the present application.
  • FIG. 8 is a structural block diagram of a device for generating a user report according to an embodiment of the present application.
  • FIG. 9 is a schematic diagram of a terminal device provided by another embodiment of the present application.
  • the execution subject of the process is a terminal device, which includes but is not limited to: servers, computers, smart phones, tablet computers, and other devices capable of executing the method for generating user reports.
  • Fig. 1 shows an implementation flow chart of the method for generating a user report provided by the first embodiment of the present application, and the details are as follows:
  • the terminal device may be a server of the user database, and the server may be connected to the distributed microphone module through a communication link.
  • the communication link may be a physical link for wired communication, or may be through a local area network or the Internet.
  • the microphone module can be deployed in the same area as the terminal device, or distributed in various interview locations to collect voice signals generated during the interview.
  • the microphone module is specifically a microphone array.
  • the microphone array contains multiple microphone devices. During the process of collecting voice signals, the microphone array can obtain information about the current interview scene from multiple different angles. The voice signal is filtered and shaped through multiple voice signals to obtain a target signal for voice recognition.
  • a microphone array composed of a certain number of microphones to collect voice signals to sample and process the spatial characteristics of the sound field. It is used in the complex environment of the interview environment and can effectively solve noise, reverberation, vocal interference, echo, etc. The problem is to improve the signal quality of the voice signal collection, so that when the text information is subsequently output, the success rate of the text information conversion can be improved.
  • the terminal device may be set with an interview time period. If the terminal device detects that the current time has reached the preset interview start time, the microphone module is turned on to obtain the voice signal of the current interview scene through the microphone module. In addition, when the terminal device detects that the current time reaches the preset interview end time, the microphone module is turned off, and all the voice signals collected during the interview time period are converted into text information. Since during the meeting, the user's speech is not continuous, but intermittent, the terminal device can be configured with a start decibel value and an end decibel value.
  • the microphone module When the microphone module detects that the decibel value of the current interview scene is greater than the start decibel value, it will Start to collect the voice signal, and when the decibel value is less than the end decibel value, end the collection of the voice signal, take each collected voice signal as a conversation paragraph in the conversation process, and output the corresponding conversation text for each conversation paragraph
  • the interview process multiple conversational paragraphs are generated between the target user and the interviewer based on the question-and-answer process.
  • the terminal device can recognize the emotional feature value of each conversational paragraph, and generate all the conversational texts generated during the entire conversation. Personality analysis report of target users.
  • the terminal device may perform the output operation of text information after receiving a segment of voice signal, and after detecting the end of the current interview (for example, reaching the preset interview end time or detecting the pre-interview) If no voice signal is received within the waiting time), the operation of S102 is executed based on the text information corresponding to all the collected voice signals, that is, the collection operation is performed in parallel with the voice recognition operation; the terminal device can also collect the current conference All the voice information of is stored in the database, and after the interview is over, the operation of S102 is executed.
  • the terminal device may be provided with a voice recognition algorithm.
  • the terminal device may parse the voice signal through the voice recognition algorithm and output text information corresponding to the voice signal. This achieves the purpose of voice recognition, automatically records the interview content, and obtains the target.
  • the terminal device can determine the interview language used in the interview process, and adjust the voice recognition algorithm based on the interview language, thereby improving the accuracy of recognition.
  • the manner of determining the language of the interview may be: obtaining user information of the target user participating in the interview, the user information including information such as the user's household registration or residential address; and determining the language of the interview based on the household register or residential address of the target user.
  • the terminal device can divide the conversation text into multiple conversation segments based on the preset maximum number of sentences, and each conversation segment contains no more than the preset maximum number of sentences.
  • the conversation duration is long, the amount of generated conversation text is relatively large.
  • the terminal device can generate a corresponding sentence selection box based on the maximum number of sentences, and traverse the conversation text based on the sentence selection box to select consecutive conversation segments of multiple sentences, thereby stabilizing the number of sentences recognized each time. The consistency of the identification parameters is improved.
  • the manner of converting the voice signal into conversational text may specifically be: parsing the voice signal, and extracting the waveform characteristics and pitch characteristics corresponding to each frame of the voice signal.
  • the waveform characteristics and pitch characteristics corresponding to each frame of speech signal are sequentially input into the trained speech recognition model.
  • the speech recognition model is specifically trained based on the standard waveforms and pitch waveforms corresponding to all candidate characters, and the similarity with each candidate character can be calculated by importing the speech signal of each frame into the aforementioned speech recognition model.
  • the candidate character with the highest similarity is selected as the text corresponding to the speech signal of the frame, and the conversation text corresponding to the speech signal is generated based on the text of all frames.
  • semantic analysis is performed on the conversation text to obtain conversation keywords corresponding to the conversation text and conversation tags corresponding to each of the keywords, and a conversation content set is generated.
  • the terminal device may be equipped with a semantic recognition algorithm, which can perform semantic analysis on the conversation text, and extract the conversation keywords contained in the aforementioned conversation text.
  • the process of extracting conversation keywords by the semantic recognition algorithm can be specifically as follows: the conversation text is divided into words, divided into multiple phrases containing several characters, and each phrase contains at least one character and no more than 4 characters;
  • the part-of-speech recognition of phrases can filter invalid phrases that are not related to emotions. For example, some connectives have less relevance to the analysis of emotional personality, such as the connectives "and", "and” and “union”, etc. There are some particles such as " ⁇ ", " ⁇ ” and " ⁇ ".
  • the terminal device After the terminal device filters out invalid phrases, the effective phrases containing user emotions are obtained, and the above effective phrases are recognized as conversation keywords; optionally
  • the terminal device stores a key dictionary and judges whether the valid phrase is in the key dictionary. If it exists, the valid phrase is recognized as a conversation keyword; otherwise, the valid phrase is recognized as an invalid phrase.
  • the terminal device may configure a corresponding session tag for the session keyword, and the session tag is used to indicate the feature value of the session keyword in the preset word dimension.
  • the conversation tag can be used to mark the part of speech of the conversation keyword, such as the conversation keyword "today".
  • the conversation label can be set to "noun”
  • the conversation label Can be set to "time qualifier” and so on.
  • different session tags can be configured for session keywords.
  • the number of the above-mentioned session tags can be one, or two or more, which is not limited here.
  • the conversation between the interviewer and the interviewer is as follows: "Interviewer: Hello, please introduce yourself. Interviewer: Hello, interviewer. My name is Zhang San. I am from Shenzhen. graduated from university. Good at testing. Interviewer: What do you know about our position?”
  • the number of conversation texts is 3
  • i is the number of conversation texts.
  • the conversation sequence of "Hello, please introduce yourself” is 1.
  • each dialogue contains the corresponding number of sentences.
  • “Hello, please introduce themselves” statement includes the number is 2, namely, "Hello” and "Please introduce yourself” At this point, N i 2.
  • the automatic tag recognition algorithm before determining the tag corresponding to the keyword, can be trained to maximize the value of the maximization function, and the automatic recognition can be performed at this time.
  • the label recognition algorithm has been adjusted, where the maximization function can be expressed as:
  • represents model parameters.
  • a conversation word vector corresponding to each of the conversation keywords in the conversation content set is obtained, and an emotional feature value corresponding to the voice signal is determined based on each of the conversation word vectors.
  • the terminal device may generate a conversation word vector corresponding to the conversation keyword according to each conversation keyword and the corresponding conversation label in the conversation content set.
  • the method for generating the above-mentioned conversational word vector may be as follows: the terminal device is configured with a key dictionary, and each candidate keyword in the key dictionary is configured with a corresponding word number, and the conversation keyword is identified in the above-mentioned
  • the word number in the key dictionary determines the value of the first dimension based on the word number; correspondingly, the terminal device can generate a tag dictionary and determine the second dimension value of the conversation keyword by querying the tag number of the conversation tag in the tag dictionary.
  • a conversation word vector is generated based on the first dimension value and the second dimension value.
  • the method for generating the above-mentioned conversational word vector may also be: obtaining the parameter values of the conversation keyword in multiple parts of speech dimensions, generating a multi-dimensional vector, and correspondingly, obtaining the conversation label in multiple parts of speech dimensions.
  • the parameter value can also generate a multi-dimensional vector about the conversation tag, and merge the multi-dimensional vector of the conversation keyword with the multi-dimensional vector of the tag key to obtain the above-mentioned conversation word vector.
  • the terminal device may be configured with an emotion recognition network, and the terminal device imports the emotion recognition network in sequence according to the appearance order of each session keyword, and imports the preset end mark after all the session keywords are input.
  • the emotion recognition network outputs the emotional feature value corresponding to the above-mentioned conversational text, that is, the speech signal.
  • the aforementioned emotional feature value may include scores in multiple emotional dimensions, such as an emotional magnitude dimension and a positive degree dimension.
  • the generated personality analysis report of the target user is stored in the blockchain network, and the data information can be shared between different platforms through the storage of the blockchain, and the data can also be prevented from being tampered with.
  • Blockchain is a new application mode of computer technology such as distributed data storage, point-to-point transmission, consensus mechanism, and encryption algorithm.
  • Blockchain essentially a decentralized database, is a series of data blocks associated with cryptographic methods. Each data block contains a batch of network transaction information for verification. The validity of the information (anti-counterfeiting) and the generation of the next block.
  • the blockchain can include the underlying platform of the blockchain, the platform product service layer, and the application service layer.
  • the terminal device can generate a user portrait of the target user according to the emotional characteristics corresponding to all conversational content, determine the probability score corresponding to each personality type, and finally select the personality type with the highest probability score as the personality type of the target user. And generate the personality analysis report of the target user mentioned above.
  • the terminal device can also record the probability scores of all personality types in the personality analysis report, so that the interview manager can determine the potential personality characteristics of the target user based on the personality analysis report, which improves the richness of the content of the personality analysis report.
  • the method for generating a user report collects the voice signal of the target user during a conversation with the target user, converts the voice signal into the corresponding conversation text, and responds to the conversation.
  • the text is semantically analyzed to obtain the corresponding conversational content collection, and based on the conversational word vector of each conversation keyword in the conversational content collection, the emotional feature value corresponding to the voice signal is generated, and based on the emotional feature value of all voice signals, the personality of the target user is determined Type, and generate a personality analysis report on the target user, so that the target user’s language can be used to determine the personality during the conversation with the target user, and the purpose of automatically outputting the analysis report is realized.
  • this embodiment does not rely on the interviewer or the conversation object to manually fill in or subjectively judge, and does not require the user to spend extra time writing a personality analysis report on the target user, thereby greatly reducing user operations, and
  • the above process can determine the emotional characteristic value through the voice signals at different stages in the conversation process, instead of using a single utterance or sentence to judge the personality, thereby improving the accuracy of the personality analysis report.
  • Fig. 2 shows a specific implementation flow chart of a method S103 for generating a user report provided by the second embodiment of the present application.
  • S103 in a method for generating a user report provided in this embodiment includes: S1031 to S1036, which are detailed as follows:
  • the obtaining the conversation word vector corresponding to each of the conversation keywords in the conversation content set, and determining the emotional feature value corresponding to the voice signal based on each of the conversation word vectors includes:
  • the terminal device is configured with a knowledge graph, which contains multiple knowledge nodes, and there are corresponding association relationships between different knowledge nodes, thereby forming a network connected by multiple knowledge nodes, that is, the above Knowledge graph.
  • the terminal device can determine the knowledge node associated with the session keyword on the above-mentioned knowledge graph, and identify other nodes adjacent to the associated knowledge node, that is, other knowledge nodes with an association relationship, as the associated entity of the session keyword.
  • the terminal device may determine the weighted weight of the above-mentioned concern entity according to the confidence of the association relationship between the knowledge node associated with the session keyword and the associated entity.
  • the word concept vector of the conversation keyword is generated according to the weighted weight of all the associated entities.
  • c(t) is the above-mentioned word concept vector
  • g(t) is the total number of associated entities contained in the session keyword
  • c k is the word vector of the k-th associated entity of the above-mentioned session keyword
  • w k is the above-mentioned The weighted weight of the k-th associated entity of the session keyword.
  • the word concept vector after the word concept vector corresponding to the conversation keyword is calculated, the word concept vector can be converted into a word feature vector by linear change, and the specific conversion method can be:
  • the conversation text may contain multiple conversation sentences.
  • the terminal device may divide the conversation keywords based on the conversation sentences described by the conversation keywords to obtain multiple conversation keyword groups, and all conversation keywords in each conversation keyword group correspond to the same conversation sentence.
  • the terminal device can encapsulate the word concept vectors belonging to the same conversation sentence to generate a sentence probability vector corresponding to the conversation sentence.
  • the above-mentioned dialogue update vector sentence is used to characterize the emotional characteristics of the conversation sentence in this application. Therefore, the terminal device can separately import the sentence concept vector of each conversation sentence into the first attention algorithm to obtain the dialogue update vector.
  • S1035 encapsulate the sentence concept vectors of all conversation sentences of the conversation text, generate the conversation concept vectors of the conversation text, and import the conversation concept vectors into the second attention model to generate the conversation Text concept vector for text.
  • the terminal device can determine the overall emotional characteristics of the entire conversation text according to the contextual connections between different sentences. Therefore, the sentence concept vectors of all conversational sentences can be encapsulated to obtain the conversation concept vector, and the conversation concept vector can be imported into the aforementioned second attention model to obtain the text concept vector.
  • the text approximate vector can be specifically expressed as:
  • the emotional feature value is determined according to the dialogue update vector and the text concept vector.
  • the terminal device can import the dialogue update vector and the text concept vector into the third attention model to obtain the emotion concept vector corresponding to the conversation text.
  • the emotional concept vector can be specifically:
  • R i is the above-mentioned emotional concept vector
  • Update the vector for the dialogue The terminal device can import the foregoing emotional concept vector into a preset pooling layer, perform emotional feature extraction, and obtain the emotional feature value corresponding to the foregoing emotional concept vector.
  • the pooling layer can be expressed as:
  • p is the above-mentioned emotional feature value
  • W 3 ⁇ R d*q , b 3 ⁇ R q represents the model parameter
  • q represents the number of classes.
  • the conversational content is extended by obtaining the associated entities of the conversation keywords, and the conversation update vector based on a single sentence and the text concept vector based on all sentences are determined respectively to determine the emotional feature value of the target user. Determine the user's emotional characteristics from multiple dimensions, thereby improving the accuracy of the emotional characteristics.
  • Fig. 3 shows a specific implementation flow chart of a method S1031 for generating a user report provided by the third embodiment of the present application.
  • S1031 in a method for generating a user report provided in this embodiment includes: S301 to S303, which are detailed as follows:
  • the determining the associated entity of each of the session keywords in the preset knowledge graph and obtaining the weighted weight corresponding to each of the associated entities includes:
  • the association confidence between the different associated nodes can be determined. For example, if two knowledge nodes have a co-occurrence relationship in most of the text (the co-occurrence relationship is that multiple knowledge nodes appear in the same sentence at the same time), the correlation between the above-mentioned knowledge nodes has a higher confidence; , If two knowledge nodes only have a co-occurrence relationship in a small amount of text, the confidence of the association between the above-mentioned knowledge nodes is low. According to the confidence of the association between the knowledge node and the associated entity associated with the session keyword, the above-mentioned association strength factor can be obtained.
  • the terminal device may include a conversion algorithm of the correlation strength factor, and import the correlation confidence corresponding to the associated entity into the conversion algorithm to generate the aforementioned correlation strength factor.
  • the terminal device can be equipped with an emotion measurement algorithm, which can convert words into a computer-recognizable emotion intensity factor.
  • the terminal device can import the associated entity into the aforementioned emotion measurement algorithm, and output the emotion intensity factor corresponding to the associated entity.
  • a weighted weight of the associated entity is constructed based on the emotional intensity factor and the associated intensity factor.
  • the terminal device can generate the weighted weight of the associated entity according to the emotion factor and the correlation strength factor, and the weighted weight includes the closeness of the association with the session keyword and the emotional feature, which is convenient for the subsequent emotional feature value.
  • the weighting weight may specifically be:
  • w k is the weighted weight corresponding to the k-th associated entity
  • rel k is the associated intensity factor corresponding to the k-th associated entity
  • aff k is the emotional intensity factor of the k-th associated entity
  • ⁇ k is the k-th associated entity.
  • the weighted weight corresponding to the associated entity when calculating the emotional feature value is determined.
  • FIG. 4 shows a specific implementation flowchart of a method S301 for generating a user report provided by the fourth embodiment of the present application.
  • a method S301 for generating a user report provided by this embodiment includes: S3011 to S3013, which are detailed as follows:
  • the obtaining the correlation strength factor between each of the associated entities and the session keywords includes:
  • the association confidence between the associated entity and the session keyword is determined.
  • the confidence level of the association relationship between each knowledge node may be recorded in the knowledge graph, and the terminal device marks the session keyword and the associated entity in the knowledge graph to determine the confidence level of the association relationship between the two.
  • the confidence of the correlation is identified as the confidence of the correlation between the above two. Among them, the more the number of co-occurrences between the associated entity and the session keyword, the higher the corresponding confidence of the association; conversely, the less the number of co-occurrences between the two, the lower the confidence of the corresponding association.
  • the conversation sentence associated with the conversation keyword is imported into a preset pooling layer, a sentence vector of the conversation sentence associated with each of the conversation keywords is generated, and the conversation keyword is determined based on the sentence vector
  • the conversational text vector of the segment; the conversational text vector is specifically:
  • CR(X i ) is the conversation text vector of the conversation keyword, and the conversation text number where the conversation keyword is located is i; Is the sentence vector of the conversation sentence where the conversation keyword is located, and the sentence number of the conversation sentence in the conversation text is j; the M is a preset correlation coefficient.
  • the conversation text contains multiple conversation sentences.
  • the terminal device can be configured with the number of contacts M.
  • the maximum number of sessions that need to be uniformly recognized can be determined based on the correlation coefficient M.
  • the correlation strength factor is calculated based on the conversation text vector and the correlation confidence; the correlation strength factor is specifically:
  • rel k is the correlation strength factor of the k-th session keyword
  • c k is the correlation confidence of the k-th associated entity of the session keyword
  • max-min(s k ) is the session keyword
  • the terminal device may include multiple different emotion measurement algorithms, and the emotional parameter values of the associated entities determined by different emotion measurement algorithms may be different.
  • the terminal device may determine the related entities according to different emotion measurement algorithms.
  • the emotional range that is, the above-mentioned max-min(s k ), the emotional range of the related entity and the above two parameters are imported into a preset correlation strength conversion algorithm to obtain the correlation strength value of the related entity.
  • the association between different conversational sentences in the entire conversational text is considered, so that the accuracy of the correlation strength factor can be improved.
  • FIG. 5 shows a specific implementation flowchart of a method S302 for generating a user report provided by the fifth embodiment of the present application.
  • a method S302 for generating a user report provided in this embodiment includes: S3021 to S3023, which are detailed as follows:
  • the determining the emotion intensity factor of each associated entity based on a preset emotion measurement algorithm includes:
  • the terminal device can determine the emotional intensity factor in different ways according to the different emotional attributes of the associated entities. For example, if the demonstrative pronoun "I" does not contain emotional characteristics, the corresponding emotional attribute is a non-affective type; the adjective "great” contains a certain degree of emotional characteristics, and the corresponding emotional attribute is an emotional type. Based on this, the terminal device can identify the emotional attribute of each associated entity. If the emotional attribute of the associated entity is a non-emotional type, the operation of S3022 is performed; conversely, if the emotional type of the associated entity is an emotional type, the operation of S3023 is performed.
  • the affective intensity factor is configured as a preset default value.
  • the terminal device can configure all non-emotional related entities with a fixed value of emotional intensity factor, and the value of the emotional intensity factor can be configured to be 0.5.
  • the emotion intensity factor of the conversation keyword is calculated through a preset emotion conversion algorithm; the emotion intensity factor is specifically:
  • aff k is the emotional intensity factor of the k-th associated entity
  • VAD(c k ) is the positive emotional score of the k-th associated entity
  • A(c k ) is the k-th associated entity The emotional magnitude score of the associated entity.
  • the emotional intensity factor is specifically composed of two different emotional dimensions, divided into a positive emotional dimension and an emotional amplitude dimension, where the positive emotional dimension is specifically used to identify whether the corresponding emotional feature of the entity is positive, if the degree of positive
  • the positive emotional score corresponding to "laugh” is positive, while the positive emotional score corresponding to "cry” is negative, and the emotional positive score corresponding to "optimism” is higher than the emotional positive score of "acceptance”
  • the emotional amplitude score is used to identify the emotional fluctuation amplitude of the entity, for example, the emotional amplitude score of "laugh” will be lower than the emotional amplitude score of "large”.
  • the terminal device can determine the corresponding emotional scores of each associated entity in the above two dimensions through a preset emotional measurement algorithm, and obtain the corresponding emotional intensity factor. among them, It is based on the norm of 2.
  • the emotion attribute of the associated entity is identified, and the calculation method of the corresponding emotion intensity factor is selected, thereby improving the accuracy of the emotion intensity factor.
  • FIG. 6 shows a specific implementation flow chart of a method S1034 for generating a user report provided by the sixth embodiment of the present application.
  • S1034 in a method for generating a user report provided in this embodiment includes: S601 to S603, which are detailed as follows:
  • the respectively importing the sentence concept vector of each of the conversational sentences into the first attention algorithm to obtain the dialogue update vector of each of the conversational sentences includes:
  • the sentence concept vector of the conversation sentence is linearly changed to obtain a linear vector containing h endpoints; where h is the preset number of endpoints.
  • the terminal device can perform a linear transformation on the sentence concept of the conversational sentence, and project the sentence concept vector into h more endpoints to obtain a linear vector about the sentence concept vector.
  • the above-mentioned value of h may be a preset linear transformation parameter of the first attention algorithm, or may be changed based on the text amount of the conversational text.
  • the linear vector is imported into the multi-head self-attention layer of the first attention algorithm to obtain the attention vector of the conversation sentence; the attention vector is specifically:
  • the terminal device may import the linear vector obtained by the foregoing calculation to the foregoing multi-head attention layer, and the attention layer includes three nodes.
  • the terminal device can calculate the product between the linear vector and the transposition of the linear vector, and process the multiplied vector through the softmax function, and finally multiply the linear vector again to achieve triple iteration to improve feature extraction Accuracy.
  • a dialogue update vector of the conversation sentence is generated based on the attention vector; the dialogue update vector is specifically:
  • W 1 , W 2 , b 1 and b 2 are model parameters of the first attention model.
  • the terminal device can import the generated attention vector into the feedforward layer of the first shocking network to obtain the dialogue update vector corresponding to the conversation sentence.
  • the feedforward layer can first perform an inverse linear transformation on the attention vector, transform the attention vector containing multiple endpoints to a vector containing a single endpoint, and then perform subsequent operations.
  • the interviewer after adding the emotion judgment method based on the new NLP transformer, the interviewer can quickly judge certain personality characteristics of the candidate through the candidate's answer, and give necessary and reasonable follow-up questions.
  • AI intelligent interview can judge the emotion of the candidate based on the answer to the candidate, and then judge the personality of the candidate.
  • the interviewer can analyze the candidate's personality characteristics after the interview is completed, and use it as the basis for selecting candidates.
  • FIG. 7 shows a specific implementation flowchart of a method S104 for generating a user report provided by the seventh embodiment of the present application.
  • S104 in a method for generating a user report provided in this embodiment includes: S1041-S1043, which are detailed as follows:
  • the generating a personality analysis report of the target user based on the emotional characteristic values of all voice signals includes:
  • an emotion waveform diagram of the target user is generated according to the emotion characteristic value of each voice signal.
  • the terminal device can mark each emotional feature value on a preset coordinate axis according to each conversation text, that is, the generation sequence of the voice signal, and connect each emotional feature value in turn to obtain the target user in the entire conversation process. Corresponding sentiment waveform in.
  • the emotion waveform diagram is matched with the standard personality waveform diagrams of each candidate personality to determine the user personality of the target user.
  • the terminal device can calculate the deviation value between the standard waveform diagrams of each candidate personality of the emotion waveform diagram, and calculate the matching degree between the target user and each candidate personality based on the inverse of the deviation value, and select the matching degree The highest candidate personality is taken as the user personality of the target user.
  • multiple candidate personalities with a matching degree greater than a preset matching threshold can also be selected as the user personality of the target user.
  • the personality analysis report is obtained based on the user's personality.
  • the terminal device can obtain the standard language segment corresponding to each user's personality, and generate a personality analysis report based on the above-mentioned standard language segment.
  • the personality analysis report is generated, thereby improving the generation efficiency of the personality analysis report.
  • FIG. 8 shows a structural block diagram of a device for generating a user report according to an embodiment of the present application, and each unit included in the device for generating a user report is used to execute each step in the embodiment corresponding to FIG. 1.
  • FIG. 8 shows a structural block diagram of a device for generating a user report according to an embodiment of the present application, and each unit included in the device for generating a user report is used to execute each step in the embodiment corresponding to FIG. 1.
  • FIG. 8 shows a structural block diagram of a device for generating a user report according to an embodiment of the present application, and each unit included in the device for generating a user report is used to execute each step in the embodiment corresponding to FIG. 1.
  • FIG. 8 shows a structural block diagram of a device for generating a user report according to an embodiment of the present application, and each unit included in the device for generating a user report is used to execute each step in the embodiment corresponding to FIG. 1.
  • FIG. 8 shows a structural block diagram of a device for
  • the device for generating the user report includes:
  • the conversation text obtaining unit 81 is configured to obtain multiple voice signals generated by the target user during a conversation, and convert each of the voice signals into corresponding conversation text;
  • the conversation content collection generating unit 82 is configured to perform semantic analysis on the conversation text to obtain the conversation keywords corresponding to the conversation text and the conversation tags corresponding to each of the keywords, and generate a conversation content collection;
  • the emotion feature value determining unit 83 is configured to obtain the conversation word vector corresponding to each of the conversation keywords in the conversation content set, and determine the emotion feature value corresponding to the voice signal based on each of the conversation word vectors;
  • the personality analysis report generating unit 84 is configured to generate a personality analysis report of the target user based on the emotional feature values of all voice signals.
  • the emotional feature value determining unit 83 includes:
  • a weighted weight determining unit configured to determine the associated entity of each of the session keywords in the preset knowledge graph, and obtain the weighted weight corresponding to each of the associated entities
  • a word concept vector generating unit configured to generate the word concept vector of the conversation keyword according to the weighted weights of all the associated entities
  • the sentence concept vector generating unit is used to encapsulate all the word concept vectors belonging to the same conversation sentence based on the conversation sentence to which each of the conversation keywords belongs to generate the sentence concept vector of the conversation sentence; the conversation sentence Is obtained after sentence division of the conversation text;
  • a dialogue update vector generating unit configured to respectively import the sentence concept vector of each of the conversational sentences into the first attention algorithm to obtain the dialogue update vector of each of the conversational sentences;
  • the text concept vector generating unit is used to encapsulate the sentence concept vectors of all conversation sentences of the conversation text, generate the conversation concept vectors of the conversation text, and import the conversation concept vectors into the second attention model, Generating a text concept vector of the conversation text;
  • the emotional feature value calculation unit is configured to determine the emotional feature value according to the dialogue update vector and the text concept vector.
  • the weighting weight determining unit includes:
  • An association strength factor determination unit configured to obtain an association strength factor between each of the associated entities and the session keywords
  • the emotion intensity factor determination unit is configured to determine the emotion intensity factor of each associated entity based on a preset emotion measurement algorithm
  • the weighted weight calculation unit is configured to construct the weighted weight of the associated entity based on the emotional intensity factor and the associated intensity factor.
  • the correlation strength factor determination unit includes:
  • An association confidence degree determining unit configured to determine the association confidence degree between the associated entity and the session keyword based on the knowledge graph
  • the conversational text vector determining unit is used to import the conversational sentences associated with the conversational keywords into the preset pooling layer, generate the sentence vectors of the conversational sentences associated with each of the conversational keywords, and determine all the conversational sentences based on the sentence vectors.
  • the conversational text vector of the segment where the conversational keywords are located; the conversational text vector is specifically:
  • CR(X i ) is the conversation text vector of the conversation keyword, and the conversation text number where the conversation keyword is located is i; Is the sentence vector of the conversation sentence where the conversation keyword is located, the sentence number of the conversation sentence in the conversation text is j; the M is a preset correlation coefficient;
  • the correlation strength factor calculation unit is configured to calculate the correlation strength factor based on the conversational text vector and the correlation confidence; the correlation strength factor is specifically:
  • rel k is the correlation strength factor of the k-th session keyword
  • c k is the correlation confidence of the k-th associated entity of the session keyword
  • max-min(s k ) is the session keyword
  • the emotion intensity factor determination unit includes:
  • a time distribution diagram generating unit configured to generate a request time distribution diagram for the service type according to the request initiation time included in all the service requests;
  • the emotional attribute recognition unit is used to recognize the emotional attribute of the associated entity
  • a non-emotional type processing unit configured to configure the emotional intensity factor to a preset default value if the emotional attribute of the associated entity is a non-emotional type
  • the emotion type processing unit is configured to calculate the emotion intensity factor of the conversation keyword by using a preset emotion conversion algorithm if the emotion attribute of the associated entity is an emotion type; the emotion intensity factor is specifically :
  • aff k is the emotional intensity factor of the k-th associated entity
  • VAD(c k ) is the positive emotional score of the k-th associated entity
  • A(c k ) is the k-th associated entity The emotional magnitude score of the associated entity.
  • the dialog update vector generating unit includes:
  • the linear vector generating unit is used to linearly change the sentence concept vector of the conversation sentence to obtain a linear vector containing h endpoints; wherein, the h is the preset number of endpoints;
  • the attention vector generating unit is configured to import the linear vector into the multi-head self-attention layer of the first attention algorithm to obtain the attention vector of the conversation sentence; the attention vector is specifically:
  • the dialogue update vector determining unit is configured to generate a dialogue update vector of the conversation sentence based on the attention vector; the dialogue update vector is specifically:
  • W 1 , W 2 , b 1 and b 2 are model parameters of the first attention model.
  • the personality analysis report generating unit 84 includes:
  • An emotion waveform diagram generating unit configured to generate the emotion waveform diagram of the target user according to the emotion characteristic value of each of the voice signals
  • a user personality determination unit configured to match the emotion waveform diagram with the standard personality waveform diagrams of each candidate personality to determine the user personality of the target user;
  • the personality analysis report output unit is configured to obtain the personality analysis report based on the user's personality.
  • the user report generation device provided by the embodiment of the present application also does not rely on the interviewer or the conversation object to manually fill in or subjectively judge, and does not require the user to spend extra time writing a personality analysis report on the target user, thereby greatly reducing user operations.
  • the above process can determine the emotional characteristic value through the voice signals at different stages in the conversation process, instead of using a single utterance or sentence to judge the personality, so that the accuracy of the personality analysis report can be improved.
  • FIG. 9 is a schematic diagram of a terminal device provided by another embodiment of the present application.
  • the terminal device 9 of this embodiment includes: a processor 90, a memory 91, and a computer program 92 stored in the memory 91 and running on the processor 90, such as a program for generating user reports .
  • the processor 90 executes the computer program 92, the steps in the foregoing method for generating user reports are implemented, such as S101 to S104 shown in FIG. 1.
  • the processor 90 executes the computer program 92, the functions of the units in the foregoing device embodiments, such as the functions of the modules 81 to 84 shown in FIG. 8, are realized.
  • the computer program 92 may be divided into one or more units, and the one or more units are stored in the memory 91 and executed by the processor 90 to complete the application.
  • the one or more units may be a series of computer program instruction segments capable of completing specific functions, and the instruction segments are used to describe the execution process of the computer program 92 in the terminal device 9.
  • the computer program 92 may be divided into a conversation text acquisition unit, a conversation content collection generation unit, an emotional feature value determination unit, and a personality analysis report generation unit, and the specific functions of each unit are as described above.
  • the terminal device 9 may be a computing device such as a desktop computer, a notebook, a palmtop computer, and a cloud terminal device.
  • the terminal device may include, but is not limited to, a processor 90 and a memory 91.
  • FIG. 9 is only an example of the terminal device 9 and does not constitute a limitation on the terminal device 9. It may include more or less components than shown in the figure, or a combination of certain components, or different components.
  • the terminal device may also include input and output devices, network access devices, buses, and so on.
  • the so-called processor 90 may be a central processing unit (Central Processing Unit, CPU), other general-purpose processors, digital signal processors (Digital Signal Processor, DSP), application specific integrated circuits (Application Specific Integrated Circuit, ASIC), Ready-made programmable gate array (Field-Programmable Gate Array, FPGA) or other programmable logic devices, discrete gates or transistor logic devices, discrete hardware components, etc.
  • the general-purpose processor may be a microprocessor or the processor may also be any conventional processor or the like.
  • the memory 91 may be an internal storage unit of the terminal device 9, for example, a hard disk or a memory of the terminal device 9.
  • the memory 91 may also be an external storage device of the terminal device 9, such as a plug-in hard disk equipped on the terminal device 9, a smart memory card (Smart Media Card, SMC), or a Secure Digital (SD). Card, Flash Card, etc.
  • the memory 91 may also include both an internal storage unit of the terminal device 9 and an external storage device.
  • the memory 91 is used to store the computer program and other programs and data required by the terminal device.
  • the memory 91 can also be used to temporarily store data that has been output or will be output.
  • the embodiments of the present application also provide a computer-readable storage medium, where the computer-readable storage medium stores a computer program, and when the computer program is executed by a processor, the steps in each of the foregoing method embodiments can be realized.
  • the computer-readable storage medium may be non-volatile or volatile.
  • the computer-readable storage medium includes: U disk, mobile hard disk, read-only memory (ROM, Read-Only Memory), random access memory (RAM, Random Access Memory), magnetic disks or optical disks, etc., which can store program codes Medium.
  • the functional units in the various embodiments of the present application may be integrated into one processing unit, or each unit may exist alone physically, or two or more units may be integrated into one unit.
  • the above-mentioned integrated unit can be implemented in the form of hardware or software functional unit.

Landscapes

  • Engineering & Computer Science (AREA)
  • Business, Economics & Management (AREA)
  • Human Resources & Organizations (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Entrepreneurship & Innovation (AREA)
  • Strategic Management (AREA)
  • Data Mining & Analysis (AREA)
  • Economics (AREA)
  • General Engineering & Computer Science (AREA)
  • Marketing (AREA)
  • Computational Linguistics (AREA)
  • Tourism & Hospitality (AREA)
  • Quality & Reliability (AREA)
  • Databases & Information Systems (AREA)
  • Operations Research (AREA)
  • Artificial Intelligence (AREA)
  • Educational Administration (AREA)
  • Mathematical Physics (AREA)
  • General Business, Economics & Management (AREA)
  • Development Economics (AREA)
  • Human Computer Interaction (AREA)
  • Game Theory and Decision Science (AREA)
  • Health & Medical Sciences (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • General Health & Medical Sciences (AREA)
  • Machine Translation (AREA)

Abstract

A user report generating method and device, comprising: acquiring a plurality of voice signals generated by a target user during a session, and converting the voice signals into a corresponding session text (S101); implementing semantic analysis of the session text to obtain session keywords corresponding to the session text and a session label corresponding to the keywords, and generating a session content set (S102); acquiring a session word vector corresponding to the session keywords in the session content set and, on the basis of the session word vectors, determining an emotion feature value corresponding to the voice signals (S103); and, on the basis of the emotion feature values of all of the voice signals, generating a personality analysis report of the target user (S104). The present method does not require the user to spend additional time writing a personality analysis report relating to the target user and can thereby greatly reduce user operations, and determines emotion feature values based on the voice signals at different stages during the session, increasing the accuracy of the personality analysis report.

Description

一种用户报告的生成方法及终端设备Method for generating user report and terminal equipment
本申请申明享有2020年05月14日递交的申请号为202010406546.2、名称为“一种用户报告的生成方法及终端设备”中国专利申请的优先权,该中国专利申请的整体内容以参考的方式结合在本申请中。This application affirms that it enjoys the priority of the Chinese patent application with the application number 202010406546.2 and the name "A method and terminal device for generating user reports" filed on May 14, 2020. The entire content of the Chinese patent application is incorporated by reference In this application.
技术领域Technical field
本申请属于人工智能技术领域,尤其涉及一种用户报告的生成方法及终端设备。This application belongs to the field of artificial intelligence technology, and in particular relates to a method for generating a user report and a terminal device.
背景技术Background technique
随着企业规模的不断扩大,员工的数量也随之增加,因而发明人意识到如何能够高效地对面试人员进行筛选,确定面试人员的性格特性等,则直接影响面试效率以及影响决策速度,而发明人发现通过用户分析报告,则能够快速了解面试用户的情况,提高大大提高面试效率。With the continuous expansion of the scale of the enterprise, the number of employees has also increased. Therefore, the inventor realized how to efficiently screen interviewers and determine the personality characteristics of the interviewers, which directly affects the efficiency of the interview and the speed of decision-making. The inventor found that through the user analysis report, the situation of the interviewed user can be quickly understood, and the interview efficiency can be greatly improved.
现有的用户分析报告生成技术,发明人发现主要是依赖面试官对面试人员进行性格分析,通过收集面试人员对于预设题目的回答,主观认定面试人员的性格特征,并生成用户分析报告,发明人意识到现有的用户分析报告依靠人力完成,生成效率较低,从而降低了人员管理效率。With the existing user analysis report generation technology, the inventor found that it mainly relies on the interviewer to analyze the interviewer’s personality. By collecting the interviewer’s answers to preset questions, the interviewer’s personality characteristics are subjectively determined, and the user analysis report is generated. People realize that the existing user analysis reports are completed by manpower, and the generation efficiency is low, thereby reducing the efficiency of personnel management.
技术问题technical problem
有鉴于此,本申请实施例提供了一种用户报告的生成方法及终端设备,以解决现有的用户报告的生成技术,依靠人力完成,报告生成效率较低,从而降低了人员管理效率的问题。In view of this, the embodiment of the present application provides a method and terminal device for generating user reports to solve the problem of existing user report generation technology, which relies on manpower to complete, and the report generation efficiency is low, thereby reducing the efficiency of personnel management. .
技术解决方案Technical solutions
本申请实施例的第一方面提供了一种用户报告的生成方法,包括:The first aspect of the embodiments of the present application provides a method for generating a user report, including:
获取目标用户在会话过程中产生的多个语音信号,并将各个所述语音信号转换为对应的会话文本;Acquiring multiple voice signals generated by the target user during the conversation, and converting each of the voice signals into corresponding conversation text;
对所述会话文本进行语义分析,得到所述会话文本对应的会话关键词以及各个所述关键词对应的会话标签,生成会话内容集合;Performing semantic analysis on the conversation text, obtaining conversation keywords corresponding to the conversation text and conversation tags corresponding to each of the keywords, and generating a conversation content set;
获得所述会话内容集合内的各个所述会话关键词对应的会话词向量,并基于各个所述会话词向量确定所述语音信号对应的情感特征值;Obtaining a conversation word vector corresponding to each of the conversation keywords in the conversation content set, and determining an emotional feature value corresponding to the voice signal based on each of the conversation word vectors;
基于所有语音信号的所述情感特征值,生成所述目标用户的性格分析报告。Based on the emotional feature values of all voice signals, a personality analysis report of the target user is generated.
有益效果Beneficial effect
本申请实施例通过在与目标用户进行会话的过程中,采集目标用户的语音信号,并将语音信号转换为对应的会话文本,并对会话文本进行语义分析得到对应的会话内容合集,并基于会话内容合集内各个会话关键词的会话词向量,生成语音信号对应的情感特征值,并基于所有语音信号的情感特征值,确定目标用户的性格类型,并生成关于目标用户的性格分析报告,从而能够在与目标用户会话的过程中,通过目标用户的语言确定性格,实现了自动输出分析报告的目的。与现有的用户报告技术相比,本实施例不依赖面试人员或会话对象进行手动填写或主观判断,无需用户花费额外时间撰写关于目标用户的性格分析报告,从而能够大大减少用户的操作,并且上述的过程可以通过在会话过程中不同阶段的语音信号,确定情感特征值,并非以单一的话语或语句进行性格判断,从而能够提高性格分析报告的准确性。The embodiment of the application collects the voice signal of the target user during the conversation with the target user, converts the voice signal into the corresponding conversation text, and performs semantic analysis on the conversation text to obtain the corresponding conversation content collection, which is based on the conversation The conversation word vector of each conversation keyword in the content collection generates the emotional feature value corresponding to the voice signal, and based on the emotional feature value of all voice signals, determines the personality type of the target user, and generates a personality analysis report about the target user, so as to be able to During the conversation with the target user, the target user’s language is used to determine the personality, which achieves the purpose of automatically outputting an analysis report. Compared with the existing user report technology, this embodiment does not rely on the interviewer or the conversation object to manually fill in or subjectively judge, and does not require the user to spend extra time writing a personality analysis report on the target user, thereby greatly reducing user operations, and The above process can determine the emotional characteristic value through the voice signals at different stages in the conversation process, instead of using a single utterance or sentence to judge the personality, thereby improving the accuracy of the personality analysis report.
附图说明Description of the drawings
图1是本申请第一实施例提供的一种用户报告的生成方法的实现流程图;FIG. 1 is an implementation flowchart of a method for generating a user report provided by the first embodiment of the present application;
图2是本申请第二实施例提供的一种用户报告的生成方法S103具体实现流程图;2 is a specific implementation flow chart of a method S103 for generating a user report provided by the second embodiment of the present application;
图3是本申请第三实施例提供的一种用户报告的生成方法S1031具体实现流程图;FIG. 3 is a specific implementation flow chart of a method S1031 for generating a user report provided by the third embodiment of the present application;
图4是本申请第四实施例提供的一种用户报告的生成方法S301具体实现流程图;FIG. 4 is a specific implementation flow chart of a method S301 for generating a user report provided by the fourth embodiment of the present application;
图5是本申请第五实施例提供的一种用户报告的生成方法S302具体实现流程图;FIG. 5 is a specific implementation flowchart of a method S302 for generating a user report provided by the fifth embodiment of the present application;
图6是本申请第六实施例提供的一种用户报告的生成方法S1034具体实现流程图;6 is a specific implementation flow chart of a method S1034 for generating a user report provided by the sixth embodiment of the present application;
图7是本申请第七实施例提供的一种用户报告的生成方法S104具体实现流程图;FIG. 7 is a specific implementation flowchart of a method S104 for generating a user report provided by the seventh embodiment of the present application;
图8是本申请一实施例提供的一种用户报告的生成设备的结构框图;FIG. 8 is a structural block diagram of a device for generating a user report according to an embodiment of the present application;
图9是本申请另一实施例提供的一种终端设备的示意图。FIG. 9 is a schematic diagram of a terminal device provided by another embodiment of the present application.
本发明的实施方式Embodiments of the present invention
在本申请实施例中,流程的执行主体为终端设备,该终端设备包括但不限于:服务器、计算机、智能手机以及平板电脑等能够执行用户报告的生成方法的设备。图1示出了本申请第一实施例提供的用户报告的生成方法的实现流程图,详述如下:In the embodiment of the present application, the execution subject of the process is a terminal device, which includes but is not limited to: servers, computers, smart phones, tablet computers, and other devices capable of executing the method for generating user reports. Fig. 1 shows an implementation flow chart of the method for generating a user report provided by the first embodiment of the present application, and the details are as follows:
在S101中,获取目标用户在会话过程中产生的多个语音信号,并将各个所述语音信号转换为对应的会话文本。In S101, multiple voice signals generated by a target user during a conversation are acquired, and each of the voice signals is converted into a corresponding conversation text.
在本实施例中,终端设备可以为用户数据库的服务器,该服务器可以通过通信链路与分布式麦克风模块相连,该通信链路可以为有线通信的实体链路,也可以为通过局域网或互联网等方式建立的虚拟链路。该麦克风模块可以与终端设备部署于同一区域,也可以分布式部署于各个面试场所,用于采集面试过程中产生的语音信号。In this embodiment, the terminal device may be a server of the user database, and the server may be connected to the distributed microphone module through a communication link. The communication link may be a physical link for wired communication, or may be through a local area network or the Internet. The virtual link established by the method. The microphone module can be deployed in the same area as the terminal device, or distributed in various interview locations to collect voice signals generated during the interview.
可选地,在本实施例中,该麦克风模块具体为一麦克风阵列,麦克风阵列内包含有多个麦克风装置,麦克风阵列在采集语音信号的过程中,可以从多个不同角度获取当前面试场景的语音信号,并通过多个语音信号进行滤波整形,得到用于进行语音识别的目标信号。由一定数量的麦克风组成麦克风阵列采集语音信号,用来对声场的空间特性进行采样并处理的系统,应用于在面试环境的复杂环境下,能够有效解决噪音、混响、人声干扰、回声等问题,提高了语音信号采集的信号质量,从而在后续输出文字信息时,能够提高文字信息转换的成功率。Optionally, in this embodiment, the microphone module is specifically a microphone array. The microphone array contains multiple microphone devices. During the process of collecting voice signals, the microphone array can obtain information about the current interview scene from multiple different angles. The voice signal is filtered and shaped through multiple voice signals to obtain a target signal for voice recognition. A microphone array composed of a certain number of microphones to collect voice signals to sample and process the spatial characteristics of the sound field. It is used in the complex environment of the interview environment and can effectively solve noise, reverberation, vocal interference, echo, etc. The problem is to improve the signal quality of the voice signal collection, so that when the text information is subsequently output, the success rate of the text information conversion can be improved.
在本实施例中,终端设备可以设置有面试时间段,若终端设备检测到当前时刻到达预设的面试启动时刻,则开启麦克风模块,以通过麦克风模块获取当前所处的面试场景的语音信号。并且,在终端设备检测到当前时刻到达预设的面试结束时刻,则关闭麦克风模块,并将该面试时间段内的采集到的所有语音信号转换为文字信息。由于在会议过程中,用户发言并非连续性的,而是间断性的,终端设备可以配置有启动分贝值以及结束分贝值,在麦克风模块检测到当前面试场景的分贝值大于启动分贝值时,会开始采集语音信号,并在分贝值小于结束分贝值时,结束采集语音信号,将每个采集到的一段语音信号,作为会话过程中的一个会话段落,并为每个会话段落输出对应的会话文本,由于在面试过程中,目标用户与面试官之间基于问答过程产生多个会话段落,终端设备可以分别识别各个会话段落的情感特征值,并根据在整个会话过程中产生的所有会话文本,生成目标用户的性格分析报告。In this embodiment, the terminal device may be set with an interview time period. If the terminal device detects that the current time has reached the preset interview start time, the microphone module is turned on to obtain the voice signal of the current interview scene through the microphone module. In addition, when the terminal device detects that the current time reaches the preset interview end time, the microphone module is turned off, and all the voice signals collected during the interview time period are converted into text information. Since during the meeting, the user's speech is not continuous, but intermittent, the terminal device can be configured with a start decibel value and an end decibel value. When the microphone module detects that the decibel value of the current interview scene is greater than the start decibel value, it will Start to collect the voice signal, and when the decibel value is less than the end decibel value, end the collection of the voice signal, take each collected voice signal as a conversation paragraph in the conversation process, and output the corresponding conversation text for each conversation paragraph In the interview process, multiple conversational paragraphs are generated between the target user and the interviewer based on the question-and-answer process. The terminal device can recognize the emotional feature value of each conversational paragraph, and generate all the conversational texts generated during the entire conversation. Personality analysis report of target users.
可选地,在本实施例中,终端设备可以在接收到一段语音信号后,则执行文字信息的输出操作,并在检测到当前面试结束后(例如到达预设的面试结束时间或检测到预设的等待时长内没有接收到语音信号),基于所有采集到的语音信号所对应的文字信息,执行S102的操作,即采集操作与语音识别操作并行执行;终端设备也可以将当前会议的采集到的所有语音信息存储在数据库内,并在面试结束后,执行S102的操作。Optionally, in this embodiment, the terminal device may perform the output operation of text information after receiving a segment of voice signal, and after detecting the end of the current interview (for example, reaching the preset interview end time or detecting the pre-interview) If no voice signal is received within the waiting time), the operation of S102 is executed based on the text information corresponding to all the collected voice signals, that is, the collection operation is performed in parallel with the voice recognition operation; the terminal device can also collect the current conference All the voice information of is stored in the database, and after the interview is over, the operation of S102 is executed.
在本实施例中,终端设备可以设置有语音识别算法,终端设备可以通过语音识别算法 对语音信号进行解析,输出语音信号对应的文字信息,实现了语音识别的目的,自动记录面试内容,获得目标用户在会话过程中的会话文本。可选地,终端设备在进行语音识别的过程中,可以确定面试过程中使用的面试语种,并基于面试语种调整语音识别算法,从而提高识别的准确率。具体地,确定面试语种的方式可以为:获取参与面试的目标用户的用户信息,所述用户信息包含用户户籍或居住地址等信息;基于目标用户的户籍或居住地址,确定面试语种。In this embodiment, the terminal device may be provided with a voice recognition algorithm. The terminal device may parse the voice signal through the voice recognition algorithm and output text information corresponding to the voice signal. This achieves the purpose of voice recognition, automatically records the interview content, and obtains the target. The conversation text of the user during the conversation. Optionally, in the process of voice recognition, the terminal device can determine the interview language used in the interview process, and adjust the voice recognition algorithm based on the interview language, thereby improving the accuracy of recognition. Specifically, the manner of determining the language of the interview may be: obtaining user information of the target user participating in the interview, the user information including information such as the user's household registration or residential address; and determining the language of the interview based on the household register or residential address of the target user.
在一种可能的实现方式中,终端设备可以基于预设的最大语句数量,将会话文本划分为多个会话语段,每个会话语段包含的语句数量不大于预设的最大语句数量,在会话时长较长时,生成的会话文本的数量量较大,通过对会话文本进行划分,能够提高后续识别操作的效率,保证标记数量的稳定。当然,终端设备可以基于最大语句数量生成对应的语句选取框,基于该语句选取框在会话文本上进行遍历选取多个语句连续的会话语段,从而能够将每次识别的语句个数稳定,实现了识别参量的一致性。In a possible implementation, the terminal device can divide the conversation text into multiple conversation segments based on the preset maximum number of sentences, and each conversation segment contains no more than the preset maximum number of sentences. When the conversation duration is long, the amount of generated conversation text is relatively large. By dividing the conversation text, the efficiency of subsequent recognition operations can be improved, and the number of marks can be stabilized. Of course, the terminal device can generate a corresponding sentence selection box based on the maximum number of sentences, and traverse the conversation text based on the sentence selection box to select consecutive conversation segments of multiple sentences, thereby stabilizing the number of sentences recognized each time. The consistency of the identification parameters is improved.
在一种可能的实现方式中,将语音信号转换为会话文本的方式具体可以为:对语音信号进行解析,提取每一帧语音信号对应的波形特征和音调特征。将每一帧语音信号对应的波形特征和音调特征顺序输入训练完的语音识别模型中。该语音识别模型具体基于所有候选字符对应的标准波形以及音调波形训练得到,可以通过将每一帧的语音信号导入上述语音识别模型,可以计算得到与各个候选字符之间的相似度。选取相似度最高的候选字符作为该帧语音信号对应的文字,基于所有帧的文字,生成该语音信号对应的会话文本。In a possible implementation manner, the manner of converting the voice signal into conversational text may specifically be: parsing the voice signal, and extracting the waveform characteristics and pitch characteristics corresponding to each frame of the voice signal. The waveform characteristics and pitch characteristics corresponding to each frame of speech signal are sequentially input into the trained speech recognition model. The speech recognition model is specifically trained based on the standard waveforms and pitch waveforms corresponding to all candidate characters, and the similarity with each candidate character can be calculated by importing the speech signal of each frame into the aforementioned speech recognition model. The candidate character with the highest similarity is selected as the text corresponding to the speech signal of the frame, and the conversation text corresponding to the speech signal is generated based on the text of all frames.
在S102中,对所述会话文本进行语义分析,得到所述会话文本对应的会话关键词以及各个所述关键词对应的会话标签,生成会话内容集合。In S102, semantic analysis is performed on the conversation text to obtain conversation keywords corresponding to the conversation text and conversation tags corresponding to each of the keywords, and a conversation content set is generated.
在本实施例中,终端设备可以配置有语义识别算法,可以对会话文本进行语义分析,提取上述会话文本包含的会话关键词。该语义识别算法提取会话关键词的过程具体可以为:对会话文本进行词语划分,划分为包含若干字符的多个词组,每个词组至少包含一个字符,且不大于4个字符;终端设备对各个词组的词性进行识别,可以过滤与情感不相关的无效词组,举例性地,部分连接词与情感性格等分析的关联度较少,例如连接词“以及”、“和”以及“并”等,有部分助词“的”、“地”以及“得”等,终端设备对无效词组进行滤除后,则得到包含有用户情感相关的有效词组,将上述有效词组识别为会话关键词;可选地,终端设备存储有关键词典,判断上述有效词组是否在上述关键词典内,若存在,则识别该有效词组为会话关键词,反之,则识别该有效词组为无效词组。In this embodiment, the terminal device may be equipped with a semantic recognition algorithm, which can perform semantic analysis on the conversation text, and extract the conversation keywords contained in the aforementioned conversation text. The process of extracting conversation keywords by the semantic recognition algorithm can be specifically as follows: the conversation text is divided into words, divided into multiple phrases containing several characters, and each phrase contains at least one character and no more than 4 characters; The part-of-speech recognition of phrases can filter invalid phrases that are not related to emotions. For example, some connectives have less relevance to the analysis of emotional personality, such as the connectives "and", "and" and "union", etc. There are some particles such as "的", "地" and "得". After the terminal device filters out invalid phrases, the effective phrases containing user emotions are obtained, and the above effective phrases are recognized as conversation keywords; optionally The terminal device stores a key dictionary and judges whether the valid phrase is in the key dictionary. If it exists, the valid phrase is recognized as a conversation keyword; otherwise, the valid phrase is recognized as an invalid phrase.
在本实施例中,终端设备可以为会话关键词配置对应的会话标签,该会话标签用于标示该会话关键词在预设的词语维度的特征值。举例性,该会话标签可以用于标记该会话关键词的词性,例如“今天”这一会话关键词,基于词性分类的情况下会话标签可以设置“名词”,基于词语内容分类的情况下会话标签可以设置为“时间限定词”等。基于不同的划分方式以及情感识别过程的需要,可以为会话关键词配置不同的会话标签。上述会话标签的个数可以一个,也可以为两个或以上,在此不做限定。将所有会话关键词以会话标签进行封装,得到上述的会话内容集合,示例性地,该会话内容集合可以表示为
Figure PCTCN2020119300-appb-000001
其中,i=1,...,N;j=1,...,N i;上述N为在整个会话过程中包含的语音信号的总数,即会话文本的个数;而N i则表示第i个会话文本内包含的语句个数。
In this embodiment, the terminal device may configure a corresponding session tag for the session keyword, and the session tag is used to indicate the feature value of the session keyword in the preset word dimension. For example, the conversation tag can be used to mark the part of speech of the conversation keyword, such as the conversation keyword "today". In the case of part-of-speech classification, the conversation label can be set to "noun", and in the case of word content classification, the conversation label Can be set to "time qualifier" and so on. Based on different division methods and the needs of the emotion recognition process, different session tags can be configured for session keywords. The number of the above-mentioned session tags can be one, or two or more, which is not limited here. Encapsulate all session keywords with session tags to obtain the above-mentioned set of session content. Illustratively, the set of session content can be expressed as
Figure PCTCN2020119300-appb-000001
Where, i = 1, ..., N ; j = 1, ..., N i; the N speech signal included in the total for the entire session, i.e. the session number of text; N i indicates the The number of sentences contained in the i-th session text.
举例性地,面试官与面试人员的对话如下“面试官:你好,请进行自我介绍。面试人员:面试官你好。我的名字是张三。我来自深圳。毕业于大学。擅长测试。面试官:请问你对我们岗位的认识是?”其中,上述会话过程中存在3段语音信号,即会话文本的个数为3,用i表示会话文本的个数。例如“你好,请进行自我介绍”的会话顺序为1。而每个对话包含对应的语句数。例如“你好,请进行自我介绍”包含的语句数为2,分别为“你好”以 及“请进行自我介绍”,此时,N i为2。 For example, the conversation between the interviewer and the interviewer is as follows: "Interviewer: Hello, please introduce yourself. Interviewer: Hello, interviewer. My name is Zhang San. I am from Shenzhen. Graduated from university. Good at testing. Interviewer: What do you know about our position?” Among them, there are 3 speech signals in the above conversation process, that is, the number of conversation texts is 3, and i is the number of conversation texts. For example, the conversation sequence of "Hello, please introduce yourself" is 1. And each dialogue contains the corresponding number of sentences. For example, "Hello, please introduce themselves" statement includes the number is 2, namely, "Hello" and "Please introduce yourself" At this point, N i 2.
进一步地,作为本申请的另一实施例,在S102之前还可以包括:在确定关键词对应的标签之前,可以自动标签识别算法进行训练,使得最大化函数的取值最大,此时可以识别自动化标签识别算法已调整完毕,其中,最大化函数具体可以表示为:Further, as another embodiment of the present application, before S102, it may also include: before determining the tag corresponding to the keyword, the automatic tag recognition algorithm can be trained to maximize the value of the maximization function, and the automatic recognition can be performed at this time. The label recognition algorithm has been adjusted, where the maximization function can be expressed as:
Figure PCTCN2020119300-appb-000002
Figure PCTCN2020119300-appb-000002
其中,θ表示模型参数。Among them, θ represents model parameters.
在S103中,获得所述会话内容集合内的各个所述会话关键词对应的会话词向量,并基于各个所述会话词向量确定所述语音信号对应的情感特征值。In S103, a conversation word vector corresponding to each of the conversation keywords in the conversation content set is obtained, and an emotional feature value corresponding to the voice signal is determined based on each of the conversation word vectors.
在本实施例中,终端设备可以根据该会话内容集合内,各个会话关键词以及对应的会话标签,生成该会话关键词对应的会话词向量。在一种可能的实现方式中,生成上述会话词向量的方式可以为:终端设备配置有关键词典,并为关键词典内的各个候选关键词配置有对应的词编号,识别该会话关键词在上述关键词典内的词编号,基于词编号确定第一维度数值;对应地,终端设备可以为生成标签字典,通过查询会话标签在上述标签字典的标签编号,确定上述会话关键词的第二维度数值,基于第一维度数值和第二维度数值生成会话词向量。In this embodiment, the terminal device may generate a conversation word vector corresponding to the conversation keyword according to each conversation keyword and the corresponding conversation label in the conversation content set. In a possible implementation manner, the method for generating the above-mentioned conversational word vector may be as follows: the terminal device is configured with a key dictionary, and each candidate keyword in the key dictionary is configured with a corresponding word number, and the conversation keyword is identified in the above-mentioned The word number in the key dictionary determines the value of the first dimension based on the word number; correspondingly, the terminal device can generate a tag dictionary and determine the second dimension value of the conversation keyword by querying the tag number of the conversation tag in the tag dictionary. A conversation word vector is generated based on the first dimension value and the second dimension value.
在一种可能的实现方式中,生成上述会话词向量的方式还可以为:获取会话关键词在多个词性维度的参量值,生成一个多维向量,对应地,获取会话标签在多个词性维度的参量值,同样可以生成关于会话标签的多维向量,将会话关键词的多维向量与标签关键词的多维向量进行合并,得到上述的会话词向量。In a possible implementation, the method for generating the above-mentioned conversational word vector may also be: obtaining the parameter values of the conversation keyword in multiple parts of speech dimensions, generating a multi-dimensional vector, and correspondingly, obtaining the conversation label in multiple parts of speech dimensions. The parameter value can also generate a multi-dimensional vector about the conversation tag, and merge the multi-dimensional vector of the conversation keyword with the multi-dimensional vector of the tag key to obtain the above-mentioned conversation word vector.
在本实施例中,终端设备可以配置有情感识别网络,终端设备根据各个会话关键词的出现次序,依次导入到该情感识别网络内,并在所有会话关键词输入完成后导入预设的结束标示符,情感识别网络则输出关于上述会话文本,即语音信号对应的情感特征值。具体地,上述情感特征值可以包含在多个情感维度的得分,例如情感幅度维度以及积极程度维度等。In this embodiment, the terminal device may be configured with an emotion recognition network, and the terminal device imports the emotion recognition network in sequence according to the appearance order of each session keyword, and imports the preset end mark after all the session keywords are input. The emotion recognition network outputs the emotional feature value corresponding to the above-mentioned conversational text, that is, the speech signal. Specifically, the aforementioned emotional feature value may include scores in multiple emotional dimensions, such as an emotional magnitude dimension and a positive degree dimension.
在S104中,基于所有语音信号的所述情感特征值,生成所述目标用户的性格分析报告。In S104, based on the emotional feature values of all voice signals, a personality analysis report of the target user is generated.
在一实施例中,将生成的目标用户的性格分析报告存储于区块链网络中,通过区块链存储,实现数据信息在不同平台之间的共享,也可防止数据被篡改。In an embodiment, the generated personality analysis report of the target user is stored in the blockchain network, and the data information can be shared between different platforms through the storage of the blockchain, and the data can also be prevented from being tampered with.
区块链是分布式数据存储、点对点传输、共识机制、加密算法等计算机技术的新型应用模式。区块链(Blockchain),本质上是一个去中心化的数据库,是一串使用密码学方法相关联产生的数据块,每一个数据块中包含了一批次网络交易的信息,用于验证其信息的有效性(防伪)和生成下一个区块。区块链可以包括区块链底层平台、平台产品服务层以及应用服务层。Blockchain is a new application mode of computer technology such as distributed data storage, point-to-point transmission, consensus mechanism, and encryption algorithm. Blockchain, essentially a decentralized database, is a series of data blocks associated with cryptographic methods. Each data block contains a batch of network transaction information for verification. The validity of the information (anti-counterfeiting) and the generation of the next block. The blockchain can include the underlying platform of the blockchain, the platform product service layer, and the application service layer.
在本实施例中,终端设备可以根据所有会话内容对应的情感特征,生成目标用户的用户画像,确定各个性格类型对应的概率得分,最后选取概率得分最高的一个性格类型作为目标用户的性格类型,并生成上述目标用户的性格分析报告。可选地,终端设备还可以将所有性格类型的概率得分记录在性格分析报告中,从而面试管理人员可以根据该性格分析报告确定目标用户的潜在性格特性,提高了性格分析报告内容的丰富度。In this embodiment, the terminal device can generate a user portrait of the target user according to the emotional characteristics corresponding to all conversational content, determine the probability score corresponding to each personality type, and finally select the personality type with the highest probability score as the personality type of the target user. And generate the personality analysis report of the target user mentioned above. Optionally, the terminal device can also record the probability scores of all personality types in the personality analysis report, so that the interview manager can determine the potential personality characteristics of the target user based on the personality analysis report, which improves the richness of the content of the personality analysis report.
以上可以看出,本申请实施例提供的一种用户报告的生成方法通过在与目标用户进行会话的过程中,采集目标用户的语音信号,并将语音信号转换为对应的会话文本,并对会话文本进行语义分析得到对应的会话内容合集,并基于会话内容合集内各个会话关键词的会话词向量,生成语音信号对应的情感特征值,并基于所有语音信号的情感特征值,确定目标用户的性格类型,并生成关于目标用户的性格分析报告,从而能够在与目标用户会话 的过程中,通过目标用户的语言确定性格,实现了自动输出分析报告的目的。与现有的用户报告技术相比,本实施例不依赖面试人员或会话对象进行手动填写或主观判断,无需用户花费额外时间撰写关于目标用户的性格分析报告,从而能够大大减少用户的操作,并且上述的过程可以通过在会话过程中不同阶段的语音信号,确定情感特征值,并非以单一的话语或语句进行性格判断,从而能够提高性格分析报告的准确性。It can be seen from the above that the method for generating a user report provided by the embodiment of the present application collects the voice signal of the target user during a conversation with the target user, converts the voice signal into the corresponding conversation text, and responds to the conversation. The text is semantically analyzed to obtain the corresponding conversational content collection, and based on the conversational word vector of each conversation keyword in the conversational content collection, the emotional feature value corresponding to the voice signal is generated, and based on the emotional feature value of all voice signals, the personality of the target user is determined Type, and generate a personality analysis report on the target user, so that the target user’s language can be used to determine the personality during the conversation with the target user, and the purpose of automatically outputting the analysis report is realized. Compared with the existing user report technology, this embodiment does not rely on the interviewer or the conversation object to manually fill in or subjectively judge, and does not require the user to spend extra time writing a personality analysis report on the target user, thereby greatly reducing user operations, and The above process can determine the emotional characteristic value through the voice signals at different stages in the conversation process, instead of using a single utterance or sentence to judge the personality, thereby improving the accuracy of the personality analysis report.
图2示出了本申请第二实施例提供的一种用户报告的生成方法S103的具体实现流程图。参见图2,相对于图1所述实施例,本实施例提供的一种用户报告的生成方法中S103包括:S1031~S1036,具体详述如下:Fig. 2 shows a specific implementation flow chart of a method S103 for generating a user report provided by the second embodiment of the present application. Referring to FIG. 2, compared with the embodiment described in FIG. 1, S103 in a method for generating a user report provided in this embodiment includes: S1031 to S1036, which are detailed as follows:
进一步地,所述获得所述会话内容集合内的各个所述会话关键词对应的会话词向量,并基于各个所述会话词向量确定所述语音信号对应的情感特征值,包括:Further, the obtaining the conversation word vector corresponding to each of the conversation keywords in the conversation content set, and determining the emotional feature value corresponding to the voice signal based on each of the conversation word vectors includes:
在S1031中,确定各个所述会话关键词在预设的知识图谱内的关联实体,并获取各个所述关联实体对应的加权权重。In S1031, determine the associated entity of each of the session keywords in the preset knowledge graph, and obtain the weighted weight corresponding to each of the associated entities.
在本实施例中,终端设备配置有知识图谱,该知识图谱包含有多个知识节点,不同的知识节点之间存在对应的关联关系,从而构成一个由多个知识节点相互连接的网络,即上述的知识图谱。终端设备可以判断会话关键词在上述知识图谱上关联的知识节点,并将与该关联的知识节点相邻的其他节点,即存在关联关系的其他知识节点,识别为会话关键词的关联实体。In this embodiment, the terminal device is configured with a knowledge graph, which contains multiple knowledge nodes, and there are corresponding association relationships between different knowledge nodes, thereby forming a network connected by multiple knowledge nodes, that is, the above Knowledge graph. The terminal device can determine the knowledge node associated with the session keyword on the above-mentioned knowledge graph, and identify other nodes adjacent to the associated knowledge node, that is, other knowledge nodes with an association relationship, as the associated entity of the session keyword.
在一种可能的实现方式中,终端设备可以根据会话关键词关联的知识节点与关联实体之间关联关系的置信度,确定上述挂念实体的加权权重。In a possible implementation manner, the terminal device may determine the weighted weight of the above-mentioned concern entity according to the confidence of the association relationship between the knowledge node associated with the session keyword and the associated entity.
在S1032中,根据所有所述关联实体的所述加权权重,生成所述会话关键词的词概念向量。In S1032, the word concept vector of the conversation keyword is generated according to the weighted weight of all the associated entities.
在本实施例中,由于基于知识图谱确定关联实体并没有与会话关键词所在会话语句的上下文概念以及情感相关的特征,而加权权重在计算时考虑了上下文关联性以及情感特征,因此可以转换为包含上述两个特征的概念向量。计算的方式具体为:In this embodiment, since it is determined based on the knowledge graph that the associated entity does not have features related to the contextual concept and emotion of the conversational sentence where the conversation keyword is located, and the weighted weight is calculated considering the contextual relevance and emotional features, it can be converted to A concept vector containing the above two features. The specific calculation method is:
Figure PCTCN2020119300-appb-000003
Figure PCTCN2020119300-appb-000003
其中,c(t)为上述词概念向量;g(t)为所述会话关键词包含的关联实体的总数;c k为上述会话关键词的第k个关联实体的词向量,w k为上述的会话关键词的第k个关联实体的加权权重。 Wherein, c(t) is the above-mentioned word concept vector; g(t) is the total number of associated entities contained in the session keyword; c k is the word vector of the k-th associated entity of the above-mentioned session keyword, and w k is the above-mentioned The weighted weight of the k-th associated entity of the session keyword.
优选地,作为本申请的另一实施例,在计算得到会话关键词对应的词概念向量后,可以通过线性变化的方式,将词概念向量转换为词特征向量,具体转换方式可以为:Preferably, as another embodiment of the present application, after the word concept vector corresponding to the conversation keyword is calculated, the word concept vector can be converted into a word feature vector by linear change, and the specific conversion method can be:
Figure PCTCN2020119300-appb-000004
Figure PCTCN2020119300-appb-000004
其中,
Figure PCTCN2020119300-appb-000005
为上述会话关键词的词特征向量;W为模型参数,模型参数W∈R d*2d;t为会话关键词所在语句的语句向量;Embed(t)为嵌入的会话关键词的大小;Post(t)为在会话语句的位置编码;R d为会话关键词的词向量大小。
among them,
Figure PCTCN2020119300-appb-000005
Is the word feature vector of the above-mentioned session keywords; W is the model parameter, and the model parameter W∈R d*2d ; t is the sentence vector of the sentence where the session keyword is located; Embed(t) is the size of the embedded session keyword; Post( t) is the coding at the position of the conversation sentence; R d is the word vector size of the conversation keyword.
在S1033中,基于各个所述会话关键词的所属的会话语句,对属于同一所述会话语句的所有词概念向量进行封装,生成所述会话语句的语句概念向量;所述会话语句是对所述会话文本进行语句划分后得到的。In S1033, based on the conversational sentence to which each of the conversational keywords belongs, encapsulate all word concept vectors belonging to the same conversational sentence to generate the sentence concept vector of the conversational sentence; The conversation text is obtained after sentence division.
在本实施例中,会话文本内可以包含有多个会话语句。终端设备可以基于会话关键词所述的会话语句对会话关键词进行划分,得到多个会话关键词组,每个会话关键词组内的所有会话关键词对应同一会话语句。终端设备可以将属于同一会话语句的词概念向量进行封装,生成关于该会话语句对应的语句概率向量。In this embodiment, the conversation text may contain multiple conversation sentences. The terminal device may divide the conversation keywords based on the conversation sentences described by the conversation keywords to obtain multiple conversation keyword groups, and all conversation keywords in each conversation keyword group correspond to the same conversation sentence. The terminal device can encapsulate the word concept vectors belonging to the same conversation sentence to generate a sentence probability vector corresponding to the conversation sentence.
在S1034中,分别将各个所述会话语句的所述语句概念向量导入到第一注意力算法,得到各个所述会话语句的对话更新向量。In S1034, the sentence concept vector of each conversation sentence is imported into the first attention algorithm to obtain the dialogue update vector of each conversation sentence.
在本实施例中,上述对话更新向量语句用于表征会话语句本申请的情感特征,因此,终端设备可以分别将各个会话语句的语句概念向量导入到第一注意力算法中,得到对话更新向量。In this embodiment, the above-mentioned dialogue update vector sentence is used to characterize the emotional characteristics of the conversation sentence in this application. Therefore, the terminal device can separately import the sentence concept vector of each conversation sentence into the first attention algorithm to obtain the dialogue update vector.
在S1035中,将所述会话文本的所有会话语句的所述语句概念向量进行封装,生成所述会话文本的对话概念向量,并将所述会话概念向量导入第二注意力模型,生成所述会话文本的文本概念向量。In S1035, encapsulate the sentence concept vectors of all conversation sentences of the conversation text, generate the conversation concept vectors of the conversation text, and import the conversation concept vectors into the second attention model to generate the conversation Text concept vector for text.
在本实施例中,由于第一注意力模型具体用于确定单一语句的情感特征,终端设备可以根据不同语句之间的上下文的联系,确定整个会话文本整体的情感特征。因此可以将所有会话语句的语句概念向量进行封装,得到对话概念向量,并将对话概念向量导入到上述的第二注意力模型内,得到文本概念向量。其中,文本概向量具体可以表示为:In this embodiment, since the first attention model is specifically used to determine the emotional characteristics of a single sentence, the terminal device can determine the overall emotional characteristics of the entire conversation text according to the contextual connections between different sentences. Therefore, the sentence concept vectors of all conversational sentences can be encapsulated to obtain the conversation concept vector, and the conversation concept vector can be imported into the aforementioned second attention model to obtain the text concept vector. Among them, the text approximate vector can be specifically expressed as:
Figure PCTCN2020119300-appb-000006
Figure PCTCN2020119300-appb-000006
Figure PCTCN2020119300-appb-000007
Figure PCTCN2020119300-appb-000007
FF(x)=max(0,W 1x+b 1)W 2+b 2 FF(x)=max(0,W 1 x+b 1 )W 2 +b 2
其中,为文本概念向量,为第i个会话文本的对话概念向量;W 1、W 2、b 1以及b 2为所述第二注意力模型的模型参数;d s为基于线性变换的端点数h确定的系数值,d s=d/h。L(x)为基于端点数h的线性变换;L’(x)为基于端点数h的逆线性变换。 Where is the text concept vector, which is the dialogue concept vector of the i-th conversational text; W 1 , W 2 , b 1 and b 2 are the model parameters of the second attention model; d s is the number of endpoints based on linear transformation The coefficient value determined by h, d s =d/h. L(x) is a linear transformation based on the number of endpoints h; L'(x) is an inverse linear transformation based on the number of endpoints h.
在S1036中,根据所述对话更新向量以及所述文本概念向量,确定所述情感特征值。In S1036, the emotional feature value is determined according to the dialogue update vector and the text concept vector.
在本实施例中,终端设备可以将对话更新向量以及文本概念向量导入到第三注意力模型,得到会话文本对应的情感概念向量。该情感概念向量具体为可以为:In this embodiment, the terminal device can import the dialogue update vector and the text concept vector into the third attention model to obtain the emotion concept vector corresponding to the conversation text. The emotional concept vector can be specifically:
Figure PCTCN2020119300-appb-000008
Figure PCTCN2020119300-appb-000008
其中,R i为上述情感概念向量;
Figure PCTCN2020119300-appb-000009
为对话更新向量。终端设备可以对上述的情感概念向量导入预设的池化层,进行情感特征提取,得到上述情感概念向量对应的情感特征值,该池化层可以表示为:
Among them, R i is the above-mentioned emotional concept vector;
Figure PCTCN2020119300-appb-000009
Update the vector for the dialogue. The terminal device can import the foregoing emotional concept vector into a preset pooling layer, perform emotional feature extraction, and obtain the emotional feature value corresponding to the foregoing emotional concept vector. The pooling layer can be expressed as:
O=max_pool(R i) O=max_pool(R i )
p=softmax(O*W 3+b 3) p=softmax(O*W 3 +b 3 )
其中,p为上述的情感特征值;W 3∈R d*q,b 3∈R q表示模型参数,q表示类数。 Among them, p is the above-mentioned emotional feature value; W 3 ∈ R d*q , b 3 ∈ R q represents the model parameter, and q represents the number of classes.
在本申请实施例中,通过获取会话关键词的关联实体,对会话内容进行延伸,并分别确定基于单个语句的对话更新向量以及基于所有语句的文本概念向量,确定目标用户的情感特征值,能够从多个维度确定用户的情感特征,从而提高了情感特征的准确性。In the embodiment of this application, the conversational content is extended by obtaining the associated entities of the conversation keywords, and the conversation update vector based on a single sentence and the text concept vector based on all sentences are determined respectively to determine the emotional feature value of the target user. Determine the user's emotional characteristics from multiple dimensions, thereby improving the accuracy of the emotional characteristics.
图3示出了本申请第三实施例提供的一种用户报告的生成方法S1031的具体实现流程 图。参见图3,相对于图2所述的实施例,本实施例提供的一种用户报告的生成方法中S1031包括:S301~S303,具体详述如下:Fig. 3 shows a specific implementation flow chart of a method S1031 for generating a user report provided by the third embodiment of the present application. Referring to FIG. 3, compared to the embodiment described in FIG. 2, S1031 in a method for generating a user report provided in this embodiment includes: S301 to S303, which are detailed as follows:
进一步地,所述确定各个所述会话关键词在预设的知识图谱内的关联实体,并获取各个所述关联实体对应的加权权重,包括:Further, the determining the associated entity of each of the session keywords in the preset knowledge graph and obtaining the weighted weight corresponding to each of the associated entities includes:
在S301中,获取各个所述关联实体与所述会话关键词之间的关联强度因子。In S301, obtain the correlation strength factor between each of the associated entities and the session keywords.
在本实施例中,根据不同知识节点之间的关联紧密度,可以确定不同关联节点之间的关联置信度。例如某两个知识节点在大部分的文本中存在共现关系(共现关系即同时出现在多个知识节点同时出现在同一语句内),则上述知识节点之间的关联置信度较高;反之,若两个知识节点只在少量的文本中存在共现关系,则上述知识节点之间的关联置信度较低。根据会话关键词关联的知识节点与关联实体之间的关联置信度,可以得到上述的关联强度因子。In this embodiment, according to the closeness of the association between the different knowledge nodes, the association confidence between the different associated nodes can be determined. For example, if two knowledge nodes have a co-occurrence relationship in most of the text (the co-occurrence relationship is that multiple knowledge nodes appear in the same sentence at the same time), the correlation between the above-mentioned knowledge nodes has a higher confidence; , If two knowledge nodes only have a co-occurrence relationship in a small amount of text, the confidence of the association between the above-mentioned knowledge nodes is low. According to the confidence of the association between the knowledge node and the associated entity associated with the session keyword, the above-mentioned association strength factor can be obtained.
在一种可能的实现方式中,终端设备可以包含有关联强度因子的转换算法,将关联实体对应的关联置信度导入上述转换算法中,生成上述关联强度因子。In a possible implementation manner, the terminal device may include a conversion algorithm of the correlation strength factor, and import the correlation confidence corresponding to the associated entity into the conversion algorithm to generate the aforementioned correlation strength factor.
在S302中,基于预设的情感度量算法,确定各个所述关联实体的情感强度因子。In S302, based on a preset emotion measurement algorithm, the emotion intensity factor of each associated entity is determined.
在本实施例中,由于不同的词语具有对应的情感特征,例如“笑容”这个词语在情感上是较为积极的词语,而“哭泣”这个词语在情感上是较为消极的词语,可以根据不同词语对应的内容以及意义,可以转换为对应的情感强度因子。终端设备可以配置有情感度量算法,可以将词语转换为计算机可识别的情感强度因子。在该情况下,终端设备可以将关联实体导入到上述的情感度量算法,输出关联实体对应的情感强度因子。In this embodiment, because different words have corresponding emotional characteristics, for example, the word "smile" is a more positive word emotionally, while the word "cry" is a negative word emotionally, and can be based on different words The corresponding content and meaning can be converted into corresponding emotional intensity factors. The terminal device can be equipped with an emotion measurement algorithm, which can convert words into a computer-recognizable emotion intensity factor. In this case, the terminal device can import the associated entity into the aforementioned emotion measurement algorithm, and output the emotion intensity factor corresponding to the associated entity.
在S303中,基于所述情感强度因子以及所述关联强度因子,构建所述关联实体的加权权重。In S303, a weighted weight of the associated entity is constructed based on the emotional intensity factor and the associated intensity factor.
在本实施例中,终端设备可以根据情感因子以及关联强度因子,生成关联实体的加权权重,该加权权重则包含有与会话关键词之间的关联紧密度以及情感特征,便于后续的情感特征值的确定。其中,该加权权重具体为可以为:In this embodiment, the terminal device can generate the weighted weight of the associated entity according to the emotion factor and the correlation strength factor, and the weighted weight includes the closeness of the association with the session keyword and the emotional feature, which is convenient for the subsequent emotional feature value. Of ok. Wherein, the weighting weight may specifically be:
w k=λ k*rel k+(1-λ k)*aff k w kk *rel k +(1-λ k )*aff k
其中,w k为第k个关联实体对应的加权权重;rel k为所述第k个关联实体对应的关联强度因子,aff k为第k个关联实体的情感强度因子,λ k为第k个关联实体的预设参数。 Where w k is the weighted weight corresponding to the k-th associated entity; rel k is the associated intensity factor corresponding to the k-th associated entity, aff k is the emotional intensity factor of the k-th associated entity, and λ k is the k-th associated entity. The preset parameters of the associated entity.
在本申请实施例中,通过计算关联实体与会话关键词之间的关联强度以及该关键词本申请的情感特征,确定该关联实体在计算情感特征值时对应的加权权重,关联度越高,则对应的加权权重越高,本申请的情感特征对于后续的会话文本的情感特征值的贡献则越大,从而能够提高情感特征值的准确性。In the embodiment of the present application, by calculating the correlation strength between the associated entity and the session keyword and the emotional feature of the keyword in this application, the weighted weight corresponding to the associated entity when calculating the emotional feature value is determined. The higher the degree of association, The higher the corresponding weighted weight, the greater the contribution of the emotional feature of the present application to the emotional feature value of the subsequent conversational text, so that the accuracy of the emotional feature value can be improved.
图4示出了本申请第四实施例提供的一种用户报告的生成方法S301的具体实现流程图。参见图4,相对于图3所述实施例,本实施例提供的一种用户报告的生成方法S301包括:S3011~S3013,具体详述如下:FIG. 4 shows a specific implementation flowchart of a method S301 for generating a user report provided by the fourth embodiment of the present application. Referring to FIG. 4, compared with the embodiment described in FIG. 3, a method S301 for generating a user report provided by this embodiment includes: S3011 to S3013, which are detailed as follows:
进一步地,所述获取各个所述关联实体与所述会话关键词之间的关联强度因子,包括:Further, the obtaining the correlation strength factor between each of the associated entities and the session keywords includes:
在S3011中,基于所述知识图谱,确定所述关联实体与所述会话关键词之间的关联置信度。In S3011, based on the knowledge graph, the association confidence between the associated entity and the session keyword is determined.
在本实施例中,知识图谱内可以记录有各个知识节点之间关联关系的置信度,终端设备在知识图谱中标记出会话关键词以及关联实体,确定两者之间的关联关系的置信度,将该关联关系的置信度识别为上述两者之间的关联置信度。其中,关联实体与会话关键词之间的共现次数越多,则对应的关联置信度越高;反之,若上述两者之间的共现次数越少,则对应的关联置信度越低。In this embodiment, the confidence level of the association relationship between each knowledge node may be recorded in the knowledge graph, and the terminal device marks the session keyword and the associated entity in the knowledge graph to determine the confidence level of the association relationship between the two. The confidence of the correlation is identified as the confidence of the correlation between the above two. Among them, the more the number of co-occurrences between the associated entity and the session keyword, the higher the corresponding confidence of the association; conversely, the less the number of co-occurrences between the two, the lower the confidence of the corresponding association.
在S3012中,将所述会话关键词关联的会话语句导入到预设的池化层,生成各个所述 会话关键词关联的会话语句的语句向量,并基于所述语句向量确定所述会话关键词所在语段的会话文本向量;所述会话文本向量具体为:In S3012, the conversation sentence associated with the conversation keyword is imported into a preset pooling layer, a sentence vector of the conversation sentence associated with each of the conversation keywords is generated, and the conversation keyword is determined based on the sentence vector The conversational text vector of the segment; the conversational text vector is specifically:
Figure PCTCN2020119300-appb-000010
Figure PCTCN2020119300-appb-000010
其中,CR(X i)为所述会话关键词的所述会话文本向量,所述会话关键词所在的会话文本编号为i;
Figure PCTCN2020119300-appb-000011
为所述会话关键词所在的会话语句的语句向量,所述会话语句在所述会话文本中的语句编号为j;所述M为预设的关联系数。
Wherein, CR(X i ) is the conversation text vector of the conversation keyword, and the conversation text number where the conversation keyword is located is i;
Figure PCTCN2020119300-appb-000011
Is the sentence vector of the conversation sentence where the conversation keyword is located, and the sentence number of the conversation sentence in the conversation text is j; the M is a preset correlation coefficient.
在本实施例中,会话文本包含有多个会话语句,假设会话关键词所在的会话语句为
Figure PCTCN2020119300-appb-000012
则与会话关键词存在关联关系的关联语句则为
Figure PCTCN2020119300-appb-000013
Figure PCTCN2020119300-appb-000014
的M个会话语句,其中M为预设的关联系数。为了控制终端设备的数据处理量,终端设备可以配置有关联系数M,在进行情感特征识别的过程中,可以基于关联系数M确定所需统一识别的最大会话个数。将M个会话语句对应的语句向量导入到上述的文本向量转换函数,则可以确定以会话关键词为基准的会话文本对应的会话文本向量。
In this embodiment, the conversation text contains multiple conversation sentences. Assume that the conversation sentence where the conversation keyword is
Figure PCTCN2020119300-appb-000012
Then the associated sentence that has an association relationship with the session keyword is
Figure PCTCN2020119300-appb-000013
to
Figure PCTCN2020119300-appb-000014
M conversational sentences in, where M is the preset correlation coefficient. In order to control the data processing volume of the terminal device, the terminal device can be configured with the number of contacts M. In the process of emotional feature recognition, the maximum number of sessions that need to be uniformly recognized can be determined based on the correlation coefficient M. By importing the sentence vectors corresponding to the M conversational sentences into the above-mentioned text-to-vector conversion function, the conversational text vector corresponding to the conversational text based on the conversational keywords can be determined.
在S3013中,基于所述会话文本向量以及所述关联置信度,计算所述关联强度因子;所述关联强度因子具体为:In S3013, the correlation strength factor is calculated based on the conversation text vector and the correlation confidence; the correlation strength factor is specifically:
rel k=max-min(s k)*|cos(CR(X i),c k)| rel k =max-min(s k )*|cos(CR(X i ),c k )|
其中,rel k为第k个会话关键词的所述关联强度因子;c k为所述会话关键词第k个关联实体的关联置信度;max-min(s k)为所述会话关键词第k个关联实体对应的情感极差。 Where rel k is the correlation strength factor of the k-th session keyword; c k is the correlation confidence of the k-th associated entity of the session keyword; max-min(s k ) is the session keyword The emotions corresponding to k associated entities are extremely poor.
在本实施例中,终端设备可以包含有多个不同的情感度量算法,不同的情感度量算法所确定关联实体的情感参数值可能会存在差异,终端设备可以根据不同的情感度量算法关于关联实体的情感极差,即上述的max-min(s k),将关联实体的情感极差以及上述两个参量导入预设的关联强度转换算法,得到关联实体的关联强度值。 In this embodiment, the terminal device may include multiple different emotion measurement algorithms, and the emotional parameter values of the associated entities determined by different emotion measurement algorithms may be different. The terminal device may determine the related entities according to different emotion measurement algorithms. The emotional range, that is, the above-mentioned max-min(s k ), the emotional range of the related entity and the above two parameters are imported into a preset correlation strength conversion algorithm to obtain the correlation strength value of the related entity.
在本申请实施例中,通过确定会话文本向量,在计算关联实体的过程中,考虑整个会话文本中不同会话语句之间的关联性,从而能够提高了关联强度因子的准确性。In the embodiment of the present application, by determining the conversational text vector, in the process of calculating the associated entity, the association between different conversational sentences in the entire conversational text is considered, so that the accuracy of the correlation strength factor can be improved.
图5示出了本申请第五实施例提供的一种用户报告的生成方法S302的具体实现流程图。参见图5,相对于图3所述实施例,本实施例提供的一种用户报告的生成方法S302包括:S3021~S3023,具体详述如下:FIG. 5 shows a specific implementation flowchart of a method S302 for generating a user report provided by the fifth embodiment of the present application. Referring to FIG. 5, compared to the embodiment described in FIG. 3, a method S302 for generating a user report provided in this embodiment includes: S3021 to S3023, which are detailed as follows:
进一步地,所述基于预设的情感度量算法,确定各个所述关联实体的情感强度因子,包括:Further, the determining the emotion intensity factor of each associated entity based on a preset emotion measurement algorithm includes:
在S3021中,识别所述关联实体的情感属性。In S3021, the emotional attributes of the associated entity are identified.
在本实施例中,终端设备可以根据关联实体的情感属性的不同,采用不同的方式确定情感强度因子。举例性地,“我”这个指示代词不包含情感特征,则对应的情感属性为非情感类型;“伟大”这个形容词包含一定程度的情感特性,则对应的情感属性为情感类型。基于此,终端设备可以识别各个关联实体的情感属性,若关联实体的情感属性为非情感类型,则执行S3022的操作;反之,若关联实体的情感类型为情感类型,则执行S3023的操作。In this embodiment, the terminal device can determine the emotional intensity factor in different ways according to the different emotional attributes of the associated entities. For example, if the demonstrative pronoun "I" does not contain emotional characteristics, the corresponding emotional attribute is a non-affective type; the adjective "great" contains a certain degree of emotional characteristics, and the corresponding emotional attribute is an emotional type. Based on this, the terminal device can identify the emotional attribute of each associated entity. If the emotional attribute of the associated entity is a non-emotional type, the operation of S3022 is performed; conversely, if the emotional type of the associated entity is an emotional type, the operation of S3023 is performed.
在S3022中,若所述关联实体的所述情感属性为非情感类型,则将所述情感强度因子配置为预设的默认值。In S3022, if the affective attribute of the associated entity is a non-emotional type, the affective intensity factor is configured as a preset default value.
在本实施例中,终端设备可以将所有非情感类型的关联实体配置固定数值的情感强度因子,该情感强度因子的数值可以配置为0.5。In this embodiment, the terminal device can configure all non-emotional related entities with a fixed value of emotional intensity factor, and the value of the emotional intensity factor can be configured to be 0.5.
在S3023中,若所述关联实体的所述情感属性为情感类型,则通过预设的情感转换算法,计算所述会话关键词的所述情感强度因子;所述情感强度因子具体为:In S3023, if the emotion attribute of the associated entity is an emotion type, the emotion intensity factor of the conversation keyword is calculated through a preset emotion conversion algorithm; the emotion intensity factor is specifically:
Figure PCTCN2020119300-appb-000015
Figure PCTCN2020119300-appb-000015
其中,aff k为第k个所述关联实体的情感强度因子;VAD(c k)为所述第k个所述关联实体的积极情感分值;A(c k)为所述第k个所述关联实体的情感幅度分值。 Where aff k is the emotional intensity factor of the k-th associated entity; VAD(c k ) is the positive emotional score of the k-th associated entity; A(c k ) is the k-th associated entity The emotional magnitude score of the associated entity.
在本实施例中,情感强度因子具体由两个不同的情感维度构成,分比为积极情感维度以及情感幅度维度,其中积极情感维度具体用于标识该实体对应的情感特征是否积极,若积极程度越高,则对应的情感分值越高。举例性地,“笑”对应的积极情感分值为正数,而“哭”对应的积极情感分值为负数,而“乐观”对应的情感积极分值会比“接纳”的情感积极分值要高;而情感幅度分值即用于标识该实体的情感波动幅度,例如“笑”的情感幅度分值会低于“大小”的情感幅度分值。终端设备可以通过预设的情感度量算法,确定各个关联实体在上述两个维度对应的情感分值,并得到对应的情感强度因子。其中,
Figure PCTCN2020119300-appb-000016
为基于2的范数。
In this embodiment, the emotional intensity factor is specifically composed of two different emotional dimensions, divided into a positive emotional dimension and an emotional amplitude dimension, where the positive emotional dimension is specifically used to identify whether the corresponding emotional feature of the entity is positive, if the degree of positive The higher the value, the higher the corresponding emotional score. For example, the positive emotional score corresponding to "laugh" is positive, while the positive emotional score corresponding to "cry" is negative, and the emotional positive score corresponding to "optimism" is higher than the emotional positive score of "acceptance" The emotional amplitude score is used to identify the emotional fluctuation amplitude of the entity, for example, the emotional amplitude score of "laugh" will be lower than the emotional amplitude score of "large". The terminal device can determine the corresponding emotional scores of each associated entity in the above two dimensions through a preset emotional measurement algorithm, and obtain the corresponding emotional intensity factor. among them,
Figure PCTCN2020119300-appb-000016
It is based on the norm of 2.
在本申请实施例中,通过识别关联实体的情感属性,选取对应的情感强度因子的计算方式,从而提高了情感强度因子的准确性。In the embodiment of the present application, the emotion attribute of the associated entity is identified, and the calculation method of the corresponding emotion intensity factor is selected, thereby improving the accuracy of the emotion intensity factor.
图6示出了本申请第六实施例提供的一种用户报告的生成方法S1034的具体实现流程图。参见图6,相对于图2所述实施例,本实施例提供的一种用户报告的生成方法中S1034包括:S601~S603,具体详述如下:FIG. 6 shows a specific implementation flow chart of a method S1034 for generating a user report provided by the sixth embodiment of the present application. Referring to FIG. 6, compared to the embodiment described in FIG. 2, S1034 in a method for generating a user report provided in this embodiment includes: S601 to S603, which are detailed as follows:
进一步地,所述分别将各个所述会话语句的所述语句概念向量导入到第一注意力算法,得到各个所述会话语句的对话更新向量,包括:Further, the respectively importing the sentence concept vector of each of the conversational sentences into the first attention algorithm to obtain the dialogue update vector of each of the conversational sentences includes:
在S601中,对所述会话语句的语句概念向量进行线性变化,得到包含h个端点的线性向量;其中,所述h为预设的端点个数。In S601, the sentence concept vector of the conversation sentence is linearly changed to obtain a linear vector containing h endpoints; where h is the preset number of endpoints.
在本实施例中,终端设备可以对会话语句的语句概念进行线性变换,将会语句概念向量投射多h个端点中,得到关于语句概念向量的线性向量。其中,上述h的数值可以为第一注意力算法的预设线性变换参量,也可以基于会话文本的文本量进行变更。In this embodiment, the terminal device can perform a linear transformation on the sentence concept of the conversational sentence, and project the sentence concept vector into h more endpoints to obtain a linear vector about the sentence concept vector. Wherein, the above-mentioned value of h may be a preset linear transformation parameter of the first attention algorithm, or may be changed based on the text amount of the conversational text.
在S602中,将所述线性向量导入到所述第一注意力算法的多头自注意力层,得到所述会话语句的注意力向量;所述注意力向量具体为:In S602, the linear vector is imported into the multi-head self-attention layer of the first attention algorithm to obtain the attention vector of the conversation sentence; the attention vector is specifically:
Figure PCTCN2020119300-appb-000017
Figure PCTCN2020119300-appb-000017
其中,
Figure PCTCN2020119300-appb-000018
为第i个会话文本中第n个所述会话语句的所述注意力向量;
Figure PCTCN2020119300-appb-000019
为所述线性向量;d s为基于所述线性向量的端点数h确定的系数值。
among them,
Figure PCTCN2020119300-appb-000018
Is the attention vector of the nth conversation sentence in the ith conversation text;
Figure PCTCN2020119300-appb-000019
Is the linear vector; d s is a coefficient value determined based on the number of endpoints h of the linear vector.
在本实施例中,终端设备可以将上述计算得到的线性向量导入到上述的多头注意力层,该注意力层包含三个节点。首先,终端设备可以计算线性向量与线性向量的转置之间的乘积,并将乘积后的向量通过softmax函数进行处理,最后再次与该线性向量相乘,从而能够实现三重迭代,以提高特征提取的准确性。In this embodiment, the terminal device may import the linear vector obtained by the foregoing calculation to the foregoing multi-head attention layer, and the attention layer includes three nodes. First, the terminal device can calculate the product between the linear vector and the transposition of the linear vector, and process the multiplied vector through the softmax function, and finally multiply the linear vector again to achieve triple iteration to improve feature extraction Accuracy.
在S603中,基于所述注意力向量生成所述会话语句的对话更新向量;所述对话更新向量具体为:In S603, a dialogue update vector of the conversation sentence is generated based on the attention vector; the dialogue update vector is specifically:
Figure PCTCN2020119300-appb-000020
Figure PCTCN2020119300-appb-000020
其中,W 1、W 2、b 1以及b 2为所述第一注意力模型的模型参数。 Wherein, W 1 , W 2 , b 1 and b 2 are model parameters of the first attention model.
在本实施例中,终端设备可以将生成的注意力向量导入第一震惊网络的前馈层,得到会话语句对应的对话更新向量。该前馈层可以先对注意力向量进行逆线性变换,将包含多个端点的注意力向量变换到包含单一端点的向量后,再进行后续的操作。In this embodiment, the terminal device can import the generated attention vector into the feedforward layer of the first shocking network to obtain the dialogue update vector corresponding to the conversation sentence. The feedforward layer can first perform an inverse linear transformation on the attention vector, transform the attention vector containing multiple endpoints to a vector containing a single endpoint, and then perform subsequent operations.
在本申请实施例中,在加入了基于新型NLP变压器的情感判断方法后,面试官可以通过候选人的回答迅速对候选人的某些性格特征进行判断,并给出必要和合理的追问。在实际的AI面试应用中,因为判断更精准,硬件的应答速度也得到了提高,所以不仅节省了硬件空间,而且提高了运行速度和面试体验。AI智能面试可以根据对候选人的回答判断候选人的情感,进而判断候选人的性格,面试官可以在面试完成后进行分析,了解候选人性格特征的分布情况,作为选拔候选人的依据。In the embodiment of the application, after adding the emotion judgment method based on the new NLP transformer, the interviewer can quickly judge certain personality characteristics of the candidate through the candidate's answer, and give necessary and reasonable follow-up questions. In the actual AI interview application, because the judgment is more accurate, the response speed of the hardware has also been improved, so it not only saves the hardware space, but also improves the running speed and interview experience. AI intelligent interview can judge the emotion of the candidate based on the answer to the candidate, and then judge the personality of the candidate. The interviewer can analyze the candidate's personality characteristics after the interview is completed, and use it as the basis for selecting candidates.
图7示出了本申请第七实施例提供的一种用户报告的生成方法S104的具体实现流程图。参见图7,相对于图1至图6任一所述实施例,本实施例提供的一种用户报告的生成方法中S104包括:S1041-S1043,具体详述如下:FIG. 7 shows a specific implementation flowchart of a method S104 for generating a user report provided by the seventh embodiment of the present application. Referring to FIG. 7, relative to any one of the embodiments described in FIG. 1 to FIG. 6, S104 in a method for generating a user report provided in this embodiment includes: S1041-S1043, which are detailed as follows:
进一步地,所述基于所有语音信号的所述情感特征值,生成所述目标用户的性格分析报告,包括:Further, the generating a personality analysis report of the target user based on the emotional characteristic values of all voice signals includes:
在S1041中,根据各个所述语音信号的情感特征值,生成所述目标用户的情感波形图。In S1041, an emotion waveform diagram of the target user is generated according to the emotion characteristic value of each voice signal.
在本实施例中,终端设备可以根据各个会话文本,即语音信号的生成次序,在预设的坐标轴上标记出各个情感特征值,并依次连接各个情感特征值,得到目标用户在整个会话过程中对应的情感波形图。In this embodiment, the terminal device can mark each emotional feature value on a preset coordinate axis according to each conversation text, that is, the generation sequence of the voice signal, and connect each emotional feature value in turn to obtain the target user in the entire conversation process. Corresponding sentiment waveform in.
在S1042中,将所述情感波形图与各个候选性格的标准性格波形图进行匹配,确定所述目标用户的用户性格。In S1042, the emotion waveform diagram is matched with the standard personality waveform diagrams of each candidate personality to determine the user personality of the target user.
在本实施例中,终端设备可以计算情感波形图各个候选性格的标准波形图之间的偏差值,并基于该偏差值的倒数计算得到目标用户与各个候选性格之间的匹配度,选取匹配度最高的候选性格作为目标用户的用户性格,当然,也可以选取匹配度大于预设的匹配阈值的多个候选性格作为目标用户的用户性格。In this embodiment, the terminal device can calculate the deviation value between the standard waveform diagrams of each candidate personality of the emotion waveform diagram, and calculate the matching degree between the target user and each candidate personality based on the inverse of the deviation value, and select the matching degree The highest candidate personality is taken as the user personality of the target user. Of course, multiple candidate personalities with a matching degree greater than a preset matching threshold can also be selected as the user personality of the target user.
在S1043中,基于所述用户性格得到所述性格分析报告。In S1043, the personality analysis report is obtained based on the user's personality.
在本实施例中,终端设备可以获取各个用户性格对应的标准语段,基于上述标准语段生成性格分析报告。In this embodiment, the terminal device can obtain the standard language segment corresponding to each user's personality, and generate a personality analysis report based on the above-mentioned standard language segment.
在本申请实施例中,通过生成目标用户的情感波形图,并从候选性格中识别出目标用户的用户性格,生成性格分析报告,从而能够提高了性格分析报告的生成效率。In the embodiment of the present application, by generating the emotional waveform diagram of the target user, and identifying the user personality of the target user from the candidate personality, the personality analysis report is generated, thereby improving the generation efficiency of the personality analysis report.
应理解,上述实施例中各步骤的序号的大小并不意味着执行顺序的先后,各过程的执行顺序应以其功能和内在逻辑确定,而不应对本申请实施例的实施过程构成任何限定。It should be understood that the size of the sequence number of each step in the foregoing embodiment does not mean the order of execution, and the execution sequence of each process should be determined by its function and internal logic, and should not constitute any limitation to the implementation process of the embodiment of the present application.
图8示出了本申请一实施例提供的一种用户报告的生成设备的结构框图,该用户报告的生成设备包括的各单元用于执行图1对应的实施例中的各步骤。具体请参阅图8与图1所对应的实施例中的相关描述。为了便于说明,仅示出了与本实施例相关的部分。FIG. 8 shows a structural block diagram of a device for generating a user report according to an embodiment of the present application, and each unit included in the device for generating a user report is used to execute each step in the embodiment corresponding to FIG. 1. For details, please refer to the relevant description in the embodiment corresponding to FIG. 8 and FIG. 1. For ease of description, only the parts related to this embodiment are shown.
参见图8,所述用户报告的生成设备包括:Referring to Figure 8, the device for generating the user report includes:
会话文本获取单元81,用于获取目标用户在会话过程中产生的多个语音信号,并将各个所述语音信号转换为对应的会话文本;The conversation text obtaining unit 81 is configured to obtain multiple voice signals generated by the target user during a conversation, and convert each of the voice signals into corresponding conversation text;
会话内容集合生成单元82,用于对所述会话文本进行语义分析,得到所述会话文本对应的会话关键词以及各个所述关键词对应的会话标签,生成会话内容集合;The conversation content collection generating unit 82 is configured to perform semantic analysis on the conversation text to obtain the conversation keywords corresponding to the conversation text and the conversation tags corresponding to each of the keywords, and generate a conversation content collection;
情感特征值确定单元83,用于获得所述会话内容集合内的各个所述会话关键词对应的 会话词向量,并基于各个所述会话词向量确定所述语音信号对应的情感特征值;The emotion feature value determining unit 83 is configured to obtain the conversation word vector corresponding to each of the conversation keywords in the conversation content set, and determine the emotion feature value corresponding to the voice signal based on each of the conversation word vectors;
性格分析报告生成单元84,用于基于所有语音信号的所述情感特征值,生成所述目标用户的性格分析报告。The personality analysis report generating unit 84 is configured to generate a personality analysis report of the target user based on the emotional feature values of all voice signals.
可选地,所述情感特征值确定单元83包括:Optionally, the emotional feature value determining unit 83 includes:
加权权重确定单元,用于确定各个所述会话关键词在预设的知识图谱内的关联实体,并获取各个所述关联实体对应的加权权重;A weighted weight determining unit, configured to determine the associated entity of each of the session keywords in the preset knowledge graph, and obtain the weighted weight corresponding to each of the associated entities;
词概念向量生成单元,用于根据所有所述关联实体的所述加权权重,生成所述会话关键词的词概念向量;A word concept vector generating unit, configured to generate the word concept vector of the conversation keyword according to the weighted weights of all the associated entities;
语句概念向量生成单元,用于基于各个所述会话关键词的所属的会话语句,对属于同一所述会话语句的所有词概念向量进行封装,生成所述会话语句的语句概念向量;所述会话语句是对所述会话文本进行语句划分后得到的;The sentence concept vector generating unit is used to encapsulate all the word concept vectors belonging to the same conversation sentence based on the conversation sentence to which each of the conversation keywords belongs to generate the sentence concept vector of the conversation sentence; the conversation sentence Is obtained after sentence division of the conversation text;
对话更新向量生成单元,用于分别将各个所述会话语句的所述语句概念向量导入到第一注意力算法,得到各个所述会话语句的对话更新向量;A dialogue update vector generating unit, configured to respectively import the sentence concept vector of each of the conversational sentences into the first attention algorithm to obtain the dialogue update vector of each of the conversational sentences;
文本概念向量生成单元,用于将所述会话文本的所有会话语句的所述语句概念向量进行封装,生成所述会话文本的对话概念向量,并将所述会话概念向量导入第二注意力模型,生成所述会话文本的文本概念向量;The text concept vector generating unit is used to encapsulate the sentence concept vectors of all conversation sentences of the conversation text, generate the conversation concept vectors of the conversation text, and import the conversation concept vectors into the second attention model, Generating a text concept vector of the conversation text;
情感特征值计算单元,用于根据所述对话更新向量以及所述文本概念向量,确定所述情感特征值。The emotional feature value calculation unit is configured to determine the emotional feature value according to the dialogue update vector and the text concept vector.
可选地,所述加权权重确定单元包括:Optionally, the weighting weight determining unit includes:
关联强度因子确定单元,用于获取各个所述关联实体与所述会话关键词之间的关联强度因子;An association strength factor determination unit, configured to obtain an association strength factor between each of the associated entities and the session keywords;
情感强度因子确定单元,用于基于预设的情感度量算法,确定各个所述关联实体的情感强度因子;The emotion intensity factor determination unit is configured to determine the emotion intensity factor of each associated entity based on a preset emotion measurement algorithm;
加权权重计算单元,用于基于所述情感强度因子以及所述关联强度因子,构建所述关联实体的加权权重。The weighted weight calculation unit is configured to construct the weighted weight of the associated entity based on the emotional intensity factor and the associated intensity factor.
可选地,所述关联强度因子确定单元包括:Optionally, the correlation strength factor determination unit includes:
关联置信度确定单元,用于基于所述知识图谱,确定所述关联实体与所述会话关键词之间的关联置信度;An association confidence degree determining unit, configured to determine the association confidence degree between the associated entity and the session keyword based on the knowledge graph;
会话文本向量确定单元,用于将所述会话关键词关联的会话语句导入到预设的池化层,生成各个所述会话关键词关联的会话语句的语句向量,并基于所述语句向量确定所述会话关键词所在语段的会话文本向量;所述会话文本向量具体为:The conversational text vector determining unit is used to import the conversational sentences associated with the conversational keywords into the preset pooling layer, generate the sentence vectors of the conversational sentences associated with each of the conversational keywords, and determine all the conversational sentences based on the sentence vectors. The conversational text vector of the segment where the conversational keywords are located; the conversational text vector is specifically:
Figure PCTCN2020119300-appb-000021
Figure PCTCN2020119300-appb-000021
其中,CR(X i)为所述会话关键词的所述会话文本向量,所述会话关键词所在的会话文本编号为i;
Figure PCTCN2020119300-appb-000022
为所述会话关键词所在的会话语句的语句向量,所述会话语句在所述会话文本中的语句编号为j;所述M为预设的关联系数;
Wherein, CR(X i ) is the conversation text vector of the conversation keyword, and the conversation text number where the conversation keyword is located is i;
Figure PCTCN2020119300-appb-000022
Is the sentence vector of the conversation sentence where the conversation keyword is located, the sentence number of the conversation sentence in the conversation text is j; the M is a preset correlation coefficient;
关联强度因子计算单元,用于基于所述会话文本向量以及所述关联置信度,计算所述关联强度因子;所述关联强度因子具体为:The correlation strength factor calculation unit is configured to calculate the correlation strength factor based on the conversational text vector and the correlation confidence; the correlation strength factor is specifically:
rel k=max-min(s k)*|cos(CR(X i),c k)| rel k =max-min(s k )*|cos(CR(X i ),c k )|
其中,rel k为第k个会话关键词的所述关联强度因子;c k为所述会话关键词第k个关联实体的关联置信度;max-min(s k)为所述会话关键词第k个关联实体对应的情感极差。 Where rel k is the correlation strength factor of the k-th session keyword; c k is the correlation confidence of the k-th associated entity of the session keyword; max-min(s k ) is the session keyword The emotions corresponding to k associated entities are extremely poor.
可选地,所述情感强度因子确定单元包括:Optionally, the emotion intensity factor determination unit includes:
时间分布图生成单元,用于根据所有所述服务请求包含的请求发起时间,生成关于所述服务类型的请求时间分布图;A time distribution diagram generating unit, configured to generate a request time distribution diagram for the service type according to the request initiation time included in all the service requests;
情感属性识别单元,用于识别所述关联实体的情感属性;The emotional attribute recognition unit is used to recognize the emotional attribute of the associated entity;
非情感类型处理单元,用于若所述关联实体的所述情感属性为非情感类型,则将所述情感强度因子配置为预设的默认值;A non-emotional type processing unit, configured to configure the emotional intensity factor to a preset default value if the emotional attribute of the associated entity is a non-emotional type;
情感类型处理单元,用于若所述关联实体的所述情感属性为情感类型,则通过预设的情感转换算法,计算所述会话关键词的所述情感强度因子;所述情感强度因子具体为:The emotion type processing unit is configured to calculate the emotion intensity factor of the conversation keyword by using a preset emotion conversion algorithm if the emotion attribute of the associated entity is an emotion type; the emotion intensity factor is specifically :
Figure PCTCN2020119300-appb-000023
Figure PCTCN2020119300-appb-000023
其中,aff k为第k个所述关联实体的情感强度因子;VAD(c k)为所述第k个所述关联实体的积极情感分值;A(c k)为所述第k个所述关联实体的情感幅度分值。 Where aff k is the emotional intensity factor of the k-th associated entity; VAD(c k ) is the positive emotional score of the k-th associated entity; A(c k ) is the k-th associated entity The emotional magnitude score of the associated entity.
可选地,所述对话更新向量生成单元包括:Optionally, the dialog update vector generating unit includes:
线性向量生成单元,用于对所述会话语句的语句概念向量进行线性变化,得到包含h个端点的线性向量;其中,所述h为预设的端点个数;The linear vector generating unit is used to linearly change the sentence concept vector of the conversation sentence to obtain a linear vector containing h endpoints; wherein, the h is the preset number of endpoints;
注意力向量生成单元,用于将所述线性向量导入到所述第一注意力算法的多头自注意力层,得到所述会话语句的注意力向量;所述注意力向量具体为:The attention vector generating unit is configured to import the linear vector into the multi-head self-attention layer of the first attention algorithm to obtain the attention vector of the conversation sentence; the attention vector is specifically:
Figure PCTCN2020119300-appb-000024
Figure PCTCN2020119300-appb-000024
其中,
Figure PCTCN2020119300-appb-000025
为第i个会话文本中第n个所述会话语句的所述注意力向量;
Figure PCTCN2020119300-appb-000026
为所述线性向量;d s为基于所述线性向量的端点数h确定的系数值;
among them,
Figure PCTCN2020119300-appb-000025
Is the attention vector of the nth conversation sentence in the ith conversation text;
Figure PCTCN2020119300-appb-000026
Is the linear vector; d s is a coefficient value determined based on the number of endpoints h of the linear vector;
对话更新向量确定单元,用于基于所述注意力向量生成所述会话语句的对话更新向量;所述对话更新向量具体为:The dialogue update vector determining unit is configured to generate a dialogue update vector of the conversation sentence based on the attention vector; the dialogue update vector is specifically:
Figure PCTCN2020119300-appb-000027
Figure PCTCN2020119300-appb-000027
其中,W 1、W 2、b 1以及b 2为所述第一注意力模型的模型参数。 Wherein, W 1 , W 2 , b 1 and b 2 are model parameters of the first attention model.
可选地,所述性格分析报告生成单元84包括:Optionally, the personality analysis report generating unit 84 includes:
情感波形图生成单元,用于根据各个所述语音信号的情感特征值,生成所述目标用户的情感波形图;An emotion waveform diagram generating unit, configured to generate the emotion waveform diagram of the target user according to the emotion characteristic value of each of the voice signals;
用户性格确定单元,用于将所述情感波形图与各个候选性格的标准性格波形图进行匹配,确定所述目标用户的用户性格;A user personality determination unit, configured to match the emotion waveform diagram with the standard personality waveform diagrams of each candidate personality to determine the user personality of the target user;
性格分析报告输出单元,用于基于所述用户性格得到所述性格分析报告。The personality analysis report output unit is configured to obtain the personality analysis report based on the user's personality.
因此,本申请实施例提供的用户报告的生成设备同样不依赖面试人员或会话对象进行手动填写或主观判断,无需用户花费额外时间撰写关于目标用户的性格分析报告,从而能够大大减少用户的操作,并且上述的过程可以通过在会话过程中不同阶段的语音信号,确定情感特征值,并非以单一的话语或语句进行性格判断,从而能够提高性格分析报告的准确性。Therefore, the user report generation device provided by the embodiment of the present application also does not rely on the interviewer or the conversation object to manually fill in or subjectively judge, and does not require the user to spend extra time writing a personality analysis report on the target user, thereby greatly reducing user operations. In addition, the above process can determine the emotional characteristic value through the voice signals at different stages in the conversation process, instead of using a single utterance or sentence to judge the personality, so that the accuracy of the personality analysis report can be improved.
图9是本申请另一实施例提供的一种终端设备的示意图。如图9所示,该实施例的终端设备9包括:处理器90、存储器91以及存储在所述存储器91中并可在所述处理器90 上运行的计算机程序92,例如用户报告的生成程序。所述处理器90执行所述计算机程序92时实现上述各个用户报告的生成方法实施例中的步骤,例如图1所示的S101至S104。或者,所述处理器90执行所述计算机程序92时实现上述各装置实施例中各单元的功能,例如图8所示模块81至84功能。FIG. 9 is a schematic diagram of a terminal device provided by another embodiment of the present application. As shown in FIG. 9, the terminal device 9 of this embodiment includes: a processor 90, a memory 91, and a computer program 92 stored in the memory 91 and running on the processor 90, such as a program for generating user reports . When the processor 90 executes the computer program 92, the steps in the foregoing method for generating user reports are implemented, such as S101 to S104 shown in FIG. 1. Alternatively, when the processor 90 executes the computer program 92, the functions of the units in the foregoing device embodiments, such as the functions of the modules 81 to 84 shown in FIG. 8, are realized.
示例性的,所述计算机程序92可以被分割成一个或多个单元,所述一个或者多个单元被存储在所述存储器91中,并由所述处理器90执行,以完成本申请。所述一个或多个单元可以是能够完成特定功能的一系列计算机程序指令段,该指令段用于描述所述计算机程序92在所述终端设备9中的执行过程。例如,所述计算机程序92可以被分割成会话文本获取单元、会话内容集合生成单元、情感特征值确定单元以及性格分析报告生成单元,各单元具体功能如上所述。Exemplarily, the computer program 92 may be divided into one or more units, and the one or more units are stored in the memory 91 and executed by the processor 90 to complete the application. The one or more units may be a series of computer program instruction segments capable of completing specific functions, and the instruction segments are used to describe the execution process of the computer program 92 in the terminal device 9. For example, the computer program 92 may be divided into a conversation text acquisition unit, a conversation content collection generation unit, an emotional feature value determination unit, and a personality analysis report generation unit, and the specific functions of each unit are as described above.
所述终端设备9可以是桌上型计算机、笔记本、掌上电脑及云端终端设备等计算设备。所述终端设备可包括,但不仅限于,处理器90、存储器91。本领域技术人员可以理解,图9仅仅是终端设备9的示例,并不构成对终端设备9的限定,可以包括比图示更多或更少的部件,或者组合某些部件,或者不同的部件,例如所述终端设备还可以包括输入输出设备、网络接入设备、总线等。The terminal device 9 may be a computing device such as a desktop computer, a notebook, a palmtop computer, and a cloud terminal device. The terminal device may include, but is not limited to, a processor 90 and a memory 91. Those skilled in the art can understand that FIG. 9 is only an example of the terminal device 9 and does not constitute a limitation on the terminal device 9. It may include more or less components than shown in the figure, or a combination of certain components, or different components. For example, the terminal device may also include input and output devices, network access devices, buses, and so on.
所称处理器90可以是中央处理单元(Central Processing Unit,CPU),还可以是其他通用处理器、数字信号处理器(Digital Signal Processor,DSP)、专用集成电路(Application Specific Integrated Circuit,ASIC)、现成可编程门阵列(Field-Programmable Gate Array,FPGA)或者其他可编程逻辑器件、分立门或者晶体管逻辑器件、分立硬件组件等。通用处理器可以是微处理器或者该处理器也可以是任何常规的处理器等。The so-called processor 90 may be a central processing unit (Central Processing Unit, CPU), other general-purpose processors, digital signal processors (Digital Signal Processor, DSP), application specific integrated circuits (Application Specific Integrated Circuit, ASIC), Ready-made programmable gate array (Field-Programmable Gate Array, FPGA) or other programmable logic devices, discrete gates or transistor logic devices, discrete hardware components, etc. The general-purpose processor may be a microprocessor or the processor may also be any conventional processor or the like.
所述存储器91可以是所述终端设备9的内部存储单元,例如终端设备9的硬盘或内存。所述存储器91也可以是所述终端设备9的外部存储设备,例如所述终端设备9上配备的插接式硬盘,智能存储卡(Smart Media Card,SMC),安全数字(Secure Digital,SD)卡,闪存卡(Flash Card)等。进一步地,所述存储器91还可以既包括所述终端设备9的内部存储单元也包括外部存储设备。所述存储器91用于存储所述计算机程序以及所述终端设备所需的其他程序和数据。所述存储器91还可以用于暂时地存储已经输出或者将要输出的数据。The memory 91 may be an internal storage unit of the terminal device 9, for example, a hard disk or a memory of the terminal device 9. The memory 91 may also be an external storage device of the terminal device 9, such as a plug-in hard disk equipped on the terminal device 9, a smart memory card (Smart Media Card, SMC), or a Secure Digital (SD). Card, Flash Card, etc. Further, the memory 91 may also include both an internal storage unit of the terminal device 9 and an external storage device. The memory 91 is used to store the computer program and other programs and data required by the terminal device. The memory 91 can also be used to temporarily store data that has been output or will be output.
本申请实施例还提供了一种计算机可读存储介质,所述计算机可读存储介质存储有计算机程序,所述计算机程序被处理器执行时实现可实现上述各个方法实施例中的步骤。所述计算机可读存储介质可以是非易失性,也可以是易失性。所述计算机可读存储介质包括:U盘、移动硬盘、只读存储器(ROM,Read-Only Memory)、随机存取存储器(RAM,Random Access Memory)、磁碟或者光盘等各种可以存储程序代码的介质。The embodiments of the present application also provide a computer-readable storage medium, where the computer-readable storage medium stores a computer program, and when the computer program is executed by a processor, the steps in each of the foregoing method embodiments can be realized. The computer-readable storage medium may be non-volatile or volatile. The computer-readable storage medium includes: U disk, mobile hard disk, read-only memory (ROM, Read-Only Memory), random access memory (RAM, Random Access Memory), magnetic disks or optical disks, etc., which can store program codes Medium.
另外,在本申请各个实施例中的各功能单元可以集成在一个处理单元中,也可以是各个单元单独物理存在,也可以两个或两个以上单元集成在一个单元中。上述集成的单元既可以采用硬件的形式实现,也可以采用软件功能单元的形式实现。In addition, the functional units in the various embodiments of the present application may be integrated into one processing unit, or each unit may exist alone physically, or two or more units may be integrated into one unit. The above-mentioned integrated unit can be implemented in the form of hardware or software functional unit.
以上所述实施例仅用以说明本申请的技术方案,而非对其限制;尽管参照前述实施例对本申请进行了详细的说明,本领域的普通技术人员应当理解:其依然可以对前述各实施例所记载的技术方案进行修改,或者对其中部分技术特征进行等同替换;而这些修改或者替换,并不使相应技术方案的本质脱离本申请各实施例技术方案的精神和范围,均应包含在本申请的保护范围之内。The above-mentioned embodiments are only used to illustrate the technical solutions of the present application, not to limit them; although the present application has been described in detail with reference to the foregoing embodiments, those of ordinary skill in the art should understand that they can still implement the foregoing The technical solutions recorded in the examples are modified, or some of the technical features are equivalently replaced; these modifications or replacements do not cause the essence of the corresponding technical solutions to deviate from the spirit and scope of the technical solutions of the embodiments of the application, and should be included in Within the scope of protection of this application.

Claims (20)

  1. 一种用户报告的生成方法,其中,包括:A method for generating user reports, including:
    获取目标用户在会话过程中产生的多个语音信号,并将各个所述语音信号转换为对应的会话文本;Acquiring multiple voice signals generated by the target user during the conversation, and converting each of the voice signals into corresponding conversation text;
    对所述会话文本进行语义分析,得到所述会话文本对应的会话关键词以及各个所述关键词对应的会话标签,生成会话内容集合;Performing semantic analysis on the conversation text, obtaining conversation keywords corresponding to the conversation text and conversation tags corresponding to each of the keywords, and generating a conversation content set;
    获得所述会话内容集合内的各个所述会话关键词对应的会话词向量,并基于各个所述会话词向量确定所述语音信号对应的情感特征值;Obtaining a conversation word vector corresponding to each of the conversation keywords in the conversation content set, and determining an emotional feature value corresponding to the voice signal based on each of the conversation word vectors;
    基于所有语音信号的所述情感特征值,生成所述目标用户的性格分析报告。Based on the emotional feature values of all voice signals, a personality analysis report of the target user is generated.
  2. 根据权利要求1所述的生成方法,其中,所述获得所述会话内容集合内的各个所述会话关键词对应的会话词向量,并基于各个所述会话词向量确定所述语音信号对应的情感特征值,包括:The generating method according to claim 1, wherein said obtaining a conversation word vector corresponding to each of the conversation keywords in the conversation content set, and determining the emotion corresponding to the speech signal based on each of the conversation word vectors Characteristic values, including:
    确定各个所述会话关键词在预设的知识图谱内的关联实体,并获取各个所述关联实体对应的加权权重;Determine the associated entity of each of the session keywords in the preset knowledge graph, and obtain the weighted weight corresponding to each of the associated entities;
    根据所有所述关联实体的所述加权权重,生成所述会话关键词的词概念向量;Generating the word concept vector of the conversation keyword according to the weighted weights of all the associated entities;
    基于各个所述会话关键词的所属的会话语句,对属于同一所述会话语句的所有词概念向量进行封装,生成所述会话语句的语句概念向量;所述会话语句是对所述会话文本进行语句划分后得到的;Based on the conversational sentence to which each of the conversational keywords belongs, encapsulate all word concept vectors belonging to the same conversational sentence to generate the sentence concept vector of the conversational sentence; the conversational sentence is a sentence for the conversational text Obtained after division;
    分别将各个所述会话语句的所述语句概念向量导入到第一注意力算法,得到各个所述会话语句的对话更新向量;Importing the sentence concept vector of each of the conversational sentences into the first attention algorithm to obtain the dialogue update vector of each of the conversational sentences;
    将所述会话文本的所有会话语句的所述语句概念向量进行封装,生成所述会话文本的对话概念向量,并将所述会话概念向量导入第二注意力模型,生成所述会话文本的文本概念向量;Encapsulate the sentence concept vectors of all the conversation sentences of the conversation text to generate the conversation concept vector of the conversation text, and import the conversation concept vector into the second attention model to generate the text concept of the conversation text vector;
    根据所述对话更新向量以及所述文本概念向量,确定所述情感特征值。The emotional feature value is determined according to the dialogue update vector and the text concept vector.
  3. 根据权利要求2所述的生成方法,其中,所述确定各个所述会话关键词在预设的知识图谱内的关联实体,并获取各个所述关联实体对应的加权权重,包括:The generating method according to claim 2, wherein the determining the associated entity of each of the session keywords in a preset knowledge graph and obtaining the weighted weight corresponding to each of the associated entities comprises:
    获取各个所述关联实体与所述会话关键词之间的关联强度因子;Acquiring the correlation strength factor between each of the associated entities and the session keywords;
    基于预设的情感度量算法,确定各个所述关联实体的情感强度因子;Determine the emotional intensity factor of each associated entity based on a preset emotion measurement algorithm;
    基于所述情感强度因子以及所述关联强度因子,构建所述关联实体的加权权重。Based on the emotion intensity factor and the association intensity factor, a weighted weight of the associated entity is constructed.
  4. 根据权利要求3所述的生成方法,其中,所述获取各个所述关联实体与所述会话关键词之间的关联强度因子,包括:The generating method according to claim 3, wherein said obtaining the correlation strength factor between each of said associated entities and said session keywords comprises:
    基于所述知识图谱,确定所述关联实体与所述会话关键词之间的关联置信度;Based on the knowledge graph, determining the association confidence between the associated entity and the session keyword;
    将所述会话关键词关联的会话语句导入到预设的池化层,生成各个所述会话关键词关联的会话语句的语句向量,并基于所述语句向量确定所述会话关键词所在语段的会话文本向量;所述会话文本向量具体为:Import the conversational sentences associated with the conversation keywords into the preset pooling layer, generate the sentence vectors of the conversational sentences associated with each of the conversation keywords, and determine based on the sentence vectors the segment of the conversation keywords. Conversation text vector; the conversation text vector is specifically:
    Figure PCTCN2020119300-appb-100001
    Figure PCTCN2020119300-appb-100001
    其中,CR(X i)为所述会话关键词的所述会话文本向量,所述会话关键词所在的会话文本编号为i;
    Figure PCTCN2020119300-appb-100002
    为所述会话关键词所在的会话语句的语句向量,所述会话语句在所述会话文本中的语句编号为j;所述M为预设的关联系数;
    Wherein, CR(X i ) is the conversation text vector of the conversation keyword, and the conversation text number where the conversation keyword is located is i;
    Figure PCTCN2020119300-appb-100002
    Is the sentence vector of the conversation sentence where the conversation keyword is located, the sentence number of the conversation sentence in the conversation text is j; the M is a preset correlation coefficient;
    基于所述会话文本向量以及所述关联置信度,计算所述关联强度因子;所述关联强度因子具体为:Based on the conversational text vector and the correlation confidence, the correlation strength factor is calculated; the correlation strength factor is specifically:
    rel k=max-min(s k)*|cos(CR(X i),c k)| rel k =max-min(s k )*|cos(CR(X i ),c k )|
    其中,rel k为第k个会话关键词的所述关联强度因子;c k为所述会话关键词第k个关联实体的关联置信度;max-min(s k)为所述会话关键词第k个关联实体对应的情感极差。 Where rel k is the correlation strength factor of the k-th session keyword; c k is the correlation confidence of the k-th associated entity of the session keyword; max-min(s k ) is the session keyword The emotions corresponding to k associated entities are extremely poor.
  5. 根据权利要求3所述的生成方法,其中,所述基于预设的情感度量算法,确定各个所述关联实体的情感强度因子,包括:The generating method according to claim 3, wherein the determining the emotion intensity factor of each of the associated entities based on a preset emotion measurement algorithm comprises:
    识别所述关联实体的情感属性;Identifying the emotional attributes of the associated entity;
    若所述关联实体的所述情感属性为非情感类型,则将所述情感强度因子配置为预设的默认值;If the emotional attribute of the associated entity is a non-emotional type, configure the emotional intensity factor as a preset default value;
    若所述关联实体的所述情感属性为情感类型,则通过预设的情感转换算法,计算所述会话关键词的所述情感强度因子;所述情感强度因子具体为:If the emotion attribute of the associated entity is an emotion type, the emotion intensity factor of the conversation keyword is calculated through a preset emotion conversion algorithm; the emotion intensity factor is specifically:
    Figure PCTCN2020119300-appb-100003
    Figure PCTCN2020119300-appb-100003
    其中,aff k为第k个所述关联实体的情感强度因子;VAD(c k)为所述第k个所述关联实体的积极情感分值;A(c k)为所述第k个所述关联实体的情感幅度分值。 Where aff k is the emotional intensity factor of the k-th associated entity; VAD(c k ) is the positive emotional score of the k-th associated entity; A(c k ) is the k-th associated entity The emotional magnitude score of the associated entity.
  6. 根据权利要求2所述的生成方法,其中,所述分别将各个所述会话语句的所述语句概念向量导入到第一注意力算法,得到各个所述会话语句的对话更新向量,包括:The generating method according to claim 2, wherein the respectively importing the sentence concept vector of each of the conversational sentences into a first attention algorithm to obtain the dialogue update vector of each of the conversational sentences comprises:
    对所述会话语句的语句概念向量进行线性变化,得到包含h个端点的线性向量;其中,所述h为预设的端点个数;Linearly change the sentence concept vector of the conversation sentence to obtain a linear vector containing h endpoints; wherein, h is the preset number of endpoints;
    将所述线性向量导入到所述第一注意力算法的多头自注意力层,得到所述会话语句的注意力向量;所述注意力向量具体为:The linear vector is imported into the multi-head self-attention layer of the first attention algorithm to obtain the attention vector of the conversation sentence; the attention vector is specifically:
    Figure PCTCN2020119300-appb-100004
    Figure PCTCN2020119300-appb-100004
    其中,
    Figure PCTCN2020119300-appb-100005
    为第i个会话文本中第n个所述会话语句的所述注意力向量;
    Figure PCTCN2020119300-appb-100006
    为所述线性向量;d s为基于所述线性向量的端点数h确定的系数值;
    among them,
    Figure PCTCN2020119300-appb-100005
    Is the attention vector of the nth conversation sentence in the ith conversation text;
    Figure PCTCN2020119300-appb-100006
    Is the linear vector; d s is a coefficient value determined based on the number of endpoints h of the linear vector;
    基于所述注意力向量生成所述会话语句的对话更新向量;所述对话更新向量具体为:The dialogue update vector of the conversation sentence is generated based on the attention vector; the dialogue update vector is specifically:
    Figure PCTCN2020119300-appb-100007
    Figure PCTCN2020119300-appb-100007
    其中,W 1、W 2、b 1以及b 2为所述第一注意力模型的模型参数。 Wherein, W 1 , W 2 , b 1 and b 2 are model parameters of the first attention model.
  7. 根据权利要求1-6任一项所述的生成方法,其中,所述基于所有语音信号的所述情感特征值,生成所述目标用户的性格分析报告,包括:The generating method according to any one of claims 1 to 6, wherein the generating the personality analysis report of the target user based on the emotional characteristic values of all voice signals comprises:
    根据各个所述语音信号的情感特征值,生成所述目标用户的情感波形图;Generating the emotional waveform diagram of the target user according to the emotional feature value of each of the voice signals;
    将所述情感波形图与各个候选性格的标准性格波形图进行匹配,确定所述目标用户的用户性格;Matching the emotion waveform diagram with the standard personality waveform diagrams of each candidate personality to determine the user personality of the target user;
    基于所述用户性格得到所述性格分析报告。The personality analysis report is obtained based on the user's personality.
  8. 一种用户报告的生成设备,其中,包括:A device for generating user reports, including:
    会话文本获取单元,用于获取目标用户在会话过程中产生的多个语音信号,并将各个 所述语音信号转换为对应的会话文本;The conversation text obtaining unit is used to obtain multiple voice signals generated by the target user in the conversation process, and convert each of the voice signals into corresponding conversation text;
    会话内容集合生成单元,用于对所述会话文本进行语义分析,得到所述会话文本对应的会话关键词以及各个所述关键词对应的会话标签,生成会话内容集合;A conversation content collection generating unit, configured to perform semantic analysis on the conversation text to obtain conversation keywords corresponding to the conversation text and conversation tags corresponding to each of the keywords, and generate a conversation content collection;
    情感特征值确定单元,用于获得所述会话内容集合内的各个所述会话关键词对应的会话词向量,并基于各个所述会话词向量确定所述语音信号对应的情感特征值;An emotion feature value determining unit, configured to obtain a conversation word vector corresponding to each of the conversation keywords in the conversation content set, and determine the emotion feature value corresponding to the voice signal based on each of the conversation word vectors;
    性格分析报告生成单元,用于基于所有语音信号的所述情感特征值,生成所述目标用户的性格分析报告。The personality analysis report generating unit is configured to generate the personality analysis report of the target user based on the emotional characteristic values of all voice signals.
  9. 根据权利要求8所述的生成设备,其中,所述情感特征值确定单元包括:8. The generating device according to claim 8, wherein the emotional feature value determining unit comprises:
    加权权重确定单元,用于确定各个所述会话关键词在预设的知识图谱内的关联实体,并获取各个所述关联实体对应的加权权重;A weighted weight determining unit, configured to determine the associated entity of each of the session keywords in the preset knowledge graph, and obtain the weighted weight corresponding to each of the associated entities;
    词概念向量生成单元,用于根据所有所述关联实体的所述加权权重,生成所述会话关键词的词概念向量;A word concept vector generating unit, configured to generate the word concept vector of the conversation keyword according to the weighted weights of all the associated entities;
    语句概念向量生成单元,用于基于各个所述会话关键词的所属的会话语句,对属于同一所述会话语句的所有词概念向量进行封装,生成所述会话语句的语句概念向量;所述会话语句是对所述会话文本进行语句划分后得到的;The sentence concept vector generating unit is used to encapsulate all the word concept vectors belonging to the same conversation sentence based on the conversation sentence to which each of the conversation keywords belongs to generate the sentence concept vector of the conversation sentence; the conversation sentence Is obtained after sentence division of the conversation text;
    对话更新向量生成单元,用于分别将各个所述会话语句的所述语句概念向量导入到第一注意力算法,得到各个所述会话语句的对话更新向量;A dialogue update vector generating unit, configured to respectively import the sentence concept vector of each of the conversational sentences into the first attention algorithm to obtain the dialogue update vector of each of the conversational sentences;
    文本概念向量生成单元,用于将所述会话文本的所有会话语句的所述语句概念向量进行封装,生成所述会话文本的对话概念向量,并将所述会话概念向量导入第二注意力模型,生成所述会话文本的文本概念向量;The text concept vector generating unit is used to encapsulate the sentence concept vectors of all conversation sentences of the conversation text, generate the conversation concept vectors of the conversation text, and import the conversation concept vectors into the second attention model, Generating a text concept vector of the conversation text;
    情感特征值计算单元,用于根据所述对话更新向量以及所述文本概念向量,确定所述情感特征值。The emotional feature value calculation unit is configured to determine the emotional feature value according to the dialogue update vector and the text concept vector.
  10. 根据权利要求8所述的生成设备,其中,所述加权权重确定单元包括:The generating device according to claim 8, wherein the weighting weight determining unit comprises:
    关联强度因子确定单元,用于获取各个所述关联实体与所述会话关键词之间的关联强度因子;An association strength factor determination unit, configured to obtain an association strength factor between each of the associated entities and the session keywords;
    情感强度因子确定单元,用于基于预设的情感度量算法,确定各个所述关联实体的情感强度因子;The emotion intensity factor determination unit is configured to determine the emotion intensity factor of each associated entity based on a preset emotion measurement algorithm;
    加权权重计算单元,用于基于所述情感强度因子以及所述关联强度因子,构建所述关联实体的加权权重。The weighted weight calculation unit is configured to construct the weighted weight of the associated entity based on the emotional intensity factor and the associated intensity factor.
  11. 一种终端设备,其中,所述终端设备包括存储器、处理器以及存储在所述存储器中并可在所述处理器上运行的计算机可读指令,所述处理器执行所述计算机可读指令时执行以下步骤:A terminal device, wherein the terminal device includes a memory, a processor, and computer-readable instructions that are stored in the memory and can run on the processor, and when the processor executes the computer-readable instructions Perform the following steps:
    获取目标用户在会话过程中产生的多个语音信号,并将各个所述语音信号转换为对应的会话文本;Acquiring multiple voice signals generated by the target user during the conversation, and converting each of the voice signals into corresponding conversation text;
    对所述会话文本进行语义分析,得到所述会话文本对应的会话关键词以及各个所述关键词对应的会话标签,生成会话内容集合;Performing semantic analysis on the conversation text, obtaining conversation keywords corresponding to the conversation text and conversation tags corresponding to each of the keywords, and generating a conversation content set;
    获得所述会话内容集合内的各个所述会话关键词对应的会话词向量,并基于各个所述会话词向量确定所述语音信号对应的情感特征值;Obtaining a conversation word vector corresponding to each of the conversation keywords in the conversation content set, and determining an emotional feature value corresponding to the voice signal based on each of the conversation word vectors;
    基于所有语音信号的所述情感特征值,生成所述目标用户的性格分析报告。Based on the emotional feature values of all voice signals, a personality analysis report of the target user is generated.
  12. 根据权利要求11所述的终端设备,其中,所述获得所述会话内容集合内的各个所述会话关键词对应的会话词向量,并基于各个所述会话词向量确定所述语音信号对应的情感特征值,包括:The terminal device according to claim 11, wherein said obtaining a conversation word vector corresponding to each of the conversation keywords in the conversation content set, and determining the emotion corresponding to the voice signal based on each of the conversation word vectors Characteristic values, including:
    确定各个所述会话关键词在预设的知识图谱内的关联实体,并获取各个所述关联实体对应的加权权重;Determine the associated entity of each of the session keywords in the preset knowledge graph, and obtain the weighted weight corresponding to each of the associated entities;
    根据所有所述关联实体的所述加权权重,生成所述会话关键词的词概念向量;Generating the word concept vector of the conversation keyword according to the weighted weights of all the associated entities;
    基于各个所述会话关键词的所属的会话语句,对属于同一所述会话语句的所有词概念 向量进行封装,生成所述会话语句的语句概念向量;所述会话语句是对所述会话文本进行语句划分后得到的;Based on the conversational sentence to which each of the conversational keywords belongs, encapsulate all word concept vectors belonging to the same conversational sentence to generate the sentence concept vector of the conversational sentence; the conversational sentence is a sentence for the conversational text Obtained after division;
    分别将各个所述会话语句的所述语句概念向量导入到第一注意力算法,得到各个所述会话语句的对话更新向量;Importing the sentence concept vector of each of the conversational sentences into the first attention algorithm to obtain the dialogue update vector of each of the conversational sentences;
    将所述会话文本的所有会话语句的所述语句概念向量进行封装,生成所述会话文本的对话概念向量,并将所述会话概念向量导入第二注意力模型,生成所述会话文本的文本概念向量;Encapsulate the sentence concept vectors of all the conversation sentences of the conversation text, generate the conversation concept vectors of the conversation text, and import the conversation concept vectors into the second attention model to generate the text concept of the conversation text vector;
    根据所述对话更新向量以及所述文本概念向量,确定所述情感特征值。The emotional feature value is determined according to the dialogue update vector and the text concept vector.
  13. 根据权利要求12所述的中的终端设备,其中,所述确定各个所述会话关键词在预设的知识图谱内的关联实体,并获取各个所述关联实体对应的加权权重,包括:The terminal device according to claim 12, wherein the determining the associated entity of each of the session keywords in a preset knowledge graph and obtaining the weighted weight corresponding to each of the associated entities comprises:
    获取各个所述关联实体与所述会话关键词之间的关联强度因子;Acquiring the correlation strength factor between each of the associated entities and the session keywords;
    基于预设的情感度量算法,确定各个所述关联实体的情感强度因子;Determine the emotional intensity factor of each associated entity based on a preset emotion measurement algorithm;
    基于所述情感强度因子以及所述关联强度因子,构建所述关联实体的加权权重。Based on the emotion intensity factor and the association intensity factor, a weighted weight of the associated entity is constructed.
  14. 根据权利要求13所述的终端设备,其中,所述获取各个所述关联实体与所述会话关键词之间的关联强度因子,包括:The terminal device according to claim 13, wherein said obtaining the correlation strength factor between each of said associated entities and said session keywords comprises:
    基于所述知识图谱,确定所述关联实体与所述会话关键词之间的关联置信度;Based on the knowledge graph, determining the association confidence between the associated entity and the session keyword;
    将所述会话关键词关联的会话语句导入到预设的池化层,生成各个所述会话关键词关联的会话语句的语句向量,并基于所述语句向量确定所述会话关键词所在语段的会话文本向量;所述会话文本向量具体为:Import the conversational sentences associated with the conversational keywords into the preset pooling layer, generate sentence vectors of conversational sentences associated with each of the conversational keywords, and determine the segment of the conversational keywords based on the sentence vectors. Conversation text vector; the conversation text vector is specifically:
    Figure PCTCN2020119300-appb-100008
    Figure PCTCN2020119300-appb-100008
    其中,CR(X i)为所述会话关键词的所述会话文本向量,所述会话关键词所在的会话文本编号为i;
    Figure PCTCN2020119300-appb-100009
    为所述会话关键词所在的会话语句的语句向量,所述会话语句在所述会话文本中的语句编号为j;所述M为预设的关联系数;
    Wherein, CR(X i ) is the conversation text vector of the conversation keyword, and the conversation text number where the conversation keyword is located is i;
    Figure PCTCN2020119300-appb-100009
    Is the sentence vector of the conversation sentence where the conversation keyword is located, the sentence number of the conversation sentence in the conversation text is j; the M is a preset correlation coefficient;
    基于所述会话文本向量以及所述关联置信度,计算所述关联强度因子;所述关联强度因子具体为:The correlation strength factor is calculated based on the conversation text vector and the correlation confidence; the correlation strength factor is specifically:
    rel k=max-min(s k)*|cos(CR(X i),c k)| rel k =max-min(s k )*|cos(CR(X i ),c k )|
    其中,rel k为第k个会话关键词的所述关联强度因子;c k为所述会话关键词第k个关联实体的关联置信度;max-min(s k)为所述会话关键词第k个关联实体对应的情感极差。 Where rel k is the correlation strength factor of the k-th session keyword; c k is the correlation confidence of the k-th associated entity of the session keyword; max-min(s k ) is the session keyword The emotions corresponding to k associated entities are extremely poor.
  15. 根据权利要求13所述的终端设备,其中,所述基于预设的情感度量算法,确定各个所述关联实体的情感强度因子,包括:The terminal device according to claim 13, wherein the determining the emotion intensity factor of each of the associated entities based on a preset emotion measurement algorithm comprises:
    识别所述关联实体的情感属性;Identifying the emotional attributes of the associated entity;
    若所述关联实体的所述情感属性为非情感类型,则将所述情感强度因子配置为预设的默认值;If the emotional attribute of the associated entity is a non-emotional type, configure the emotional intensity factor as a preset default value;
    若所述关联实体的所述情感属性为情感类型,则通过预设的情感转换算法,计算所述会话关键词的所述情感强度因子;所述情感强度因子具体为:If the emotion attribute of the associated entity is an emotion type, the emotion intensity factor of the conversation keyword is calculated through a preset emotion conversion algorithm; the emotion intensity factor is specifically:
    Figure PCTCN2020119300-appb-100010
    Figure PCTCN2020119300-appb-100010
    其中,aff k为第k个所述关联实体的情感强度因子;VAD(c k)为所述第k个所述关联实 体的积极情感分值;A(c k)为所述第k个所述关联实体的情感幅度分值。 Where aff k is the emotional intensity factor of the k-th associated entity; VAD(c k ) is the positive emotional score of the k-th associated entity; A(c k ) is the k-th associated entity The emotional magnitude score of the associated entity.
  16. 一种计算机可读存储介质,所述计算机可读存储介质存储有计算机可读指令,其中,所述计算机可读指令被处理器执行时实现如下步骤:A computer-readable storage medium, the computer-readable storage medium stores computer-readable instructions, wherein, when the computer-readable instructions are executed by a processor, the following steps are implemented:
    获取目标用户在会话过程中产生的多个语音信号,并将各个所述语音信号转换为对应的会话文本;Acquiring multiple voice signals generated by the target user during the conversation, and converting each of the voice signals into corresponding conversation text;
    对所述会话文本进行语义分析,得到所述会话文本对应的会话关键词以及各个所述关键词对应的会话标签,生成会话内容集合;Performing semantic analysis on the conversation text, obtaining conversation keywords corresponding to the conversation text and conversation tags corresponding to each of the keywords, and generating a conversation content set;
    获得所述会话内容集合内的各个所述会话关键词对应的会话词向量,并基于各个所述会话词向量确定所述语音信号对应的情感特征值;Obtaining a conversation word vector corresponding to each of the conversation keywords in the conversation content set, and determining an emotional feature value corresponding to the voice signal based on each of the conversation word vectors;
    基于所有语音信号的所述情感特征值,生成所述目标用户的性格分析报告。Based on the emotional feature values of all voice signals, a personality analysis report of the target user is generated.
  17. 根据权利要求16所述的计算机可读存储介质,其中,所述获得所述会话内容集合内的各个所述会话关键词对应的会话词向量,并基于各个所述会话词向量确定所述语音信号对应的情感特征值,包括:The computer-readable storage medium according to claim 16, wherein said obtaining a conversation word vector corresponding to each of the conversation keywords in the conversation content set, and determining the voice signal based on each of the conversation word vectors Corresponding emotional feature values include:
    确定各个所述会话关键词在预设的知识图谱内的关联实体,并获取各个所述关联实体对应的加权权重;Determine the associated entity of each of the session keywords in the preset knowledge graph, and obtain the weighted weight corresponding to each of the associated entities;
    根据所有所述关联实体的所述加权权重,生成所述会话关键词的词概念向量;Generating the word concept vector of the conversation keyword according to the weighted weights of all the associated entities;
    基于各个所述会话关键词的所属的会话语句,对属于同一所述会话语句的所有词概念向量进行封装,生成所述会话语句的语句概念向量;所述会话语句是对所述会话文本进行语句划分后得到的;Based on the conversational sentence to which each of the conversational keywords belongs, encapsulate all word concept vectors belonging to the same conversational sentence to generate the sentence concept vector of the conversational sentence; the conversational sentence is a sentence for the conversational text Obtained after division;
    分别将各个所述会话语句的所述语句概念向量导入到第一注意力算法,得到各个所述会话语句的对话更新向量;Importing the sentence concept vector of each of the conversational sentences into the first attention algorithm to obtain the dialogue update vector of each of the conversational sentences;
    将所述会话文本的所有会话语句的所述语句概念向量进行封装,生成所述会话文本的对话概念向量,并将所述会话概念向量导入第二注意力模型,生成所述会话文本的文本概念向量;Encapsulate the sentence concept vectors of all the conversation sentences of the conversation text, generate the conversation concept vectors of the conversation text, and import the conversation concept vectors into the second attention model to generate the text concept of the conversation text vector;
    根据所述对话更新向量以及所述文本概念向量,确定所述情感特征值。The emotional feature value is determined according to the dialogue update vector and the text concept vector.
  18. 根据权利要求17所述的计算机可读存储介质,其中,所述确定各个所述会话关键词在预设的知识图谱内的关联实体,并获取各个所述关联实体对应的加权权重,包括:18. The computer-readable storage medium according to claim 17, wherein said determining the associated entity of each of the session keywords in a preset knowledge graph and obtaining the weighted weight corresponding to each of the associated entities comprises:
    获取各个所述关联实体与所述会话关键词之间的关联强度因子;Acquiring the correlation strength factor between each of the associated entities and the session keywords;
    基于预设的情感度量算法,确定各个所述关联实体的情感强度因子;Determine the emotional intensity factor of each associated entity based on a preset emotion measurement algorithm;
    基于所述情感强度因子以及所述关联强度因子,构建所述关联实体的加权权重。Based on the emotion intensity factor and the association intensity factor, a weighted weight of the associated entity is constructed.
  19. 根据权利要求18所述的计算机可读存储介质,其中,所述获取各个所述关联实体与所述会话关键词之间的关联强度因子,包括:18. The computer-readable storage medium according to claim 18, wherein said obtaining the correlation strength factor between each of said associated entities and said session keywords comprises:
    基于所述知识图谱,确定所述关联实体与所述会话关键词之间的关联置信度;Based on the knowledge graph, determining the association confidence between the associated entity and the session keyword;
    将所述会话关键词关联的会话语句导入到预设的池化层,生成各个所述会话关键词关联的会话语句的语句向量,并基于所述语句向量确定所述会话关键词所在语段的会话文本向量;所述会话文本向量具体为:Import the conversational sentences associated with the conversational keywords into the preset pooling layer, generate sentence vectors of conversational sentences associated with each of the conversational keywords, and determine the segment of the conversational keywords based on the sentence vectors. Conversation text vector; the conversation text vector is specifically:
    Figure PCTCN2020119300-appb-100011
    Figure PCTCN2020119300-appb-100011
    其中,CR(X i)为所述会话关键词的所述会话文本向量,所述会话关键词所在的会话文本编号为i;
    Figure PCTCN2020119300-appb-100012
    为所述会话关键词所在的会话语句的语句向量,所述会话语句在所述会话文本中的语句编号为j;所述M为预设的关联系数;
    Wherein, CR(X i ) is the conversation text vector of the conversation keyword, and the conversation text number where the conversation keyword is located is i;
    Figure PCTCN2020119300-appb-100012
    Is the sentence vector of the conversation sentence where the conversation keyword is located, the sentence number of the conversation sentence in the conversation text is j; the M is a preset correlation coefficient;
    基于所述会话文本向量以及所述关联置信度,计算所述关联强度因子;所述关联强度因子具体为:The correlation strength factor is calculated based on the conversation text vector and the correlation confidence; the correlation strength factor is specifically:
    rel k=max-min(s k)*|cos(CR(X i),c k)| rel k =max-min(s k )*|cos(CR(X i ),c k )|
    其中,rel k为第k个会话关键词的所述关联强度因子;c k为所述会话关键词第k个关联实体的关联置信度;max-min(s k)为所述会话关键词第k个关联实体对应的情感极差。 Where rel k is the correlation strength factor of the k-th session keyword; c k is the correlation confidence of the k-th associated entity of the session keyword; max-min(s k ) is the session keyword The emotions corresponding to k associated entities are extremely poor.
  20. 如权利要求18所述的计算机可读存储介质,其中,所述基于预设的情感度量算法,确定各个所述关联实体的情感强度因子,包括:18. The computer-readable storage medium of claim 18, wherein the determining the emotional intensity factor of each of the associated entities based on a preset emotion measurement algorithm comprises:
    识别所述关联实体的情感属性;Identifying the emotional attributes of the associated entity;
    若所述关联实体的所述情感属性为非情感类型,则将所述情感强度因子配置为预设的默认值;If the emotional attribute of the associated entity is a non-emotional type, configure the emotional intensity factor as a preset default value;
    若所述关联实体的所述情感属性为情感类型,则通过预设的情感转换算法,计算所述会话关键词的所述情感强度因子;所述情感强度因子具体为:If the emotion attribute of the associated entity is an emotion type, the emotion intensity factor of the conversation keyword is calculated through a preset emotion conversion algorithm; the emotion intensity factor is specifically:
    Figure PCTCN2020119300-appb-100013
    Figure PCTCN2020119300-appb-100013
    其中,aff k为第k个所述关联实体的情感强度因子;VAD(c k)为所述第k个所述关联实体的积极情感分值;A(c k)为所述第k个所述关联实体的情感幅度分值。 Where aff k is the emotional intensity factor of the k-th associated entity; VAD(c k ) is the positive emotional score of the k-th associated entity; A(c k ) is the k-th associated entity The emotional magnitude score of the associated entity.
PCT/CN2020/119300 2020-05-14 2020-09-30 User report generating method and terminal device WO2021114841A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN202010406546.2 2020-05-14
CN202010406546.2A CN111694940A (en) 2020-05-14 2020-05-14 User report generation method and terminal equipment

Publications (1)

Publication Number Publication Date
WO2021114841A1 true WO2021114841A1 (en) 2021-06-17

Family

ID=72477356

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2020/119300 WO2021114841A1 (en) 2020-05-14 2020-09-30 User report generating method and terminal device

Country Status (2)

Country Link
CN (1) CN111694940A (en)
WO (1) WO2021114841A1 (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN116898441A (en) * 2022-08-25 2023-10-20 北京聆心智能科技有限公司 Character testing method and device based on man-machine conversation and electronic equipment

Families Citing this family (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111694940A (en) * 2020-05-14 2020-09-22 平安科技(深圳)有限公司 User report generation method and terminal equipment
TWI741937B (en) * 2021-01-20 2021-10-01 橋良股份有限公司 Judgment system for suitability of talents and implementation method thereof
CN115438142B (en) * 2021-06-02 2023-07-11 戎易商智(北京)科技有限公司 Conversational interactive data analysis report system
CN113962210A (en) * 2021-11-24 2022-01-21 黄河勘测规划设计研究院有限公司 Intelligent report compiling method based on NLP technology
CN114124860A (en) * 2021-11-26 2022-03-01 中国联合网络通信集团有限公司 Session management method, device, equipment and storage medium
CN116522958A (en) * 2023-07-04 2023-08-01 京东科技信息技术有限公司 Session sample generation method, model training method, emotion recognition method and device
CN117171308A (en) * 2023-07-28 2023-12-05 至本医疗科技(上海)有限公司 Method, device and medium for generating scientific research data analysis response information

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20150332599A1 (en) * 2014-05-19 2015-11-19 Educational Testing Service Systems and Methods for Determining the Ecological Validity of An Assessment
CN105095183A (en) * 2014-05-22 2015-11-25 株式会社日立制作所 Text emotional tendency determination method and system
CN105893344A (en) * 2016-03-28 2016-08-24 北京京东尚科信息技术有限公司 User semantic sentiment analysis-based response method and device
US20170243134A1 (en) * 2016-02-10 2017-08-24 RapportBoost.ai Optimization System and Method for Chat-Based Conversations
CN109254993A (en) * 2017-07-07 2019-01-22 北京掌沃云视媒文化传媒有限公司 A kind of text based personality data analysing method and system
CN109766452A (en) * 2019-01-18 2019-05-17 北京工业大学 A kind of character personality analysis method based on social data
CN111694940A (en) * 2020-05-14 2020-09-22 平安科技(深圳)有限公司 User report generation method and terminal equipment

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20150332599A1 (en) * 2014-05-19 2015-11-19 Educational Testing Service Systems and Methods for Determining the Ecological Validity of An Assessment
CN105095183A (en) * 2014-05-22 2015-11-25 株式会社日立制作所 Text emotional tendency determination method and system
US20170243134A1 (en) * 2016-02-10 2017-08-24 RapportBoost.ai Optimization System and Method for Chat-Based Conversations
CN105893344A (en) * 2016-03-28 2016-08-24 北京京东尚科信息技术有限公司 User semantic sentiment analysis-based response method and device
CN109254993A (en) * 2017-07-07 2019-01-22 北京掌沃云视媒文化传媒有限公司 A kind of text based personality data analysing method and system
CN109766452A (en) * 2019-01-18 2019-05-17 北京工业大学 A kind of character personality analysis method based on social data
CN111694940A (en) * 2020-05-14 2020-09-22 平安科技(深圳)有限公司 User report generation method and terminal equipment

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN116898441A (en) * 2022-08-25 2023-10-20 北京聆心智能科技有限公司 Character testing method and device based on man-machine conversation and electronic equipment
CN116898441B (en) * 2022-08-25 2024-03-22 北京聆心智能科技有限公司 Character testing method and device based on man-machine conversation and electronic equipment

Also Published As

Publication number Publication date
CN111694940A (en) 2020-09-22

Similar Documents

Publication Publication Date Title
WO2021114841A1 (en) User report generating method and terminal device
WO2022095380A1 (en) Ai-based virtual interaction model generation method and apparatus, computer device and storage medium
CN109509470B (en) Voice interaction method and device, computer readable storage medium and terminal equipment
CN110457432B (en) Interview scoring method, interview scoring device, interview scoring equipment and interview scoring storage medium
WO2020147428A1 (en) Interactive content generation method and apparatus, computer device, and storage medium
Wu et al. Emotion recognition from text using semantic labels and separable mixture models
WO2021000497A1 (en) Retrieval method and apparatus, and computer device and storage medium
CN113205817B (en) Speech semantic recognition method, system, device and medium
US9483582B2 (en) Identification and verification of factual assertions in natural language
WO2022178969A1 (en) Voice conversation data processing method and apparatus, and computer device and storage medium
CN107886949A (en) A kind of content recommendation method and device
CN112233698B (en) Character emotion recognition method, device, terminal equipment and storage medium
CN111401077A (en) Language model processing method and device and computer equipment
CN110851650B (en) Comment output method and device and computer storage medium
CN113807103B (en) Recruitment method, device, equipment and storage medium based on artificial intelligence
CN113314119A (en) Voice recognition intelligent household control method and device
WO2023124648A1 (en) Text summary generation method and apparatus, device and storage medium
CN114328817A (en) Text processing method and device
CN111126084B (en) Data processing method, device, electronic equipment and storage medium
Zhang Voice keyword retrieval method using attention mechanism and multimodal information fusion
CN113254620B (en) Response method, device and equipment based on graph neural network and storage medium
WO2024093578A1 (en) Voice recognition method and apparatus, and electronic device, storage medium and computer program product
CN109472032A (en) A kind of determination method, apparatus, server and the storage medium of entity relationship diagram
WO2023207566A1 (en) Voice room quality assessment method, apparatus, and device, medium, and product
Liu et al. ProMETheus: An intelligent mobile voice meeting minutes system

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 20900648

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 20900648

Country of ref document: EP

Kind code of ref document: A1