WO2021013126A1 - 一种发送会话消息的方法与设备 - Google Patents

一种发送会话消息的方法与设备 Download PDF

Info

Publication number
WO2021013126A1
WO2021013126A1 PCT/CN2020/103032 CN2020103032W WO2021013126A1 WO 2021013126 A1 WO2021013126 A1 WO 2021013126A1 CN 2020103032 W CN2020103032 W CN 2020103032W WO 2021013126 A1 WO2021013126 A1 WO 2021013126A1
Authority
WO
WIPO (PCT)
Prior art keywords
message
user
conversation
voice
target
Prior art date
Application number
PCT/CN2020/103032
Other languages
English (en)
French (fr)
Inventor
罗剑嵘
Original Assignee
上海盛付通电子支付服务有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 上海盛付通电子支付服务有限公司 filed Critical 上海盛付通电子支付服务有限公司
Publication of WO2021013126A1 publication Critical patent/WO2021013126A1/zh

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L51/00User-to-user messaging in packet-switching networks, transmitted according to store-and-forward or real-time protocols, e.g. e-mail
    • H04L51/04Real-time or near real-time messaging, e.g. instant messaging [IM]
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L51/00User-to-user messaging in packet-switching networks, transmitted according to store-and-forward or real-time protocols, e.g. e-mail
    • H04L51/07User-to-user messaging in packet-switching networks, transmitted according to store-and-forward or real-time protocols, e.g. e-mail characterised by the inclusion of specific contents
    • H04L51/10Multimedia information
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L51/00User-to-user messaging in packet-switching networks, transmitted according to store-and-forward or real-time protocols, e.g. e-mail
    • H04L51/52User-to-user messaging in packet-switching networks, transmitted according to store-and-forward or real-time protocols, e.g. e-mail for supporting social networking services

Definitions

  • This application relates to the field of communications, and in particular to a technology for sending session messages.
  • the social application in the prior art only supports sending the voice message recorded by the user separately. For example, the user presses the record button on a conversation page of the social application to start recording the voice, and when the user lets go, the voice message recorded by the user is directly sent.
  • One purpose of this application is to provide a method and device for sending session messages.
  • a method for sending a session message including:
  • Generate an atomic conversation message and send the atomic conversation message to a second user communicating with the first user on the conversation page via a social server, wherein the atomic conversation message includes the voice message and the target Emoji message.
  • a method for presenting session messages including:
  • atomic conversation message sent by a first user via a social server, where the atomic conversation message includes a voice message of the first user and a target emoticon message corresponding to the voice message;
  • the atomic conversation message is presented in the conversation page of the first user and the second user, wherein the voice message and the target emoticon message are presented in the same message box in the conversation page.
  • a user equipment for sending a session message including:
  • a second module configured to determine the target emoticon message corresponding to the voice message in response to the triggering operation of the voice message sent by the first user
  • a three-module configured to generate an atomic conversation message, and send the atomic conversation message to a second user who communicates with the first user on the conversation page via a social server, wherein the atomic conversation message includes the The voice message and the target emoticon message.
  • a user equipment for presenting conversation messages including:
  • the two-one module is configured to receive an atomic conversation message sent by a first user via a social server, where the atomic conversation message includes a voice message of the first user and a target emoticon message corresponding to the voice message;
  • the second and second module is used to present the atomic conversation message in the conversation page of the first user and the second user, wherein the voice message and the target emoticon message are presented in the same message in the conversation page frame.
  • a device for sending a session message wherein the device includes:
  • Generate an atomic conversation message and send the atomic conversation message to a second user communicating with the first user on the conversation page via a social server, wherein the atomic conversation message includes the voice message and the target Emoji message.
  • a device for presenting session messages wherein the device includes:
  • atomic conversation message sent by a first user via a social server, where the atomic conversation message includes a voice message of the first user and a target emoticon message corresponding to the voice message;
  • the atomic conversation message is presented in the conversation page of the first user and the second user, wherein the voice message and the target emoticon message are presented in the same message box in the conversation page.
  • a computer-readable medium storing instructions, which when executed cause the system to perform the following operations:
  • Generate an atomic conversation message and send the atomic conversation message to a second user communicating with the first user on the conversation page via a social server, wherein the atomic conversation message includes the voice message and the target Emoji message.
  • a computer-readable medium storing instructions, which when executed cause the system to perform the following operations:
  • atomic conversation message sent by a first user via a social server, where the atomic conversation message includes a voice message of the first user and a target emoticon message corresponding to the voice message;
  • the atomic conversation message is presented in the conversation page of the first user and the second user, wherein the voice message and the target emoticon message are presented in the same message box in the conversation page.
  • the present application obtains the user emotion corresponding to the voice message by performing voice analysis on the voice message entered by the user, and automatically generates the expression message corresponding to the voice message according to the user emotion, and treats the voice message and the expression message as one Atomic conversation messages are sent to social objects, and presented in the same message box in the form of atomic conversation messages on the conversation page of the social objects, which can enable users to express their emotions more accurately and vividly, improve the efficiency of sending emoticons, and enhance The user experience, and can avoid the problem of sending voice messages and emoticons as two messages in a group conversation that may be interrupted by other users’ conversation messages and thus affect the smoothness of the user’s expression.
  • Fig. 1 shows a flowchart of a method for sending a session message according to some embodiments of the present application
  • Fig. 2 shows a flowchart of a method for presenting session messages according to some embodiments of the present application
  • FIG. 3 shows a flowchart of a system method for presenting conversation messages according to some embodiments of the present application
  • Fig. 4 shows a structural diagram of a device for sending session messages according to some embodiments of the present application
  • Fig. 5 shows a structural diagram of a device for presenting session messages according to some embodiments of the present application
  • Figure 6 shows an exemplary system that can be used to implement the various embodiments described in this application
  • FIG. 7 shows a schematic diagram of presenting session messages according to some embodiments of the present application.
  • FIG. 8 shows a schematic diagram of presenting a session message according to some embodiments of the present application.
  • the terminal, the device of the service network, and the trusted party all include one or more processors (CPU), input/output interfaces, network interfaces, and memory.
  • processors CPU
  • input/output interfaces network interfaces
  • memory volatile and non-volatile memory
  • the memory may include non-permanent memory in computer readable media, random access memory (RAM) and/or non-volatile memory, such as read-only memory (ROM) or flash memory (flash RAM). Memory is an example of computer readable media.
  • RAM random access memory
  • ROM read-only memory
  • flash RAM flash memory
  • Computer-readable media include permanent and non-permanent, removable and non-removable media, and information storage can be realized by any method or technology.
  • the information can be computer-readable instructions, data structures, program modules, or other data.
  • Examples of computer storage media include, but are not limited to, phase change memory (PRAM), static random access memory (SRAM), dynamic random access memory (DRAM), other types of random access memory (RAM), read-only memory (ROM), electrically erasable programmable read-only memory (EEPROM), flash memory or other memory technology, CD-ROM, digital versatile disc (DVD) or other optical storage, Magnetic cassettes, magnetic tape storage or other magnetic storage devices or any other non-transmission media can be used to store information that can be accessed by computing devices.
  • PRAM phase change memory
  • SRAM static random access memory
  • DRAM dynamic random access memory
  • RAM random access memory
  • ROM read-only memory
  • EEPROM electrically erasable programmable read-only memory
  • flash memory or other memory technology
  • CD-ROM compact disc
  • the equipment referred to in this application includes but is not limited to user equipment, network equipment, or equipment formed by the integration of user equipment and network equipment through a network.
  • the user equipment includes, but is not limited to, any mobile electronic product that can perform human-computer interaction with the user (for example, human-computer interaction through a touchpad), such as a smart phone, a tablet computer, etc., and the mobile electronic product can adopt any operation System, such as android operating system, iOS operating system, etc.
  • the network device includes an electronic device that can automatically perform numerical calculation and information processing in accordance with pre-set or stored instructions.
  • Its hardware includes, but is not limited to, a microprocessor, an application specific integrated circuit (ASIC), and programmable logic.
  • the network device includes, but is not limited to, a computer, a network host, a single network server, a set of multiple network servers, or a cloud composed of multiple servers; here, the cloud is composed of a large number of computers or network servers based on Cloud Computing, Among them, cloud computing is a type of distributed computing, a virtual supercomputer composed of a group of loosely coupled computer sets.
  • the network includes, but is not limited to, the Internet, a wide area network, a metropolitan area network, a local area network, a VPN network, and a wireless ad hoc network (Ad Hoc network).
  • the device may also be a program running on the user equipment, network equipment, or user equipment and network equipment, network equipment, touch terminal or a device formed by integrating network equipment and touch terminal through a network.
  • the emoticon message is input and sent to the social object as a new conversation message.
  • the operation is cumbersome and due to Possible network delays and other factors will cause social objects to fail to receive emoticon messages in time, and affect the expression of user emotions corresponding to the voice messages.
  • the voice messages and emoticons may be talked to by other users.
  • the message is interrupted, which affects the smoothness of the user’s expression.
  • the voice message and the emoticon message are presented as two separate conversation messages on the conversation page of the social object. It is not easy for the social object to combine the voice message and the emoticon message well. Get up, it will affect the social object's understanding of the user emotion corresponding to the voice message.
  • this application obtains the user emotion corresponding to the voice message by performing voice analysis on the voice message entered by the user, and automatically generates the expression message corresponding to the voice message according to the user emotion, and uses the voice message and the expression message as an atom Conversation messages are sent to social objects and presented in the same message box in the form of atomic conversation messages on the conversation page of the social object.
  • This allows users to express their emotions more accurately and vividly, and reduces the need for users to send voice messages.
  • the operation of inputting and sending emoticons improves the efficiency of sending emoticons, reduces the cumbersomeness of sending emoticons, enhances the user experience, and can avoid sending voice messages and emoticons as two messages in a group conversation.
  • voice messages and emoticons are presented as an atomic conversation message on the conversation page of the social object, which can make the social object better
  • the voice message and emoticon message are combined to better understand the user's emotions corresponding to the voice message.
  • Fig. 1 shows a flowchart of a method for sending a session message according to an embodiment of the present application.
  • the method includes step S11, step S12, and step S13.
  • step S11 the user equipment responds to the first user’s voice input triggering operation on the conversation page to start recording the voice message;
  • step S12 the user equipment responds to the first user’s triggering operation of sending the voice message, Determine the target emoticon message corresponding to the voice message;
  • step S13 the user equipment generates an atomic conversation message, and sends the atomic conversation message to the second user communicating with the first user on the conversation page via the social server.
  • the user wherein the atomic conversation message includes the voice message and the target emoticon message.
  • step S11 the user equipment responds to a voice input trigger operation of the first user on the conversation page, and starts to record a voice message.
  • the voice input trigger operation includes, but is not limited to, clicking on the voice input button of the conversation page, pressing and holding the voice input area of the conversation page without releasing the finger, certain predetermined gesture operation, and so on. For example, the first user's finger presses and does not release the voice input area of the conversation page, and starts to record the voice message.
  • step S12 the user equipment determines a target emoticon message corresponding to the voice message in response to the first user's triggering operation of sending the voice message.
  • the sending trigger operation of the voice message includes, but is not limited to, clicking the voice sending button on the conversation page, clicking an emoticon on the conversation page, pressing the finger on the voice input area of the conversation page to start recording the voice and then releasing the finger to leave. Screen, a predetermined gesture operation, etc.
  • the target emoticon message includes but is not limited to the id corresponding to the emoticon, the url link corresponding to the emoticon, the character string generated by Base64 encoding the emoticon image, the InputStream byte input stream corresponding to the emoticon image, and the specific character string corresponding to the emoticon (for example, arrogant emoticon)
  • the corresponding specific character string is "[arrogance]"" and so on.
  • the user clicks on the voice sending button on the conversation page, and performs voice analysis on the voice message "Voice v1" that has been entered to obtain the user emotion corresponding to the voice message “Voice v1", and matching the emotion corresponding to the user emotion “Emotion e1” ", the expression “emoji e1” is used as the target expression corresponding to the voice message "voice v1", and the corresponding target expression message “e1” is generated according to the target expression "emoji e1".
  • step S13 the user equipment generates an atomic conversation message, and sends the atomic conversation message to a second user communicating with the first user on the conversation page via the social server, wherein the atomic conversation message includes all The voice message and the target emoticon message.
  • the second user may be a social user who has a one-to-one conversation with the first user, or may be multiple social users in a group conversation.
  • the first user encapsulates the voice message and the emoticon message into an atomic conversation message Sent to the second user, the voice message and emoticon message are either all successfully sent or all failed to be sent, and are presented in the same message box as an atomic conversation message on the conversation page of the second user, which can avoid being in a group conversation Sending the voice message and the emoticon message as two messages may cause the problem of being interrupted by other users' conversation messages and affecting the smoothness of the user's expression.
  • the voice message is "voice v1" and the target emoticon message is "e1”
  • an atomic conversation message "voice:'voice v1', emoticon:'e1'” is generated, and the atomic conversation message is sent to the social server through
  • the social server sends the atomic conversation message to the second user device used by the second user who communicates with the first user on the conversation page.
  • the determining the target emoticon message corresponding to the voice message includes step S121 (not shown), step S122 (not shown), and step S123 (not shown).
  • step S121 the user The device performs voice analysis on the voice message to determine the emotional feature corresponding to the voice message; in step S122, the user equipment matches and obtains the target expression corresponding to the emotional feature according to the emotional feature; in step S123, The user equipment generates a target expression message corresponding to the voice message according to the target expression.
  • the emotional characteristics include, but are not limited to, emotions such as "laugh”, “crying”, “excitement”, or a combination of multiple different emotions (for example, "crying before laughing", etc.).
  • the local cache, file, database of the user equipment or matching from the corresponding social server obtains the target expression corresponding to the emotional feature, and then generates the corresponding target expression message according to the target expression. For example, perform voice analysis on the voice message "Voice v1”, determine that the emotional feature corresponding to the voice message “Voice v1” is “excited”, and match the target expression "emoji” corresponding to the "excited” emotional feature in the local database of the user device e1", and generate the corresponding target emoticon message "e1" according to the target emoticon "emoticon e1".
  • the step S121 includes step S1211 (not shown) and step S1212 (not shown).
  • the user equipment performs voice analysis on the voice message to extract the voice information The voice feature; in step S1212, the user equipment determines the emotional feature corresponding to the voice feature according to the voice feature.
  • speech features include, but are not limited to, semantics, speech speed, intonation, and so on.
  • the user equipment performs voice analysis on the voice message "Voice v1”, and extracts the semantics of the voice message "Voice v1" as "I am so happy to pay today", the speech rate is "4 words per second", and the intonation is the previous Low to high, language momentum rises. According to semantics, speaking speed, and intonation, the emotional characteristic is determined to be "excited.”
  • the step S122 includes: the user equipment matches one or more pre-stored emotional features in the emoticon library according to the emotional feature to obtain a matching value corresponding to one or more pre-stored emotional features, wherein,
  • the expression library stores a mapping relationship between pre-stored emotional features and corresponding expressions; obtains the pre-stored emotional features with the highest matching value and the matching value reaches a predetermined matching threshold, and determines the expression corresponding to the pre-stored emotional feature as the target expression.
  • the emoticon library may be maintained by the user equipment on the user equipment side, or maintained by the server on the server side. The user equipment obtains emoticons from the response results returned by the server by sending a request to the server to obtain the emoticon library. Library.
  • the pre-stored emotional features in the expression library include “happy”, “sad”, and “fear”, and the predetermined matching threshold is 70. If the emotional feature is "excited”, match the emotional feature with the pre-stored emotional feature to obtain a match The values are 80, 10, and 20 respectively, where “happy” is the pre-stored emotional feature with the best matching value and the matching value reaches the predetermined matching threshold.
  • the expression corresponding to "happy” is determined as the target expression, or if the emotional feature is " Calm”, match the emotional feature with the pre-stored emotional feature, and the matching values obtained are 30, 20, and 10 respectively. Among them, the matching value of "happy" is the highest but the matching value does not reach the predetermined matching threshold, the matching fails and cannot be obtained The target expression corresponding to the emotional feature "excited”.
  • the step S122 includes step S1221 (not shown) and step S1222 (not shown).
  • step S1221 the user equipment matches and obtains one corresponding to the emotional feature according to the emotional feature. Or multiple expressions; in step S1222, the user equipment obtains the target expression selected by the first user from the one or more expressions. For example, according to the emotional feature "happy”, multiple expressions including “emoticon e1", “emoticon e2", and “emoticon e3" corresponding to the emotional feature "happy” are obtained by matching, and these multiple expressions are presented in the conversation On the page, the target emoticon “emoji e1” selected by the first user from the multiple emoticons is then obtained.
  • the step S1221 includes: the user equipment matches one or more pre-stored emotional features in the emoticon library according to the emotional feature to obtain each pre-stored emotion in the one or more pre-stored emotional features The matching value corresponding to the feature, wherein the expression library stores the mapping relationship between the pre-stored emotional feature and the corresponding expression; the one or more pre-stored emotional features are ranked from high to low according to the matching value corresponding to each pre-stored emotional feature
  • the expressions corresponding to the predetermined number of pre-stored emotional features arranged in the front are determined as one or more expressions corresponding to the emotional features.
  • the pre-stored emotional features in the expression library include "happy", “excited”, “sad", and "fear".
  • the emotional feature "excited” is matched with the pre-stored emotional features in the expression library, and the corresponding matching value 80 is obtained. , 90, 10, 20, arrange the pre-stored emotional features in the order of matching value from high to low to get “excited", “happy”, “fear”, and “sad”. The two pre-stored emotional features "excited” will be ranked first "And “happy” are determined as expressions corresponding to the emotional feature "excited”.
  • the voice features include but are not limited to:
  • the semantic feature includes, but is not limited to, the actual meaning of a certain voice that the computer can understand.
  • the semantic feature may be "I am happy to be paid today", “I am sad to fail an exam”, etc.
  • the speaking rate feature includes, but is not limited to, the vocabulary capacity included in a certain voice per unit time.
  • the speaking rate feature can be "4 words per second", “100 words per minute” "Wait.
  • intonation features include, but are not limited to, the rise and fall of the pitch of a certain voice, for example, flat tone, high-rise tone, lowered tone, zigzag tone, etc., among which, flat tone is a smooth and soothing tone.
  • Obvious rise and fall changes are generally used for statements, explanations and explanations without special feelings. They can also express feelings such as dignity, seriousness, grief, and indifference; high rise is low in the front and high in the back, and the language momentum rises. It is generally used to express doubts. Rhetorical questions, surprises, calls, etc.; the lowering tone is high in the front and low in the back, and the momentum gradually decreases.
  • twists and turns are intonation bending , Or first rise and then fall, or first fall and then rise, often aggravate, prolong the part that needs to be highlighted, and cause twists and turns. It is often used to express exaggeration, irony, disgust, irony, and doubt.
  • the step S13 includes: the user equipment submits to the first user a request regarding whether the target emoticon message is sent to a second user communicating with the first user on the conversation page; if The request is approved by the first user, an atomic conversation message is generated, and the atomic conversation message is sent to the second user via a social server, where the atomic conversation message includes the voice message and the target Emoticon message; if the request is rejected by the first user, the voice message is sent to the second user via a social server.
  • the text prompt message "Confirm whether to send the target emoticon message" is presented on the conversation page, and the "Confirm" button and the "Cancel” button are presented below the text prompt message.
  • the method further includes: the user equipment acquiring at least one of the personal information of the first user and one or more emoticons sent in history by the first user; wherein, the step S122 includes: According to the emotional feature and combining at least one of the personal information of the first user and one or more expressions sent by the first user in history, a target expression corresponding to the emotional feature is obtained by matching. For example, if the personal information of the first user includes "gender is female", then it will be matched first to obtain a cute target expression, or if the personal information of the first user includes "hobby is watching anime", then the first user's personal information includes "hobby is watching anime”, then it will be matched first to obtain targets with an anime style. expression.
  • the step S122 includes: the user equipment determines the emotional change trend corresponding to the emotional feature according to the emotional feature; according to the emotional change trend, matching to obtain a plurality of emotional change trends corresponding to the emotional feature Target expressions and presentation sequence information corresponding to the multiple target expressions; wherein, the step S123 includes: generating the voice message corresponding to the multiple target expressions and presentation sequence information corresponding to the multiple target expressions Target emoticon message.
  • the emotion change trend includes, but is not limited to, the change sequence of multiple emotions and the start time and duration of each emotion.
  • the presentation order information includes, but is not limited to, the time when each target expression is presented relative to the start of the voice message. Points and the length of time presented.
  • the emotional change trend is to cry first and then laugh
  • the first to fifth second of the voice message is crying
  • the sixth to tenth second of the voice message is laughter matching to obtain the target expression corresponding to crying is "emoji e1”
  • the corresponding target expression is "Expression e2”
  • the presentation order information is "Expression e1” from the 1st to the 5th second of the voice message
  • the target emoticon message corresponding to the message is "e1: 1 second to 5 seconds, e2: 6 seconds to 10 seconds".
  • Fig. 2 shows a flowchart of a method for presenting session messages according to an embodiment of the present application.
  • the method includes step S21 and step S22.
  • step S21 the user equipment receives the atomic conversation message sent by the first user via the social server, where the atomic conversation message includes the voice message of the first user and the target emoticon message corresponding to the voice message; in step S22 The user equipment presents the atomic conversation message in the conversation page of the first user and the second user, wherein the voice message and the target emoticon message are presented in the same message box in the conversation page.
  • step S21 the user equipment receives an atomic conversation message sent by a first user via a social server, where the atomic conversation message includes a voice message of the first user and a target emoticon message corresponding to the voice message. For example, receiving an atomic conversation message "voice:'voice v1', expression:'e1'" sent by the first user via the server, where the atomic conversation message includes the voice message "voice v1" and the target expression message corresponding to the voice message "E1".
  • step S22 the user equipment presents the atomic conversation message in the conversation page of the first user and the second user, wherein the voice message and the target emoticon message are presented in the same conversation page.
  • Message Box the corresponding target expression is found through the target expression message, and the voice message and the target expression are displayed in the same message box.
  • the target emoticon is "e1”
  • "e1" is the id of the target emoticon. Use this id to find the corresponding target emoticon e1 locally or from the server, and display the voice message "voice v1" and the target emoticon e1 in the same A message box, where the target expression e1 can be displayed at any position in the message box relative to the voice message "Voice v1".
  • the target emoticon message is generated on the first user equipment according to the voice message.
  • the target emoticon message "e1" is automatically generated on the first user equipment according to the voice message "Voice v1".
  • the method further includes: the user equipment detects whether the voice message and the target emoticon message have been successfully received; wherein, the step S22 includes: if the voice message and the target emoticon message Have been successfully received, the atomic conversation message is presented in the conversation page of the first user and the second user, wherein the voice message and the target emoticon message are presented in the same message box in the conversation page ; Otherwise, ignore the atomic conversation message.
  • the voice message "Voice v1" and the target emoticon message "e1" are successfully received, if they are received successfully, the voice message and the target emoticon message are displayed in the same message box, otherwise, if only the target emoticon message is received , If the voice message is not received, or only the voice message is received but the target emoticon message is not received, the received voice message or target emoticon message will not be displayed in the message box, and the received voice will be deleted from the user device Message or target emoticon message.
  • the display position of the target emoticon message relative to the voice message in the same message box is relative to the selected moment of the target emoticon message in the recording period information of the voice message.
  • the location matches.
  • the target emoticon message is selected after the voice message is entered. Accordingly, the target emoticon message is also displayed at the end of the voice message.
  • the target emoticon message is selected when the voice message is halfway through.
  • the target emoticon message is also displayed in the middle of the voice message.
  • the method further includes: the user equipment determines that the target emoticon message and the voice message are in the recording period information of the voice message according to the relative position of the target emoticon message at the selected moment of time.
  • the target emoticon message is selected at one-third of the time the voice message is entered, it is determined that the display position of the target emoticon message is relative to one-third of the display length of the voice message, and relative to the message box
  • the target emoticon message is displayed at a position of one third of the display length of the voice message.
  • the method further includes: the user equipment plays the atomic session message in response to the second user's play triggering operation of the atomic session message.
  • said playing the atomic conversation message may include: playing the voice message; and presenting the target emoticon message on the conversation page in a second presentation mode, wherein the target emoticon message is in the voice The message is presented in the same message box in the first presentation mode before being played. For example, if the second user clicks on the voice message presented on the conversation page, it will start to play the voice message in the atomic conversation message. At this time, if the target emoticon message has a background sound, the target emoticon message can be played while the voice message is being played. Background sound.
  • the first presentation mode includes, but is not limited to, a bubble in a message box, an icon or thumbnail in the message box, or a general indicator (for example, a small red dot) to indicate After the voice message is played, a corresponding expression will be presented.
  • the second presentation method includes but is not limited to a picture or animation displayed anywhere on the conversation page, or, it may also be a dynamic effect of a message box bubble. For example, before the voice message is played, the target emoticon message is displayed in the message box as a smaller "smile" icon. After the voice message is played, the target emoticon message is displayed in a larger "smile" picture.
  • the presentation mode is displayed in the middle of the conversation page.
  • the target emoticon message is presented on the conversation page in the form of a message box bubble.
  • the target emoticon message is displayed as a message box bubble dynamic The presentation of the effect is presented in the conversation page.
  • the second presentation mode is adapted to the current playback content or playback speed in the voice message.
  • the animation frequency of the target expression information in the second presentation mode is adapted to the current playback content or the playback speed in the voice message.
  • the target expression Information is presented with a higher animation frequency.
  • the method further includes: the user equipment responds to the second user's conversion text trigger operation of the voice message, converting the voice message into text information, wherein the target emoticon message is The display position in the text information matches the display position of the target emoticon message relative to the voice message.
  • the target emoticon message is displayed at the end of the voice message, and the user long presses the voice message, the voice message will be converted into text information, and the target emoticon message is also displayed at the end of the text message.
  • the target emoticon message is displayed in the middle of the voice message. If the user long presses the voice message, the operation menu will be displayed on the conversation page. Click the "Convert text" button in the operation menu to display the voice message Converted into text information, and the target emoticon message is also displayed in the middle of the text information.
  • the step S22 includes: the user equipment obtains multiple target expressions matching the voice message and presentation order information corresponding to the multiple target expressions according to the target expression message;
  • the atomic conversation message is presented on a conversation page between a user and a second user, wherein the multiple target emoticons are presented in the same message box in the conversation page as the voice message according to the presentation order information.
  • the target emoticon message is "e1: 1 second to 5 seconds, e2: 6 seconds to 10 seconds", where the target expression corresponding to e1 is "emoticon e1" and the target expression corresponding to e2 is "emoticon e2".
  • the target emoticons obtained by the target emoticon message that match the voice message are “emoticons e1” and “emoticons e2”, and the presentation order information is to present “emoticons e1” from the first second to the fifth second of the voice message, and at the sixth second of the voice message. From the second to the 10th second, "Emotion e2" is displayed. If the total duration of the voice message is 15 seconds, then "Emotion e1" is displayed in the message box at one-third of the display length of the voice message. "Emotion e2" is displayed at a position of two-thirds of the display length of the voice message in.
  • FIG. 3 shows a flowchart of a system method for presenting conversation messages according to some embodiments of the present application
  • step S31 the first user equipment responds to the first user's voice input triggering operation on the conversation page to start recording a voice message.
  • Step S31 is the same as or similar to the foregoing step S11, and will not be repeated here;
  • step S32 the first user equipment determines the target emoticon message corresponding to the voice message in response to the first user's triggering operation of sending the voice message, and step S32 is the same as or similar to the foregoing step S12.
  • step S33 the first user equipment generates an atomic conversation message, and sends the atomic conversation message to a second user communicating with the first user on the conversation page via the social server, Wherein, the atomic conversation message includes the voice message and the target emoticon message, and step S33 is the same as or similar to the aforementioned step S13, and will not be repeated here;
  • step S34 the second user equipment receives the first user via social media The atomic conversation message sent by the server, where the atomic conversation message includes the voice message of the first user and the target emoticon message corresponding to the voice message, step S34 is the same or similar to the foregoing step S21, and will not be repeated here;
  • step S35 the second user equipment presents the atomic conversation message in the conversation page of the first user and the second user, wherein the voice message and the target emoticon message are in the conversation page Presented in the same message box, step S35 is the same or similar to the foregoing step S22, and will not be repeated here.
  • FIG. 4 shows a device for sending a session message according to an embodiment of the present application.
  • the device includes a one-to-one module 11, a one-two module 12, and a one-three module 13.
  • a one-to-one module 11 is used to respond to the first user's voice input triggering operation on the conversation page to start recording a voice message;
  • a one-two module 12 is used to respond to the first user's triggering operation of sending the voice message, Determine the target emoticon message corresponding to the voice message;
  • the one-three module 13 is used to generate an atomic conversation message, and send the atomic conversation message to the second person who communicates with the first user on the conversation page via the social server.
  • the user wherein the atomic conversation message includes the voice message and the target emoticon message.
  • the one-to-one module 11 is used to respond to a voice input trigger operation of the first user on the conversation page to start recording a voice message.
  • the voice input trigger operation includes, but is not limited to, clicking on the voice input button of the conversation page, pressing and holding the voice input area of the conversation page without releasing the finger, certain predetermined gesture operation, and so on.
  • the first user's finger presses and does not release the voice input area of the conversation page, and starts to record the voice message.
  • the one-two module 12 is configured to determine the target emoticon message corresponding to the voice message in response to the triggering operation of the voice message sent by the first user.
  • the sending trigger operation of the voice message includes, but is not limited to, clicking the voice sending button on the conversation page, clicking an emoticon on the conversation page, pressing the finger on the voice input area of the conversation page to start recording the voice and then releasing the finger to leave. Screen, a predetermined gesture operation, etc.
  • the target emoticon message includes but is not limited to the id corresponding to the emoticon, the url link corresponding to the emoticon, the character string generated by Base64 encoding the emoticon image, the InputStream byte input stream corresponding to the emoticon image, and the specific character string corresponding to the emoticon (for example, arrogant emoticon)
  • the corresponding specific character string is "[arrogance]"" and so on.
  • the user clicks on the voice sending button on the conversation page, and performs voice analysis on the voice message "Voice v1" that has been entered to obtain the user emotion corresponding to the voice message “Voice v1", and matching the emotion corresponding to the user emotion “Emotion e1” ", the expression “emoji e1” is used as the target expression corresponding to the voice message "voice v1", and the corresponding target expression message “e1” is generated according to the target expression "emoji e1".
  • the first three module 13 is used to generate an atomic conversation message and send the atomic conversation message to a second user who communicates with the first user on the conversation page via a social server, wherein the atomic conversation message includes all The voice message and the target emoticon message.
  • the second user may be a social user who has a one-to-one conversation with the first user, or may be multiple social users in a group conversation.
  • the first user encapsulates the voice message and the emoticon message into an atomic conversation message Sent to the second user, the voice message and emoticon message are either all successfully sent or all failed to be sent, and are presented in the same message box as an atomic conversation message on the conversation page of the second user, which can avoid being in a group conversation Sending the voice message and the emoticon message as two messages may cause the problem of being interrupted by other users' conversation messages and affecting the smoothness of the user's expression.
  • the voice message is "voice v1" and the target emoticon message is "e1”
  • an atomic conversation message "voice:'voice v1', emoticon:'e1'” is generated, and the atomic conversation message is sent to the social server through
  • the social server sends the atomic conversation message to the second user device used by the second user who communicates with the first user on the conversation page.
  • the determination of the target expression message corresponding to the voice message includes a one-two-one module 121 (not shown), a one-two-two module 122 (not shown), and a one-two-three module 123 (not shown).
  • the one-two-one module 121 is used to perform voice analysis on the voice message to determine the emotional feature corresponding to the voice message;
  • the one-two-two module 122 is used to match and obtain the corresponding emotional feature according to the emotional feature
  • the one-two-three module 123 is used to generate the target expression message corresponding to the voice message according to the target expression.
  • the specific implementations of the one-two-one module 121, the one-two-two module 122, and the one-two-three module 123 are the same as or similar to the embodiment of steps S121, S122 and S123 in FIG. 1, so they will not be repeated here.
  • the citation method is included here.
  • the one-two-one module 121 includes a two-one-one module 1211 (not shown) and a two-one-two module 1212 (not shown).
  • the one-two-one-one module 121 is used to compare the voice
  • the message performs voice analysis to extract voice features in the voice information;
  • the one-two one-two module 1212 is used to determine the emotional feature corresponding to the voice feature according to the voice feature.
  • the specific implementation of the one-two-one-one module 1211 and the one-two-two module 1212 are the same as or similar to the embodiment of steps S1211 and S1212 in FIG. 1, so they will not be repeated here, and they are included here by reference.
  • the one-two-two module 122 is configured to: match one or more pre-stored emotional features in the emoticon library according to the emotional feature to obtain matching values corresponding to one or more pre-stored emotional features,
  • the expression library stores a mapping relationship between a pre-stored emotional feature and a corresponding expression; obtains the pre-stored emotional feature with the highest matching value and the matching value reaches a predetermined matching threshold, and determines the expression corresponding to the pre-stored emotional feature as the target expression .
  • the related operations are the same as or similar to those of the embodiment shown in FIG. 1, so they will not be repeated here, and are included here by reference.
  • the one-two-two module 122 includes a one-two-two-one module 1221 (not shown) and a one-two-two-two module 1222 (not shown).
  • the one-two-two-two module 1222 is used to obtain the target expression selected by the first user from the one or more expressions.
  • the specific implementation of the one-two-two-one module 1221 and the one-two-two-two module 1222 are the same as or similar to the embodiment of steps S1221 and S1222 in FIG. 1, so they will not be repeated here, and they are included here by reference.
  • the one-two-two-one module 1221 is configured to: match one or more pre-stored emotional features in the emoticon library according to the emotional feature to obtain each of the one or more pre-stored emotional features Matching values corresponding to a pre-stored emotional feature, wherein the expression library stores the mapping relationship between the pre-stored emotional feature and the corresponding expression; the one or more pre-stored emotional features are matched according to the matching value corresponding to each pre-stored emotional feature Arrange in the order of high to low, and determine the expressions corresponding to the predetermined number of pre-stored emotional features in front as one or more expressions corresponding to the emotional features.
  • the related operations are the same as or similar to those of the embodiment shown in FIG. 1, so they will not be repeated here, and are included here by reference.
  • the voice feature includes but is not limited to:
  • the one-three module 13 is configured to: submit to the first user a request regarding whether the target emoticon message is sent to the second user communicating with the first user on the conversation page; The request is approved by the first user, an atomic conversation message is generated, and the atomic conversation message is sent to the second user via a social server, wherein the atomic conversation message includes the voice message and the target expression Message; if the request is rejected by the first user, the voice message is sent to the second user via a social server.
  • the related operations are the same as or similar to those of the embodiment shown in FIG. 1, so they will not be repeated here, and are included here by reference.
  • the device is further configured to: obtain at least one of the personal information of the first user and one or more expressions sent by the first user in history; wherein, the one-two-two module 122 It is used to: match at least one of the one or more facial expressions sent by the first user with the personal information of the first user according to the emotional characteristic and obtain a target facial expression corresponding to the emotional characteristic.
  • the related operations are the same as or similar to those in the embodiment shown in FIG. 1, so they will not be repeated here, and they are included here by reference.
  • the device is further configured to: obtain one or more expressions sent by the first user in history; wherein, the one-two-two module 122 is configured to: according to the emotional characteristics, and combine the One or more emoticons sent in history by the first user are matched to obtain a target emoticon corresponding to the emotional feature.
  • the related operations are the same as or similar to those of the embodiment shown in FIG. 1, so they will not be repeated here, and are included here by reference.
  • the one-two-two module 122 is configured to: determine the emotional change trend corresponding to the emotional feature according to the emotional feature; according to the emotional change trend, match to obtain the corresponding emotional change trend Multiple target expressions and presentation sequence information corresponding to the multiple target expressions; wherein, the one-two-three module 123 is configured to generate according to the multiple target expressions and presentation sequence information corresponding to the multiple target expressions
  • the target emoticon message corresponding to the voice message.
  • FIG. 5 shows a device for presenting session messages according to an embodiment of the present application.
  • the device includes a two-one module 21 and a two-two module 22.
  • the two-one module 21 is configured to receive an atomic conversation message sent by a first user via a social server, where the atomic conversation message includes a voice message of the first user and a target emoticon message corresponding to the voice message; the two-two module 22 , For presenting the atomic conversation message in the conversation page of the first user and the second user, wherein the voice message and the target emoticon message are presented in the same message box in the conversation page.
  • the two-to-one module 21 is configured to receive an atomic conversation message sent by a first user via a social server, where the atomic conversation message includes a voice message of the first user and a target emoticon message corresponding to the voice message. For example, receiving an atomic conversation message "voice:'voice v1', expression:'e1'" sent by the first user via the server, where the atomic conversation message includes the voice message "voice v1" and the target expression message corresponding to the voice message "E1".
  • the two-two module 22 is configured to present the atomic conversation message in the conversation page of the first user and the second user, wherein the voice message and the target emoticon message are presented in the same conversation page Message Box.
  • the corresponding target expression is found through the target expression message, and the voice message and the target expression are displayed in the same message box.
  • the target emoticon is "e1”
  • "e1" is the id of the target emoticon. Use this id to find the corresponding target emoticon e1 locally or from the server, and display the voice message "voice v1" and the target emoticon e1 in the same A message box, where the target expression e1 can be displayed at any position in the message box relative to the voice message "Voice v1".
  • the target emoticon message is generated on the first user equipment according to the voice message.
  • the relevant target emoticon message is the same as or similar to the embodiment shown in FIG. 2, so it will not be repeated here, and it is included here by reference.
  • the device is further configured to: detect whether the voice message and the target emoticon message have been successfully received; wherein, the second-two module 22 is configured to: if the voice message and the target The emoticon messages have been successfully received, the atomic conversation message is presented in the conversation page of the first user and the second user, wherein the voice message and the target emoticon message are presented in the same conversation page Message box; otherwise, ignore the atomic conversation message.
  • the related operations are the same as or similar to those of the embodiment shown in FIG. 2, so they will not be repeated here, and are included here by reference.
  • the display position of the target emoticon message relative to the voice message in the same message box is relative to the selected moment of the target emoticon message in the recording period information of the voice message.
  • the location matches.
  • the relevant target emoticon message is the same as or similar to the embodiment shown in FIG. 2, so it will not be repeated here, and it is included here by reference.
  • the device is further configured to: determine that the target emoticon message and the voice message are in the same position according to the relative position of the selected moment of the target emoticon message in the recording period information of the voice message A relative positional relationship in a message box; the second-two module 22 is configured to: present the atomic conversation message in the conversation page of the first user and the second user according to the relative positional relationship, wherein the voice The message and the target emoticon message are presented in the same message box in the conversation page, and the display position of the target emoticon message relative to the voice message in the same message box corresponds to the relative position relationship match.
  • the related operations are the same as or similar to those of the embodiment shown in FIG. 2, so they will not be repeated here, and are included here by reference.
  • the device is further configured to: in response to the second user's play triggering operation of the atomic session message, play the atomic session message.
  • said playing the atomic conversation message may include: playing the voice message; and presenting the target emoticon message on the conversation page in a second presentation mode, wherein the target emoticon message is in the voice The message is presented in the same message box in the first presentation mode before being played.
  • the related operations are the same as or similar to those of the embodiment shown in FIG. 2, so they will not be repeated here, and are included here by reference.
  • the second presentation mode is adapted to the current playback content or playback speed in the voice message.
  • the related second presentation mode is the same as or similar to the embodiment shown in FIG. 2, so it will not be repeated here, and it is included here by reference.
  • the device is further configured to convert the voice message into text information in response to the second user’s textual conversion trigger operation on the voice message, wherein the target emoticon message is The display position in the text information matches the display position of the target emoticon message relative to the voice message.
  • the related operations are the same as or similar to those of the embodiment shown in FIG. 2, so they will not be repeated here, and are included here by reference.
  • the second-two module 22 is configured to: obtain, according to the target expression message, multiple target expressions matching the voice message and presentation order information corresponding to the multiple target expressions;
  • the atomic conversation message is presented on the conversation page of the first user and the second user, wherein the multiple target emoticons are presented in the same message box in the conversation page as the voice message according to the presentation order information.
  • the related operations are the same as or similar to those of the embodiment shown in FIG. 2, so they will not be repeated here, and are included here by reference.
  • Figure 6 shows an exemplary system that can be used to implement the various embodiments described in this application.
  • the system 300 can be used as any device in each of the described embodiments.
  • the system 300 may include one or more computer-readable media having instructions (for example, system memory or NVM/storage device 320) and be coupled with the one or more computer-readable media and configured to execute
  • the instructions are one or more processors (eg, processor(s) 305) that implement modules to perform the actions described in this application.
  • system control module 310 may include any suitable interface controller to provide at least one of the processor(s) 305 and/or any suitable device or component in communication with the system control module 310 Any appropriate interface.
  • the system control module 310 may include a memory controller module 330 to provide an interface to the system memory 315.
  • the memory controller module 330 may be a hardware module, a software module, and/or a firmware module.
  • the system memory 315 may be used to load and store data and/or instructions for the system 300, for example.
  • the system memory 315 may include any suitable volatile memory, such as a suitable DRAM.
  • the system memory 315 may include a double data rate type quad synchronous dynamic random access memory (DDR4 SDRAM).
  • DDR4 SDRAM double data rate type quad synchronous dynamic random access memory
  • system control module 310 may include one or more input/output (I/O) controllers to provide an interface to the NVM/storage device 320 and the communication interface(s) 325.
  • I/O input/output
  • NVM/storage device 320 can be used to store data and/or instructions.
  • the NVM/storage device 320 may include any suitable non-volatile memory (e.g., flash memory) and/or may include any suitable non-volatile storage device(s) (e.g., one or more hard disk drives (HDD), one or more compact disc (CD) drives and/or one or more digital versatile disc (DVD) drives).
  • suitable non-volatile memory e.g., flash memory
  • suitable non-volatile storage device(s) e.g., one or more hard disk drives (HDD), one or more compact disc (CD) drives and/or one or more digital versatile disc (DVD) drives.
  • HDD hard disk drives
  • CD compact disc
  • DVD digital versatile disc
  • the NVM/storage device 320 may include storage resources that are physically part of the device on which the system 300 is installed, or it may be accessed by the device and not necessarily be a part of the device. For example, the NVM/storage device 320 may be accessed via the communication interface(s) 325 through the network.
  • the communication interface(s) 325 may provide an interface for the system 300 to communicate through one or more networks and/or with any other suitable devices.
  • the system 300 can wirelessly communicate with one or more components of a wireless network according to any of one or more wireless network standards and/or protocols.
  • At least one of the processor(s) 305 may be packaged with the logic of one or more controllers of the system control module 310 (eg, the memory controller module 330). For one embodiment, at least one of the processor(s) 305 may be packaged with the logic of one or more controllers of the system control module 310 to form a system in package (SiP). For one embodiment, at least one of the processor(s) 305 may be integrated with the logic of one or more controllers of the system control module 310 on the same mold. For one embodiment, at least one of the processor(s) 305 may be integrated with the logic of one or more controllers of the system control module 310 on the same mold to form a system on chip (SoC).
  • SoC system on chip
  • the system 300 may be, but is not limited to, a server, a workstation, a desktop computing device, or a mobile computing device (for example, a laptop computing device, a holding computing device, a tablet computer, a netbook, etc.).
  • the system 300 may have more or fewer components and/or different architectures.
  • the system 300 includes one or more cameras, keyboards, liquid crystal display (LCD) screens (including touchscreen displays), non-volatile memory ports, multiple antennas, graphics chips, application specific integrated circuits ( ASIC) and speakers.
  • LCD liquid crystal display
  • ASIC application specific integrated circuits
  • the present application also provides a computer-readable storage medium that stores computer code, and when the computer code is executed, the method described in any of the preceding items is executed.
  • the present application also provides a computer program product.
  • the computer program product is executed by a computer device, the method described in any of the preceding items is executed.
  • This application also provides a computer device, which includes:
  • One or more processors are One or more processors;
  • Memory used to store one or more computer programs
  • the one or more processors When the one or more computer programs are executed by the one or more processors, the one or more processors are caused to implement the method as described in any one of the preceding items.
  • this application can be implemented in software and/or a combination of software and hardware, for example, it can be implemented by an application specific integrated circuit (ASIC), a general purpose computer or any other similar hardware device.
  • the software program of the present application may be executed by a processor to realize the steps or functions described above.
  • the software program (including related data structure) of the present application can be stored in a computer-readable recording medium, such as RAM memory, magnetic or optical drive or floppy disk and similar devices.
  • some steps or functions of the present application may be implemented by hardware, for example, as a circuit that cooperates with a processor to execute each step or function.
  • the computer program instructions in the computer-readable medium include but are not limited to source files, executable files, installation package files, etc.
  • the manner in which computer program instructions are executed by the computer includes but not Limited to: the computer directly executes the instruction, or the computer compiles the instruction and then executes the corresponding compiled program, or the computer reads and executes the instruction, or the computer reads and installs the instruction before executing the corresponding post-installation program.
  • the computer-readable medium may be any available computer-readable storage medium or communication medium that can be accessed by a computer.
  • Communication media includes media by which communication signals containing, for example, computer-readable instructions, data structures, program modules, or other data are transmitted from one system to another system.
  • Communication media can include conductive transmission media (such as cables and wires (for example, optical fiber, coaxial, etc.)) and wireless (unguided transmission) media that can propagate energy waves, such as sound, electromagnetic, RF, microwave, and infrared .
  • Computer readable instructions, data structures, program modules or other data may be embodied as, for example, a modulated data signal in a wireless medium such as a carrier wave or similar mechanism such as embodied as part of spread spectrum technology.
  • modulated data signal refers to a signal whose one or more characteristics have been altered or set in such a way as to encode information in the signal. Modulation can be analog, digital or hybrid modulation techniques.
  • a computer-readable storage medium may include volatile, non-volatile, nonvolatile, and nonvolatile, and may be implemented in any method or technology for storing information such as computer-readable instructions, data structures, program modules, or other data. Removable and non-removable media.
  • computer-readable storage media include, but are not limited to, volatile memory, such as random access memory (RAM, DRAM, SRAM); and non-volatile memory, such as flash memory, various read-only memories (ROM, PROM, EPROM) , EEPROM), magnetic and ferromagnetic/ferroelectric memory (MRAM, FeRAM); and magnetic and optical storage devices (hard disks, tapes, CDs, DVDs); or other currently known media or future developments that can be stored for computer systems Computer readable information/data used.
  • volatile memory such as random access memory (RAM, DRAM, SRAM
  • non-volatile memory such as flash memory, various read-only memories (ROM, PROM, EPROM) , EEPROM), magnetic and ferromagnetic/ferroelectric memory (MRAM, FeRAM); and magnetic and optical storage devices (hard disks, tapes, CDs, DVDs); or other currently known media or future developments that can be stored for computer systems Computer readable information/data used.
  • volatile memory such as random access memory (RAM, DRAM,
  • an embodiment according to the present application includes a device including a memory for storing computer program instructions and a processor for executing the program instructions, wherein, when the computer program instructions are executed by the processor, trigger
  • the operation of the device is based on the aforementioned methods and/or technical solutions according to multiple embodiments of the present application.

Abstract

本申请的目的是提供一种发送会话消息的方法与设备,该方法包括:响应于第一用户在会话页面的语音输入触发操作,开始录入语音消息;响应于所述第一用户对所述语音消息的发送触发操作,确定所述语音消息对应的目标表情消息;生成原子会话消息,并将所述原子会话消息经由社交服务器发送至在所述会话页面与所述第一用户通信的第二用户,其中,所述原子会话消息包括所述语音消息和所述目标表情消息。本申请可以使用户更准确生动地表达自己的情绪,提高表情消息的发送效率,增强用户的体验,并且能够避免在群会话中将语音消息和表情消息作为两条消息发送而可能导致的被其他用户的会话消息冲断从而影响用户的表达顺畅性的问题。

Description

一种发送会话消息的方法与设备
本申请是以CN申请号为201910667026.4,申请日为2019.07.23的申请为基础,并主张其优先权,该CN申请的公开内容在此作为整体引入本申请中
技术领域
本申请涉及通信领域,尤其涉及一种用于发送会话消息的技术。
背景技术
随着时代的发展,用户可在社交应用的会话页面中向参与会话的其他成员发送消息,如文本、表情、语音等。然而,现有技术的社交应用中仅支持单独发送用户录制的语音消息,例如,用户在社交应用的一个会话页面中按下录制按钮开始录制语音,当用户松手时直接发送用户录入的语音消息。
发明内容
本申请的一个目的是提供一种发送会话消息的方法与设备。
根据本申请的一个方面,提供了一种发送会话消息的方法,该方法包括:
响应于第一用户在会话页面的语音输入触发操作,开始录入语音消息;
响应于所述第一用户对所述语音消息的发送触发操作,确定所述语音消息对应的目标表情消息;
生成原子会话消息,并将所述原子会话消息经由社交服务器发送至在所述会话页面与所述第一用户通信的第二用户,其中,所述原子会话消息包括所述语音消息和所述目标表情消息。
根据本申请的另一个方面,提供了一种呈现会话消息的方法,该方法包括:
接收第一用户经由社交服务器发送的原子会话消息,其中所述原子会话消息包括所述第一用户的语音消息及所述语音消息对应的目标表情消息;
在所述第一用户与第二用户的会话页面中呈现所述原子会话消息,其中,所述语音消息与所述目标表情消息在所述会话页面中呈现于同一个消息框。
根据本申请的一个方面,提供了一种发送会话消息的用户设备,该设备包括:
一一模块,用于响应于第一用户在会话页面的语音输入触发操作,开始录入语音消息;
一二模块,用于响应于所述第一用户对所述语音消息的发送触发操作,确定所述语音消息对应的目标表情消息;
一三模块,用于生成原子会话消息,并将所述原子会话消息经由社交服务器发送至在所述会话页面与所述第一用户通信的第二用户,其中,所述原子会话消息包括所述语音消息和所述目标表情消息。
根据本申请的另一个方面,提供了一种呈现会话消息的用户设备,该设备包括:
二一模块,用于接收第一用户经由社交服务器发送的原子会话消息,其中所述原子会话消息包括所述第一用户的语音消息及所述语音消息对应的目标表情消息;
二二模块,用于在所述第一用户与第二用户的会话页面中呈现所述原子会话消息,其中,所述语音消息与所述目标表情消息在所述会话页面中呈现于同一个消息框。
根据本申请的一个方面,提供了一种发送会话消息的设备,其中,该设备包括:
响应于第一用户在会话页面的语音输入触发操作,开始录入语音消息;
响应于所述第一用户对所述语音消息的发送触发操作,确定所述语音消息对应的目标表情消息;
生成原子会话消息,并将所述原子会话消息经由社交服务器发送至在所述会话页面与所述第一用户通信的第二用户,其中,所述原子会话消息包括所述语音消息和所述目标表情消息。
根据本申请的另一个方面,提供了一种呈现会话消息的设备,其中,该设备包括:
接收第一用户经由社交服务器发送的原子会话消息,其中所述原子会话消息包括所述第一用户的语音消息及所述语音消息对应的目标表情消息;
在所述第一用户与第二用户的会话页面中呈现所述原子会话消息,其中,所述语音消息与所述目标表情消息在所述会话页面中呈现于同一个消息框。
根据本申请的一个方面,提供了一种存储指令的计算机可读介质,所述指令在被执行时使得系统进行如下操作:
响应于第一用户在会话页面的语音输入触发操作,开始录入语音消息;
响应于所述第一用户对所述语音消息的发送触发操作,确定所述语音消息对应的目标表情消息;
生成原子会话消息,并将所述原子会话消息经由社交服务器发送至在所述会话页面与所述第一用户通信的第二用户,其中,所述原子会话消息包括所述语音消息和所述目标表情消息。
根据本申请的一个方面,提供了一种存储指令的计算机可读介质,所述指令在被执行时使得系统进行如下操作:
接收第一用户经由社交服务器发送的原子会话消息,其中所述原子会话消息包括所述第一用户的语音消息及所述语音消息对应的目标表情消息;
在所述第一用户与第二用户的会话页面中呈现所述原子会话消息,其中,所述语音消息与所述目标表情消息在所述会话页面中呈现于同一个消息框。
与现有技术相比,本申请通过对用户录入的语音消息进行语音分析,得到语音消息对应的用户情绪,并根据用户情绪自动生成语音消息对应的表情消息,并将语音消息和表情消息作为一个原子会话消息发送给社交对象,以在社交对象的会话页面中以原子会话消息的形式呈现在同一个消息框中,可以使用户更准确生动地表达自己的情绪,提高表情消息的发送效率,增强用户的体验,并且能够避免在群会话中将语音消息和表情消息作为两条消息发送而可能导致的被其他用户的会话消息冲断从而影响用户的表达顺畅性的问题。
附图说明
通过阅读参照以下附图所作的对非限制性实施例所作的详细描述,本申请的其它特征、目的和优点将会变得更明显:
图1示出根据本申请一些实施例的一种发送会话消息的方法流程图;
图2示出根据本申请一些实施例的一种呈现会话消息的方法流程图;
图3示出根据本申请一些实施例的一种呈现会话消息的系统方法流程图;
图4示出根据本申请一些实施例的一种发送会话消息的设备结构图;
图5示出根据本申请一些实施例的一种呈现会话消息的设备结构图;
图6示出可被用于实施本申请中所述各个实施例的示例性系统;
图7示出根据本申请一些实施例的一种呈现会话消息的呈现示意图;
图8示出根据本申请一些实施例的一种呈现会话消息的呈现示意图;
附图中相同或相似的附图标记代表相同或相似的部件。
具体实施方式
下面结合附图对本申请作进一步详细描述。
在本申请一个典型的配置中,终端、服务网络的设备和可信方均包括一个或多个处理器(CPU)、输入/输出接口、网络接口和内存。
内存可能包括计算机可读介质中的非永久性存储器,随机存取存储器(RAM)和/或非易失性内存等形式,如只读存储器(ROM)或闪存(flash RAM)。内存是计算机可读介质的示例。
计算机可读介质包括永久性和非永久性、可移动和非可移动媒体可以由任何方法或技术来实现信息存储。信息可以是计算机可读指令、数据结构、程序的模块或其他数据。计算机的存储介质的例子包括,但不限于相变内存(PRAM)、静态随机存取存储器(SRAM)、动态随机存取存储器(DRAM)、其他类型的随机存取存储器(RAM)、只读存储器(ROM)、电可擦除可编程只读存储器(EEPROM)、快闪记忆体或其他内存技术、只读光盘只读存储器(CD-ROM)、数字多功能光盘(DVD)或其他光学存储、磁盒式磁带,磁带磁盘存储或其他磁性存储设备或任何其他非传输介质,可用于存储可以被计算设备访问的信息。
本申请所指设备包括但不限于用户设备、网络设备、或用户设备与网络设备通过网络相集成所构成的设备。所述用户设备包括但不限于任何一种可与用户进行人机交互(例如通过触摸板进行人机交互)的移动电子产品,例如智能手机、平板电脑等,所述移动电子产品可以采用任意操作系统,如android操作系统、iOS操作系统等。其中,所述网络设备包括一种能够按照事先设定或存储的指令,自动进行数值计算和信息处理的电子设备,其硬件包括但不限于微处理器、专用集成电路(ASIC)、可编程逻辑器件(PLD)、现场可编程门阵列(FPGA)、数字信号处理器(DSP)、嵌入式设备等。所述网络设备包括但不限于计算机、网络主机、单个网络服务器、多个网络服务器集或多个服务器构成的云;在此,云由基于云计算(Cloud Computing)的大量计算机或网络服务器构成,其中,云计算是分布式计算的一种,由一群松散耦合的计算机集组成的一个虚拟超级计算机。所述网络包括但不限于互联网、广域网、城域网、局域网、VPN网络、无线自组织网络(Ad Hoc网络)等。优选地,所述设备还可以是运行于所述用户设备、网络设备、或用户设备与网络设备、网络设备、触摸终端或网络设备与触摸终端通过网络相集成所构成的设备上的程 序。
当然,本领域技术人员应能理解上述设备仅为举例,其他现有的或今后可能出现的设备如可适用于本申请,也应包含在本申请保护范围以内,并在此以引用方式包含于此。
在本申请的描述中,“多个”的含义是两个或者更多,除非另有明确具体的限定。
在现有技术中,用户如果想要在语音消息中加入表情,通常只能在录入并发送完语音消息后,输入表情消息并作为一条新的会话消息发送给社交对象,操作比较繁琐,并且由于可能存在的网络延迟等因素,会造成社交对象不能及时收到表情消息,影响语音消息对应的用户情绪的表达,进一步地,在群会话中,语音消息和表情消息之间可能被其他用户的会话消息冲断,从而影响用户的表达顺畅性,同时,语音消息和表情消息作为两条独立的会话消息呈现在社交对象的会话页面中,不容易使社交对象很好地把语音消息和表情消息结合起来,会影响社交对象对语音消息对应的用户情绪的理解。
与现在技术相比,本申请通过对用户录入的语音消息进行语音分析,得到语音消息对应的用户情绪,并根据用户情绪自动生成语音消息对应的表情消息,并将语音消息和表情消息作为一个原子会话消息发送给社交对象,以在社交对象的会话页面中以原子会话消息的形式呈现在同一个消息框中,可以使用户更准确生动地表达自己的情绪,减少了需要用户在发送语音消息后输入表情消息并发送的操作,提高了表情消息的发送效率,减少了表情消息发送的繁琐性,增强了用户的体验,并且能够避免在群会话中将语音消息和表情消息作为两条消息发送而可能导致的被其他用户的会话消息冲断从而影响用户的表达顺畅性的问题,同时,语音消息和表情消息作为一条原子会话消息呈现在社交对象的会话页面中,可以使社交对象更好地把语音消息和表情消息结合起来,从而更好地理解语音消息对应的用户情绪。
图1示出了根据本申请一个实施例的一种发送会话消息的方法流程图,该方法包括步骤S11、步骤S12和步骤S13。在步骤S11中,用户设备响应于第一用户在会话页面的语音输入触发操作,开始录入语音消息;在步骤S12中,用户设备响应于所述第一用户对所述语音消息的发送触发操作,确定所述语音消息对应的目标表情消息;在步骤S13中,用户设备生成原子会话消息,并将所述原子会话消息经由社交服务器发送至在所述会话页面与所述第一用户通信的第二用户,其中,所述原子会话消息包括所述语音消息和所述目标表情消息。
在步骤S11中,用户设备响应于第一用户在会话页面的语音输入触发操作,开始录入语音消息。在一些实施例中,语音输入触发操作包括但不限于点击会话页面的语音输入按钮、手指按住会话页面的语音输入区域不释放、某个预定的手势操作等。例如,第一用户手指按住会话页面的语音输入区域不释放,即开始录入语音消息。
在步骤S12中,用户设备响应于所述第一用户对所述语音消息的发送触发操作,确定所述语音消息对应的目标表情消息。在一些实施例中,语音消息的发送触发操作包括但不限于点击会话页面上的语音发送按钮、点击会话页面上的某个表情、手指按住会话页面的语音输入区域开始录入语音后手指释放离开屏幕、某个预定的手势操作等。目标表情消息包括但不限于表情对应的id、表情对应的url链接、表情图片通过Base64编码后生成的字符串、表情图片对应的InputStream字节输入流、表情对应的特定字符串(例如,傲慢表情对应的特定字符串是“【傲慢】”)等。例如,用户点击会话页面上的语音发送按钮,通过对录入完成的语音消息“语音v1”进行语音分析,得到语音消息“语音v1”对应的用户情绪,匹配得到该用户情绪对应的表情“表情e1”,将表情“表情e1”作为语音消息“语音v1”对应的目标表情,并根据目标表情“表情e1”生成对应的目标表情消息“e1”。
在步骤S13中,用户设备生成原子会话消息,并将所述原子会话消息经由社交服务器发送至在所述会话页面与所述第一用户通信的第二用户,其中,所述原子会话消息包括所述语音消息和所述目标表情消息。在一些实施例中,第二用户可以是与第一用户一对一会话的社交用户,也可以是群会话中的多个社交用户,第一用户将语音消息和表情消息封装成一个原子会话消息发送到第二用户,语音消息和表情消息或者全部发送成功,或者全部发送失败,并在第二用户的会话页面中以原子会话消息的形式呈现在同一个消息框中,可以避免在群会话中将语音消息和表情消息作为两条消息发送而可能导致的被其他用户的会话消息冲断从而影响用户的表达顺畅性的问题。例如,语音消息是“语音v1”,目标表情消息是“e1”,生成原子会话消息“语音:‘语音v1’,表情:‘e1’”,并将该原子会话消息发送到社交服务器,并经由社交服务器将该原子会话消息发送到在会话页面与第一用户通信的第二用户所使用的第二用户设备。
在一些实施例中,所述确定所述语音消息对应的目标表情消息,包括步骤S121(未示出)、步骤S122(未示出)和步骤S123(未示出),在步骤S121中,用户设备对所述语音消息进行语音分析,确定所述语音消息对应的情感特征;在步骤S122中,用户设备根据所述情感特征,匹配获得与所述情感特征对应的目标表情;在步骤S123中,用户设 备根据所述目标表情,生成所述语音消息对应的目标表情消息。在一些实施例中,情感特征包括但不限于“笑”、“哭”、“激动”等情绪,或者多个不同情绪的组合(例如,“先哭后笑”等),根据情感特征,从用户设备本地的缓存、文件、数据库或者从对应的社交服务器匹配获取情感特征对应的目标表情,然后根据目标表情生成对应的目标表情消息。例如,对语音消息“语音v1”进行语音分析,确定语音消息“语音v1”对应的情感特征是“激动”,并在用户设备本地的数据库中匹配获得“激动”情感特征对应的目标表情“表情e1”,并根据目标表情“表情e1”生成对应的目标表情消息“e1”。
在一些实施例中,所述步骤S121包括步骤S1211(未示出)和步骤S1212(未示出),在步骤S1211中,用户设备对所述语音消息进行语音分析,提取出所述语音信息中的语音特征;在步骤S1212中,用户设备根据所述语音特征,确定所述语音特征对应的情感特征。在一些实施例中,语音特征包括但不限于语义、语速、语调等。例如,用户设备对语音消息“语音v1”进行语音分析,提取出语音消息“语音v1”的语义是“今天发工资了好开心”、语速是“每秒钟4个字”、语调是前低后高、语势上升,根据语义、语速、语调,确定情感特征是“激动”。
在一些实施例中,所述步骤S122包括:用户设备根据所述情感特征,与表情库中的一个或多个预存情感特征进行匹配,得到一个或多个预存情感特征对应的匹配值,其中,所述表情库中存储有预存情感特征与对应表情的映射关系;获取匹配值最高且所述匹配值达到预定匹配阈值的预存情感特征,将所述预存情感特征对应的表情确定为目标表情。在一些实施例中,表情库可以在用户设备端,由用户设备维护,也可以在服务器端,由服务器维护,用户设备通过向服务器发出获取表情库的请求,在服务器返回的响应结果中获取表情库。例如,表情库中的预存情感特征包括“高兴”、“难过”、“害怕”,预定匹配阈值是70,若情感特征是“激动”,将该情感特征与预存情感特征进行匹配,得到的匹配值分别是80、10、20,其中,“高兴”是匹配值最该且匹配值达到预定匹配阈值的预存情感特征,将“高兴”对应的表情确定为目标表情,或者,若情感特征是“平静”,将该情感特征与预存情感特征进行匹配,得到的匹配值分别是30、20、10,其中,“高兴”的匹配值最高但匹配值没有达到预定匹配阈值,则匹配失败,无法获取情感特征“激动”对应的目标表情。
在一些实施例中,所述步骤S122包括步骤S1221(未示出)和步骤S1222(未示出),在步骤S1221中,用户设备根据所述情感特征,匹配获得与所述情感特征对应的一个或多个表情;在步骤S1222中,用户设备获取所述第一用户自所述一个或多个表情中选择的目 标表情。例如,根据情感特征“高兴”,匹配获得与情感特征“高兴”对应的包括“表情e1”、“表情e2”、“表情e3”在内的多个表情,并将这多个表情呈现在会话页面上,然后获取第一用户从这多个表情中选择的目标表情“表情e1”。
在一些实施例中,所述步骤S1221包括:用户设备根据所述情感特征,与表情库中的一个或多个预存情感特征进行匹配,得到所述一个或多个预存情感特征中每个预存情感特征对应的匹配值,其中,所述表情库中存储有预存情感特征与对应表情的映射关系;将所述一个或多个预存情感特征按每个预存情感特征对应的匹配值按从高到低的顺序排列,将排在前面的预定数量的预存情感特征对应的表情确定为与所述情感特征对应的一个或多个表情。例如,表情库中的预存情感特征包括“高兴”、“兴奋”、“难过”、“害怕”,将情感特征“激动”与表情库中的预存情感特征进行匹配,分别得到对应的匹配值80、90、10、20,将预存情感特征按照匹配值从高到低的顺序排列得到“兴奋”、“高兴”、“害怕”、“难过”,将排在前面两位的预存情感特征“兴奋”和“高兴”确定为与情感特征“激动”对应的表情。
在一些实施例中,所述语音特征包括但不限于:
1)语义特征
在一些实施例中,语义特征包括但不限于计算机所能够理解的某个语音实际所要表达的含义,例如,语义特征可以是“今天发工资了好开心”、“考试不及格很难过”等。
2)语速特征
在一些实施例中,语速特征包括但不限于某个语音在单位时间内所包括的词汇容量的多少,例如,语速特征可以是“每秒钟4个字”、“每分钟100个字”等。
3)语调特征
在一些实施例中,语调特征包括但不限于某个语音的音调的高低升降,例如,平直调、高升调、降抑调、曲折调等,其中,平直调是语势平稳舒缓,没有明显的升降变化,一般用于不带特殊感情的陈述、解释和说明,还可以表示庄重、严肃、悲痛、冷淡等感情;高升调是前低后高,语势上升,一般用来表示疑问、反问、惊讶、号召等语气;降抑调是前高后低,语势渐降,一般用于陈述句、感叹句、祈使句,表示肯定、感叹、自信、赞叹、祝福等感情;曲折调是语调弯曲,或先升后降,或先降后升,往往把需要突出的部分加重、拖长并造成曲折,常用来表示夸张、讽刺、厌恶、反语、怀疑等语气。
4)上述任意语音特征的组合
在一些实施例中,所述步骤S13包括:用户设备向所述第一用户提交关于所述目标表 情消息是否发送至在所述会话页面与所述第一用户通信的第二用户的请求;若所述请求被所述第一用户批准,生成原子会话消息,并将所述原子会话消息经由社交服务器发送至所述第二用户,其中,所述原子会话消息包括所述语音消息和所述目标表情消息;若所述请求被所述第一用户拒绝,将所述语音消息经由社交服务器发送至所述第二用户。例如,在发送语音消息前在会话页面上呈现“确认是否发送目标表情消息”的文字提示信息,并在该文字提示信息下方呈现“确认”按钮和“取消”按钮,若用户点击“确认”按钮,则将语音消息和目标表情消息封装成原子会话消息经由社交服务器发送给第二用户,若用户点击“取消”按钮,则单独将语音消息经由社交服务器发送给第二用户。
在一些实施例中,所述方法还包括:用户设备获取所述第一用户的个人信息与所述第一用户历史发送的一个或多个表情中至少一项;其中,所述步骤S122包括:根据所述情感特征,并结合所述第一用户的个人信息与所述第一用户历史发送的一个或多个表情中至少一项,匹配获得与所述情感特征对应的目标表情。例如,第一用户的个人信息包括“性别是女性”,则优先匹配获取比较可爱的目标表情,或者,第一用户的个人信息包括“兴趣爱好是看动漫”则优先匹配获取偏动漫风格的目标表情。又例如,在所有与该情感特征匹配的多个表情中,“表情e1”是第一用户历史发送次数最多的表情,则确定“表情e1”是该情感特征对应的目标表情,或者“表情e2”是第一用户在最近一周时间内发送次数最多的表情,则确定“表情e2”是该情感特征对应的目标表情。
在一些实施例中,所述步骤S122包括:用户设备根据所述情感特征,确定所述情感特征对应的情感变化趋势;根据所述情感变化趋势,匹配获得与所述情感变化趋势对应的多个目标表情以及所述多个目标表情对应的呈现顺序信息;其中,所述步骤S123包括:根据所述多个目标表情以及所述多个目标表情对应的呈现顺序信息,生成所述语音消息对应的目标表情消息。在一些实施例中,情感变化趋势包括但不限于多个情感的变化顺序以及每个情感的开始时间、持续时间,呈现顺序信息包括但不限于每个目标表情相对于语音消息的开始呈现的时间点以及呈现的时间长度。例如,情感变化趋势是先哭后笑、语音消息的第1秒到第5秒是哭,语音消息的第6秒到第10秒是笑匹配获得哭对应的目标表情是“表情e1”,笑对应的目标表情是“表情e2”,呈现顺序信息是语音消息的第1秒到第5秒呈现“表情e1”、语音消息的第6秒到第10秒呈现“表情e2”,以此生成语音消息对应的目标表情消息“e1:1秒-5秒,e2:6秒-10秒”。
图2示出了根据本申请一个实施例的一种呈现会话消息的方法流程图,该方法包 括步骤S21和步骤S22。在步骤S21中,用户设备接收第一用户经由社交服务器发送的原子会话消息,其中所述原子会话消息包括所述第一用户的语音消息及所述语音消息对应的目标表情消息;在步骤S22中,用户设备在所述第一用户与第二用户的会话页面中呈现所述原子会话消息,其中,所述语音消息与所述目标表情消息在所述会话页面中呈现于同一个消息框。
在步骤S21中,用户设备接收第一用户经由社交服务器发送的原子会话消息,其中所述原子会话消息包括所述第一用户的语音消息及所述语音消息对应的目标表情消息。例如,接收第一用户经由服务器发送的原子会话消息“语音:‘语音v1’,表情:‘e1’”,其中,该原子会话消息包括语音消息“语音v1”和该语音消息对应的目标表情消息“e1”。
在步骤S22中,用户设备在所述第一用户与第二用户的会话页面中呈现所述原子会话消息,其中,所述语音消息与所述目标表情消息在所述会话页面中呈现于同一个消息框。在一些实施例中,通过目标表情消息来找到对应的目标表情,并将语音消息和目标表情显示在同一个消息框中。例如,目标表情是“e1”,“e1”是目标表情的id,通过此id从用户设备本地或服务器中找到对应的目标表情e1,并将语音消息“语音v1”和目标表情e1显示在同一个消息框中,其中,目标表情e1可以显示消息框中相对于语音消息“语音v1”的任意位置处。
在一些实施例中,所述目标表情消息是在第一用户设备上根据所述语音消息生成的。例如,目标表情消息“e1”是在第一用户设备上根据语音消息“语音v1”自动生成的。
在一些实施例中,所述方法还包括:用户设备检测所述语音消息与所述目标表情消息是否均已成功接收;其中,所述步骤S22包括:若所述语音消息与所述目标表情消息均已成功接收,在所述第一用户与第二用户的会话页面中呈现所述原子会话消息,其中,所述语音消息与所述目标表情消息在所述会话页面中呈现于同一个消息框;否则,忽略所述原子会话消息。例如,检测语音消息“语音v1”和目标表情消息“e1”是否都成功接收,若都成功接收,则在同一个消息框中显示语音消息和目标表情消息,否则,若只接收到了目标表情消息,未接收到语音消息,或者,只接收了语音消息,未接收到目标表情消息,则不会在消息框中显示接收到的语音消息或目标表情消息,并从用户设备上删除接收到的语音消息或目标表情消息。
在一些实施例中,所述目标表情消息在所述同一个消息框中相对所述语音消息的显示位置,与所述目标表情消息的被选择时刻在所述语音消息的录制时段信息中的相对位置相 匹配。例如,目标表情消息是在语音消息录入完成后才被选择的,相应地,也将目标表情消息显示在语音消息的末尾位置,又例如,目标表情消息是在语音消息录到一半的时候被选择的,相应地,也将目标表情消息显示在语音消息的正中间位置。
在一些实施例中,所述方法还包括:用户设备根据所述目标表情消息的被选择时刻在所述语音消息的录制时段信息中的相对位置,确定所述目标表情消息与所述语音消息在同一个消息框中的相对位置关系;所述步骤S22包括:用户设备根据所述相对位置关系在所述第一用户与第二用户的会话页面中呈现所述原子会话消息,其中,所述语音消息与所述目标表情消息在所述会话页面中呈现于所述同一个消息框,所述目标表情消息在所述同一个消息框中相对所述语音消息的显示位置与所述相对位置关系相匹配。例如,根据目标表情消息在语音消息录入到三分之一的时刻被选择,确定目标表情消息的显示位置是相对于语音消息的显示长度的三分之一的位置,并在消息框中相对于语音消息的显示长度的三分之一的位置处显示目标表情消息。
在一些实施例中,所述方法还包括:用户设备响应于所述第二用户对所述原子会话消息的播放触发操作,播放所述原子会话消息。其中,所述播放所述原子会话消息,可以包括:播放所述语音消息;并将所述目标表情消息以第二呈现方式呈现于所述会话页面,其中,所述目标表情消息在所述语音消息被播放前以第一呈现方式呈现于所述同一个消息框。例如,第二用户点击会话页面上呈现的语音消息,会开始播放原子会话消息中的语音消息,此时,若目标表情消息带有背景声音,可以在播放语音消息的同时,播放目标表情消息中的背景声音。在一些实施例中,第一呈现方式包括但不限于消息框的气泡、消息框中的一个图标或缩略图,或者,还可以是一个通用指示符(例如,一个小红点),用于指示这条语音消息播放后会呈现对应的表情,第二呈现方式包括但不限于在会话页面的任意位置显示的图片或动画,或者,还可以是,消息框气泡的动态效果。例如,语音消息在被播放之前,目标表情消息以一个较小的“微笑”图标的呈现方式显示在消息框中,在语音消息被播放之后,目标表情消息以一个较大的“微笑”图片的呈现方式显示在会话页面的正中间的位置处。如图7所示,在语音消息被播放前,目标表情消息以消息框气泡的呈现方式呈现在会话页面中,如图8所示,在语音消息被播放后,目标表情消息以消息框气泡动态效果的呈现方式呈现在会话页面中。
在一些实施例中,所述第二呈现方式与所述语音消息中的当前播放内容或播放语速相适应。例如,目标表情信息在第二呈现方式下的动画频率与所述语音消息中的当前播放内 容或播放语速相适应,例如,当前播放内容为较紧急内容或播放语速较快时,目标表情信息以较高动画频率呈现。本领域技术人员应能理解,可通过语音识别或语义分析等方式确定语音消息的当前播放内容是否紧急或当前播放语速的快慢,例如,涉及“火警”或“报警”等字词的应为较紧急内容,或者,若语音消息的当前语速高于该用户的平均语速,则确定该语音消息的当前播放语速较快。
在一些实施例中,所述方法还包括:用户设备响应于所述第二用户对所述语音消息的转换文字触发操作,将所述语音消息转换为文本信息,其中,所述目标表情消息在所述文本信息中的显示位置与所述目标表情消息相对所述语音消息的显示位置相匹配。例如,在消息框中,目标表情消息显示在语音消息的末尾,用户长按该语音消息,会将该语音消息转换为文本信息,并且将目标表情消息也同样显示在该文本信息的末尾,又例如,在消息框中,目标表情消息显示在语音消息的正中间,用户长按该语音消息,会在会话页面上呈现操作菜单,点击操作菜单中的“转换文字”按钮,会将该语音消息转换为文本信息,并且将目标表情消息也同样显示在该文本信息的正中间位置处。
在一些实施例中,所述步骤S22包括:用户设备根据所述目标表情消息获得与所述语音消息相匹配的多个目标表情以及所述多个目标表情对应的呈现顺序信息;在所述第一用户与第二用户的会话页面中呈现所述原子会话消息,其中,所述多个目标表情根据所述呈现顺序信息与所述语音消息在所述会话页面中呈现于同一个消息框。例如,目标表情消息是“e1:1秒-5秒,e2:6秒-10秒”,其中,e1对应的目标表情是“表情e1”,e2对应的目标表情是“表情e2”,根据该目标表情消息获得与语音消息相匹配的目标表情是“表情e1”和“表情e2”,呈现顺序信息是在语音消息的第1秒到第5秒呈现“表情e1”、在语音消息的第6秒到第10秒呈现“表情e2”,若语音消息的总时长是15秒,则在消息框中相对于语音消息的显示长度的三分之一的位置处显示“表情e1”,在消息框中相对于语音消息的显示长度的三分之二的位置处显示“表情e2”。
图3示出根据本申请一些实施例的一种呈现会话消息的系统方法流程图;
如图3所述,在步骤S31中,第一用户设备响应于第一用户在会话页面的语音输入触发操作,开始录入语音消息,步骤S31和前述步骤S11相同或者相似,在此不再赘述;在步骤S32中,所述第一用户设备响应于所述第一用户对所述语音消息的发送触发操作,确定所述语音消息对应的目标表情消息,步骤S32和前述步骤S12相同或者相似,在此不再赘述;在步骤S33中,所述第一用户设备生成原子会话消息,并将所述原子会话消息 经由社交服务器发送至在所述会话页面与所述第一用户通信的第二用户,其中,所述原子会话消息包括所述语音消息和所述目标表情消息,步骤S33和前述步骤S13相同或者相似,在此不再赘述;在步骤S34中,第二用户设备接收第一用户经由社交服务器发送的原子会话消息,其中,所述原子会话消息包括所述第一用户的语音消息及所述语音消息对应的目标表情消息,步骤S34和前述步骤S21相同或者相似,在此不再赘述;在步骤S35中,所述第二用户设备在所述第一用户与第二用户的会话页面中呈现所述原子会话消息,其中,所述语音消息与所述目标表情消息在所述会话页面中呈现于同一个消息框,步骤S35和前述步骤S22相同或者相似,在此不再赘述。
图4出了根据本申请一个实施例的一种发送会话消息的设备,该设备包括一一模块11、一二模块12和一三模块13。一一模块11,用于响应于第一用户在会话页面的语音输入触发操作,开始录入语音消息;一二模块12,用于响应于所述第一用户对所述语音消息的发送触发操作,确定所述语音消息对应的目标表情消息;一三模块13,用于生成原子会话消息,并将所述原子会话消息经由社交服务器发送至在所述会话页面与所述第一用户通信的第二用户,其中,所述原子会话消息包括所述语音消息和所述目标表情消息。
一一模块11,用于响应于第一用户在会话页面的语音输入触发操作,开始录入语音消息。在一些实施例中,语音输入触发操作包括但不限于点击会话页面的语音输入按钮、手指按住会话页面的语音输入区域不释放、某个预定的手势操作等。例如,第一用户手指按住会话页面的语音输入区域不释放,即开始录入语音消息。
一二模块12,用于响应于所述第一用户对所述语音消息的发送触发操作,确定所述语音消息对应的目标表情消息。在一些实施例中,语音消息的发送触发操作包括但不限于点击会话页面上的语音发送按钮、点击会话页面上的某个表情、手指按住会话页面的语音输入区域开始录入语音后手指释放离开屏幕、某个预定的手势操作等。目标表情消息包括但不限于表情对应的id、表情对应的url链接、表情图片通过Base64编码后生成的字符串、表情图片对应的InputStream字节输入流、表情对应的特定字符串(例如,傲慢表情对应的特定字符串是“【傲慢】”)等。例如,用户点击会话页面上的语音发送按钮,通过对录入完成的语音消息“语音v1”进行语音分析,得到语音消息“语音v1”对应的用户情绪,匹配得到该用户情绪对应的表情“表情e1”,将表情“表情e1”作为语音消息“语音v1”对应的目标表情,并根据目标表情“表情e1”生成对应的目标表情消息“e1”。
一三模块13,用于生成原子会话消息,并将所述原子会话消息经由社交服务器发送 至在所述会话页面与所述第一用户通信的第二用户,其中,所述原子会话消息包括所述语音消息和所述目标表情消息。在一些实施例中,第二用户可以是与第一用户一对一会话的社交用户,也可以是群会话中的多个社交用户,第一用户将语音消息和表情消息封装成一个原子会话消息发送到第二用户,语音消息和表情消息或者全部发送成功,或者全部发送失败,并在第二用户的会话页面中以原子会话消息的形式呈现在同一个消息框中,可以避免在群会话中将语音消息和表情消息作为两条消息发送而可能导致的被其他用户的会话消息冲断从而影响用户的表达顺畅性的问题。例如,语音消息是“语音v1”,目标表情消息是“e1”,生成原子会话消息“语音:‘语音v1’,表情:‘e1’”,并将该原子会话消息发送到社交服务器,并经由社交服务器将该原子会话消息发送到在会话页面与第一用户通信的第二用户所使用的第二用户设备。
在一些实施例中,所述确定所述语音消息对应的目标表情消息,包括一二一模块121(未示出)、一二二模块122(未示出)和一二三模块123(未示出),一二一模块121用于对所述语音消息进行语音分析,确定所述语音消息对应的情感特征;一二二模块122用于根据所述情感特征,匹配获得与所述情感特征对应的目标表情;一二三模块123用于根据所述目标表情,生成所述语音消息对应的目标表情消息。在此,一二一模块121、一二二模块122和一二三模块123的具体实现方式与图1中有关步骤S121、S122和S123的实施例相同或相近,故不再赘述,在此以引用方式包含于此。
在一些实施例中,所述一二一模块121包括一二一一模块1211(未示出)和一二一二模块1212(未示出),一二一一模块121用于对所述语音消息进行语音分析,提取出所述语音信息中的语音特征;一二一二模块1212用于根据所述语音特征,确定所述语音特征对应的情感特征。在此,一二一一模块1211和一二一二模块1212的具体实现方式与图1中有关步骤S1211和S1212的实施例相同或相近,故不再赘述,在此以引用方式包含于此。
在一些实施例中,所述一二二模块122用于:根据所述情感特征,与表情库中的一个或多个预存情感特征进行匹配,得到一个或多个预存情感特征对应的匹配值,其中,所述表情库中存储有预存情感特征与对应表情的映射关系;获取匹配值最高且所述匹配值达到预定匹配阈值的预存情感特征,将所述预存情感特征对应的表情确定为目标表情。在此,相关操作与图1所示实施例相同或相近,故不再赘述,在此以引用方式包含于此。
在一些实施例中,所述一二二模块122包括一二二一模块1221(未示出)和一二二 二模块1222(未示出),一二二一模块1221用于根据所述情感特征,匹配获得与所述情感特征对应的一个或多个表情;一二二二模块1222用于获取所述第一用户自所述一个或多个表情中选择的目标表情。在此,一二二一模块1221和一二二二模块1222的具体实现方式与图1中有关步骤S1221和S1222的实施例相同或相近,故不再赘述,在此以引用方式包含于此。
在一些实施例中,所述一二二一模块1221用于:根据所述情感特征,与表情库中的一个或多个预存情感特征进行匹配,得到所述一个或多个预存情感特征中每个预存情感特征对应的匹配值,其中,所述表情库中存储有预存情感特征与对应表情的映射关系;将所述一个或多个预存情感特征按每个预存情感特征对应的匹配值按从高到低的顺序排列,将排在前面的预定数量的预存情感特征对应的表情确定为与所述情感特征对应的一个或多个表情。在此,相关操作与图1所示实施例相同或相近,故不再赘述,在此以引用方式包含于此。
在一些实施例中,所述语音特征包括但不限于:
1)语义特征
2)语速特征
3)语调特征
4)上述任意语音特征的组合
在此,相关语音特征与图1所示实施例相同或相近,故不再赘述,在此以引用方式包含于此。
在一些实施例中,一三模块13用于:向所述第一用户提交关于所述目标表情消息是否发送至在所述会话页面与所述第一用户通信的第二用户的请求;若所述请求被所述第一用户批准,生成原子会话消息,并将所述原子会话消息经由社交服务器发送至所述第二用户,其中,所述原子会话消息包括所述语音消息和所述目标表情消息;若所述请求被所述第一用户拒绝,将所述语音消息经由社交服务器发送至所述第二用户。在此,相关操作与图1所示实施例相同或相近,故不再赘述,在此以引用方式包含于此。
在一些实施例中,所述设备还用于:获取所述第一用户的个人信息与所述第一用户历史发送的一个或多个表情中至少一项;其中,所述一二二模块122用于:根据所述情感特征,并结合所述第一用户的个人信息与所述第一用户历史发送的一个或多个表情中至少一项,匹配获得与所述情感特征对应的目标表情。在此,相关操作与图1所示实施例相同或 相近,故不再赘述,在此以引用方式包含于此。
在一些实施例中,所述设备还用于:获取所述第一用户历史发送的一个或多个表情;其中,所述一二二模块122用于:根据所述情感特征,并结合所述第一用户历史发送的一个或多个表情,匹配获得与所述情感特征对应的目标表情。在此,相关操作与图1所示实施例相同或相近,故不再赘述,在此以引用方式包含于此。
在一些实施例中,所述一二二模块122用于:根据所述情感特征,确定所述情感特征对应的情感变化趋势;根据所述情感变化趋势,匹配获得与所述情感变化趋势对应的多个目标表情以及所述多个目标表情对应的呈现顺序信息;其中,所述一二三模块123用于:根据所述多个目标表情以及所述多个目标表情对应的呈现顺序信息,生成所述语音消息对应的目标表情消息。在此,相关操作与图1所示实施例相同或相近,故不再赘述,在此以引用方式包含于此。
图5出了根据本申请一个实施例的一种呈现会话消息的设备,该设备包括二一模块21和二二模块22。二一模块21,用于接收第一用户经由社交服务器发送的原子会话消息,其中所述原子会话消息包括所述第一用户的语音消息及所述语音消息对应的目标表情消息;二二模块22,用于在所述第一用户与第二用户的会话页面中呈现所述原子会话消息,其中,所述语音消息与所述目标表情消息在所述会话页面中呈现于同一个消息框。
二一模块21,用于接收第一用户经由社交服务器发送的原子会话消息,其中所述原子会话消息包括所述第一用户的语音消息及所述语音消息对应的目标表情消息。例如,接收第一用户经由服务器发送的原子会话消息“语音:‘语音v1’,表情:‘e1’”,其中,该原子会话消息包括语音消息“语音v1”和该语音消息对应的目标表情消息“e1”。
二二模块22,用于在所述第一用户与第二用户的会话页面中呈现所述原子会话消息,其中,所述语音消息与所述目标表情消息在所述会话页面中呈现于同一个消息框。在一些实施例中,通过目标表情消息来找到对应的目标表情,并将语音消息和目标表情显示在同一个消息框中。例如,目标表情是“e1”,“e1”是目标表情的id,通过此id从用户设备本地或服务器中找到对应的目标表情e1,并将语音消息“语音v1”和目标表情e1显示在同一个消息框中,其中,目标表情e1可以显示消息框中相对于语音消息“语音v1”的任意位置处。
在一些实施例中,所述目标表情消息是在第一用户设备上根据所述语音消息生成的。在此,相关目标表情消息与图2所示实施例相同或相近,故不再赘述,在此以引用方式包含于此。
在一些实施例中,所述设备还用于:检测所述语音消息与所述目标表情消息是否均已成功接收;其中,所述二二模块22用于:若所述语音消息与所述目标表情消息均已成功接收,在所述第一用户与第二用户的会话页面中呈现所述原子会话消息,其中,所述语音消息与所述目标表情消息在所述会话页面中呈现于同一个消息框;否则,忽略所述原子会话消息。在此,相关操作与图2所示实施例相同或相近,故不再赘述,在此以引用方式包含于此。
在一些实施例中,所述目标表情消息在所述同一个消息框中相对所述语音消息的显示位置,与所述目标表情消息的被选择时刻在所述语音消息的录制时段信息中的相对位置相匹配。在此,相关目标表情消息与图2所示实施例相同或相近,故不再赘述,在此以引用方式包含于此。
在一些实施例中,所述设备还用于:根据所述目标表情消息的被选择时刻在所述语音消息的录制时段信息中的相对位置,确定所述目标表情消息与所述语音消息在同一个消息框中的相对位置关系;所述二二模块22用于:根据所述相对位置关系在所述第一用户与第二用户的会话页面中呈现所述原子会话消息,其中,所述语音消息与所述目标表情消息在所述会话页面中呈现于所述同一个消息框,所述目标表情消息在所述同一个消息框中相对所述语音消息的显示位置与所述相对位置关系相匹配。在此,相关操作与图2所示实施例相同或相近,故不再赘述,在此以引用方式包含于此。
在一些实施例中,所述设备还用于:响应于所述第二用户对所述原子会话消息的播放触发操作,播放所述原子会话消息。其中,所述播放所述原子会话消息,可以包括:播放所述语音消息;并将所述目标表情消息以第二呈现方式呈现于所述会话页面,其中,所述目标表情消息在所述语音消息被播放前以第一呈现方式呈现于所述同一个消息框。在此,相关操作与图2所示实施例相同或相近,故不再赘述,在此以引用方式包含于此。
在一些实施例中,所述第二呈现方式与所述语音消息中的当前播放内容或播放语速相适应。在此,相关第二呈现方式与图2所示实施例相同或相近,故不再赘述,在此以引用方式包含于此。
在一些实施例中,所述设备还用于:响应于所述第二用户对所述语音消息的转换文字触发操作,将所述语音消息转换为文本信息,其中,所述目标表情消息在所述文本信息中的显示位置与所述目标表情消息相对所述语音消息的显示位置相匹配。在此,相关操作与图2所示实施例相同或相近,故不再赘述,在此以引用方式包含于此。
在一些实施例中,所述二二模块22用于:根据所述目标表情消息获得与所述语音消息相匹配的多个目标表情以及所述多个目标表情对应的呈现顺序信息;在所述第一用户与第二用户的会话页面中呈现所述原子会话消息,其中,所述多个目标表情根据所述呈现顺序信息与所述语音消息在所述会话页面中呈现于同一个消息框。在此,相关操作与图2所示实施例相同或相近,故不再赘述,在此以引用方式包含于此。
图6出了可被用于实施本申请中所述的各个实施例的示例性系统。
如图6示在一些实施例中,系统300能够作为各所述实施例中的任意一个设备。在一些实施例中,系统300可包括具有指令的一个或多个计算机可读介质(例如,系统存储器或NVM/存储设备320)以及与该一个或多个计算机可读介质耦合并被配置为执行指令以实现模块从而执行本申请中所述的动作的一个或多个处理器(例如,(一个或多个)处理器305)。
对于一个实施例,系统控制模块310可包括任意适当的接口控制器,以向(一个或多个)处理器305中的至少一个和/或与系统控制模块310通信的任意适当的设备或组件提供任意适当的接口。
系统控制模块310可包括存储器控制器模块330,以向系统存储器315提供接口。存储器控制器模块330可以是硬件模块、软件模块和/或固件模块。
系统存储器315可被用于例如为系统300加载和存储数据和/或指令。对于一个实施例,系统存储器315可包括任意适当的易失性存储器,例如,适当的DRAM。在一些实施例中,系统存储器315可包括双倍数据速率类型四同步动态随机存取存储器(DDR4SDRAM)。
对于一个实施例,系统控制模块310可包括一个或多个输入/输出(I/O)控制器,以向NVM/存储设备320及(一个或多个)通信接口325提供接口。
例如,NVM/存储设备320可被用于存储数据和/或指令。NVM/存储设备320可包括任意适当的非易失性存储器(例如,闪存)和/或可包括任意适当的(一个或多个)非易失性存储设备(例如,一个或多个硬盘驱动器(HDD)、一个或多个光盘(CD)驱动器和/或一个或多个数字通用光盘(DVD)驱动器)。
NVM/存储设备320可包括在物理上作为系统300被安装在其上的设备的一部分的存储资源,或者其可被该设备访问而不必作为该设备的一部分。例如,NVM/存储设备320可通过网络经由(一个或多个)通信接口325进行访问。
(一个或多个)通信接口325可为系统300提供接口以通过一个或多个网络和/或与任意其他适当的设备通信。系统300可根据一个或多个无线网络标准和/或协议中的任意标准和/或协议来与无线网络的一个或多个组件进行无线通信。
对于一个实施例,(一个或多个)处理器305中的至少一个可与系统控制模块310的一个或多个控制器(例如,存储器控制器模块330)的逻辑封装在一起。对于一个实施例,(一个或多个)处理器305中的至少一个可与系统控制模块310的一个或多个控制器的逻辑封装在一起以形成系统级封装(SiP)。对于一个实施例,(一个或多个)处理器305中的至少一个可与系统控制模块310的一个或多个控制器的逻辑集成在同一模具上。对于一个实施例,(一个或多个)处理器305中的至少一个可与系统控制模块310的一个或多个控制器的逻辑集成在同一模具上以形成片上系统(SoC)。
在各个实施例中,系统300可以但不限于是:服务器、工作站、台式计算设备或移动计算设备(例如,膝上型计算设备、持有计算设备、平板电脑、上网本等)。在各个实施例中,系统300可具有更多或更少的组件和/或不同的架构。例如,在一些实施例中,系统300包括一个或多个摄像机、键盘、液晶显示器(LCD)屏幕(包括触屏显示器)、非易失性存储器端口、多个天线、图形芯片、专用集成电路(ASIC)和扬声器。
本申请还提供了一种计算机可读存储介质,所述计算机可读存储介质存储有计算机代码,当所述计算机代码被执行时,如前任一项所述的方法被执行。
本申请还提供了一种计算机程序产品,当所述计算机程序产品被计算机设备执行时,如前任一项所述的方法被执行。
本申请还提供了一种计算机设备,所述计算机设备包括:
一个或多个处理器;
存储器,用于存储一个或多个计算机程序;
当所述一个或多个计算机程序被所述一个或多个处理器执行时,使得所述一个或多个处理器实现如前任一项所述的方法。
需要注意的是,本申请可在软件和/或软件与硬件的组合体中被实施,例如,可采用专用集成电路(ASIC)、通用目的计算机或任何其他类似硬件设备来实现。在一个实施例中,本申请的软件程序可以通过处理器执行以实现上文所述步骤或功能。同样地,本申请的软件程序(包括相关的数据结构)可以被存储到计算机可读记录介质中,例如,RAM存储器,磁或光驱动器或软磁盘及类似设备。另外,本申请的一些步骤 或功能可采用硬件来实现,例如,作为与处理器配合从而执行各个步骤或功能的电路。
另外,本申请的一部分可被应用为计算机程序产品,例如计算机程序指令,当其被计算机执行时,通过该计算机的操作,可以调用或提供根据本申请的方法和/或技术方案。本领域技术人员应能理解,计算机程序指令在计算机可读介质中的存在形式包括但不限于源文件、可执行文件、安装包文件等,相应地,计算机程序指令被计算机执行的方式包括但不限于:该计算机直接执行该指令,或者该计算机编译该指令后再执行对应的编译后程序,或者该计算机读取并执行该指令,或者该计算机读取并安装该指令后再执行对应的安装后程序。在此,计算机可读介质可以是可供计算机访问的任意可用的计算机可读存储介质或通信介质。
通信介质包括藉此包含例如计算机可读指令、数据结构、程序模块或其他数据的通信信号被从一个系统传送到另一系统的介质。通信介质可包括有导的传输介质(诸如电缆和线(例如,光纤、同轴等))和能传播能量波的无线(未有导的传输)介质,诸如声音、电磁、RF、微波和红外。计算机可读指令、数据结构、程序模块或其他数据可被体现为例如无线介质(诸如载波或诸如被体现为扩展频谱技术的一部分的类似机制)中的已调制数据信号。术语“已调制数据信号”指的是其一个或多个特征以在信号中编码信息的方式被更改或设定的信号。调制可以是模拟的、数字的或混合调制技术。
作为示例而非限制,计算机可读存储介质可包括以用于存储诸如计算机可读指令、数据结构、程序模块或其它数据的信息的任何方法或技术实现的易失性和非易失性、可移动和不可移动的介质。例如,计算机可读存储介质包括,但不限于,易失性存储器,诸如随机存储器(RAM,DRAM,SRAM);以及非易失性存储器,诸如闪存、各种只读存储器(ROM,PROM,EPROM,EEPROM)、磁性和铁磁/铁电存储器(MRAM,FeRAM);以及磁性和光学存储设备(硬盘、磁带、CD、DVD);或其它现在已知的介质或今后开发的能够存储供计算机系统使用的计算机可读信息/数据。
在此,根据本申请的一个实施例包括一个装置,该装置包括用于存储计算机程序指令的存储器和用于执行程序指令的处理器,其中,当该计算机程序指令被该处理器执行时,触发该装置运行基于前述根据本申请的多个实施例的方法和/或技术方案。
对于本领域技术人员而言,显然本申请不限于上述示范性实施例的细节,而且在不背离本申请的精神或基本特征的情况下,能够以其他的具体形式实现本申请。因此,无论从哪一点来看,均应将实施例看作是示范性的,而且是非限制性的,本申请的范 围由所附权利要求而不是上述说明限定,因此旨在将落在权利要求的等同要件的含义和范围内的所有变化涵括在本申请内。不应将权利要求中的任何附图标记视为限制所涉及的权利要求。此外,显然“包括”一词不排除其他单元或步骤,单数不排除复数。装置权利要求中陈述的多个单元或装置也可以由一个单元或装置通过软件或者硬件来实现。第一,第二等词语用来表示名称,而并不表示任何特定的顺序。

Claims (19)

  1. 一种发送会话消息的方法,用于第一用户设备,其特征在于,所述方法包括:
    响应于第一用户在会话页面的语音输入触发操作,开始录入语音消息;
    响应于所述第一用户对所述语音消息的发送触发操作,确定所述语音消息对应的目标表情消息;
    生成原子会话消息,并将所述原子会话消息经由社交服务器发送至在所述会话页面与所述第一用户通信的第二用户,其中,所述原子会话消息包括所述语音消息和所述目标表情消息。
  2. 根据权利要求1所述的方法,其特征在于,所述确定所述语音消息对应的目标表情消息,包括:
    对所述语音消息进行语音分析,确定所述语音消息对应的情感特征;
    根据所述情感特征,匹配获得与所述情感特征对应的目标表情;
    根据所述目标表情,生成所述语音消息对应的目标表情消息。
  3. 根据权利要求2所述的方法,其特征在于,所述对所述语音消息进行语音分析,确定所述语音消息对应的情感特征,包括:
    对所述语音消息进行语音分析,提取出所述语音信息中的语音特征;
    根据所述语音特征,确定所述语音特征对应的情感特征。
  4. 根据权利要求2或3所述的方法,其特征在于,所述根据所述情感特征,匹配获得与所述情感特征对应的目标表情,包括:
    根据所述情感特征,与表情库中的一个或多个预存情感特征进行匹配,得到一个或多个预存情感特征对应的匹配值,其中,所述表情库中存储有预存情感特征与对应表情的映射关系;
    获取匹配值最高且所述匹配值达到预定匹配阈值的预存情感特征,将所述预存情感特征对应的表情确定为目标表情。
  5. 根据权利要求2或3所述的方法,其特征在于,所述根据所述情感特征,匹配获得与所述情感特征对应的目标表情,包括:
    根据所述情感特征,匹配获得与所述情感特征对应的一个或多个表情;
    获取所述第一用户自所述一个或多个表情中选择的目标表情。
  6. 根据权利要求5所述的方法,其特征在于,所述根据所述情感特征,匹配获得与所述情感特征对应的一个或多个表情,包括:
    根据所述情感特征,与表情库中的一个或多个预存情感特征进行匹配,得到所述一个或多个预存情感特征中每个预存情感特征对应的匹配值,其中,所述表情库中存储有预存情感特征与对应表情的映射关系;
    将所述一个或多个预存情感特征按每个预存情感特征对应的匹配值按从高到低的顺序排列,将排在前面的预定数量的预存情感特征对应的表情确定为与所述情感特征对应的一个或多个表情。
  7. 根据权利要求2所述的方法,其特征在于,所述方法还包括:
    获取所述第一用户的个人信息与所述第一用户历史发送的一个或多个表情中至少一项;
    其中,所述根据所述情感特征,匹配获得与所述情感特征对应的目标表情,包括:
    根据所述情感特征,并结合所述第一用户的个人信息与所述第一用户历史发送的一个或多个表情中至少一项,匹配获得与所述情感特征对应的目标表情。
  8. 根据权利要求2所述的方法,其特征在于,所述根据所述情感特征,匹配获得与所述情感特征对应的目标表情,包括:
    根据所述情感特征,确定所述情感特征对应的情感变化趋势;
    根据所述情感变化趋势,匹配获得与所述情感变化趋势对应的多个目标表情以及所述多个目标表情对应的呈现顺序信息;
    其中,所述根据所述目标表情,生成所述语音消息对应的目标表情消息,包括:
    根据所述多个目标表情以及所述多个目标表情对应的呈现顺序信息,生成所述语音消息对应的目标表情消息。
  9. 一种呈现会话消息的方法,用于第二用户设备,其特征在于,所述方法包括:
    接收第一用户经由社交服务器发送的原子会话消息,其中所述原子会话消息包括所述第一用户的语音消息及所述语音消息对应的目标表情消息;
    在所述第一用户与第二用户的会话页面中呈现所述原子会话消息,其中,所述语音消息与所述目标表情消息在所述会话页面中呈现于同一个消息框。
  10. 根据权利要求9所述的方法,其特征在于,所述目标表情消息是在第一用户设 备上根据所述语音消息生成的。
  11. 根据权利要求10所述的方法,其特征在于,所述方法还包括:
    检测所述语音消息与所述目标表情消息是否均已成功接收;
    其中,所述在所述第一用户与第二用户的会话页面中呈现所述原子会话消息,其中,所述语音消息与所述目标表情消息在所述会话页面中呈现于同一个消息框,包括:
    若所述语音消息与所述目标表情消息均已成功接收,在所述第一用户与第二用户的会话页面中呈现所述原子会话消息,其中,所述语音消息与所述目标表情消息在所述会话页面中呈现于同一个消息框;否则,忽略所述原子会话消息。
  12. 根据权利要求10或11所述的方法,其特征在于,所述目标表情消息在所述同一个消息框中相对所述语音消息的显示位置,与所述目标表情消息的被选择时刻在所述语音消息的录制时段信息中的相对位置相匹配。
  13. 根据权利要求12所述的方法,其特征在于,所述方法还包括:
    根据所述目标表情消息的被选择时刻在所述语音消息的录制时段信息中的相对位置,确定所述目标表情消息与所述语音消息在同一个消息框中的相对位置关系;
    其中,所述在所述第一用户与第二用户的会话页面中呈现所述原子会话消息,其中,所述语音消息与所述目标表情消息在所述会话页面中呈现于同一个消息框,包括:
    根据所述相对位置关系在所述第一用户与第二用户的会话页面中呈现所述原子会话消息,其中,所述语音消息与所述目标表情消息在所述会话页面中呈现于所述同一个消息框,所述目标表情消息在所述同一个消息框中相对所述语音消息的显示位置与所述相对位置关系相匹配。
  14. 根据权利要求9至13中任一项所述的方法,其特征在于,所述方法还包括:
    响应于所述第二用户对所述原子会话消息的播放触发操作,播放所述原子会话消息;
    其中,所述播放所述原子会话消息,包括:
    播放所述语音消息;并将所述目标表情消息以第二呈现方式呈现于所述会话页面,其中,所述目标表情消息在所述语音消息被播放前以第一呈现方式呈现于所述同一个消息框。
  15. 根据权利要求9至14中任一项所述的方法,其特征在于,所述方法还包括:
    响应于所述第二用户对所述语音消息的转换文字触发操作,将所述语音消息转换为 文本信息,其中,所述目标表情消息在所述文本信息中的显示位置与所述目标表情消息相对所述语音消息的显示位置相匹配。
  16. 根据权利要求9所述的方法,其特征在于,所述在所述第一用户与第二用户的会话页面中呈现所述原子会话消息,其中,所述语音消息与所述目标表情消息在所述会话页面中呈现于同一个消息框,包括:
    根据所述目标表情消息获得与所述语音消息相匹配的多个目标表情以及所述多个目标表情对应的呈现顺序信息;
    在所述第一用户与第二用户的会话页面中呈现所述原子会话消息,其中,所述多个目标表情根据所述呈现顺序信息与所述语音消息在所述会话页面中呈现于同一个消息框。
  17. 一种呈现会话消息的方法,其特征在于,所述方法包括:
    第一用户设备响应于第一用户在会话页面的语音输入触发操作,开始录入语音消息;
    所述第一用户设备响应于所述第一用户对所述语音消息的发送触发操作,确定所述语音消息对应的目标表情消息;
    所述第一用户设备生成原子会话消息,并将所述原子会话消息经由社交服务器发送至在所述会话页面与所述第一用户通信的第二用户,其中,所述原子会话消息包括所述语音消息和所述目标表情消息;
    第二用户设备接收第一用户经由社交服务器发送的原子会话消息,其中,所述原子会话消息包括所述第一用户的语音消息及所述语音消息对应的目标表情消息;
    所述第二用户设备在所述第一用户与第二用户的会话页面中呈现所述原子会话消息,其中,所述语音消息与所述目标表情消息在所述会话页面中呈现于同一个消息框。
  18. 一种发送会话消息的设备,其特征在于,所述设备包括:
    处理器;以及
    被安排成存储计算机可执行指令的存储器,所述可执行指令在被执行时使所述处理器执行如权利要求1至16中任一项所述方法的操作。
  19. 一种存储指令的计算机可读介质,所述指令在被执行时使得系统进行如权利要求1至16中任一项所述方法的操作。
PCT/CN2020/103032 2019-07-23 2020-07-20 一种发送会话消息的方法与设备 WO2021013126A1 (zh)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN201910667026.4A CN110311858B (zh) 2019-07-23 2019-07-23 一种发送会话消息的方法与设备
CN201910667026.4 2019-07-23

Publications (1)

Publication Number Publication Date
WO2021013126A1 true WO2021013126A1 (zh) 2021-01-28

Family

ID=68081704

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2020/103032 WO2021013126A1 (zh) 2019-07-23 2020-07-20 一种发送会话消息的方法与设备

Country Status (2)

Country Link
CN (1) CN110311858B (zh)
WO (1) WO2021013126A1 (zh)

Families Citing this family (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110311858B (zh) * 2019-07-23 2022-06-07 上海盛付通电子支付服务有限公司 一种发送会话消息的方法与设备
CN110943908A (zh) * 2019-11-05 2020-03-31 上海盛付通电子支付服务有限公司 语音消息发送方法、电子设备及介质
CN112235183B (zh) * 2020-08-29 2021-11-12 上海量明科技发展有限公司 通信消息处理方法、设备及即时通信客户端
CN114780190B (zh) * 2022-04-13 2023-12-22 脸萌有限公司 消息处理方法、装置、电子设备及存储介质
CN115460166A (zh) * 2022-09-06 2022-12-09 网易(杭州)网络有限公司 即时语音通信方法、装置、电子设备及存储介质

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102830977A (zh) * 2012-08-21 2012-12-19 上海量明科技发展有限公司 即时通信录制中添加插入型数据的方法、客户端及系统
CN105989165A (zh) * 2015-03-04 2016-10-05 深圳市腾讯计算机系统有限公司 在即时聊天工具中播放表情信息的方法、装置及系统
CN106161215A (zh) * 2016-08-31 2016-11-23 维沃移动通信有限公司 一种信息发送方法及移动终端
CN106888158A (zh) * 2017-02-28 2017-06-23 努比亚技术有限公司 一种即时通信方法和装置
CN107516533A (zh) * 2017-07-10 2017-12-26 阿里巴巴集团控股有限公司 一种会话信息处理方法、装置、电子设备
CN109859776A (zh) * 2017-11-30 2019-06-07 阿里巴巴集团控股有限公司 一种语音编辑方法以及装置
CN110311858A (zh) * 2019-07-23 2019-10-08 上海盛付通电子支付服务有限公司 一种发送会话消息的方法与设备

Family Cites Families (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106383648A (zh) * 2015-07-27 2017-02-08 青岛海信电器股份有限公司 一种智能终端语音显示的方法和装置
JP2019504413A (ja) * 2015-12-29 2019-02-14 エム・ゼット・アイ・ピィ・ホールディングス・リミテッド・ライアビリティ・カンパニーMz Ip Holdings, Llc 絵文字を提案するためのシステムおよび方法
CN106899486B (zh) * 2016-06-22 2020-09-25 阿里巴巴集团控股有限公司 一种消息显示方法及装置
CN106789581A (zh) * 2016-12-23 2017-05-31 广州酷狗计算机科技有限公司 即时通讯方法、装置及系统
CN107040452B (zh) * 2017-02-08 2020-08-04 浙江翼信科技有限公司 一种信息处理方法、装置和计算机可读存储介质

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102830977A (zh) * 2012-08-21 2012-12-19 上海量明科技发展有限公司 即时通信录制中添加插入型数据的方法、客户端及系统
CN105989165A (zh) * 2015-03-04 2016-10-05 深圳市腾讯计算机系统有限公司 在即时聊天工具中播放表情信息的方法、装置及系统
CN106161215A (zh) * 2016-08-31 2016-11-23 维沃移动通信有限公司 一种信息发送方法及移动终端
CN106888158A (zh) * 2017-02-28 2017-06-23 努比亚技术有限公司 一种即时通信方法和装置
CN107516533A (zh) * 2017-07-10 2017-12-26 阿里巴巴集团控股有限公司 一种会话信息处理方法、装置、电子设备
CN109859776A (zh) * 2017-11-30 2019-06-07 阿里巴巴集团控股有限公司 一种语音编辑方法以及装置
CN110311858A (zh) * 2019-07-23 2019-10-08 上海盛付通电子支付服务有限公司 一种发送会话消息的方法与设备

Also Published As

Publication number Publication date
CN110311858A (zh) 2019-10-08
CN110311858B (zh) 2022-06-07

Similar Documents

Publication Publication Date Title
WO2021013126A1 (zh) 一种发送会话消息的方法与设备
WO2021013125A1 (zh) 一种发送会话消息的方法与设备
US11755296B2 (en) Computer device and method for facilitating an interactive conversational session with a digital conversational character
JP2018181330A (ja) 複数のインテリジェントパーソナルアシスタントサービスに対する管理レイヤ
JP2016524190A (ja) 環境を認識した対話ポリシーおよび応答生成
JP6467554B2 (ja) メッセージ送信方法、メッセージ処理方法及び端末
CN110234032B (zh) 一种语音技能创建方法及系统
JP2016510452A (ja) アクションを決定する際の非言語コミュニケーションの使用
US10973458B2 (en) Daily cognitive monitoring of early signs of hearing loss
KR20080019255A (ko) 상호작용 멀티미디어 프리젠테이션을 위한 상태 기초타이밍
JP2019015951A (ja) 電子機器のウェイクアップ方法、装置、デバイス及びコンピュータ可読記憶媒体
WO2022142619A1 (zh) 一种私密音视频通话的方法与设备
US8868419B2 (en) Generalizing text content summary from speech content
WO2016179128A1 (en) Techniques to automatically generate bookmarks for media files
US20230197070A1 (en) Language Model Prediction of API Call Invocations and Verbal Responses
US10901688B2 (en) Natural language command interface for application management
WO2021218535A1 (zh) 一种ui控件生成、触发方法及终端
JP2020529680A (ja) 通話中の感情を認識し、認識された感情を活用する方法およびシステム
WO2021147930A1 (zh) 一种拼接消息的方法与设备
WO2024016901A1 (zh) 基于歌词的信息提示方法、装置、设备、介质及产品
KR20140111574A (ko) 오디오 명령에 따른 동작을 수행하는 장치 및 방법
JP7331044B2 (ja) 情報処理方法、装置、システム、電子機器、記憶媒体およびコンピュータプログラム
WO2022142618A1 (zh) 一种通过虚拟会议机器人执行指令的方法与设备
WO2023121681A1 (en) Automated text-to-speech pronunciation editing for long form text documents
CN113590871A (zh) 一种音频分类方法、装置及计算机可读存储介质

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 20845005

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 20845005

Country of ref document: EP

Kind code of ref document: A1