TW201909171A - Session information processing method and apparatus, and electronic device - Google Patents

Session information processing method and apparatus, and electronic device Download PDF

Info

Publication number
TW201909171A
TW201909171A TW107116262A TW107116262A TW201909171A TW 201909171 A TW201909171 A TW 201909171A TW 107116262 A TW107116262 A TW 107116262A TW 107116262 A TW107116262 A TW 107116262A TW 201909171 A TW201909171 A TW 201909171A
Authority
TW
Taiwan
Prior art keywords
information
processed
session
content
decoration
Prior art date
Application number
TW107116262A
Other languages
Chinese (zh)
Other versions
TWI671739B (en
Inventor
周琪
Original Assignee
香港商阿里巴巴集團服務有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 香港商阿里巴巴集團服務有限公司 filed Critical 香港商阿里巴巴集團服務有限公司
Publication of TW201909171A publication Critical patent/TW201909171A/en
Application granted granted Critical
Publication of TWI671739B publication Critical patent/TWI671739B/en

Links

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L21/00Processing of the speech or voice signal to produce another audible or non-audible signal, e.g. visual or tactile, in order to modify its quality or its intelligibility
    • G10L21/06Transformation of speech into a non-audible representation, e.g. speech visualisation or speech processing for tactile aids
    • G10L21/10Transforming into visible information
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L51/00User-to-user messaging in packet-switching networks, transmitted according to store-and-forward or real-time protocols, e.g. e-mail
    • H04L51/04Real-time or near real-time messaging, e.g. instant messaging [IM]
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L51/00User-to-user messaging in packet-switching networks, transmitted according to store-and-forward or real-time protocols, e.g. e-mail
    • H04L51/04Real-time or near real-time messaging, e.g. instant messaging [IM]
    • H04L51/043Real-time or near real-time messaging, e.g. instant messaging [IM] using or handling presence information

Landscapes

  • Engineering & Computer Science (AREA)
  • Signal Processing (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Health & Medical Sciences (AREA)
  • Computational Linguistics (AREA)
  • Quality & Reliability (AREA)
  • Data Mining & Analysis (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Human Computer Interaction (AREA)
  • Physics & Mathematics (AREA)
  • Acoustics & Sound (AREA)
  • Multimedia (AREA)
  • User Interface Of Digital Computer (AREA)

Abstract

Disclosed in the embodiments of the present description are a session information processing method and apparatus, and an electronic device. The method comprises: acquiring voice information to be processed, and recognizing mood information expressed by said voice information; then, according to information content comprising the mood content, invoking corresponding decorative configuration information, and further using the decorative configuration information to render a display carrier of said voice information, such as an audio information bar, and then displaying the rendered audio information bar.

Description

會話資訊處理方法、裝置、電子設備Session information processing method, device, electronic device

本說明書實施例屬於資訊資料處理技術領域,尤其涉及一種會話資訊處理方法、裝置、電子設備。The embodiments of the present specification belong to the field of information processing technologies, and in particular, to a session information processing method, apparatus, and electronic device.

隨著網際網路和智慧終端機的快速發展,使用者聊天互動的方式也多種多樣。如即時通訊軟體中的文本、語音、視頻聊天等。   其中,語音聊天是當前使用者常用也較為重要的聊天方式之一,主要通過音頻為資訊載體進行雙方或多方的資訊互動。因為音頻資訊自身的載體特性,語音聊天中的語言資訊不像文本文字一樣可以直接閱讀,通過視覺快速感知資訊內容。在會話聊天中需要播放出音頻資訊才能知道其具體的資訊內容或含義,例如圖1中所示的會話介面中,語音聊天的語言資訊在會話中展示的是單一的條狀樣式,使用者需要通過點擊播放該語音資訊來獲取其含義。對於使用者來說,展示在會話介面中的音頻資訊相當於一個黑盒子,使用者完全不知道其表述的資訊內容,通常需要通過點擊音頻資訊進行播放收聽。因此,需要一種語音資訊內容展示更直觀、趣味性更高、使用者體驗更好的會話資訊處理方式。With the rapid development of the Internet and smart terminals, there are many ways for users to interact and interact. Such as text, voice, video chat, etc. in instant messaging software. Among them, voice chat is one of the most common and important chat methods for current users. It mainly uses audio as the information carrier to carry out information interaction between two or more parties. Because of the carrier characteristics of the audio information itself, the language information in the voice chat can be read directly like the text text, and the information content is quickly perceived through the visual. In the conversation chat, audio information needs to be played to know its specific information content or meaning. For example, in the conversation interface shown in FIG. 1, the language information of the voice chat is displayed in a single strip style in the session, and the user needs Get the meaning by clicking on the voice message. For the user, the audio information displayed in the session interface is equivalent to a black box, and the user does not know the information content of the expression, and usually needs to play and listen by clicking the audio information. Therefore, there is a need for a voice information content display method that is more intuitive, more interesting, and has a better user experience.

本說明書實施例目的在於提供一種會話資訊處理方法、裝置、電子設備,可以通過分析音頻內的關鍵詞或者感情資訊,對音頻訊息體進行額外的描述或裝飾,讓音頻資訊的內容可以從視覺上展示,提高聊天趣味性,提高使用者體驗。   本說明書實施例提供的一種會話資訊處理方法、裝置、電子設備是包括如下的方式實現的:   一種會話資訊處理方法,所述方法包括:   獲取待處理語音資訊,識別所述待處理語音資訊的資訊內容,所述資訊內容至少包括所述待處理語音資訊表達的情緒資訊;   獲取與所述資訊內容相匹配的裝飾配置資訊;   利用所述裝飾配置資訊對所述待處理語音資訊的展示載體進行渲染;   展示所述渲染後的展示載體。   一種會話資訊處理方法,所述方法包括:   獲取待處理語音資訊,識別所述待處理語音資訊的資訊內容,所述資訊內容至少包括待處理語音資訊中包含的關鍵詞資訊;   獲取與所述資訊內容相匹配的裝飾配置資訊;   利用所述裝飾配置資訊對所述待處理語音資訊的展示載體進行渲染;   展示所述渲染後的展示載體。   一種會話資訊處理方法,所述方法包括:   獲取待處理語音資訊,識別所述待處理語音資訊的資訊內容,所述資訊內容至少包括待處理語音資訊中包含的關鍵詞資訊、待處理語音資訊表達的情緒資訊中的一種;   獲取與所述資訊內容相匹配的裝飾配置資訊;   利用所述裝飾配置資訊對所述待處理語音資訊的展示載體進行渲染;   展示所述渲染後的展示載體。   一種會話資訊處理方法,所述方法包括:   獲取待處理會話資訊,識別所述待處理會話資訊的資訊內容,所述資訊內容至少包括待處理會話資訊中包含的關鍵詞資訊、待處理會話資訊表達的情緒資訊中的一種;   獲取與識別出的所述資訊內容相匹配的裝飾配置資訊;   利用所述裝飾配置資訊對所述待處理會話資訊的展示載體進行渲染;   展示所述渲染後的展示載體。   一種會話資訊處理裝置,所述裝置包括:   情緒識別模組,用於獲取待處理語音資訊,識別所述待處理語音資訊的資訊內容,所述資訊內容至少包括所述待處理語音資訊表達的情緒資訊;   裝飾匹配模組,用於獲取與所述資訊內容相匹配的裝飾配置資訊;   渲染模組,用於利用所述裝飾配置資訊對所述待處理語音資訊的展示載體進行渲染;   效果展示模組,用於展示所述渲染後的展示載體。   一種會話資訊處理裝置,所述裝置包括:   關鍵詞識別模組,用於獲取待處理語音資訊,識別所述待處理語音資訊的資訊內容,所述資訊內容至少包括所述待處理語音資訊包含的關鍵詞資訊;   裝飾匹配模組,用於獲取與所述資訊內容相匹配的裝飾配置資訊;   渲染模組,用於利用所述裝飾配置資訊對所述待處理語音資訊的展示載體進行渲染;   效果展示模組,用於展示所述渲染後的展示載體。   一種會話資訊處理裝置,所述裝置包括:   資訊內容識別模組,用於獲取待處理語音資訊,識別所述待處理語音資訊的資訊內容,所述資訊內容至少包括待處理語音資訊中包含的關鍵詞資訊、待處理語音資訊表達的情緒資訊中的一種;   裝飾匹配模組,用於獲取與所述資訊內容相匹配的裝飾配置資訊;   渲染模組,用於利用所述裝飾配置資訊對所述待處理語音資訊的展示載體進行渲染;   效果展示模組,用於展示所述渲染後的展示載體。   一種會話資訊處理裝置,所述裝置包括:   資訊內容識別模組,用於獲取待處理會話資訊,識別所述待處理會話資訊的資訊內容,所述資訊內容至少包括待處理會話資訊中包含的關鍵詞資訊、待處理會話資訊表達的情緒資訊中的一種;   裝飾匹配模組,用於獲取與識別出的所述資訊內容相匹配的裝飾配置資訊;   渲染模組,用於利用所述裝飾配置資訊對所述待處理會話資訊的展示載體進行渲染;   效果展示模組,用於展示所述渲染後的展示載體。   一種會話資訊處理裝置,包括處理器以及用於儲存處理器可執行指令的記憶體,所述處理器執行所述指令時實現:   獲取待處理會話資訊,識別所述待處理會話資訊的資訊內容,所述資訊內容至少包括待處理會話資訊中包含的關鍵詞資訊、待處理會話資訊表達的情緒資訊中的一種;   獲取與識別出的所述資訊內容相匹配的裝飾配置資訊;   利用所述裝飾配置資訊對所述待處理會話資訊的展示載體進行渲染;   展示所述渲染後的展示載體。   一種電子設備,包括至少一個處理器以及使用者儲存處理器可執行指令的記憶體,所述處理器執行所述指令時實現:   獲取待處理會話資訊,識別所述待處理會話資訊的資訊內容,所述資訊內容至少包括待處理會話資訊中包含的關鍵詞資訊、待處理會話資訊表達的情緒資訊中的一種;   獲取與識別出的所述資訊內容相匹配的裝飾配置資訊;   利用所述裝飾配置資訊對所述待處理會話資訊的展示載體進行渲染;   展示所述渲染後的展示載體。   本說明書一個或多個實施例提供的一種會話資訊處理方法、裝置、電子設備,可以獲取待處理語音資訊,識別所述待處理語音資訊表達的情緒資訊。然後可以根據包括所述情緒內容的資訊內容調取相應的裝飾配置資訊,進而利用所述裝飾配置資訊對所述待處理語音資訊的展示載體(例如音頻資訊條)進行渲染後進行展示。這樣根據語音資訊表達的情感對展示載體進行渲染後展示,不僅可以表達或部分表達了語音資訊的內容,還增加了語音資訊聊天時的趣味性。利用本說明書實施例提供的實施方案,可以通過分析音頻內熱詞或者感情,對音頻訊息體進行額外的描述或裝飾,讓音頻資訊的內容可以從視覺上展示,提高聊天趣味性,提高使用者體驗。The embodiment of the present specification aims to provide a session information processing method, device, and electronic device, which can additionally describe or decorate an audio message body by analyzing keywords or emotional information in the audio, so that the content of the audio information can be visually observed. Showcase, improve chat fun and improve user experience. A session information processing method, apparatus, and electronic device provided by the embodiments of the present disclosure are implemented as follows: A session information processing method, the method includes: acquiring pending voice information, and identifying information about the to-be-processed voice information Contenting, the information content includes at least the emotion information expressed by the to-be-processed voice information; acquiring decoration configuration information that matches the information content; and rendering the display carrier of the to-be-processed voice information by using the decoration configuration information ; Display the rendered display carrier. A session information processing method, the method includes: acquiring a to-be-processed voice information, and identifying the information content of the to-be-processed voice information, where the information content includes at least keyword information included in the to-be-processed voice information; The decoration configuration information matching the content; rendering the display carrier of the to-be-processed voice information by using the decoration configuration information; and displaying the rendered display carrier. A session information processing method, the method includes: acquiring a to-be-processed voice information, and identifying the information content of the to-be-processed voice information, where the information content includes at least keyword information included in the to-be-processed voice information, and the voice information to be processed And obtaining a decoration configuration information that matches the information content; and using the decoration configuration information to render the display carrier of the to-be-processed voice information; and displaying the rendered display carrier. A session information processing method, the method includes: acquiring session information to be processed, and identifying information content of the session information to be processed, where the information content includes at least keyword information included in the session information to be processed, and information about the session information to be processed. And obtaining a decoration configuration information that matches the identified information content; rendering the display carrier of the to-be-processed session information by using the decoration configuration information; and displaying the rendered display carrier . A session information processing device, the device includes: an emotion recognition module, configured to acquire voice information to be processed, and identify information content of the to-be-processed voice information, where the information content includes at least an emotion expressed by the to-be-processed voice information a decoration matching module, configured to obtain decoration configuration information that matches the information content; a rendering module, configured to render the display carrier of the to-be-processed voice information by using the decoration configuration information; a group for displaying the rendered display carrier. A session information processing device, the device includes: a keyword recognition module, configured to acquire voice information to be processed, and identify information content of the to-be-processed voice information, where the information content includes at least the voice information to be processed The keyword matching information is used to obtain the decoration configuration information that matches the information content; the rendering module is configured to use the decoration configuration information to render the display carrier of the to-be-processed voice information; a display module for displaying the rendered display carrier. A session information processing device, the device includes: an information content identification module, configured to acquire voice information to be processed, and identify information content of the to-be-processed voice information, where the information content includes at least a key included in the voice information to be processed a word information, one of emotional information expressed by the voice information to be processed; a decoration matching module, configured to obtain decoration configuration information matching the information content; and a rendering module, configured to use the decoration configuration information The display carrier of the to-be-processed voice information is rendered; and the effect display module is configured to display the rendered display carrier. A session information processing device, the device includes: an information content identification module, configured to acquire session information to be processed, and identify information content of the session information to be processed, where the information content includes at least a key included in the session information to be processed a word information, one of emotional information expressed by the session information to be processed; a decoration matching module, configured to acquire decoration configuration information that matches the identified information content; and a rendering module configured to use the decoration configuration information The display carrier of the to-be-processed session information is rendered; and the effect display module is configured to display the rendered display carrier. A session information processing device includes a processor and a memory for storing processor-executable instructions. When the processor executes the instructions, the processor: obtains pending session information, and identifies information content of the to-be-processed session information. The information content includes at least one of keyword information included in the session information to be processed and emotion information expressed by the session information to be processed; acquiring decoration configuration information matching the identified information content; using the decoration configuration The information is used to render the display carrier of the to-be-processed session information; and the rendered display carrier is displayed. An electronic device includes at least one processor and a memory for a user to store processor-executable instructions, and when the processor executes the instructions, the method implements: acquiring pending session information, and identifying information content of the to-be-processed session information, The information content includes at least one of keyword information included in the session information to be processed and emotion information expressed by the session information to be processed; acquiring decoration configuration information matching the identified information content; using the decoration configuration The information is used to render the display carrier of the to-be-processed session information; and the rendered display carrier is displayed. A session information processing method, apparatus, and electronic device provided by one or more embodiments of the present disclosure may acquire voice information to be processed, and identify emotion information expressed by the voice information to be processed. Then, the corresponding decoration configuration information may be retrieved according to the information content including the emotional content, and then the display carrier (for example, an audio information bar) of the to-be-processed voice information is rendered and displayed by using the decoration configuration information. In this way, according to the emotion expressed by the voice information, the display carrier is rendered and displayed, which not only can express or partially express the content of the voice information, but also increase the interest of the voice information chat. With the implementation provided by the embodiments of the present specification, the audio message body can be additionally described or decorated by analyzing the hot words or feelings in the audio, so that the content of the audio information can be visually displayed, improving the chat interest and improving the user experience. .

為了使本技術領域的人員更好地理解本說明書中的技術方案,下面將結合本說明書一個或多個實施例中的附圖,對本說明書一個或多個實施例中的技術方案進行清楚、完整地描述,顯然,所描述的實施例僅僅是說明書一部分實施例,而不是全部的實施例。基於說明書一個或多個實施例,本領域普通技術人員在沒有作出進步性勞動前提下所獲得的所有其他實施例,都應當屬於本說明書實施例方案保護的範圍。   圖2是本說明書提供的所述一種會話資訊處理方法實施例的流程示意圖。雖然本說明書提供了如下述實施例或附圖所示的方法操作步驟或裝置結構,但基於常規或者無需進步性的勞動在所述方法或裝置中可以包括更多或者部分合併後更少的操作步驟或模組單元。在邏輯性上不存在必要因果關係的步驟或結構中,這些步驟的執行順序或裝置的模組結構不限於本說明書實施例或附圖所示的執行順序或模組結構。所述的方法或模組結構的在實際中的裝置、伺服器或終端產品應用時,可以按照實施例或者附圖所示的方法或模組結構進行循序執行或者並存執行(例如並行處理器或者多執行緒處理的環境、甚至包括分散式處理、伺服器集群的實施環境)。   本說明書實施例提供的實施方案可以用於多種具有會話功能的終端裝置中,如移動智慧終端機、車載設備、PC (personal computer,個人電腦)、穿戴設備、業務系統的伺服器等。所述的終端裝置可以安裝或構建有即時通訊應用、臨時會話等,本說明書實施例提供的實施方案可以用於包括基於上述終端裝置的會話應用場景中。本說明書實施例可以通過對語音資訊進行關鍵詞或情感識別,然後展示在音頻資訊展示載體上。具體的可以包括通過對單條或多條語音資訊進行分析,根據識別演算法,判斷語音資訊發送方的情緒。然後可以根據情緒資訊,對音頻資訊的展示載體樣式進行二次渲染(如:改變訊息體顏色,添加Emoji表情頭等),大大提高聊天趣味性和使用者體驗,並可以對音頻資訊的表達資訊內容進行展示或部分展示。具體的一個實施例如圖2所示,本說明書提供的一種會話資訊處理方法的一種實施例中,所述方法可以包括:   S2:獲取待處理語音資訊,識別所述待處理語音資訊的資訊內容,所述資訊內容至少包括所述待處理語音資訊表達的情緒資訊。   在本實施例中,使用者在會話介面中輸入一段語音資訊後,可以獲取該語音資訊,作為本實施例進行會話資訊處理的待處理語音資訊。使用者輸入的語音資訊可以先由於處理進行進一步處理後,如下述的利用裝飾配置資訊進行渲染後一同展示在會話介面中,也可以先展示該段語音資訊的訊息體(這裡可以指包含語音資訊的展示載體,如會話介面中的音頻資訊條),然後再對該訊息體進行渲染後展示動畫效果等。   本實施例應用場景中,獲取待處理語音資訊後,可以識別所述待處理語音資訊中的資訊內容,其中識別的資訊內中可以至少包括所述待處理語音資訊所表達的情緒資訊。識別出的情緒資訊具體的可以採用特徵詞或識別符來表示,如“高興”、“悲傷”、“驚喜”等。所述待處理語音資訊中情感資訊的識別方式可以採用多種演算法或模型進行處理。本實施例中,對語言資訊的分析處理可以包括提取音頻資訊特徵,可以通過識別演算法,識別說話者的情緒,如喜悅、憤怒等。本說明書實施例提供所述方法的一個實施例中,所述識別所述待處理語音資訊的資訊內容可以包括:   S20:提取所述待處理語音資訊的音頻特徵資料,將所述音頻特徵資料輸入預先構建的情緒識別模型,通過所述情緒識別模型確定所述待處理語音資訊所屬的情緒類別。   具體的實現方式中,可以從數位化的音頻信號中提取出特徵信號,如語速、音調、頻率等,可以對這些特徵資訊進行建模,通過多個樣本進行模型訓練,優化模型。模型訓練完成後可以用來作為分類器,對新輸入的待處理語音資訊的音頻信號反映出來的情緒進行分類,確定所述待處理語音資訊表達的情緒資訊,如“喜悅”、“憤怒”。所述的情緒識別模型可以包括採用選取的識別演算法或網路模型經過樣本資料訓練後得到的可以識別音頻資訊所屬情緒分類的演算法。例如可以選取KNN(k- NearestNeighbor,K最鄰近)分類演算法,或者深度神經網路/卷積神經網路及其變換/變種後的神經網路等。   本說明書提供的一個或多個實施例中,提取的所述音頻特徵資料包括所述待處理語音資訊的生物的情感特徵資料、語氣特徵資料中的至少一種。   所述的生物的情感特徵資料可以包括生氣、憤怒、悲傷、喜悅等情感,所述的語氣特徵資料可以包括疑問、陳述、反問、感歎等語氣語態上的特徵資料。在一個實施例中,提取音頻特徵資料可以包括可以反映人的情感方面的特徵資料,如基於語速、音調、頻率等反映出人的情緒,具體的一個示例中,如語速快、音調高可以反映發言人的情緒較為高漲,反之,語速慢、音調低可以反映發言人的情緒較為低落。其他的實施例中,也可以從語法的角度獲取語氣特徵資料,作為音頻特徵資料。例如從待處理語音資訊中的來判斷發言人是疑惑還是感歎。當然,也可以將上述兩者結合,通過提取的音調特徵資料以及主謂賓順序判斷發明人的情緒。具體的可以根據應用環境或資料處理需求等設置相應的所要提取的音頻特徵資料的種類。   S4:獲取與所述資訊內容相匹配的裝飾配置資訊。   本說明書的一個實施例中,可以預先設置裝飾配置資料庫,所述裝飾配置資料庫可以儲存對音頻資訊的展示載體進行裝飾的裝飾配置資訊,如音頻資訊條變色等。所述的裝飾配置資料庫可以設置在本地終端,如伺服器本機存放區,或者客戶端本機存放區,也可以儲存在遠端服務端中。這樣,在獲取所述待處理語音資訊中的資訊內容後,可以從所述裝飾配置資料庫中獲取與識別出的所述資訊內容相匹配的裝飾配置資訊。   本實施例應用場景中,所述資訊內容可以包括情緒資訊。一些實施例中,識別出的資訊內容可以包括多個或多種情緒資訊,如“高興”、“喜歡”、“驚訝”等傾向於正向積極的情緒類別。因此,本說明書提供的一個或多個實施例中,可以對識別出的多個情緒資訊的識別結果進行合併,轉換成相應的請求參數。可以通過統一、規範化的請求參數調取相應的裝飾配置資訊。具體的,一個實施例中,所述獲取與識別出的所述資訊內容相匹配的裝飾配置資訊可以包括:   S40:將所述待處理語音資訊中情緒資訊的識別結果進行合併,生成請求參數;   S42:獲取基於所述請求參數匹配出的裝飾配置資訊。   如上述“高興”、“喜歡”、“驚訝”可以合併為包括“happy”、“surprise”兩個情緒類別的請求參數,然後可以從所述裝飾配置資料庫中匹配出一個或兩個相對應的裝飾配置資訊。當然,其他的實施方式中,也可以根據設置的處理規則將上述三個情緒類別的資訊合併為一個請求參數“happy”,根據該請求參數調取“happy”的裝飾配置資訊。   本說明書實施例所述的裝飾配置資訊可以包含(但不限於)文字描述、Emoji表情、背景顏色、動畫效果等。一個情緒資訊或一種情緒資訊或一個請求參數可以對應一個或多個裝飾配置資訊,可以設置有相應的映射關係,如“高興”的情緒可以對應“高興1”、“高興2”、“高興3”三個裝飾配置資訊,具體的使用哪個裝飾配置資訊,可以按序或者隨機或者其他定義的方式執行。   因此,本說明書的一個或多個實施例中,所述裝飾配置資訊可以包括下述中的至少一種:   文字描述、表情、變更背景顏色、在所述展示載體預設位置播放動畫效果、所述展示載體的變形動畫效果。   利用包括上述類型的裝飾配置資訊,可以大幅增加會話的趣味性,改變單一的語音資訊條的展示方式。並且,設置的裝飾配置資訊與語音資訊中的所表達的情緒資訊相關聯、匹配,可以通過所述裝飾配置資訊展示或部分展示出語音資訊的內容,提高使用者體驗。   S6:利用所述裝飾配置資訊對所述待處理語音資訊的展示載體進行渲染。   獲取裝飾配置資訊後,可以利用所述裝飾配置資訊對展示載體進行渲染。所說的展示載體,可以包括例如音頻資訊條的方式,也可以包括其他的在會話介面中用於表示使用者發出的資訊內容為語音資訊的訊息載體,如音頻資訊條、音頻資訊框、音頻資訊視窗、音頻符號等。所述的音頻資訊視窗可以包括單獨使用者播放音頻資訊的視窗,例如,當使用者輸入語音資訊後,可以在會話介面的指定位置或單獨彈出音頻資訊視窗用於播放語音資訊,此時可以利用所述裝飾配置資訊來渲染所述音頻資訊視窗。當然,所述展示載體也可以是音頻符號或其他形式,如喇叭音頻符合、標記有“語音”的按鍵、“voice”字元等。因此,所述方法的其他實施例中,所述展示載體可以包括:   音頻資訊條、音頻資訊視窗、音頻符號中的至少一種。   S8:展示所述渲染後的展示載體。   可以在會話介面中展示所述渲染後的展示載體。例如利用所述裝飾配置資訊對語音資訊條渲染後呈現“彩虹變換”的動畫效果。   具體的一個實施場景中,可以設置在所述展示載體預設位置播放動畫效果,如圖3所示,圖3是本說明書實施例所述方法的一個實施例場景示意圖。在圖3中,根據使用者輸入的語音資訊識別出使用者為“疑問”時,可以在音頻資訊條的水準右側播放“三個問號”的短動畫,每個問號符號可以間隔預定時間彈出。   另一個實施場景中,所述裝飾配置資訊可以包括所述展示載體的變形動畫效果。如圖4所示,圖4是本說明書實施例所述方法的另一個實施例場景示意圖。圖4中,使用者輸入語音資訊1後,識別出使用者的情緒資訊為“高興”,則可以將該使用者輸入的語音資訊1的音頻資訊條設置為多種顏色指定頻率變換的“彩虹變換”的動畫效果,呈現出喜悅的氣氛,與使用者輸入的語音資訊所表達的情緒相對應。相應的,如果識別出情緒資訊為“悲傷”,則可以使用藍色與紫色較慢頻率變換的“憂鬱”的裝飾裝配資訊,如圖5所示。圖5是本說明書實施例所述方法的另一個實施例場景示意圖。   本說明書一個或多個實施例提供的一種會話資訊處理方法,可以獲取待處理語音資訊,識別所述待處理語音資訊表達的情緒資訊。然後可以根據包括所述情緒內容的資訊內容調取相應的裝飾配置資訊,進而利用所述裝飾配置資訊對所述待處理語音資訊的展示載體(例如音頻資訊條)進行渲染後進行展示。這樣根據語音資訊表達的情感對展示載體進行渲染後展示,不僅可以表達或部分表達了語音資訊的內容,還增加了語音資訊聊天時的趣味性。利用本說明書實施例提供的實施方案,可以通過分析音頻內的感情資訊,對音頻訊息體進行額外的描述或裝飾,讓音頻資訊的內容可以從視覺上展示,提高聊天趣味性,提高使用者體驗。   本說明書一個或多個實施例提供的所述會話資訊處理方法中,還可以從語音資訊中包含的關鍵詞資訊這一維度來識別語音資訊內容,然後根據包含所述關鍵詞的資訊內容匹配出相應的裝飾配置資訊,進一步的對語音資訊的展示載體進行渲染後展示在會話介面中。具體的,如圖6所示,圖6是本說明書提供的所述方法另一個實施例的方法流程示意圖。本說明書提供的所述方法的另一個實施例中,所述方法可以包括:   S120:獲取待處理語音資訊,識別所述待處理語音資訊的資訊內容,所述資訊內容至少包括待處理語音資訊中包含的關鍵詞資訊;   S140:獲取與識別出的所述資訊內容相匹配的裝飾配置資訊;   S160:利用所述裝飾配置資訊對所述待處理語音資訊的展示載體進行渲染;   S180:展示所述渲染後的展示載體。   具體的處理過程中,可以判斷待處理語音資訊未處理時,進行語音轉文字處理,對生成的文本根據規則(如提取網路熱詞、或者主謂賓等句法元素)提取關鍵詞。所述識別所述待處理語音資訊的資訊內容可以包括:   將所述待處理語音資訊轉換為文字文本;   將所述文字文本與預設詞庫中的關鍵詞進行匹配,提取所包含的關鍵詞資訊。   所述的關鍵詞可以包括提前篩選出來詞語(如:網路流行詞彙),這些詞語可以儲存在預設詞庫中。本說明書的一個實施例中,可以對語音進行文本轉化後,判斷文本中的詞是否包含有所述預測詞庫中的關鍵詞,如果有,則提取出該關鍵詞。提取關鍵詞的演算法可以採用通過字串搜索實現,具體的演算法可以包括全搜、KMP(一種改進的字串匹配演算法,由D.E.Knuth,J.H.Morris和V.R.Pratt提出)、Boyer More(一種字串匹配演算法)等經典演算法。   其他的實施方式中,所述識別所述待處理語音資訊的資訊內容也可以包括:   將所述待處理語音資訊轉換為文字文本,根據採用的文本語法規則提取所述文字文本中的關鍵詞。   例如根據轉換後的文字文本的主謂賓等語法元素來識別提取關鍵詞。   當然,參照前述基於識別待處理語音中的情感資訊的實施方式,本說明書中利用語音資訊轉換成文字文本後識別關鍵詞的實施方式還可以包括其他的實施例,具體的一個實施例中,所述獲取與識別出的所述資訊內容相匹配的裝飾配置資訊可以包括:   將所述待處理語音資訊中關鍵詞資訊的識別結果進行合併,生成請求參數;   獲取基於所述請求參數匹配出的裝飾配置資訊。   以及,所述獲取基於所述請求參數匹配出的裝飾配置資訊可以包括:   將所述請求參數在本地預設的裝飾配置資料庫中進行匹配,獲取裝飾配置資訊;   或者,   向遠端服務端發送所述請求參數,獲取所述遠端服務端根據所述請求參數匹配得到的裝飾配置資訊。   根據前述實施例描述,還可以包括其他更多實施方式,如所述裝飾配置資訊的種類等。本說明書中的各個實施例均採用遞進的方式描述,各個實施例之間相同相似的部分互相參見即可,每個實施例重點說明的都是與其他實施例的不同之處。具體的可以參照相關實施例的描述,在此不做一一贅述。   本說明書一個或多個實施例提供的一種會話資訊處理方法,可以獲取待處理語音資訊,將所述待處理語音資訊轉換成文本資訊,提取文本資訊中的關鍵詞。然後可以根據包括所述關鍵詞的資訊內容調取相應的裝飾配置資訊,進而利用所述裝飾配置資訊對所述待處理語音資訊的展示載體(例如音頻資訊條)進行渲染後進行展示。這樣根據語音資訊所包含的關鍵詞調取相應的裝飾配置資訊對展示載體進行渲染後展示,不僅可以表達或部分表達了語音資訊的內容,還增加了語音資訊聊天時的趣味性。利用本說明書實施例提供的實施方案,可以通過分析音頻內的關鍵詞,對音頻訊息體進行額外的描述或裝飾,讓音頻資訊的內容可以從視覺上展示,提高聊天趣味性,提高使用者體驗。   本說明書提供的所述方法的另一種實施例式中,可以將上述基於待處理語音資訊中的關鍵詞資訊、情緒資訊的實施方式合併,從這兩個維度中的至少一個識別待處理語音資訊的資訊內容,然後匹配出裝飾配置資訊,對訊息載體渲染後展示。例如,可以同時識別待處理語音資訊中的關鍵詞資訊和所表達的情緒資訊。因此,本說明書的一個或多個實施例提供的一種會話資訊處理方法中,所述方法包括:   獲取待處理語音資訊,識別所述待處理語音資訊的資訊內容,所述資訊內容至少包括待處理語音資訊中包含的關鍵詞資訊、待處理語音資訊表達的情緒資訊中的一種;   獲取與識別出的所述資訊內容相匹配的裝飾配置資訊;   利用所述裝飾配置資訊對所述待處理語音資訊的展示載體進行渲染;   展示所述渲染後的展示載體。   本說明書一個或多個實施例中,當採用上述識別待處理語音資訊中包括待處理語音資訊中包含的關鍵詞資訊、待處理語音資訊表達的情緒資訊的資訊內容時,可以將所述待處理語音資訊中關鍵詞資訊、情緒資訊的識別結果進行合併,生成請求參數。相應的,所述獲取與所述資訊內容相匹配的裝飾配置資訊包括獲取基於所述請求參數匹配出的裝飾配置資訊。如圖7所示,圖7是本說明書提供的所述方法的另一個實施例應用場景中的處理流程示意圖。   這樣,待處理語音資訊中關鍵詞的識別結合與情緒資訊的識別結果進行合併,可以優化識別的資訊內容結果,可以匹配更加準確的匹配符合待處理語音資訊內容的裝飾配置資訊,提高展示效果,提高使用者體驗。   進一步的,基於本說明書實施例的創新思想,在會話資訊處理場景中,處理的資料物件可以不限於語音資訊,其他的實施場景中也可以包括文字、圖像或視頻等資訊。基於本說明書上述實施方案,在處理文字、圖像或視頻的會話資訊時,可以識別這些文字、圖像或視頻中的關鍵詞資訊、所表達的情緒資訊中的至少一種的資訊內容,然後基於識別出的資訊內容匹配相應的裝飾配置資訊,對會話資訊的展示載體進行渲染後展示在會話介面中。因此,本說明書還提供一種會話資訊處理方法,所述方法可以包括:   S220:獲取待處理會話資訊,識別所述待處理會話資訊的資訊內容,所述資訊內容至少包括待處理會話資訊中包含的關鍵詞資訊、待處理會話資訊表達的情緒資訊中的一種;   S240:獲取與識別出的所述資訊內容相匹配的裝飾配置資訊;   S260:利用所述裝飾配置資訊對所述待處理會話資訊的展示載體進行渲染;   S280:展示所述渲染後的展示載體。   圖8是本說明書提供的所述方法另一種實施例的方法流程示意圖。同樣的,上述所述的其他資料來源的會話處理方法根據語音資訊處理方法實施例的描述還可以包括其他的實施方式,具體的實現方式可以參照相關方法實施例的描述,在此不作一一贅述。   本說明書實施例提供的會話資訊處理方法不僅可以用於語音資訊的處理,渲染的訊息體還可以包括文本文字、視頻等。其解決問題的實現方案與前述語音資訊的處理相似,因此本說明書實施例具體的實施可以參見前述語音資訊處理方法的實施,重複之處不再贅述。   上述對本說明書特定實施例進行了描述。其它實施例在所附權利要求書的範圍內。在一些情況下,在申請專利範圍中記載的動作或步驟可以按照不同於實施例中的順序來執行並且仍然可以實現期望的結果。另外,在附圖中描繪的過程不一定要求示出的特定順序或者連續順序才能實現期望的結果。在某些實施方式中,多工處理和並行處理也是可以的或者可能是有利的。   本說明書一個或多個實施例提供的一種會話資訊處理方法,可以獲取待處理語音資訊,識別所述待處理語音資訊表達的情緒資訊。然後可以根據包括所述情緒內容的資訊內容調取相應的裝飾配置資訊,進而利用所述裝飾配置資訊對所述待處理語音資訊的展示載體(例如音頻資訊條)進行渲染後進行展示。這樣根據語音資訊表達的情感對展示載體進行渲染後展示,不僅可以表達或部分表達了語音資訊的內容,還增加了語音資訊聊天時的趣味性。利用本說明書實施例提供的實施方案,可以通過分析音頻內熱詞或者感情,對音頻訊息體進行額外的描述或裝飾,讓音頻資訊的內容可以從視覺上展示,提高聊天趣味性,提高使用者體驗。   基於上述所述的會話資訊處理方法,本說明書一個或多個實施例還提供一種會話資訊處理裝置。所述的裝置可以包括使用了本說明書實施例所述方法的系統(包括分散式系統)、軟體(應用)、模組、組件、伺服器、客戶端等並結合必要的實施硬體的裝置。基於同一創新構思,本說明書實施例提供的一個或多個實施例中的裝置如下面的實施例所述。由於裝置解決問題的實現方案與方法相似,因此本說明書實施例具體的裝置的實施可以參見前述方法的實施,重複之處不再贅述。以下所使用的,術語“單元”或者“模組”可以實現預定功能的軟體和/或硬體的組合。儘管以下實施例所描述的裝置較佳地以軟體來實現,但是硬體,或者軟體和硬體的組合的實現也是可能並被構想的。具體的,圖9是本說明書提供的一種會話資訊處理裝置實施例的模組結構示意圖,如圖9所示,所述裝置可以包括:   情緒識別模組101,可以用於獲取待處理語音資訊,識別所述待處理語音資訊的資訊內容,所述資訊內容至少包括所述待處理語音資訊表達的情緒資訊;   裝飾匹配模組102,可以用於獲取與所述資訊內容相匹配的裝飾配置資訊;   渲染模組103,可以用於利用所述裝飾配置資訊對所述待處理語音資訊的展示載體進行渲染;   效果展示模組104,可以用於展示所述渲染後的展示載體。   當然,參照前述方法實施例描述,所述裝置的其他實施例中,所述情緒識別模組101可以包括:   情緒識別模型單元,可以用於提取所述待處理語音資訊的音頻特徵資料,將所述音頻特徵資料輸入預先構建的情緒識別模型,以及,通過所述情緒識別模型確定所述待處理語音資訊所屬的情緒類別。   以及另一個實施例中,提取的所述音頻特徵資料包括所述待處理語音資訊的生物的情感特徵資料、語氣特徵資料中的至少一種。   其他的裝置實施例中,也可以通過識別語音資訊中的關鍵詞來匹配裝飾配置資訊。圖10是本說明書提供的另一種會話資訊處理裝置實施例的模組結構示意圖,如圖10所示,所述裝置可以包括:   關鍵詞識別模組201,可以用於獲取待處理語音資訊,識別所述待處理語音資訊的資訊內容,所述資訊內容至少包括所述待處理語音資訊包含的關鍵詞資訊;   裝飾匹配模組202,可以用於獲取與所述資訊內容相匹配的裝飾配置資訊;   渲染模組203,可以用於利用所述裝飾配置資訊對所述待處理語音資訊的展示載體進行渲染;   效果展示模組204,可以用於展示所述渲染後的展示載體。   當然,參照前述方法實施例描述,所述裝置的其他實施例中,所述情緒識別模組201可以包括:   文本轉換單元2011,可以用於將所述待處理語音資訊轉換為文字文本;   詞庫匹配單元2012,可以用於將所述文字文本與預設詞庫中的關鍵詞進行匹配,提取所包含的關鍵詞資訊。   所述一種會話資訊處理裝置其他的實施例中,所述關鍵詞識別模組201可以包括:   文本轉換單元2011,可以用於將所述待處理語音資訊轉換為文字文本;   語法匹配單元2013,可以用於將所述待處理語音資訊轉換為文字文本,根據使用的文本語法規則提取所述文字文本中的關鍵詞。   當然,所述裝置的其他實施例中,所述關鍵詞識別裝置可以設置包括所述詞庫匹配單元和所述語法匹配單元,使用其中一種或兩種方式來來識別語音資訊中的關鍵詞。如圖11所示,圖11是本說明書提供的所述裝置中所述關鍵詞識別模組一個實施例的模組結構示意圖。   根據前述相應方法實施例所述,上述所述基於語音資訊中關鍵詞或情緒識別的裝置還可以包括其他的實施例。所述裝置還可以包括:   特徵合併單元,可以用於將所述待處理語音資訊中關鍵詞資訊或情緒資訊的識別結果進行合併,生成請求參數;   相應的,所述裝飾匹配模組可以基於所述請求參數匹配裝飾配置資訊。   其他的實施例中,所述裝飾配置資訊可以包括下述中的至少一種:   文字描述、表情、變更背景顏色、在所述展示載體預設位置播放動畫效果、所述展示載體的變形動畫效果。以及其他的實施例中,所述展示載體可以包括:   音頻資訊條、音頻資訊視窗、音頻符合中的至少一種。   所述裝置的另一種實施例中,可以設置包括識別語音資訊中的關鍵詞的處理模組和識別語音資訊表達的情緒資訊的處理模組。所述裝置可以基於其中任意一個模組或者使用上述兩個模組來識別待處理語音資訊的資訊內容。因此,本說明書還提供另一個會話資訊處理裝置的實施例,所述裝置可以包括:   資訊內容識別模組301,可以用於獲取待處理語音資訊,識別所述待處理語音資訊的資訊內容,所述資訊內容至少包括待處理語音資訊中包含的關鍵詞資訊、待處理語音資訊表達的情緒資訊中的一種;   裝飾匹配模組302,可以用於獲取與所述資訊內容相匹配的裝飾配置資訊;   渲染模組303,可以用於利用所述裝飾配置資訊對所述待處理語音資訊的展示載體進行渲染;   效果展示模組304,可以用於展示所述渲染後的展示載體。   圖12是本說明書提供的所述裝置的另一個實施例的模組結構示意圖。如圖12所示,另一種會話資訊處理裝置的實施例中,如果使用識別語音資訊關鍵詞和情緒資訊的處理方式時,所述裝置還可以包括:   特徵合併單元305,可以用於將所述待處理語音資訊中關鍵詞資訊和情緒資訊的識別結果進行合併,生成請求參數;   相應的,所述裝飾匹配模組302可以基於所述請求參數匹配裝飾配置資訊。   需要說明的,上述所述的裝置根據方法實施例的描述還可以包括其他的實施方式,具體的實現方式可以參照相關方法實施例的描述,在此不作一一贅述。   本說明書一個或多個實施例提供的一種會話資訊處理裝置,可以獲取待處理語音資訊,識別所述待處理語音資訊表達的情緒資訊。然後可以根據包括所述情緒內容的資訊內容調取相應的裝飾配置資訊,進而利用所述裝飾配置資訊對所述待處理語音資訊的展示載體(例如音頻資訊條)進行渲染後進行展示。這樣根據語音資訊表達的情感對展示載體進行渲染後展示,不僅可以表達或部分表達了語音資訊的內容,還增加了語音資訊聊天時的趣味性。利用本說明書實施例提供的實施方案,可以通過分析音頻內熱詞或者感情,對音頻訊息體進行額外的描述或裝飾,讓音頻資訊的內容可以從視覺上展示,提高聊天趣味性,提高使用者體驗。   另一種會話資訊處理裝置的實施例中,識別的處理的資訊源還可以包括文字、圖像、視頻等,可以識別出文字或視頻等會話資訊中包含的關鍵詞或情緒資訊,然後匹配相應的裝飾配置資訊來渲染訊息體。具體的一個實施例如圖13所示,圖13是本說明書提供的所述裝置的另一個實施例的模組結構示意圖,所述裝置可以包括:   資訊內容識別模組401,可以用於獲取待處理會話資訊,識別所述待處理會話資訊的資訊內容,所述資訊內容至少包括待處理會話資訊中包含的關鍵詞資訊、待處理會話資訊表達的情緒資訊中的一種;   裝飾匹配模組402,可以用於獲取與識別出的所述資訊內容相匹配的裝飾配置資訊;   渲染模組403,可以用於利用所述裝飾配置資訊對所述待處理會話資訊的展示載體進行渲染;   效果展示模組404,可以用於展示所述渲染後的展示載體。   同樣的,上述所述的其他資料來源的會話處理方法根據語音資訊處理方法實施例的描述還可以包括其他的實施方式,其解決問題的實現方案與前述語音資訊的處理相似,因此本說明書一個或多個實施例具體的實施可以參見前述語音資訊處理方法的實施,重複之處不再贅述。   本說明書實施例提供的上述會話資訊處理方法或裝置可以在電腦中由處理器執行相應的程式指令來實現,如使用windows作業系統的c++語言在PC端實現,或其他例如使用android、iOS系統程式設計語言在智慧終端機實現,以及基於量子電腦的處理邏輯實現等。本說明書提供的一種會話資訊處理裝置的另一種實施例中,可以包括處理器以及用於儲存處理器可執行指令的記憶體,所述處理器執行所述指令時實現:   獲取待處理語音資訊,識別所述待處理語音資訊的資訊內容,所述資訊內容至少包括待處理語音資訊中包含的關鍵詞資訊、待處理語音資訊表達的情緒資訊中的一種;   獲取與所述資訊內容相匹配的裝飾配置資訊;   利用所述裝飾配置資訊對所述待處理語音資訊的展示載體進行渲染;   展示所述渲染後的展示載體。   所示裝置的另一種實施例中,可以包括處理器以及用於儲存處理器可執行指令的記憶體,所述處理器執行所述指令時實現:   獲取待處理會話資訊,識別所述待處理會話資訊的資訊內容,所述資訊內容至少包括待處理會話資訊中包含的關鍵詞資訊、待處理會話資訊表達的情緒資訊中的一種;   獲取與識別出的所述資訊內容相匹配的裝飾配置資訊;   利用所述裝飾配置資訊對所述待處理會話資訊的展示載體進行渲染;   展示所述渲染後的展示載體。   所述待處理會話資訊可以包括語音資訊、文本資訊、圖像資訊、視頻資訊中的至少一種。   需要說明的,上述所述的裝置根據方法實施例的描述還可以包括其他的實施方式,具體的實現方式可以參照相關方法實施例的描述,在此不作一一贅述。   本說明書實施例提供的一種會話資訊處理裝置,可以獲取待處理語音資訊,識別所述待處理語音資訊表達的情緒資訊。然後可以根據包括所述情緒內容的資訊內容調取相應的裝飾配置資訊,進而利用所述裝飾配置資訊對所述待處理語音資訊的展示載體(例如音頻資訊條)進行渲染後進行展示。這樣根據語音資訊表達的情感對展示載體進行渲染後展示,不僅可以表達或部分表達了語音資訊的內容,還增加了語音資訊聊天時的趣味性。利用本說明書實施例提供的實施方案,可以通過分析音頻內熱詞或者感情,對音頻訊息體進行額外的描述或裝飾,讓音頻資訊的內容可以從視覺上展示,提高聊天趣味性,提高使用者體驗。   上述所述的方法或裝置可以用於多種具有會話功能的電子設備中,如移動智慧終端機、車載設備、PC(personal computer,個人電腦)、智慧穿戴設備、業務系統的伺服器等,可以對展示訊息的展示載體進行二次渲染,提高聊天趣味性,提高使用者體驗。具體的,本說明書提供一種電子設備,如圖14所示,可以包括至少一個處理器以及使用者儲存處理器可執行指令的記憶體,所述處理器執行所述指令時實現:   獲取待處理會話資訊,識別所述待處理會話資訊的資訊內容,所述資訊內容至少包括待處理會話資訊中包含的關鍵詞資訊、待處理會話資訊表達的情緒資訊中的一種;   獲取與識別出的所述資訊內容相匹配的裝飾配置資訊;   利用所述裝飾配置資訊對所述待處理會話資訊的展示載體進行渲染;   展示所述渲染後的展示載體。   所述電子設備的一個實施例中,所述待處理會話資訊可以包括語音資訊、文本資訊、圖像資訊、視頻資訊中的至少一種。當前,根據不同的資料來源還可以包括其他類型的待處理會話資訊,具體的可以根據實際會話場景確定。   需要說明的是說明書上述所述的裝置或電子設備根據相關方法實施例的描述還可以包括其他的實施方式,具體的實現方式可以參照方法實施例的描述,在此不作一一贅述。本說明書中的各個實施例均採用遞進的方式描述,各個實施例之間相同相似的部分互相參見即可,每個實施例重點說明的都是與其他實施例的不同之處。尤其,對於硬體+程式類實施例而言,由於其基本相似於方法實施例,所以描述的比較簡單,相關之處參見方法實施例的部分說明即可。   上述對本說明書特定實施例進行了描述。其它實施例在所附申請專利範圍的範圍內。在一些情況下,在申請專利範圍中記載的動作或步驟可以按照不同於實施例中的順序來執行並且仍然可以實現期望的結果。另外,在附圖中描繪的過程不一定要求示出的特定順序或者連續順序才能實現期望的結果。在某些實施方式中,多工處理和並行處理也是可以的或者可能是有利的。   本說明書一個或多個實施例提供的一種會話資訊處理方法、裝置、電子設備,可以獲取待處理語音資訊,識別所述待處理語音資訊表達的情緒資訊。然後可以根據包括所述情緒內容的資訊內容調取相應的裝飾配置資訊,進而利用所述裝飾配置資訊對所述待處理語音資訊的展示載體(例如音頻資訊條)進行渲染後進行展示。這樣根據語音資訊表達的情感對展示載體進行渲染後展示,不僅可以表達或部分表達了語音資訊的內容,還增加了語音資訊聊天時的趣味性。利用本說明書實施例提供的實施方案,可以通過分析音頻內熱詞或者感情,對音頻訊息體進行額外的描述或裝飾,讓音頻資訊的內容可以從視覺上展示,提高聊天趣味性,提高使用者體驗。   儘管說明書實施例內容中提到語音到文本的識別轉換、KMP演算法的關鍵字搜索、利用神經網路識別情緒資訊、裝飾配置資訊匹配方式/展示效果等之類的資料/定義、獲取、互動、計算、判斷等描述,但是,本說明書實施例並不局限於必須是符合行業通信標準、標準電腦資料處理和儲存規則或本說明書一個或多個實施例所描述的情況。某些行業標準或者使用自訂方式或實施例描述的實施基礎上略加修改後的實施方案也可以實現上述實施例相同、等同或相近、或變形後可預料的實施效果。應用這些修改或變形後的資料獲取、儲存、判斷、處理方式等獲取的實施例,仍然可以屬於本說明書實施例的可選實施方案範圍之內。   上述對本說明書特定實施例進行了描述。其它實施例在所附申請專利範圍的範圍內。在一些情況下,在申請專利範圍中記載的動作或步驟可以按照不同於實施例中的順序來執行並且仍然可以實現期望的結果。另外,在附圖中描繪的過程不一定要求示出的特定順序或者連續順序才能實現期望的結果。在某些實施方式中,多工處理和並行處理也是可以的或者可能是有利的   在20世紀90年代,對於一個技術的改進可以很明顯地區分是硬體上的改進(例如,對二極體、電晶體、開關等電路結構的改進)還是軟體上的改進(對於方法流程的改進)。然而,隨著技術的發展,當今的很多方法流程的改進已經可以視為硬體電路結構的直接改進。設計人員幾乎都通過將改進的方法流程程式設計到硬體電路中來得到相應的硬體電路結構。因此,不能說一個方法流程的改進就不能用硬體實體模組來實現。例如,可程式設計邏輯裝置(Programmable Logic Device, PLD)(例如場可程式設計閘陣列(Field Programmable Gate Array,FPGA))就是這樣一種積體電路,其邏輯功能由使用者對裝置程式設計來確定。由設計人員自行程式設計來把一個數位系統“集成”在一片PLD上,而不需要請晶片製造廠商來設計和製作專用的積體電路晶片。而且,如今,取代手工地製作積體電路晶片,這種程式設計也多半改用“邏輯編譯器(logic compiler)”軟體來實現,它與程式開發撰寫時所用的軟體編譯器相類似,而要編譯之前的原始代碼也得用特定的程式設計語言來撰寫,此稱之為硬體描述語言(Hardware Description Language,HDL),而HDL也並非僅有一種,而是有許多種,如ABEL(Advanced Boolean Expression Language)、AHDL(Altera Hardware Description Language)、Confluence、CUPL(Cornell University Programming Language)、HDCal、JHDL(Java Hardware Description Language)、Lava、Lola、MyHDL、PALASM、RHDL (Ruby Hardware Description Language)等,目前最普遍使用的是VHDL(Very-High-Speed Integrated Circuit Hardware Description Language)與Verilog。本領域技術人員也應該清楚,只需要將方法流程用上述幾種硬體描述語言稍作邏輯程式設計並程式設計到積體電路中,就可以很容易得到實現該邏輯方法流程的硬體電路。   控制器可以按任何適當的方式實現,例如,控制器可以採取例如微處理器或處理器以及儲存可由該(微)處理器執行的電腦可讀程式碼(例如軟體或固件)的電腦可讀媒體、邏輯閘、開關、專用積體電路(Application Specific Integrated Circuit,ASIC)、可程式設計邏輯控制器和嵌入微控制器的形式,控制器的例子包括但不限於以下微控制器:ARC 625D、Atmel AT91SAM、Microchip PIC18F26K20以及Silicone Labs C8051F320,記憶體控制器還可以被實現為記憶體的控制邏輯的一部分。本領域技術人員也知道,除了以純電腦可讀程式碼方式實現控制器以外,完全可以通過將方法步驟進行邏輯程式設計來使得控制器以邏輯閘、開關、專用積體電路、可程式設計邏輯控制器和嵌入微控制器等的形式來實現相同功能。因此這種控制器可以被認為是一種硬體部件,而對其內包括的用於實現各種功能的裝置也可以視為硬體部件內的結構。或者甚至,可以將用於實現各種功能的裝置視為既可以是實現方法的軟體模組又可以是硬體部件內的結構。   上述實施例闡明的系統、裝置、模組或單元,具體可以由電腦晶片或實體實現,或者由具有某種功能的產品來實現。一種典型的實現設備為電腦。具體的,電腦例如可以為個人電腦、膝上型電腦、車載人機互動設備、蜂窩電話、相機電話、智慧型電話、個人數位助理、媒體播放機、導航設備、電子郵件設備、遊戲控制台、平板電腦、可穿戴設備或者這些設備中的任何設備的組合。   雖然本說明書一個或多個實施例提供了如實施例或流程圖所述的方法操作步驟,但基於常規或者無進步性的手段可以包括更多或者更少的操作步驟。實施例中列舉的步驟順序僅僅為眾多步驟執行順序中的一種方式,不代表唯一的執行順序。在實際中的裝置或終端產品執行時,可以按照實施例或者附圖所示的方法循序執行或者並存執行(例如並行處理器或者多執行緒處理的環境,甚至為分散式資料處理環境)。術語“包括”、“包含”或者其任何其他變體意在涵蓋非排他性的包含,從而使得包括一系列要素的過程、方法、產品或者設備不僅包括那些要素,而且還包括沒有明確列出的其他要素,或者是還包括為這種過程、方法、產品或者設備所固有的要素。在沒有更多限制的情況下,並不排除在包括所述要素的過程、方法、產品或者設備中還存在另外的相同或等同要素。   為了描述的方便,描述以上裝置時以功能分為各種模組分別描述。當然,在實施本說明書一個或多個時可以把各模組的功能在同一個或多個軟體和/或硬體中實現,也可以將實現同一功能的模組由多個子模組或子單元的組合實現等。以上所描述的裝置實施例僅僅是示意性的,例如,所述單元的劃分,僅僅為一種邏輯功能劃分,實際實現時可以有另外的劃分方式,例如多個單元或組件可以結合或者可以集成到另一個系統,或一些特徵可以忽略,或不執行。另一點,所顯示或討論的相互之間的耦合或直接耦合或通信連接可以是通過一些介面,裝置或單元的間接耦合或通信連接,可以是電性,機械或其它的形式。   本領域技術人員也知道,除了以純電腦可讀程式碼方式實現控制器以外,完全可以通過將方法步驟進行邏輯程式設計來使得控制器以邏輯閘、開關、專用積體電路、可程式設計邏輯控制器和嵌入微控制器等的形式來實現相同功能。因此這種控制器可以被認為是一種硬體部件,而對其內部包括的用於實現各種功能的裝置也可以視為硬體部件內的結構。或者甚至,可以將用於實現各種功能的裝置視為既可以是實現方法的軟體模組又可以是硬體部件內的結構。   本發明是參照根據本發明實施例的方法、設備(系統)、和電腦程式產品的流程圖和/或方塊圖來描述的。應理解可由電腦程式指令實現流程圖和/或方塊圖中的每一流程和/或方塊、以及流程圖和/或方塊圖中的流程和/或方塊的結合。可提供這些電腦程式指令到通用電腦、專用電腦、嵌入式處理機或其他可程式設計資料處理設備的處理器以產生一個機器,使得通過電腦或其他可程式設計資料處理設備的處理器執行的指令產生用於實現在流程圖一個流程或多個流程和/或方塊圖一個方塊或多個方塊中指定的功能的裝置。   這些電腦程式指令也可儲存在能引導電腦或其他可程式設計資料處理設備以特定方式工作的電腦可讀記憶體中,使得儲存在該電腦可讀記憶體中的指令產生包括指令裝置的製造品,該指令裝置實現在流程圖一個流程或多個流程和/或方塊圖一個方塊或多個方塊中指定的功能。   這些電腦程式指令也可裝載到電腦或其他可程式設計資料處理設備上,使得在電腦或其他可程式設計設備上執行一系列操作步驟以產生電腦實現的處理,從而在電腦或其他可程式設計設備上執行的指令提供用於實現在流程圖一個流程或多個流程和/或方塊圖一個方塊或多個方塊中指定的功能的步驟。   在一個典型的配置中,計算設備包括一個或多個處理器(CPU)、輸入/輸出介面、網路介面和記憶體。   記憶體可能包括電腦可讀媒體中的非永久性記憶體,隨機存取記憶體(RAM)和/或非易失性記憶體等形式,如唯讀記憶體(ROM)或快閃記憶體(flash RAM)。記憶體是電腦可讀媒體的示例。   電腦可讀媒體包括永久性和非永久性、可移動和非可移動媒體可以由任何方法或技術來實現資訊儲存。資訊可以是電腦可讀指令、資料結構、程式的模組或其他資料。電腦的儲存媒體的例子包括,但不限於相變記憶體(PRAM)、靜態隨機存取記憶體(SRAM)、動態隨機存取記憶體(DRAM)、其他類型的隨機存取記憶體(RAM)、唯讀記憶體(ROM)、電可擦除可程式設計唯讀記憶體(EEPROM)、快閃記憶體或其他記憶體技術、唯讀光碟唯讀記憶體(CD-ROM)、數位多功能光碟(DVD)或其他光學儲存、磁盒式磁帶,磁帶磁磁片儲存、石墨烯儲存或其他磁性存放裝置或任何其他非傳輸媒體,可用於儲存可以被計算設備存取的資訊。按照本文中的界定,電腦可讀媒體不包括暫存電腦可讀媒體(transitory media),如調變的資料信號和載波。   本領域技術人員應明白,本說明書一個或多個實施例可提供為方法、系統或電腦程式產品。因此,本說明書一個或多個實施例可採用完全硬體實施例、完全軟體實施例或結合軟體和硬體方面的實施例的形式。而且,本說明書一個或多個實施例可採用在一個或多個其中包含有電腦可用程式碼的電腦可用儲存媒體(包括但不限於磁碟記憶體、CD-ROM、光學記憶體等)上實施的電腦程式產品的形式。   本說明書一個或多個實施例可以在由電腦執行的電腦可執行指令的一般上下文中描述,例如程式模組。一般地,程式模組包括執行特定任務或實現特定抽象資料類型的常式、程式、物件、組件、資料結構等等。也可以在分散式運算環境中實踐本本說明書一個或多個實施例,在這些分散式運算環境中,由通過通信網路而被連接的遠端處理設備來執行任務。在分散式運算環境中,程式模組可以位於包括存放裝置在內的本地和遠端電腦儲存媒體中。   本說明書中的各個實施例均採用遞進的方式描述,各個實施例之間相同相似的部分互相參見即可,每個實施例重點說明的都是與其他實施例的不同之處。尤其,對於系統實施例而言,由於其基本相似於方法實施例,所以描述的比較簡單,相關之處參見方法實施例的部分說明即可。在本說明書的描述中,參考術語“一個實施例”、“一些實施例”、“示例”、“具體示例”、或“一些示例”等的描述意指結合該實施例或示例描述的具體特徵、結構、材料或者特點包含於本說明書的至少一個實施例或示例中。在本說明書中,對上述術語的示意性表述不必須針對的是相同的實施例或示例。而且,描述的具體特徵、結構、材料或者特點可以在任一個或多個實施例或示例中以合適的方式結合。此外,在不相互矛盾的情況下,本領域的技術人員可以將本說明書中描述的不同實施例或示例以及不同實施例或示例的特徵進行結合和組合。   以上所述僅為本說明書一個或多個實施例的實施例而已,並不用於限制本本說明書一個或多個實施例。對於本領域技術人員來說,本說明書一個或多個實施例可以有各種更改和變化。凡在本申請的精神和原理之內所作的任何修改、等同替換、改進等,均應包含在申請專利範圍之內。In order to make those skilled in the art better understand the technical solutions in the present specification, the technical solutions in one or more embodiments of the present specification will be clearly and completely described below in conjunction with the drawings in one or more embodiments of the specification. It is apparent that the described embodiments are only a part of the embodiments of the specification, and not all of the embodiments. All other embodiments obtained by a person of ordinary skill in the art based on one or more embodiments of the specification, without departing from the scope of the invention, should fall within the scope of the embodiments. FIG. 2 is a schematic flowchart diagram of an embodiment of a method for processing session information provided by the present specification. Although the present specification provides method operational steps or apparatus structures as shown in the following embodiments or figures, more or a portion of the combined operations may be included in the method or apparatus based on conventional or non-progressive labor. Step or module unit. In the steps or structures in which the necessary causal relationship does not exist logically, the execution order of the steps or the module structure of the device is not limited to the execution order or the module structure shown in the embodiment or the drawings. When the device, the server or the terminal product of the method or module structure is applied, the method or the module structure shown in the embodiment or the figure may be sequentially executed or concurrently executed (for example, a parallel processor or Multi-thread processing environment, even including distributed processing, server cluster implementation environment). The embodiments provided by the embodiments of the present specification can be used in various terminal devices having a session function, such as a mobile smart terminal, an in-vehicle device, a personal computer (PC), a wearable device, a server of a business system, and the like. The terminal device may be installed or configured with an instant messaging application, a temporary session, and the like. The implementation provided by the embodiment of the present disclosure may be used in a session application scenario based on the foregoing terminal device. The embodiment of the present specification can perform keyword or emotion recognition on the voice information, and then display it on the audio information display carrier. Specifically, the method may include analyzing the single or multiple pieces of voice information, and determining the sentiment of the voice information sender according to the recognition algorithm. Then, according to the emotional information, the display style of the audio information can be re-rendered (for example, changing the color of the message body, adding an Emoji expression head, etc.), which greatly improves the chat interest and user experience, and can express the information of the audio information. The content is displayed or partially displayed. For example, as shown in FIG. 2, in an embodiment of the method for processing session information, the method may include: S2: acquiring voice information to be processed, and identifying information content of the voice information to be processed, The information content includes at least the emotion information expressed by the to-be-processed voice information. In this embodiment, after inputting a piece of voice information in the session interface, the user can obtain the voice information as the voice information to be processed for session information processing in this embodiment. The voice information input by the user can be further processed by the processing, and then displayed in the session interface together with the decoration configuration information as shown below, or the message body of the voice information can be displayed first (here, the voice information can be included) The display carrier, such as the audio information bar in the session interface, and then render the message body to display the animation effect. In the application scenario of the embodiment, after acquiring the to-be-processed voice information, the information content in the to-be-processed voice information may be identified, where the identified information may include at least the emotion information expressed by the to-be-processed voice information. The identified emotional information can be expressed by characteristic words or identifiers, such as "happy", "sad", "surprise" and the like. The manner of identifying the sentiment information in the to-be-processed voice information may be processed by using multiple algorithms or models. In this embodiment, the analysis processing of the language information may include extracting audio information features, and identifying the speaker's emotions, such as joy, anger, etc., by recognizing the algorithm. In an embodiment of the present disclosure, the information content of the to-be-processed voice information may include: S20: extracting audio feature data of the to-be-processed voice information, and inputting the audio feature data. a pre-constructed emotion recognition model, by which the emotion category to which the to-be-processed voice information belongs is determined. In a specific implementation manner, a feature signal, such as speech rate, pitch, frequency, etc., can be extracted from the digitized audio signal, and the feature information can be modeled, and the model is trained through multiple samples to optimize the model. After the model training is completed, it can be used as a classifier to classify the emotions reflected by the newly input audio signals of the to-be-processed voice information, and determine the emotional information expressed by the to-be-processed voice information, such as “joy” and “anger”. The emotion recognition model may include an algorithm that can identify the emotion classification of the audio information after the sample data is trained by using the selected recognition algorithm or the network model. For example, KNN (k-Neest Neighbor) classification algorithm, or deep neural network/convolution neural network and its transformed/variant neural network can be selected. In one or more embodiments provided by the present specification, the extracted audio feature data includes at least one of a sentiment feature data of the biological information of the to-be-processed voice information and a tone feature data. The emotional characteristic data of the creature may include emotions such as anger, anger, sadness, joy, etc., and the tone characteristic data may include feature data in a tone state such as a question, a statement, a question, a sigh, and the like. In an embodiment, the extracted audio feature data may include feature data that can reflect the emotional aspect of the person, such as reflecting the emotion of the person based on the speech rate, pitch, frequency, etc., in a specific example, such as fast speech speed and high pitch. It can be reflected that the spokesperson’s mood is relatively high. Conversely, a slow speech rate and a low pitch can reflect the low sentiment of the spokesperson. In other embodiments, the tone feature data may also be obtained from the grammatical point of view as the audio feature data. For example, judging whether the speaker is a doubt or a sigh from the voice information to be processed. Of course, the above two can also be combined to judge the inventor's emotion through the extracted tonal feature data and the subject-predicate order. Specifically, the corresponding type of audio feature data to be extracted may be set according to an application environment or a data processing requirement. S4: Obtain decoration configuration information that matches the information content. In an embodiment of the present specification, a decoration configuration database may be preset, and the decoration configuration database may store decoration configuration information for decorating the display carrier of the audio information, such as discoloration of the audio information strip. The decorative configuration database may be set in a local terminal, such as a server local storage area, or a client local storage area, or may be stored in a remote server. In this way, after acquiring the information content in the to-be-processed voice information, the decoration configuration information matching the identified information content may be obtained from the decoration configuration database. In the application scenario of this embodiment, the information content may include emotion information. In some embodiments, the identified information content may include multiple or more emotional information, such as "happy", "like", "surprise", etc., which tend to be positive positive emotion categories. Therefore, in one or more embodiments provided by the present specification, the recognition results of the identified plurality of emotion information may be combined and converted into corresponding request parameters. The corresponding decoration configuration information can be retrieved through unified and standardized request parameters. Specifically, in an embodiment, the obtaining the decoration configuration information that matches the identified information content may include: S40: combining the recognition results of the emotion information in the to-be-processed voice information to generate a request parameter; S42: Acquire decoration configuration information matched according to the request parameter. The above "happy", "like", "surprise" can be merged into request parameters including two emotional categories "happy" and "surprise", and then one or two corresponding ones can be matched from the decorative configuration database. Decorative configuration information. Certainly, in other implementation manners, the information of the three emotion categories may be combined into one request parameter “happy” according to the set processing rule, and the decoration configuration information of “happy” is retrieved according to the request parameter. The decoration configuration information described in the embodiments of the present specification may include, but is not limited to, a text description, an Emoji expression, a background color, an animation effect, and the like. An emotional information or an emotional information or a request parameter may correspond to one or more decorative configuration information, and may be set with a corresponding mapping relationship, such as "happy" emotions may correspond to "happy 1", "happy 2", "happy 3" "Three decorative configuration information, which decoration configuration information to use, can be executed in order or randomly or in other defined ways. Therefore, in one or more embodiments of the present specification, the decoration configuration information may include at least one of the following: a text description, an expression, a background color change, an animation effect played at the display carrier preset position, Demonstrate the deformation animation effect of the carrier. By using the above-mentioned types of decorative configuration information, the interest of the session can be greatly increased, and the manner in which a single voice information bar is displayed can be changed. Moreover, the set decoration configuration information is associated with and matched with the expressed emotion information in the voice information, and the content of the voice information can be displayed or partially displayed through the decoration configuration information, thereby improving the user experience. S6: The display carrier of the to-be-processed voice information is rendered by using the decoration configuration information. After obtaining the decoration configuration information, the display carrier can be rendered by using the decoration configuration information. The display carrier may include, for example, an audio information bar, and may also include other message carriers in the session interface for indicating that the information content sent by the user is voice information, such as an audio information bar, an audio information frame, and an audio. Information window, audio symbols, etc. The audio information window may include a window for playing audio information by a separate user. For example, when the user inputs the voice information, the audio information window may be popped up at a specified position of the session interface or used to play voice information. The decoration configuration information renders the audio information window. Of course, the display carrier may also be an audio symbol or other form, such as a speaker audio match, a button labeled "voice", a "voice" character, and the like. Therefore, in other embodiments of the method, the display carrier may include: at least one of an audio information bar, an audio information window, and an audio symbol. S8: Display the rendered display carrier. The rendered display carrier can be displayed in a session interface. For example, the animation information is rendered by using the decoration configuration information to render an animation effect of “rainbow transformation”. In a specific implementation scenario, an animation effect can be set in the preset position of the display carrier, as shown in FIG. 3, which is a schematic diagram of an embodiment of the method in the embodiment of the present specification. In FIG. 3, when the user is identified as "question" according to the voice information input by the user, a short animation of "three question marks" can be played on the right side of the level of the audio information bar, and each question mark symbol can be popped up at predetermined intervals. In another implementation scenario, the decoration configuration information may include a deformation animation effect of the display carrier. As shown in FIG. 4, FIG. 4 is a schematic diagram of another embodiment of a method according to an embodiment of the present specification. In FIG. 4, after the user inputs the voice information 1 and recognizes that the user's emotion information is “happy”, the audio information bar of the voice information 1 input by the user may be set as “rainbow transformation” of multiple color specified frequency conversions. The animation effect shows an atmosphere of joy, corresponding to the emotion expressed by the voice information input by the user. Correspondingly, if the emotional information is identified as "sadness", the "folly" decorative assembly information of the blue and purple slower frequency transformation can be used, as shown in FIG. FIG. 5 is a schematic diagram of another embodiment of a method according to an embodiment of the present specification. A session information processing method provided by one or more embodiments of the present disclosure may acquire voice information to be processed and identify emotion information expressed by the voice information to be processed. Then, the corresponding decoration configuration information may be retrieved according to the information content including the emotional content, and then the display carrier (for example, an audio information bar) of the to-be-processed voice information is rendered and displayed by using the decoration configuration information. In this way, according to the emotion expressed by the voice information, the display carrier is rendered and displayed, which not only can express or partially express the content of the voice information, but also increase the interest of the voice information chat. By using the implementation provided by the embodiments of the present specification, the audio message body can be additionally described or decorated by analyzing the emotional information in the audio, so that the content of the audio information can be visually displayed, the chat interest is improved, and the user experience is improved. . In the session information processing method provided by one or more embodiments of the present specification, the voice information content may also be identified from the dimension of the keyword information included in the voice information, and then matched according to the information content including the keyword. Corresponding decoration configuration information, further rendering the voice information display carrier is displayed in the session interface. Specifically, as shown in FIG. 6, FIG. 6 is a schematic flowchart of a method of another embodiment of the method provided by the present specification. In another embodiment of the method provided by the present disclosure, the method may include: S120: Acquire voice information to be processed, and identify information content of the to-be-processed voice information, where the information content includes at least voice information to be processed. S140: acquiring decoration configuration information that matches the identified information content; S160: rendering the display carrier of the to-be-processed voice information by using the decoration configuration information; S180: displaying the Rendered display carrier. During the specific processing, it may be determined that the voice information to be processed is not processed, and the voice-to-word processing is performed, and the generated text is extracted according to rules (such as extracting network hot words or syntactic elements such as subject-predicate). The information content of the to-be-processed voice information may include: converting the to-be-processed voice information into text text; matching the text text with a keyword in a preset vocabulary, and extracting the included keyword News. The keywords may include pre-screened words (eg, online popular vocabulary), which may be stored in a preset vocabulary. In an embodiment of the present specification, after the text is converted into a voice, it is determined whether the word in the text contains the keyword in the predictive vocabulary, and if so, the keyword is extracted. The algorithm for extracting keywords can be implemented by string search. The specific algorithms can include full search, KMP (an improved string matching algorithm, by D. E. Knuth, J. H. Morris and V. R. Pratt proposed), Boyer More (a string matching algorithm) and other classic algorithms. In other implementations, the identifying the information content of the to-be-processed voice information may further include: converting the to-be-processed voice information into text text, and extracting keywords in the text text according to the adopted text grammar rules. For example, the extracted keyword is identified based on a grammatical element such as a subject-predicate of the converted text. Certainly, referring to the foregoing implementation manner for identifying the sentiment information in the to-be-processed voice, the implementation manner of using the voice information to convert the text information into the text and then identifying the keyword may further include other embodiments. In a specific embodiment, The obtaining the decoration configuration information that matches the identified information content may include: combining the recognition results of the keyword information in the to-be-processed voice information to generate a request parameter; and acquiring a decoration matched according to the request parameter Configuration information. And the obtaining the decoration configuration information that is matched based on the request parameter may include: matching the request parameter in a locally preset decoration configuration database to obtain decoration configuration information; or sending the request configuration information to the remote server. The request parameter acquires decoration configuration information obtained by the remote server according to the request parameter matching. Other embodiments, such as the kind of the decoration configuration information and the like, may also be included in accordance with the description of the foregoing embodiments. The various embodiments in the specification are described in a progressive manner, and the same or similar parts between the various embodiments may be referred to each other, and each embodiment focuses on the differences from the other embodiments. For details, refer to the description of the related embodiments, and details are not described herein. A session information processing method provided by one or more embodiments of the present disclosure may acquire voice information to be processed, convert the to-be-processed voice information into text information, and extract keywords in the text information. Then, the corresponding decoration configuration information may be retrieved according to the information content including the keyword, and then the display carrier (for example, an audio information bar) of the to-be-processed voice information is rendered and displayed by using the decoration configuration information. In this way, according to the keywords contained in the voice information, the corresponding decoration configuration information is used to render and display the display carrier, which can not only express or partially express the content of the voice information, but also increase the interest of the voice information chat. By using the embodiments provided in the embodiments of the present specification, the audio message body can be additionally described or decorated by analyzing the keywords in the audio, so that the content of the audio information can be visually displayed, the chat interest is improved, and the user experience is improved. . In another embodiment of the method provided by the present specification, the implementation manner of the keyword information and the emotion information based on the to-be-processed voice information may be combined to identify the to-be-processed voice information from at least one of the two dimensions. The information content, and then match the decoration configuration information, and display the message carrier after rendering. For example, the keyword information and the expressed emotion information in the to-be-processed voice information can be simultaneously identified. Therefore, in a session information processing method provided by one or more embodiments of the present disclosure, the method includes: acquiring voice information to be processed, and identifying information content of the to-be-processed voice information, where the information content includes at least a to-be-processed One of keyword information included in the voice information and emotion information expressed by the voice information to be processed; obtaining decoration configuration information matching the identified information content; using the decoration configuration information to the voice information to be processed The display carrier is rendered; the rendered display carrier is displayed. In one or more embodiments of the present specification, when the information about the to-be-processed voice information includes the keyword information included in the voice information to be processed and the emotion information expressed by the voice information to be processed, the to-be-processed The recognition results of the keyword information and the emotion information in the voice information are combined to generate a request parameter. Correspondingly, the obtaining the decoration configuration information that matches the information content comprises acquiring the decoration configuration information matched according to the request parameter. As shown in FIG. 7, FIG. 7 is a schematic flowchart of processing in an application scenario of another embodiment of the method provided by the present specification. In this way, the combination of the recognition of the keywords in the to-be-processed voice information and the recognition result of the emotion information can optimize the content of the identified information content, and can match the decoration configuration information that matches the content of the voice information to be processed more accurately, and improve the display effect. Improve the user experience. Further, based on the innovative idea of the embodiment of the present specification, in the session information processing scenario, the processed data object may not be limited to voice information, and other implementation scenarios may also include information such as text, image, or video. According to the above embodiment of the present specification, when processing session information of text, image or video, information content of at least one of keyword information and expressed emotion information in the text, image or video may be identified, and then based on The identified information content matches the corresponding decoration configuration information, and the presentation carrier of the session information is rendered and displayed in the session interface. Therefore, the present specification further provides a method for processing session information, the method may include: S220: Acquire session information to be processed, and identify information content of the session information to be processed, where the information content includes at least information included in the session information to be processed. a keyword information, one of the emotion information to be processed by the session information; S240: acquiring decoration configuration information that matches the identified information content; S260: using the decoration configuration information to view the to-be-processed session information Displaying the carrier for rendering; S280: displaying the rendered display carrier. FIG. 8 is a schematic flow chart of a method of another embodiment of the method provided by the present specification. The description of the method for processing the voice information processing method may also include other implementation manners. The specific implementation manners may refer to the description of the related method embodiments, and will not be described herein. . The session information processing method provided by the embodiments of the present disclosure can be used not only for the processing of voice information, but also for the text body, video, and the like. The implementation of the problem is similar to the processing of the foregoing voice information. Therefore, the specific implementation of the embodiment of the present specification can be referred to the implementation of the foregoing voice information processing method, and the repeated description is not repeated. The foregoing description of the specific embodiments of the specification has been described. Other embodiments are within the scope of the following claims. In some cases, the actions or steps recited in the scope of the claims may be performed in a different order than the embodiments and still achieve the desired results. In addition, the processes depicted in the figures are not necessarily in a particular order or in a sequential order to achieve the desired results. In some embodiments, multiplex processing and parallel processing are also possible or may be advantageous. A session information processing method provided by one or more embodiments of the present disclosure may acquire voice information to be processed and identify emotion information expressed by the voice information to be processed. Then, the corresponding decoration configuration information may be retrieved according to the information content including the emotional content, and then the display carrier (for example, an audio information bar) of the to-be-processed voice information is rendered and displayed by using the decoration configuration information. In this way, according to the emotion expressed by the voice information, the display carrier is rendered and displayed, which not only can express or partially express the content of the voice information, but also increase the interest of the voice information chat. With the implementation provided by the embodiments of the present specification, the audio message body can be additionally described or decorated by analyzing the hot words or feelings in the audio, so that the content of the audio information can be visually displayed, improving the chat interest and improving the user experience. . Based on the session information processing method described above, one or more embodiments of the present specification further provide a session information processing apparatus. The apparatus may include a system (including a decentralized system), a software (application), a module, a component, a server, a client, etc., using the method described in the embodiments of the present specification, in combination with necessary means for implementing the hardware. Based on the same innovative concept, the apparatus in one or more embodiments provided by the embodiments of the present specification is as described in the following embodiments. For the implementation of the specific device in the embodiment of the present invention, reference may be made to the implementation of the foregoing method, and the repeated description is not repeated. As used hereinafter, the term "unit" or "module" may implement a combination of software and/or hardware of a predetermined function. Although the device described in the following embodiments is preferably implemented in software, a hardware, or a combination of a soft body and a hardware, is also possible and conceivable. Specifically, FIG. 9 is a schematic structural diagram of a module of a session information processing apparatus according to the present disclosure. As shown in FIG. 9, the apparatus may include: an emotion recognition module 101, which may be used to obtain voice information to be processed. Identifying the information content of the to-be-processed voice information, the information content includes at least the emotion information expressed by the to-be-processed voice information; the decoration matching module 102 may be configured to acquire decoration configuration information that matches the information content; The rendering module 103 can be used to render the display carrier of the to-be-processed voice information by using the decoration configuration information. The effect display module 104 can be used to display the rendered display carrier. Of course, referring to the foregoing method embodiment, in other embodiments of the device, the emotion recognition module 101 may include: an emotion recognition model unit, which may be used to extract audio feature data of the to-be-processed voice information, The audio feature data is input into a pre-constructed emotion recognition model, and the emotion category to which the to-be-processed voice information belongs is determined by the emotion recognition model. And in another embodiment, the extracted audio feature data includes at least one of a sentiment feature data of the biological information of the to-be-processed voice information and a tone feature data. In other device embodiments, the decoration configuration information may also be matched by identifying keywords in the voice information. 10 is a schematic diagram of a module structure of another embodiment of a session information processing apparatus provided in the present specification. As shown in FIG. 10, the apparatus may include: a keyword identification module 201, which may be used to acquire voice information to be processed, and identify The information content of the to-be-processed voice information, the information content includes at least the keyword information included in the to-be-processed voice information; the decoration matching module 202 may be configured to obtain decoration configuration information that matches the information content; The rendering module 203 can be configured to render the display carrier of the to-be-processed voice information by using the decoration configuration information. The effect display module 204 can be used to display the rendered display carrier. Of course, referring to the foregoing method embodiment, in other embodiments of the device, the emotion recognition module 201 may include: a text conversion unit 2011, which may be used to convert the to-be-processed voice information into text text; The matching unit 2012 can be configured to match the text text with keywords in the preset vocabulary, and extract the included keyword information. In another embodiment of the session information processing apparatus, the keyword recognition module 201 may include: a text conversion unit 2011, which may be configured to convert the to-be-processed voice information into text text; the syntax matching unit 2013 may And the method is configured to convert the to-be-processed voice information into text text, and extract keywords in the text text according to used text grammar rules. Of course, in other embodiments of the apparatus, the keyword recognition apparatus may be configured to include the lexicon matching unit and the grammar matching unit, and use one or two of the methods to identify keywords in the voice information. As shown in FIG. 11, FIG. 11 is a schematic structural diagram of a module of an embodiment of the keyword recognition module in the device provided in the present specification. According to the foregoing corresponding method embodiment, the foregoing device for keyword or emotion recognition based on voice information may further include other embodiments. The device may further include: a feature merging unit, configured to combine the recognition result of the keyword information or the emotion information in the to-be-processed voice information to generate a request parameter; correspondingly, the decoration matching module may be based on The request parameter matches the decoration configuration information. In other embodiments, the decoration configuration information may include at least one of the following: a text description, an expression, a background color change, an animation effect played at the display carrier preset position, and a deformation animation effect of the display carrier. In other embodiments, the display carrier may include: at least one of an audio information bar, an audio information window, and an audio match. In another embodiment of the apparatus, a processing module including a recognition key word in the voice information and a processing module for recognizing the emotion information expressed by the voice information may be provided. The device may identify the information content of the to-be-processed voice information based on any one of the modules or using the two modules. Therefore, the present specification further provides an embodiment of another session information processing apparatus, where the apparatus may include: an information content identification module 301, which may be used to acquire voice information to be processed, and identify information content of the voice information to be processed. The information content includes at least one of the keyword information included in the voice information to be processed and the emotion information expressed by the voice information to be processed; the decoration matching module 302 may be configured to obtain the decoration configuration information that matches the information content; The rendering module 303 can be used to render the display carrier of the to-be-processed voice information by using the decoration configuration information. The effect display module 304 can be used to display the rendered display carrier. FIG. 12 is a schematic structural diagram of a module of another embodiment of the apparatus provided by the present specification. As shown in FIG. 12, in another embodiment of the session information processing apparatus, if the processing method of recognizing the voice information keyword and the emotion information is used, the apparatus may further include: a feature combining unit 305, which may be used to The recognition result of the keyword information and the emotion information in the to-be-processed voice information is combined to generate a request parameter. Correspondingly, the decoration matching module 302 can match the decoration configuration information based on the request parameter. It should be noted that the foregoing description may further include other embodiments according to the description of the method embodiments. For specific implementations, reference may be made to the description of the related method embodiments, and details are not described herein. A session information processing apparatus provided by one or more embodiments of the present disclosure may acquire voice information to be processed and identify emotion information expressed by the voice information to be processed. Then, the corresponding decoration configuration information may be retrieved according to the information content including the emotional content, and then the display carrier (for example, an audio information bar) of the to-be-processed voice information is rendered and displayed by using the decoration configuration information. In this way, according to the emotion expressed by the voice information, the display carrier is rendered and displayed, which not only can express or partially express the content of the voice information, but also increase the interest of the voice information chat. With the implementation provided by the embodiments of the present specification, the audio message body can be additionally described or decorated by analyzing the hot words or feelings in the audio, so that the content of the audio information can be visually displayed, improving the chat interest and improving the user experience. . In another embodiment of the session information processing apparatus, the identified information source may further include text, images, video, etc., and may identify keywords or emotional information contained in conversation information such as text or video, and then match corresponding Decorate configuration information to render the message body. For example, FIG. 13 is a schematic diagram of a module structure of another embodiment of the apparatus provided in the present specification. The apparatus may include: an information content identification module 401, which may be used to obtain a to-be-processed The session information is used to identify the information content of the to-be-processed session information, and the information content includes at least one of keyword information included in the session information to be processed and emotion information to be processed by the session information; the decoration matching module 402 may The rendering module 403 is configured to use the decoration configuration information to render the display carrier of the to-be-processed session information; the effect display module 404 is configured to obtain the decoration configuration information that matches the identified information content. Can be used to display the rendered display carrier. Similarly, the session processing method of the other data sources described above may further include other implementation manners according to the description of the voice information processing method embodiment, and the implementation solution of the problem is similar to the processing of the foregoing voice information, so the present specification is either For specific implementations of the multiple embodiments, reference may be made to the implementation of the foregoing voice information processing method, and the repeated description is not repeated. The foregoing session information processing method or apparatus provided by the embodiments of the present disclosure may be implemented by a processor executing a corresponding program instruction in a computer, such as using a C++ language of a Windows operating system on a PC side, or other programs such as android and iOS. The design language is implemented in smart terminals, as well as processing logic based on quantum computers. In another embodiment of the present disclosure, a session information processing apparatus may be configured to include a processor and a memory for storing processor-executable instructions, where the processor executes the instructions to: acquire voice information to be processed, Identifying the information content of the to-be-processed voice information, where the information content includes at least one of keyword information included in the voice information to be processed and emotion information expressed by the voice information to be processed; and acquiring a decoration matching the information content Configuring information; rendering the display carrier of the to-be-processed voice information by using the decoration configuration information; and displaying the rendered display carrier. In another embodiment of the illustrated apparatus, a processor and a memory for storing processor-executable instructions are implemented, the processor executing the instructions: obtaining pending session information, identifying the pending session Information content of the information, the information content includes at least one of keyword information included in the session information to be processed, and emotion information expressed by the session information to be processed; and obtaining decoration configuration information that matches the identified information content; And displaying the display carrier of the to-be-processed session information by using the decoration configuration information; and displaying the rendered display carrier. The to-be-processed session information may include at least one of voice information, text information, image information, and video information. It should be noted that the foregoing description may further include other embodiments according to the description of the method embodiments. For specific implementations, reference may be made to the description of the related method embodiments, and details are not described herein. A session information processing apparatus provided in an embodiment of the present disclosure can acquire voice information to be processed and identify emotion information expressed by the voice information to be processed. Then, the corresponding decoration configuration information may be retrieved according to the information content including the emotional content, and then the display carrier (for example, an audio information bar) of the to-be-processed voice information is rendered and displayed by using the decoration configuration information. In this way, according to the emotion expressed by the voice information, the display carrier is rendered and displayed, which not only can express or partially express the content of the voice information, but also increase the interest of the voice information chat. With the implementation provided by the embodiments of the present specification, the audio message body can be additionally described or decorated by analyzing the hot words or feelings in the audio, so that the content of the audio information can be visually displayed, improving the chat interest and improving the user experience. . The method or device described above can be used in a plurality of electronic devices having a session function, such as a mobile smart terminal device, an in-vehicle device, a personal computer (PC), a smart wearable device, a server of a business system, etc., which can be The display carrier of the display message is rendered twice to improve the chat interest and improve the user experience. Specifically, the present specification provides an electronic device, as shown in FIG. 14, which may include at least one processor and a memory for a user to store processor-executable instructions, and when the processor executes the instructions, the method is: Acquiring a pending session Information, the information content of the session information to be processed, the information content includes at least one of keyword information included in the session information to be processed and emotion information expressed by the session information to be processed; and the obtained and recognized information The decoration configuration information matched by the content; the display carrier of the to-be-processed session information is rendered by using the decoration configuration information; and the rendered display carrier is displayed. In an embodiment of the electronic device, the to-be-processed session information may include at least one of voice information, text information, image information, and video information. Currently, other types of pending session information may be included according to different data sources, and the specific information may be determined according to actual session scenarios. It should be noted that the device or the electronic device described in the foregoing description may further include other embodiments according to the description of the related method embodiments. For specific implementation manners, reference may be made to the description of the method embodiments, and details are not described herein. The various embodiments in the specification are described in a progressive manner, and the same or similar parts between the various embodiments may be referred to each other, and each embodiment focuses on the differences from the other embodiments. In particular, for the hardware + program embodiment, since it is basically similar to the method embodiment, the description is relatively simple, and the relevant parts can be referred to the description of the method embodiment. The foregoing description of the specific embodiments of the specification has been described. Other embodiments are within the scope of the appended claims. In some cases, the actions or steps recited in the scope of the claims may be performed in a different order than the embodiments and still achieve the desired results. In addition, the processes depicted in the figures are not necessarily in a particular order or in a sequential order to achieve the desired results. In some embodiments, multiplex processing and parallel processing are also possible or may be advantageous. A session information processing method, apparatus, and electronic device provided by one or more embodiments of the present disclosure may acquire voice information to be processed, and identify emotion information expressed by the voice information to be processed. Then, the corresponding decoration configuration information may be retrieved according to the information content including the emotional content, and then the display carrier (for example, an audio information bar) of the to-be-processed voice information is rendered and displayed by using the decoration configuration information. In this way, according to the emotion expressed by the voice information, the display carrier is rendered and displayed, which not only can express or partially express the content of the voice information, but also increase the interest of the voice information chat. With the implementation provided by the embodiments of the present specification, the audio message body can be additionally described or decorated by analyzing the hot words or feelings in the audio, so that the content of the audio information can be visually displayed, improving the chat interest and improving the user experience. . Although the content of the description embodiment refers to speech-to-text recognition conversion, keyword search of KMP algorithm, use of neural network to identify emotion information, decoration configuration information matching method/display effect, etc., data/definition, acquisition, interaction Descriptions of calculations, judgments, etc., however, embodiments of the present specification are not limited to being necessarily in accordance with industry communication standards, standard computer data processing and storage rules, or as described in one or more embodiments of the specification. Certain industry standards or implementations that are modified in a manner that is custom-made or that are described in the embodiments can also achieve the same, equivalent, or similar, or modified, implementation effects of the above-described embodiments. Embodiments obtained by applying these modified or modified data acquisition, storage, judgment, processing methods, etc., may still fall within the scope of alternative embodiments of the embodiments of the present specification. The foregoing description of the specific embodiments of the specification has been described. Other embodiments are within the scope of the appended claims. In some cases, the actions or steps recited in the scope of the claims may be performed in a different order than the embodiments and still achieve the desired results. In addition, the processes depicted in the figures are not necessarily in a particular order or in a sequential order to achieve the desired results. In some embodiments, multiplex processing and parallel processing are also possible or may be advantageous. In the 1990s, improvements to one technology may clearly distinguish between hardware improvements (eg, diodes). Improvements in circuit structures such as transistors, switches, etc.) or improvements in software (for improvements in method flow). However, as technology advances, many of today's method flow improvements can be seen as direct improvements in hardware circuit architecture. Designers almost always get the corresponding hardware structure by programming the improved method flow into the hardware circuit. Therefore, it cannot be said that the improvement of a method flow cannot be implemented by a hardware entity module. For example, a Programmable Logic Device (PLD) (such as a Field Programmable Gate Array (FPGA)) is an integrated circuit whose logic function is determined by the user to program the device. . Designers can program themselves to "integrate" a digital system onto a single PLD without having to ask the chip manufacturer to design and fabricate a dedicated integrated circuit die. Moreover, today, instead of manually making integrated circuit chips, this programming is mostly implemented using a "logic compiler" software, which is similar to the software compiler used in programming development. The original code before compilation must also be written in a specific programming language. This is called Hardware Description Language (HDL), and HDL is not the only one, but there are many kinds, such as ABEL (Advanced Boolean Expression Language), AHDL (Altera Hardware Description Language), Confluence, CUPL (Cornell University Programming Language), HDCal, JHDL (Java Hardware Description Language), Lava, Lola, MyHDL, PALASM, RHDL (Ruby Hardware Description Language), etc. VHDL (Very-High-Speed Integrated Circuit Hardware Description Language) and Verilog are the most commonly used at present. It should also be clear to those skilled in the art that the hardware flow of the logic method flow can be easily obtained by simply designing and programming the method flow into the integrated circuit with the above hardware description languages. The controller can be implemented in any suitable manner, for example, the controller can take the form of, for example, a microprocessor or processor and a computer readable medium storing computer readable code (eg, software or firmware) executable by the (micro)processor , logic gates, switches, Application Specific Integrated Circuits (ASICs), programmable logic controllers, and embedded microcontrollers. Examples of controllers include, but are not limited to, the following microcontrollers: ARC 625D, Atmel The AT91SAM, Microchip PIC18F26K20, and Silicone Labs C8051F320, memory controllers can also be implemented as part of the memory's control logic. Those skilled in the art also know that, besides implementing the controller in a purely computer readable code, the controller can be controlled by logic gates, switches, dedicated integrated circuits, and programmable logic by logic programming the method steps. And embedded microcontrollers and other forms to achieve the same function. Thus such a controller can be considered a hardware component, and the means for implementing various functions included therein can also be considered as a structure within the hardware component. Or even a device for implementing various functions can be considered as either a software module implementing the method or a structure within the hardware component. The system, device, module or unit illustrated in the above embodiments may be implemented by a computer chip or an entity, or by a product having a certain function. A typical implementation device is a computer. Specifically, the computer can be, for example, a personal computer, a laptop computer, a vehicle-mounted human-machine interactive device, a cellular phone, a camera phone, a smart phone, a personal digital assistant, a media player, a navigation device, an email device, a game console, A tablet, wearable device, or a combination of any of these devices. Although one or more embodiments of the present specification provide method operational steps as described in the embodiments or flowcharts, more or less operational steps may be included based on conventional or non-progressive means. The order of the steps recited in the embodiments is only one of the many steps of the order of execution, and does not represent a single order of execution. When the actual device or terminal product is executed, it may be executed sequentially or concurrently according to the method shown in the embodiment or the drawing (for example, an environment of parallel processor or multi-thread processing, or even a distributed data processing environment). The terms "comprising," "comprising," or "comprising" or "comprising" or "the" Elements, or elements that are inherent to such a process, method, product, or device. In the absence of further limitations, it is not excluded that there are additional identical or equivalent elements in the process, method, product, or device. For the convenience of description, the above devices are described as being separately divided into various modules by function. Of course, in implementing one or more of the descriptions, the functions of each module may be implemented in the same software or software and/or hardware, or the modules implementing the same function may be multiple sub-modules or sub-units. The combination of implementations and so on. The device embodiments described above are merely illustrative. For example, the division of the unit is only a logical function division. In actual implementation, there may be another division manner. For example, multiple units or components may be combined or may be integrated into Another system, or some features can be ignored or not executed. In addition, the mutual coupling or direct coupling or communication connection shown or discussed may be an indirect coupling or communication connection through some interface, device or unit, and may be in an electrical, mechanical or other form. Those skilled in the art also know that, besides implementing the controller in a purely computer readable code, the controller can be controlled by logic gates, switches, dedicated integrated circuits, and programmable logic by logic programming the method steps. And embedded microcontrollers and other forms to achieve the same function. Therefore, such a controller can be considered as a hardware component, and a device included therein for realizing various functions can also be regarded as a structure within the hardware component. Or even a device for implementing various functions can be considered as either a software module implementing the method or a structure within the hardware component. The present invention has been described with reference to flowchart illustrations and/or block diagrams of a method, apparatus (system), and computer program product according to embodiments of the invention. It will be understood that each flow and/or block of the flowcharts and/or <RTIgt; These computer program instructions can be provided to a processor of a general purpose computer, a special purpose computer, an embedded processor or other programmable data processing device to produce a machine for executing instructions by a processor of a computer or other programmable data processing device Means are generated for implementing the functions specified in one or more flows of the flowchart or in a block or blocks of the block diagram. The computer program instructions can also be stored in a computer readable memory that can boot a computer or other programmable data processing device to operate in a particular manner, such that instructions stored in the computer readable memory produce an article of manufacture including the instruction device. The instruction means implements the functions specified in one or more blocks of a flow or a flow and/or block diagram of the flowchart. These computer program instructions can also be loaded onto a computer or other programmable data processing device to perform a series of operational steps on a computer or other programmable device to produce computer-implemented processing on a computer or other programmable device. The instructions executed on the steps provide steps for implementing the functions specified in one or more flows of the flowchart or in a block or blocks of the flowchart. In a typical configuration, a computing device includes one or more processors (CPUs), input/output interfaces, a network interface, and memory. The memory may include non-permanent memory, random access memory (RAM) and/or non-volatile memory in computer readable media, such as read only memory (ROM) or flash memory ( Flash RAM). Memory is an example of a computer readable medium. Computer readable media including both permanent and non-permanent, removable and non-removable media can be stored by any method or technology. Information can be computer readable instructions, data structures, modules of programs, or other materials. Examples of computer storage media include, but are not limited to, phase change memory (PRAM), static random access memory (SRAM), dynamic random access memory (DRAM), and other types of random access memory (RAM). Read-only memory (ROM), electrically erasable programmable read-only memory (EEPROM), flash memory or other memory technology, CD-ROM only, digitally versatile A compact disc (DVD) or other optical storage, magnetic cassette, magnetic tape storage, graphene storage or other magnetic storage device or any other non-transportable medium can be used to store information that can be accessed by a computing device. As defined herein, computer readable media does not include temporary storage of computer readable media, such as modulated data signals and carrier waves. Those skilled in the art will appreciate that one or more embodiments of the present specification can be provided as a method, system, or computer program product. Thus, one or more embodiments of the present specification can take the form of a complete hardware embodiment, a full software embodiment, or an embodiment combining soft and hardware aspects. Moreover, one or more embodiments of the present specification can be implemented on one or more computer usable storage media (including but not limited to disk memory, CD-ROM, optical memory, etc.) containing computer usable code therein. The form of a computer program product. One or more embodiments of the present specification can be described in the general context of computer-executable instructions executed by a computer, such as a program module. Generally, program modules include routines, programs, objects, components, data structures, and the like that perform particular tasks or implement particular abstract data types. One or more embodiments of the present specification can also be practiced in a distributed computing environment where tasks are performed by remote processing devices that are connected through a communication network. In a decentralized computing environment, the program modules can be located in local and remote computer storage media, including storage devices. The various embodiments in the specification are described in a progressive manner, and the same or similar parts between the various embodiments may be referred to each other, and each embodiment focuses on the differences from the other embodiments. In particular, for the system embodiment, since it is basically similar to the method embodiment, the description is relatively simple, and the relevant parts can be referred to the description of the method embodiment. In the description of the present specification, the description with reference to the terms "one embodiment", "some embodiments", "example", "specific example", or "some examples" and the like means a specific feature described in connection with the embodiment or example. A structure, material or feature is included in at least one embodiment or example of the specification. In the present specification, the schematic representation of the above terms is not necessarily directed to the same embodiment or example. Furthermore, the particular features, structures, materials, or characteristics described may be combined in a suitable manner in any one or more embodiments or examples. In addition, various embodiments or examples described in the specification, as well as features of various embodiments or examples, may be combined and combined. The above description is only an embodiment of one or more embodiments of the present specification, and is not intended to limit one or more embodiments of the present specification. Various changes and modifications can be made to one or more embodiments of the present disclosure. Any modifications, equivalent substitutions, improvements, etc. made within the spirit and principles of the present application are intended to be included in the scope of the claims.

101‧‧‧情緒識別模組101‧‧‧Emotion Recognition Module

102‧‧‧裝飾匹配模組102‧‧‧Decorative Matching Module

103‧‧‧渲染模組103‧‧‧ rendering module

104‧‧‧效果展示模組104‧‧‧effect display module

201‧‧‧關鍵詞識別模組201‧‧‧Keyword Identification Module

202‧‧‧裝飾匹配模組202‧‧‧Decorative Matching Module

203‧‧‧渲染模組203‧‧‧ rendering module

204‧‧‧效果展示模組204‧‧‧effect display module

301‧‧‧資訊內容識別模組301‧‧‧Information Content Identification Module

302‧‧‧裝飾匹配模組302‧‧‧Decorative Matching Module

303‧‧‧渲染模組303‧‧‧ rendering module

304‧‧‧效果展示模組304‧‧‧effect display module

305‧‧‧特徵合併單元305‧‧‧Characteristic unit

401‧‧‧資訊內容識別模組401‧‧‧Information Content Identification Module

402‧‧‧裝飾匹配模組402‧‧‧Decorative Matching Module

403‧‧‧渲染模組403‧‧‧ rendering module

404‧‧‧效果展示模組404‧‧‧effect display module

2011‧‧‧文本轉換單元2011‧‧‧Text Conversion Unit

2012‧‧‧詞庫匹配單元2012‧‧ lexicon matching unit

2013‧‧‧語法匹配單元2013‧‧‧grammar matching unit

為了更清楚地說明本說明書實施例或現有技術中的技術方案,下面將對實施例或現有技術描述中所需要使用的附圖作簡單地介紹,顯而易見地,下面描述中的附圖僅僅是說明書中記載的一些實施例,對於本領域普通技術人員來講,在不付出進步性勞動性的前提下,還可以根據這些附圖獲得其他的附圖。   圖1是現有會話介面中一種展示語音資訊訊息體的示意圖;   圖2是本說明書提供的所述一種會話資訊處理方法實施例的流程示意圖;   圖3是本說明書實施例所述方法的一個實施例場景示意圖;   圖4是本說明書實施例所述方法的另一個實施例場景示意圖;   圖5是本說明書實施例所述方法的另一個實施例場景示意圖;   圖6是本說明書提供的所述方法另一個實施例的方法流程示意圖;   圖7是本說明書提供的所述方法的另一個實施例應用場景中的處理流程示意圖;   圖8是本說明書提供的所述方法另一種實施例的方法流程示意圖;   圖9是本說明書提供的一種會話資訊處理裝置實施例的模組結構示意圖;   圖10是本說明書提供的另一種會話資訊處理裝置實施例的模組結構示意圖;   圖11是本說明書提供的所述裝置中所述關鍵詞識別模組一個實施例的模組結構示意圖;   圖12是本說明書提供的所述裝置的另一個實施例的模組結構示意圖;   圖13是本說明書提供的所述裝置的另一個實施例的模組結構示意圖;   圖14是本說明書提供的一種電子設備的結構示意圖。In order to more clearly illustrate the embodiments of the present specification or the technical solutions in the prior art, the drawings used in the embodiments or the description of the prior art will be briefly described below. Obviously, the drawings in the following description are merely descriptions. Some of the embodiments described herein can be obtained by those of ordinary skill in the art in view of the drawings without further progressive labor. 1 is a schematic diagram of a voice message body in an existing session interface; FIG. 2 is a schematic flowchart of an embodiment of a method for processing session information provided by the present specification; FIG. 3 is an embodiment of a method according to an embodiment of the present specification. FIG. 4 is a schematic diagram of another embodiment of the method according to the embodiment of the present specification; FIG. 5 is a schematic diagram of another embodiment of the method according to the embodiment of the present specification; FIG. 7 is a schematic flowchart of a process in another embodiment of the method provided by the present specification; FIG. 8 is a schematic flowchart of a method in another embodiment of the method provided by the present specification; FIG. 9 is a schematic structural diagram of a module of an embodiment of a session information processing apparatus provided by the present specification; FIG. 10 is a schematic structural diagram of another embodiment of a session information processing apparatus provided by the present specification; Said in the device FIG. 12 is a schematic structural diagram of a module of another embodiment of the apparatus provided by the present specification; FIG. 13 is a schematic diagram of another embodiment of the apparatus provided by the present specification. FIG. 14 is a schematic structural diagram of an electronic device provided by the present specification.

Claims (35)

一種會話資訊處理方法,所述方法包括:   獲取待處理語音資訊,識別所述待處理語音資訊的資訊內容,所述資訊內容至少包括所述待處理語音資訊表達的情緒資訊;   獲取與所述資訊內容相匹配的裝飾配置資訊;   利用所述裝飾配置資訊對所述待處理語音資訊的展示載體進行渲染;   展示所述渲染後的展示載體。A session information processing method, the method includes: acquiring a to-be-processed voice information, and identifying the information content of the to-be-processed voice information, where the information content includes at least the emotion information expressed by the to-be-processed voice information; acquiring and the information The decoration configuration information matching the content; rendering the display carrier of the to-be-processed voice information by using the decoration configuration information; and displaying the rendered display carrier. 如申請專利範圍第1項所述的會話資訊處理方法,所述識別所述待處理語音資訊的資訊內容包括:   提取所述待處理語音資訊的音頻特徵資料,將所述音頻特徵資料輸入預先構建的情緒識別模型,通過所述情緒識別模型確定所述待處理語音資訊所屬的情緒類別。The session information processing method of claim 1, wherein the identifying the information content of the to-be-processed voice information comprises: extracting audio feature data of the to-be-processed voice information, and inputting the audio feature data into a pre-built An emotion recognition model, by which the emotion category to which the to-be-processed voice information belongs is determined. 如申請專利範圍第2項所述的會話資訊處理方法,提取的所述音頻特徵資料包括所述待處理語音資訊的生物的情感特徵資料、語氣特徵資料中的至少一種。The session information processing method according to claim 2, wherein the extracted audio feature data includes at least one of a sentiment feature data and a tone feature data of the to-be-processed voice information. 如申請專利範圍第1至3項中任意一項所述的會話資訊處理方法,所述獲取與識別出的所述資訊內容相匹配的裝飾配置資訊包括:   將所述待處理語音資訊中情緒資訊的識別結果進行合併,生成請求參數;   獲取基於所述請求參數匹配出的裝飾配置資訊。The method of processing the session information according to any one of the preceding claims, wherein the obtaining the decoration configuration information that matches the identified information content comprises: the emotional information in the to-be-processed voice information The recognition results are combined to generate a request parameter; and the decoration configuration information matched based on the request parameter is obtained. 如申請專利範圍第4項所述的會話資訊處理方法,所述獲取基於所述請求參數匹配出的裝飾配置資訊包括:   將所述請求參數在本地預設的裝飾配置資料庫中進行匹配,獲取裝飾配置資訊;   或者,   向遠端服務端發送所述請求參數,獲取所述遠端服務端根據所述請求參數匹配得到的裝飾配置資訊。The method for processing the session information according to claim 4, wherein the obtaining the decoration configuration information that is matched based on the request parameter comprises: matching the request parameter in a locally preset decoration configuration database, and acquiring Or decorating the configuration information; or sending the request parameter to the remote server to obtain the decoration configuration information obtained by the remote server according to the request parameter matching. 如申請專利範圍第1項所述的會話資訊處理方法,所述裝飾配置資訊包括下述中的至少一種:   文字描述、表情、變更背景顏色、在所述展示載體預設位置播放動畫效果、所述展示載體的變形動畫效果。The session information processing method according to claim 1, wherein the decoration configuration information includes at least one of the following: a text description, an expression, a background color change, an animation effect played at a preset position of the display carrier, Describe the deformation animation effect of the carrier. 如申請專利範圍第1項所述的會話資訊處理方法,所述展示載體包括:   音頻資訊條、音頻資訊視窗、音頻符號中的至少一種。The presentation information processing method of claim 1, wherein the display carrier comprises: at least one of an audio information bar, an audio information window, and an audio symbol. 一種會話資訊處理方法,所述方法包括:   獲取待處理語音資訊,識別所述待處理語音資訊的資訊內容,所述資訊內容至少包括待處理語音資訊中包含的關鍵詞資訊;   獲取與所述資訊內容相匹配的裝飾配置資訊;   利用所述裝飾配置資訊對所述待處理語音資訊的展示載體進行渲染;   展示所述渲染後的展示載體。A session information processing method, the method includes: acquiring a to-be-processed voice information, and identifying the information content of the to-be-processed voice information, where the information content includes at least keyword information included in the to-be-processed voice information; The decoration configuration information matching the content; rendering the display carrier of the to-be-processed voice information by using the decoration configuration information; and displaying the rendered display carrier. 如申請專利範圍第8項所述的會話資訊處理方法,所述識別所述待處理語音資訊的資訊內容包括:   將所述待處理語音資訊轉換為文字文本;   將所述文字文本與預設詞庫中的關鍵詞進行匹配,提取所包含的關鍵詞資訊。The session information processing method of claim 8, wherein the identifying the information content of the to-be-processed voice information comprises: converting the to-be-processed voice information into text text; and using the text text and the preset word The keywords in the library are matched to extract the keyword information contained. 如申請專利範圍第8項所述的會話資訊處理方法,所述識別所述待處理語音資訊的資訊內容包括:   將所述待處理語音資訊轉換為文字文本,根據使用的文本語法規則提取所述文字文本中的關鍵詞。The method for processing the session information according to claim 8 , wherein the identifying the information content of the to-be-processed voice information comprises: converting the to-be-processed voice information into text text, and extracting the text according to a used text grammar rule Keywords in text text. 如申請專利範圍第8至10項中任意一項所述的會話資訊處理方法,所述獲取與識別出的所述資訊內容相匹配的裝飾配置資訊包括:   將所述待處理語音資訊中關鍵詞資訊的識別結果進行合併,生成請求參數;   獲取基於所述請求參數匹配出的裝飾配置資訊。The method for processing the session information according to any one of the preceding claims, wherein the obtaining the decoration configuration information that matches the identified information content comprises: the keyword of the to-be-processed voice information The identification results of the information are combined to generate a request parameter; and the decoration configuration information matched based on the request parameter is obtained. 如申請專利範圍第11項所述的會話資訊處理方法,所述獲取基於所述請求參數匹配出的裝飾配置資訊包括:   將所述請求參數在本地預設的裝飾配置資料庫中進行匹配,獲取裝飾配置資訊;   或者,   向遠端服務端發送所述請求參數,獲取所述遠端服務端根據所述請求參數匹配得到的裝飾配置資訊。The method for processing the session information according to claim 11, wherein the obtaining the decoration configuration information that is matched based on the request parameter comprises: matching the request parameter in a locally preset decoration configuration database to obtain Or decorating the configuration information; or sending the request parameter to the remote server to obtain the decoration configuration information obtained by the remote server according to the request parameter matching. 如申請專利範圍第8項所述的會話資訊處理方法,所述裝飾配置資訊包括下述中的至少一種:   文字描述、表情、變更背景顏色、在所述展示載體預設位置播放動畫效果、所述展示載體的變形動畫效果。The session information processing method of claim 8, wherein the decoration configuration information comprises at least one of the following: a text description, an expression, a background color change, an animation effect played at a preset position of the display carrier, Describe the deformation animation effect of the carrier. 如申請專利範圍第8項所述的會話資訊處理方法,所述展示載體包括:   音頻資訊條、音頻資訊視窗、音頻符號中的至少一種。The presentation information processing method of claim 8, wherein the display carrier comprises: at least one of an audio information bar, an audio information window, and an audio symbol. 一種會話資訊處理方法,所述方法包括:   獲取待處理語音資訊,識別所述待處理語音資訊的資訊內容,所述資訊內容至少包括待處理語音資訊中包含的關鍵詞資訊、待處理語音資訊表達的情緒資訊中的一種;   獲取與所述資訊內容相匹配的裝飾配置資訊;   利用所述裝飾配置資訊對所述待處理語音資訊的展示載體進行渲染;   展示所述渲染後的展示載體。A session information processing method, the method includes: acquiring a to-be-processed voice information, and identifying the information content of the to-be-processed voice information, where the information content includes at least keyword information included in the to-be-processed voice information, and the voice information to be processed And obtaining a decoration configuration information that matches the information content; and using the decoration configuration information to render the display carrier of the to-be-processed voice information; and displaying the rendered display carrier. 如申請專利範圍第15項所述的會話資訊處理方法,當識別包括待處理語音資訊中包含的關鍵詞資訊、待處理語音資訊表達的情緒資訊的資訊內容時,所述方法還包括:   將所述待處理語音資訊中關鍵詞資訊、情緒資訊的識別結果進行合併,生成請求參數;   相應的,所述獲取與所述資訊內容相匹配的裝飾配置資訊包括獲取基於所述請求參數匹配出的裝飾配置資訊。For example, in the session information processing method described in claim 15, when the information content including the keyword information included in the to-be-processed voice information and the emotion information expressed by the voice information to be processed is identified, the method further includes: The recognition result of the keyword information and the emotion information in the processed voice information is merged to generate a request parameter; correspondingly, the obtaining the decoration configuration information matching the information content includes acquiring the decoration matched based on the request parameter Configuration information. 一種會話資訊處理方法,所述方法包括:   獲取待處理會話資訊,識別所述待處理會話資訊的資訊內容,所述資訊內容至少包括待處理會話資訊中包含的關鍵詞資訊、待處理會話資訊表達的情緒資訊中的一種;   獲取與識別出的所述資訊內容相匹配的裝飾配置資訊;   利用所述裝飾配置資訊對所述待處理會話資訊的展示載體進行渲染;   展示所述渲染後的展示載體。A session information processing method, the method includes: acquiring session information to be processed, and identifying information content of the session information to be processed, where the information content includes at least keyword information included in the session information to be processed, and information about the session information to be processed. And obtaining a decoration configuration information that matches the identified information content; rendering the display carrier of the to-be-processed session information by using the decoration configuration information; and displaying the rendered display carrier . 一種會話資訊處理裝置,所述裝置包括:   情緒識別模組,用於獲取待處理語音資訊,識別所述待處理語音資訊的資訊內容,所述資訊內容至少包括所述待處理語音資訊表達的情緒資訊;   裝飾匹配模組,用於獲取與所述資訊內容相匹配的裝飾配置資訊;   渲染模組,用於利用所述裝飾配置資訊對所述待處理語音資訊的展示載體進行渲染;   效果展示模組,用於展示所述渲染後的展示載體。A session information processing device, the device includes: an emotion recognition module, configured to acquire voice information to be processed, and identify information content of the to-be-processed voice information, where the information content includes at least an emotion expressed by the to-be-processed voice information a decoration matching module, configured to obtain decoration configuration information that matches the information content; a rendering module, configured to render the display carrier of the to-be-processed voice information by using the decoration configuration information; a group for displaying the rendered display carrier. 如申請專利範圍第18項所述的會話資訊處理裝置,所述情緒識別模組包括:   情緒識別模型單元,用於提取所述待處理語音資訊的音頻特徵資料,將所述音頻特徵資料輸入預先構建的情緒識別模型,以及通過所述情緒識別模型確定所述待處理語音資訊所屬的情緒類別。The conversation information processing device of claim 18, wherein the emotion recognition module comprises: an emotion recognition model unit, configured to extract audio feature data of the to-be-processed voice information, and input the audio feature data into advance a constructed emotion recognition model, and determining, by the emotion recognition model, an emotion category to which the to-be-processed voice information belongs. 如申請專利範圍第19項所述的會話資訊處理裝置,提取的所述音頻特徵資料包括所述待處理語音資訊的生物的情感特徵資料、語氣特徵資料中的至少一種。The conversation information processing device of claim 19, wherein the extracted audio feature data comprises at least one of a sentiment feature data of the to-be-processed voice information and a tone feature data. 如申請專利範圍第18至20項中任意一項所述的會話資訊處理裝置,所述裝置還包括:   特徵合併單元,用於將所述待處理語音資訊中情緒資訊的識別結果進行合併,生成請求參數;   相應的,所述裝飾匹配模組基於所述請求參數匹配裝飾配置資訊。The session information processing device according to any one of the preceding claims, wherein the device further comprises: a feature merging unit, configured to merge the recognition results of the emotion information in the to-be-processed voice information to generate Correspondingly, the decoration matching module matches the decoration configuration information based on the request parameter. 如申請專利範圍第18至20項中任意一項所述的會話資訊處理裝置,所述裝飾配置資訊包括下述中的至少一種:   文字描述、表情、變更背景顏色、在所述展示載體預設位置播放動畫效果、所述展示載體的變形動畫效果。The session information processing apparatus according to any one of claims 18 to 20, wherein the decoration configuration information includes at least one of the following: a text description, an expression, a background color change, and a preset on the display carrier. The position plays an animation effect, and the deformation animation effect of the display carrier. 如申請專利範圍第18至20項中任意一項所述的會話資訊處理裝置,所述展示載體包括:   音頻資訊條、音頻資訊視窗、音頻符號中的至少一種。The presentation information processing apparatus according to any one of claims 18 to 20, wherein the display carrier comprises: at least one of an audio information strip, an audio information window, and an audio symbol. 一種會話資訊處理裝置,所述裝置包括:   關鍵詞識別模組,用於獲取待處理語音資訊,識別所述待處理語音資訊的資訊內容,所述資訊內容至少包括所述待處理語音資訊包含的關鍵詞資訊;   裝飾匹配模組,用於獲取與所述資訊內容相匹配的裝飾配置資訊;   渲染模組,用於利用所述裝飾配置資訊對所述待處理語音資訊的展示載體進行渲染;   效果展示模組,用於展示所述渲染後的展示載體。A session information processing device, the device includes: a keyword recognition module, configured to acquire voice information to be processed, and identify information content of the to-be-processed voice information, where the information content includes at least the voice information to be processed The keyword matching information is used to obtain the decoration configuration information that matches the information content; the rendering module is configured to use the decoration configuration information to render the display carrier of the to-be-processed voice information; a display module for displaying the rendered display carrier. 如申請專利範圍第24項所述的會話資訊處理裝置,所述關鍵詞識別模組包括:   文本轉換單元,用於將所述待處理語音資訊轉換為文字文本;   語法匹配單元,用於將所述待處理語音資訊轉換為文字文本,根據使用的文本語法規則提取所述文字文本中的關鍵詞。The session information processing device of claim 24, wherein the keyword recognition module comprises: a text conversion unit, configured to convert the to-be-processed voice information into text text; and a syntax matching unit for The processed speech information is converted into text text, and the keywords in the text text are extracted according to the used text grammar rules. 如申請專利範圍第24項所述的會話資訊處理裝置,所述關鍵詞識別模組包括:   文本轉換單元,用於將所述待處理語音資訊轉換為文字文本;   詞庫匹配單元,用於將所述文字文本與預設詞庫中的關鍵詞進行匹配,提取所包含的關鍵詞資訊。The session information processing device of claim 24, wherein the keyword recognition module comprises: a text conversion unit, configured to convert the to-be-processed voice information into text text; a thesaurus matching unit, The text text is matched with the keywords in the preset vocabulary, and the included keyword information is extracted. 如申請專利範圍第24至26項中任意一項所述的會話資訊處理裝置,所述裝置還包括:   特徵合併單元,用於將所述待處理語音資訊中情緒資訊的識別結果進行合併,生成請求參數;   相應的,所述裝飾匹配模組基於所述請求參數匹配裝飾配置資訊。The session information processing apparatus according to any one of claims 24 to 26, further comprising: a feature merging unit, configured to merge the recognition results of the emotion information in the to-be-processed voice information to generate Correspondingly, the decoration matching module matches the decoration configuration information based on the request parameter. 如申請專利範圍第24至26項中任意一項所述的會話資訊處理裝置,所述裝飾配置資訊包括下述中的至少一種:   文字描述、表情、變更背景顏色、在所述展示載體預設位置播放動畫效果、所述展示載體的變形動畫效果。The session information processing apparatus according to any one of claims 24 to 26, wherein the decoration configuration information includes at least one of the following: a text description, an expression, a background color change, and a preset on the display carrier. The position plays an animation effect, and the deformation animation effect of the display carrier. 如申請專利範圍第24至26項中任意一項所述的會話資訊處理裝置,所述展示載體包括:   音頻資訊條、音頻資訊視窗、音頻符號中的至少一種。The presentation information processing apparatus according to any one of claims 24 to 26, wherein the display carrier comprises: at least one of an audio information strip, an audio information window, and an audio symbol. 一種會話資訊處理裝置,所述裝置包括:   資訊內容識別模組,用於獲取待處理語音資訊,識別所述待處理語音資訊的資訊內容,所述資訊內容至少包括待處理語音資訊中包含的關鍵詞資訊、待處理語音資訊表達的情緒資訊中的一種;   裝飾匹配模組,用於獲取與所述資訊內容相匹配的裝飾配置資訊;   渲染模組,用於利用所述裝飾配置資訊對所述待處理語音資訊的展示載體進行渲染;   效果展示模組,用於展示所述渲染後的展示載體。A session information processing device, the device includes: an information content identification module, configured to acquire voice information to be processed, and identify information content of the to-be-processed voice information, where the information content includes at least a key included in the voice information to be processed a word information, one of emotional information expressed by the voice information to be processed; a decoration matching module, configured to obtain decoration configuration information matching the information content; and a rendering module, configured to use the decoration configuration information The display carrier of the to-be-processed voice information is rendered; and the effect display module is configured to display the rendered display carrier. 如申請專利範圍第30項所述的會話資訊處理裝置,所述裝置還包括:   特徵合併單元,用於將所述待處理語音資訊中關鍵詞資訊和情緒資訊的識別結果進行合併,生成請求參數;   相應的,所述裝飾匹配模組基於所述請求參數匹配裝飾配置資訊。The session information processing device of claim 30, the device further comprising: a feature merging unit, configured to combine the keyword information and the emotion information in the to-be-processed voice information to generate a request parameter Correspondingly, the decoration matching module matches the decoration configuration information based on the request parameter. 一種會話資訊處理裝置,所述裝置包括:   資訊內容識別模組,用於獲取待處理會話資訊,識別所述待處理會話資訊的資訊內容,所述資訊內容至少包括待處理會話資訊中包含的關鍵詞資訊、待處理會話資訊表達的情緒資訊中的一種;   裝飾匹配模組,用於獲取與識別出的所述資訊內容相匹配的裝飾配置資訊;   渲染模組,用於利用所述裝飾配置資訊對所述待處理會話資訊的展示載體進行渲染;   效果展示模組,用於展示所述渲染後的展示載體。A session information processing device, the device includes: an information content identification module, configured to acquire session information to be processed, and identify information content of the session information to be processed, where the information content includes at least a key included in the session information to be processed a word information, one of emotional information expressed by the session information to be processed; a decoration matching module, configured to acquire decoration configuration information that matches the identified information content; and a rendering module configured to use the decoration configuration information The display carrier of the to-be-processed session information is rendered; and the effect display module is configured to display the rendered display carrier. 一種會話資訊處理裝置,包括處理器以及用於儲存處理器可執行指令的記憶體,所述處理器執行所述指令時實現:   獲取待處理會話資訊,識別所述待處理會話資訊的資訊內容,所述資訊內容至少包括待處理會話資訊中包含的關鍵詞資訊、待處理會話資訊表達的情緒資訊中的一種;   獲取與識別出的所述資訊內容相匹配的裝飾配置資訊;   利用所述裝飾配置資訊對所述待處理會話資訊的展示載體進行渲染;   展示所述渲染後的展示載體。A session information processing device includes a processor and a memory for storing processor-executable instructions. When the processor executes the instructions, the processor: obtains pending session information, and identifies information content of the to-be-processed session information. The information content includes at least one of keyword information included in the session information to be processed and emotion information expressed by the session information to be processed; acquiring decoration configuration information matching the identified information content; using the decoration configuration The information is used to render the display carrier of the to-be-processed session information; and the rendered display carrier is displayed. 如申請專利範圍第33項所述的會話資訊處理裝置,所述待處理會話資訊包括語音資訊、文本資訊、圖像資訊、視頻資訊中的至少一種。The session information processing device of claim 33, wherein the to-be-processed session information includes at least one of voice information, text information, image information, and video information. 一種電子設備,包括至少一個處理器以及使用者儲存處理器可執行指令的記憶體,所述處理器執行所述指令時實現:   獲取待處理會話資訊,識別所述待處理會話資訊的資訊內容,所述資訊內容至少包括待處理會話資訊中包含的關鍵詞資訊、待處理會話資訊表達的情緒資訊中的一種;   獲取與識別出的所述資訊內容相匹配的裝飾配置資訊;   利用所述裝飾配置資訊對所述待處理會話資訊的展示載體進行渲染;   展示所述渲染後的展示載體。An electronic device includes at least one processor and a memory for a user to store processor-executable instructions, and when the processor executes the instructions, the method implements: acquiring pending session information, and identifying information content of the to-be-processed session information, The information content includes at least one of keyword information included in the session information to be processed and emotion information expressed by the session information to be processed; acquiring decoration configuration information matching the identified information content; using the decoration configuration The information is used to render the display carrier of the to-be-processed session information; and the rendered display carrier is displayed.
TW107116262A 2017-07-10 2018-05-14 Session information processing method, device, electronic device TWI671739B (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN201710555403.6A CN107516533A (en) 2017-07-10 2017-07-10 A kind of session information processing method, device, electronic equipment
??201710555403.6 2017-07-10

Publications (2)

Publication Number Publication Date
TW201909171A true TW201909171A (en) 2019-03-01
TWI671739B TWI671739B (en) 2019-09-11

Family

ID=60722330

Family Applications (1)

Application Number Title Priority Date Filing Date
TW107116262A TWI671739B (en) 2017-07-10 2018-05-14 Session information processing method, device, electronic device

Country Status (3)

Country Link
CN (1) CN107516533A (en)
TW (1) TWI671739B (en)
WO (1) WO2019011185A1 (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
TWI684964B (en) * 2019-07-09 2020-02-11 香港商智園香港有限公司 Knowledge point mark generation system and method thereof

Families Citing this family (25)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107516533A (en) * 2017-07-10 2017-12-26 阿里巴巴集团控股有限公司 A kind of session information processing method, device, electronic equipment
CN110312161B (en) * 2018-03-20 2020-12-11 Tcl科技集团股份有限公司 Video dubbing method and device and terminal equipment
CN110798393B (en) * 2018-08-02 2021-10-26 腾讯科技(深圳)有限公司 Voiceprint bubble display method and terminal using voiceprint bubbles
CN109448728A (en) * 2018-10-29 2019-03-08 苏州工业职业技术学院 Merge the multi-party conversation method for visualizing and system of emotion recognition
CN109521932A (en) * 2018-11-06 2019-03-26 斑马网络技术有限公司 Voice control display processing method, device, vehicle, storage medium and equipment
CN109525725B (en) * 2018-11-21 2021-01-15 三星电子(中国)研发中心 Information processing method and device based on emotional state
CN109547332B (en) * 2018-11-22 2022-05-13 腾讯科技(深圳)有限公司 Communication session interaction method and device, and computer equipment
CN109768913A (en) * 2018-12-11 2019-05-17 平安科技(深圳)有限公司 Information processing method, device, computer equipment and storage medium
CN109597493B (en) * 2018-12-11 2022-05-17 科大讯飞股份有限公司 Expression recommendation method and device
CN109831572A (en) * 2018-12-14 2019-05-31 深圳壹账通智能科技有限公司 Chat picture control method, device, computer equipment and storage medium
CN109448735B (en) 2018-12-21 2022-05-20 深圳创维-Rgb电子有限公司 Method and device for adjusting video parameters based on voiceprint recognition and read storage medium
CN111914115A (en) * 2019-05-08 2020-11-10 阿里巴巴集团控股有限公司 Sound information processing method and device and electronic equipment
CN110187862A (en) * 2019-05-29 2019-08-30 北京达佳互联信息技术有限公司 Speech message display methods, device, terminal and storage medium
CN112037821A (en) * 2019-06-03 2020-12-04 阿里巴巴集团控股有限公司 Visual representation method and device of voice emotion and computer storage medium
CN110417641B (en) * 2019-07-23 2022-05-17 上海盛付通电子支付服务有限公司 Method and equipment for sending session message
CN110311858B (en) * 2019-07-23 2022-06-07 上海盛付通电子支付服务有限公司 Method and equipment for sending session message
CN112910753A (en) * 2019-12-04 2021-06-04 上海掌门科技有限公司 Voice message display method and device
CN111372029A (en) * 2020-04-17 2020-07-03 维沃移动通信有限公司 Video display method and device and electronic equipment
CN111865766B (en) * 2020-07-20 2024-02-02 博泰车联网科技(上海)股份有限公司 Interactive method, medium, equipment and system based on audio-video transmission
CN112667196A (en) * 2021-01-28 2021-04-16 百度在线网络技术(北京)有限公司 Information display method and device, electronic equipment and medium
CN112883181A (en) * 2021-02-26 2021-06-01 腾讯科技(深圳)有限公司 Session message processing method and device, electronic equipment and storage medium
CN112860213B (en) * 2021-03-09 2023-08-25 腾讯科技(深圳)有限公司 Audio processing method and device, storage medium and electronic equipment
CN113407092A (en) * 2021-06-17 2021-09-17 北京达佳互联信息技术有限公司 Text box information processing method and device, electronic equipment and storage medium
CN113595868B (en) * 2021-06-28 2023-05-16 深圳云之家网络有限公司 Voice message processing method and device based on instant messaging and computer equipment
TWI779916B (en) * 2021-11-02 2022-10-01 宏碁股份有限公司 Method and system for remote communication

Family Cites Families (21)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2002041279A (en) * 2000-07-21 2002-02-08 Megafusion Corp Agent message system
EP1326445B1 (en) * 2001-12-20 2008-01-23 Matsushita Electric Industrial Co., Ltd. Virtual television phone apparatus
JP4423327B2 (en) * 2005-02-08 2010-03-03 日本電信電話株式会社 Information communication terminal, information communication system, information communication method, information communication program, and recording medium recording the same
CN100538823C (en) * 2006-07-13 2009-09-09 英业达股份有限公司 Language aided expression system and method
CN101702316B (en) * 2009-11-20 2014-04-09 北京中星微电子有限公司 Method for converting MIDI music into color information and system thereof
TWI395201B (en) * 2010-05-10 2013-05-01 Univ Nat Cheng Kung Method and system for identifying emotional voices
CN102664007B (en) * 2012-03-27 2016-08-31 上海量明科技发展有限公司 For generating the method for character identification content, client and system
CN103543979A (en) * 2012-07-17 2014-01-29 联想(北京)有限公司 Voice outputting method, voice interaction method and electronic device
US20140163980A1 (en) * 2012-12-10 2014-06-12 Rawllin International Inc. Multimedia message having portions of media content with audio overlay
CN103634472B (en) * 2013-12-06 2016-11-23 惠州Tcl移动通信有限公司 User mood and the method for personality, system and mobile phone is judged according to call voice
CN103886869B (en) * 2014-04-09 2016-09-21 北京京东尚科信息技术有限公司 A kind of information feedback method based on speech emotion recognition and system
US20150356836A1 (en) * 2014-06-05 2015-12-10 Microsoft Corporation Conversation cues within audio conversations
CN104022942B (en) * 2014-06-26 2018-09-11 北京奇虎科技有限公司 Handle method, client, electronic equipment and the system of interactive message
US10068588B2 (en) * 2014-07-21 2018-09-04 Microsoft Technology Licensing, Llc Real-time emotion recognition from audio signals
US9786299B2 (en) * 2014-12-04 2017-10-10 Microsoft Technology Licensing, Llc Emotion type classification for interactive dialog system
KR101634086B1 (en) * 2015-01-19 2016-07-08 주식회사 엔씨소프트 Method and computer system of analyzing communication situation based on emotion information
CN105049637A (en) * 2015-08-25 2015-11-11 努比亚技术有限公司 Device and method for controlling instant communication
CN106899486B (en) * 2016-06-22 2020-09-25 阿里巴巴集团控股有限公司 Message display method and device
CN106531149B (en) * 2016-12-07 2018-02-23 腾讯科技(深圳)有限公司 Information processing method and device
CN106888158B (en) * 2017-02-28 2020-07-03 天翼爱动漫文化传媒有限公司 Instant messaging method and device
CN107516533A (en) * 2017-07-10 2017-12-26 阿里巴巴集团控股有限公司 A kind of session information processing method, device, electronic equipment

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
TWI684964B (en) * 2019-07-09 2020-02-11 香港商智園香港有限公司 Knowledge point mark generation system and method thereof

Also Published As

Publication number Publication date
TWI671739B (en) 2019-09-11
CN107516533A (en) 2017-12-26
WO2019011185A1 (en) 2019-01-17

Similar Documents

Publication Publication Date Title
TWI671739B (en) Session information processing method, device, electronic device
US20220230374A1 (en) User interface for generating expressive content
WO2022048403A1 (en) Virtual role-based multimodal interaction method, apparatus and system, storage medium, and terminal
US9767789B2 (en) Using emoticons for contextual text-to-speech expressivity
CN110830852B (en) Video content processing method and device
JP2017527926A (en) Generation of computer response to social conversation input
WO2022170848A1 (en) Human-computer interaction method, apparatus and system, electronic device and computer medium
CN110555126A (en) Automatic generation of melodies
KR101628050B1 (en) Animation system for reproducing text base data by animation
CN110599359B (en) Social contact method, device, system, terminal equipment and storage medium
CN111145777A (en) Virtual image display method and device, electronic equipment and storage medium
CN109801349B (en) Sound-driven three-dimensional animation character real-time expression generation method and system
JP6339529B2 (en) Conference support system and conference support method
KR20200056261A (en) Electronic apparatus and method for controlling thereof
Chang et al. The prompt artists
EP3155612A1 (en) Advanced recurrent neural network based letter-to-sound
WO2022242706A1 (en) Multimodal based reactive response generation
CN110148406B (en) Data processing method and device for data processing
Fang et al. Facial expression GAN for voice-driven face generation
Wang et al. Comic-guided speech synthesis
TW202022851A (en) Voice interaction method and device
CN116189279A (en) Method, device and storage medium for determining hand motion of virtual person
CN111914115A (en) Sound information processing method and device and electronic equipment
CN112331209B (en) Method and device for converting voice into text, electronic equipment and readable storage medium
CN109241331B (en) Intelligent robot-oriented story data processing method