CN116055756A - Data processing method, device and server of live broadcasting room - Google Patents

Data processing method, device and server of live broadcasting room Download PDF

Info

Publication number
CN116055756A
CN116055756A CN202211706177.4A CN202211706177A CN116055756A CN 116055756 A CN116055756 A CN 116055756A CN 202211706177 A CN202211706177 A CN 202211706177A CN 116055756 A CN116055756 A CN 116055756A
Authority
CN
China
Prior art keywords
data
target
user terminal
interaction
target user
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202211706177.4A
Other languages
Chinese (zh)
Inventor
熊威
王思梦
王博
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
China Construction Bank Corp
CCB Finetech Co Ltd
Original Assignee
China Construction Bank Corp
CCB Finetech Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by China Construction Bank Corp, CCB Finetech Co Ltd filed Critical China Construction Bank Corp
Priority to CN202211706177.4A priority Critical patent/CN116055756A/en
Publication of CN116055756A publication Critical patent/CN116055756A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/21Server components or server architectures
    • H04N21/218Source of audio or video content, e.g. local disk arrays
    • H04N21/2187Live feed
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/26Speech to text systems
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/23Processing of content or additional data; Elementary server operations; Server middleware
    • H04N21/235Processing of additional data, e.g. scrambling of additional data or processing content descriptors
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/25Management operations performed by the server for facilitating the content distribution or administrating data related to end-users or client devices, e.g. end-user or client device authentication, learning user preferences for recommending movies
    • H04N21/266Channel or content management, e.g. generation and management of keys and entitlement messages in a conditional access system, merging a VOD unicast channel into a multicast channel
    • H04N21/26606Channel or content management, e.g. generation and management of keys and entitlement messages in a conditional access system, merging a VOD unicast channel into a multicast channel for generating or managing entitlement messages, e.g. Entitlement Control Message [ECM] or Entitlement Management Message [EMM]
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/25Management operations performed by the server for facilitating the content distribution or administrating data related to end-users or client devices, e.g. end-user or client device authentication, learning user preferences for recommending movies
    • H04N21/266Channel or content management, e.g. generation and management of keys and entitlement messages in a conditional access system, merging a VOD unicast channel into a multicast channel
    • H04N21/26613Channel or content management, e.g. generation and management of keys and entitlement messages in a conditional access system, merging a VOD unicast channel into a multicast channel for generating or managing keys in general
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/47End-user applications
    • H04N21/478Supplemental services, e.g. displaying phone caller identification, shopping application
    • H04N21/4788Supplemental services, e.g. displaying phone caller identification, shopping application communicating with other users, e.g. chatting
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02DCLIMATE CHANGE MITIGATION TECHNOLOGIES IN INFORMATION AND COMMUNICATION TECHNOLOGIES [ICT], I.E. INFORMATION AND COMMUNICATION TECHNOLOGIES AIMING AT THE REDUCTION OF THEIR OWN ENERGY USE
    • Y02D30/00Reducing energy consumption in communication networks
    • Y02D30/70Reducing energy consumption in communication networks in wireless communication networks

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Databases & Information Systems (AREA)
  • Health & Medical Sciences (AREA)
  • Computational Linguistics (AREA)
  • General Engineering & Computer Science (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Human Computer Interaction (AREA)
  • Physics & Mathematics (AREA)
  • Acoustics & Sound (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Computer Security & Cryptography (AREA)
  • Information Transfer Between Computers (AREA)

Abstract

The application provides a data processing method, device and server of a live broadcast room, and is applied to the technical field of cloud computing. Based on the method, after receiving an interaction request initiated by a first user terminal, the cloud server can determine a target language matched with a target user and a target user terminal held by the target user object according to an object identification of the target user object; according to the data type of the interactive data, determining a matched target processing rule from a preset processing rule set; processing the interaction data according to the target processing rule based on the target language to obtain processed interaction data; and displaying the processed interactive data in the current live video displayed to the target user object through a live broadcasting room interface of the target user terminal. Therefore, the audience user can efficiently and conveniently interact with target user objects using different languages in a diversified communication manner in the live broadcasting room, and the interaction experience of the audience user in the live broadcasting room is improved.

Description

Data processing method, device and server of live broadcasting room
Technical Field
The application belongs to the technical field of cloud computing, and particularly relates to a data processing method, device and server of a live broadcasting room.
Background
With the rise and development of the video live broadcast industry, more and more activities such as cross-regional exhibition start to be carried out in the form of online video live broadcast.
The activities such as cross-regional exhibition are held in the form of online video live broadcast, and on one hand, a lot of convenience can be provided for users. On the other hand, the method is limited by the existing data processing method of the live broadcasting room, and because the used languages of the users in different areas are different, communication interaction between the different users in the same live broadcasting room cannot be performed well; in addition, the existing live broadcasting room cannot provide a private and safe interaction channel for users to conduct interaction related to key information, so that interaction experience of the users is relatively poor.
In view of the above problems, no effective solution has been proposed at present.
Disclosure of Invention
The data processing method, the data processing device and the server for the live broadcasting room can enable audience users to efficiently and conveniently interact with target user objects using different languages in a diversified communication mode in the live broadcasting room, and improve interaction experience of the audience users.
The application provides a data processing method of a live broadcasting room, which is applied to a cloud server and comprises the following steps:
Receiving an interaction request initiated by a first user terminal; the interaction request at least carries interaction data and object identifiers of target user objects aimed at by the interaction data;
according to the object identification of the target user object, determining a target language matched with the target user and a target user terminal held by the target user object;
according to the data type of the interactive data, determining a matched target processing rule from a preset processing rule set;
processing the interaction data according to a target processing rule based on a target language to obtain processed interaction data;
and displaying the processed interactive data in the current live video displayed to the target user object through a live broadcasting room interface of the target user terminal.
In one embodiment, the data type of the interaction data includes at least one of: text data, voice data, emoticons.
In one embodiment, when the data type of the interactive data includes voice data, processing the interactive data according to a target processing rule based on a target language to obtain processed interactive data includes:
Performing voice recognition on the voice data by using a voice recognition model to obtain corresponding text data; extracting voice characteristics of a first audience user from the voice data;
processing the text data by using a translation model matched with the target language to obtain text data based on the target language;
and processing the text data based on the target language by utilizing a preset voice synthesis model according to the voice characteristics of the first audience user to obtain corresponding synthesized voice data serving as the processed interaction data.
In one embodiment, when the data type of the interactive data includes an expression image, processing the interactive data according to a target processing rule based on a target language to obtain processed interactive data includes:
performing character detection on the expression image to determine whether meaningful text characters exist in the expression image;
under the condition that meaningful text characters exist in the expression image, the image recognition model is utilized to recognize the expression image, so that the text characters in the expression image are extracted as text data;
processing the text data by using a translation model matched with the target language to obtain text data based on the target language;
Constructing annotation data for the expression image according to the text data based on the target language;
and combining the expression image and the annotation data to obtain the expression image carrying the annotation data, and taking the expression image carrying the annotation data as the processed interaction data.
In one embodiment, in the case that the data type of the interactive data includes text data, processing the interactive data according to a target processing rule based on a target language to obtain processed interactive data includes:
processing the text data by using a translation model matched with the target language to obtain text data based on the target language;
and constructing corresponding barrage data according to the text data of the target language, and taking the barrage data as the processed interactive data.
In one embodiment, the method further comprises:
receiving a private message request initiated by a first user terminal; wherein, the private letter request at least carries an object identifier of a target user object;
responding to the private message request, and establishing a private data communication channel between the first user terminal and the target user terminal based on a live broadcasting interface according to a related encryption communication protocol; and the cloud server displays the processed interactive data in the current live video displayed to the target user object through a live broadcasting room interface of the target user terminal based on the privacy data communication channel.
In one embodiment, after establishing the private data communication channel based on the live room interface between the first user terminal and the target user terminal, the method further comprises:
encrypting the processed interaction data by utilizing public key data corresponding to the target user terminal to obtain ciphertext data of the interaction data;
sending ciphertext data of the interactive data to a target user terminal through a privacy data communication channel; the target user terminal decrypts ciphertext data of the interaction data by using the private key data to obtain the processed interaction data; and displaying the processed interactive data in the current live video displayed to the target user object through a live broadcasting room interface.
In one embodiment, before receiving the interaction request initiated by the first user terminal, the method further comprises:
receiving a connection request initiated by a first user terminal;
responding to the connection request, and establishing a first data connection with a first user terminal;
transmitting live video stream data to a first user terminal through the first data connection; and the first user terminal displays the current live video to the first user through a live broadcast room interface according to the live video stream data.
In one embodiment, while transmitting live video streaming data to the first user terminal over the first data connection, the method further comprises:
collecting characteristic parameters of a first user terminal;
determining a first language matched with a first audience user according to the characteristic parameters of the first user terminal;
detecting whether live video stream data currently transmitted through a first data connection are live video stream data based on a first language;
under the condition that live video stream data transmitted through the first data connection at present is not based on live video stream data of a first language, determining a cloud CDN cached with the live video stream data of the first language as a first target CDN;
and switching the first data connection to a first target CDN so as to transmit the live video streaming data based on the first language to a first user terminal through the first data connection.
In one embodiment, the characteristic parameter of the first user terminal comprises at least one of: the IP address of the first user terminal, default language parameters of a browser of the first user terminal, cookie data of the first user terminal.
The application also provides a data processing method of the live broadcasting room, which is applied to the first user terminal and comprises the following steps:
Displaying a current live video based on a first language to a first audience user through a live room interface;
receiving interaction data input by a first user through a live broadcasting room interface; determining a target user object aimed by the interactive data;
generating a corresponding interaction request according to the interaction data; the interaction request also carries an object identifier of the target user object;
the interaction request is sent to a cloud server; the cloud server determines a target language matched with the target user object according to the object identification of the target user object; processing the interaction data based on the target language to obtain processed interaction data; and the cloud server also displays the processed interaction data in the current live video displayed to the target user object through a live broadcasting room interface of the target user terminal.
The application also provides a data processing device of a live broadcast room, which is applied to a cloud server and comprises:
the receiving module is used for receiving an interaction request initiated by the first user terminal; the interaction request at least carries interaction data and object identifiers of target user objects aimed at by the interaction data;
The first determining module is used for determining a target language matched with the target user and a target user terminal held by the target user object according to the object identification of the target user object;
the second determining module is used for determining a matched target processing rule from a preset processing rule set according to the data type of the interactive data;
the processing module is used for processing the interaction data according to target processing rules based on target language to obtain processed interaction data;
and the touch module is used for displaying the processed interaction data in the current live video displayed to the target user object through a live broadcasting room interface of the target user terminal.
The application also provides a server comprising a processor and a memory for storing processor executable instructions, which when executed by the processor implement the relevant steps of the data processing method of the live room.
The present application also provides a computer readable storage medium having stored thereon computer instructions which when executed by a processor perform the steps of: receiving an interaction request initiated by a first user terminal; the interaction request at least carries interaction data and object identifiers of target user objects aimed at by the interaction data; according to the object identification of the target user object, determining a target language matched with the target user and a target user terminal held by the target user object; according to the data type of the interactive data, determining a matched target processing rule from a preset processing rule set; processing the interaction data according to a target processing rule based on a target language to obtain processed interaction data; and displaying the processed interactive data in the current live video displayed to the target user object through a live broadcasting room interface of the target user terminal.
The present application also provides a computer program product comprising a computer program which, when executed by a processor, implements the relevant steps of the data processing method of the live room.
Based on the data processing method, the device and the server of the live broadcasting room, after the cloud server of the cloud live broadcasting service platform receives the interaction request initiated by the first user terminal, the target language matched with the target user and the target user terminal held by the target user object can be determined according to the object identification of the target user object; according to the data type of the interactive data, determining a matched target processing rule from a preset processing rule set; processing the interaction data according to the target processing rule based on the target language to obtain processed interaction data; and displaying the processed interactive data in the current live video displayed to the target user object through a live broadcasting room interface of the target user terminal. Therefore, the audience user can efficiently and conveniently interact with target user objects using different languages in a diversified communication manner in the live broadcasting room, and the interaction experience of the audience user in the live broadcasting room is improved. Furthermore, the cloud server can also establish an exclusive privacy data communication channel with the target user object in the living broadcast room for the audience user according to the specific requirements of the audience user, and the audience user can further directly perform safer and more private communication interaction with the target user object in the living broadcast room interface based on the privacy data communication channel, so that the privacy data in the communication interaction process of the audience user is prevented from being revealed.
Drawings
For a clearer description of the embodiments of the present application, the drawings that are needed in the embodiments will be briefly described, and the drawings described below are only some embodiments described in the present application, and other drawings may be obtained according to these drawings without inventive effort for a person of ordinary skill in the art.
Fig. 1 is a flow chart of a data processing method of a live broadcast room provided by an embodiment of the application;
FIG. 2 is a schematic diagram of an embodiment of a data processing method of a live room, in which the embodiments of the present application are applied, in one example of a scenario;
FIG. 3 is a schematic diagram of an embodiment of a data processing method of a live room, in which the embodiments of the present application are applied, in one example of a scenario;
FIG. 4 is a schematic diagram of an embodiment of a data processing method of a live room, in which the embodiments of the present application are applied, in one example of a scenario;
FIG. 5 is a schematic diagram of an embodiment of a data processing method of a live room, in which the embodiments of the present application are applied, in one example of a scenario;
fig. 6 is a flow chart of a data processing method of a live room according to another embodiment of the present application;
FIG. 7 is a schematic diagram of the structural composition of a server according to one embodiment of the present application;
FIG. 8 is a schematic diagram of the structural components of a data processing apparatus of a live room according to one embodiment of the present application;
fig. 9 is a schematic structural diagram of a data processing apparatus of a live room according to another embodiment of the present application.
Detailed Description
In order to better understand the technical solutions in the present application, the following description will clearly and completely describe the technical solutions in the embodiments of the present application with reference to the drawings in the embodiments of the present application, and it is obvious that the described embodiments are only some embodiments of the present application, not all embodiments. All other embodiments, which can be made by one of ordinary skill in the art based on the embodiments herein without making any inventive effort, shall fall within the scope of the present application.
It should be noted that, in the present specification, information data related to a user is acquired and used on the premise that the user knows and agrees. And, the information data is acquired, stored, used, processed and the like, which meet the relevant regulations of national laws and regulations.
Referring to fig. 1, an embodiment of the present application provides a data processing method in a live broadcast room, where the method is specifically applied to a cloud server side. In particular implementations, the method may include the following:
s101: receiving an interaction request initiated by a first user terminal; the interaction request at least carries interaction data and object identifiers of target user objects aimed at by the interaction data;
s102: according to the object identification of the target user object, determining a target language matched with the target user and a target user terminal held by the target user object;
s103: according to the data type of the interactive data, determining a matched target processing rule from a preset processing rule set;
s104: processing the interaction data according to a target processing rule based on a target language to obtain processed interaction data;
s105: and displaying the processed interactive data in the current live video displayed to the target user object through a live broadcasting room interface of the target user terminal.
The first user terminal may be a user terminal held by a first audience user. The first viewer user may be specifically understood as any viewer user currently online in the living room.
The target user object may specifically include a user object that is in the same living room as the first viewer user, and the first viewer user object currently wants to interact separately. In particular, the target user object may include a anchor user and/or other viewer users within the same living room as the first viewer user. The target user object may include one or more user objects.
The live video can be a trans-regional exhibition live video, a trans-regional academic conference live video, a trans-regional commodity communication live video and the like. Of course, the live video listed above is only one illustrative illustration. In specific implementation, the live video may further include live video of other suitable types and content according to specific application scenarios and processing requirements. The present specification is not limited to this.
Based on the above embodiment, the cloud server may process the interaction data of the first audience user for the target user object in the same living broadcast room into the processed interaction data based on the target language matched with the target user object, and then display the processed interaction data in the current living broadcast video displayed to the target user object through the living broadcast room interface of the target user terminal, so that the audience user can efficiently and conveniently perform diversified and language barrier-free interaction communication with the target user object such as the anchor user and/or other audience users designated in the same living broadcast room in the living broadcast room, and the interaction experience of the audience user in the living broadcast room is effectively improved.
In some embodiments, referring to fig. 2, the method for processing data in the live broadcast room may be specifically applied to a cloud server side.
The cloud server specifically comprises a background server which is applied to one side of a cloud live broadcast service platform and can realize functions of data transmission, data processing and the like. Specifically, the cloud server may be, for example, an electronic device having a data operation function, a storage function, and a network interaction function. Alternatively, the cloud server may be a software program running in the electronic device that provides support for data processing, storage, and network interactions. In the present embodiment, the number of servers included in the cloud server is not particularly limited. The cloud server may be one server, several servers, or a server cluster formed by several servers.
Specifically, the cloud server may be connected to the live broadcast terminal and the plurality of user terminals in a wired or wireless manner.
The live broadcast terminal and the user terminal can specifically comprise front ends which are applied to the side of a host user and a viewer user and can realize functions of data acquisition, data transmission and the like. Specifically, the live broadcast terminal and the user terminal may be electronic devices such as a desktop computer, a tablet computer, a notebook computer, and a smart phone, for example. Alternatively, the live terminal and the user terminal may be software applications capable of running in the electronic device. For example, it may be some live APP running on a smart phone, etc.
Further, the cloud server may be further configured with a plurality of algorithm models, for example, a speech recognition model, a translation model, and an image recognition model; in addition, the cloud server may be further connected with a database, a translation terminal, and a plurality of cloud CDNs, for example, cloud CDN1, cloud CDN2, cloud CDNn, and the like.
The cloud CDN (Content Delivery Network) may specifically refer to a cloud-based content distribution network. The translation terminal can be a manual translation terminal or an automatic translation terminal based on artificial intelligence.
In specific implementation, referring to fig. 2, a host user may use a live terminal to perform live video broadcast, such as exhibition, product propaganda, and the like. The live broadcast terminal can acquire live broadcast video stream data of the anchor user in real time and upload the live broadcast video stream data to the cloud server.
After receiving the live video stream data, the cloud server can directly forward the live video stream data to the user terminal on one hand, so that a viewer user can watch live video based on the language used by the anchor user through a live broadcasting room interface of the user terminal; on the other hand, the live video stream data can be sent to a translation terminal, so that the live video stream data can be translated in real time through the translation terminal, and then the translated live video stream data based on different languages are cached in the corresponding cloud CDNs, so that audience users in different areas using different languages can relatively synchronously pull and watch live videos based on the languages used by the audience users in the same live broadcasting room through the user terminal.
Specifically, taking any one of the first audience users in the living broadcast room as an example. The current first audience user views the current live video based on the first language through a live broadcast room interface displayed by the held first user terminal. The first language may be understood as a language matching the first audience user. For example, the native language of the first viewer user, etc.
The first audience user can watch the current live video through a live broadcast room interface displayed by the first user terminal and can select a main broadcasting user and/or other audience users in the live broadcast room as target user objects; and the target user object is singly aimed at to carry out communication interaction in the live broadcasting room interface.
Specifically, a list of user objects may also be presented in the live room interface. The user object list may specifically include object identifiers of user objects such as anchor users and online audience users in the living broadcast room. The first viewer user may initiate a click operation in the list of user objects specifying a target user object to be specifically interacted with by selecting an object identification of one or more user objects.
After selecting the target user object, the live room interface may pop up the interactive data entry box. Accordingly, the first viewer user may input specific interactive data through the interactive data input box.
The first user terminal can receive the interaction data, determine a target user object aimed at by the interaction data and acquire an object identifier of the target user object; and then a corresponding interactive request can be generated. The interaction request at least can carry interaction data and object identification of the target user object.
Then, the first user terminal may send the interaction request to the cloud server. Correspondingly, the cloud server receives and acquires the first interaction request.
In some embodiments, the data type of the interaction data may specifically include at least one of the following: text data, voice data, emoticons, etc. Of course, it should be noted that the above-listed data types of the interactive data are only illustrative. In specific implementation, the interactive data may further include other suitable types of interactive data according to specific situations and processing requirements. The present specification is not limited to this.
Based on the embodiment, the audience user can freely use the interaction data of one or more different data types to interact with the target user object of the living broadcast room according to specific conditions, so that the diversified interaction requirements of the audience user can be met.
In some embodiments, in implementation, the cloud server may determine, as the target user terminal, a user terminal held by the target user object by querying a user database according to the direction identifier of the target user object.
Furthermore, the cloud server can acquire the characteristic parameters of the target user terminal according to the data connection established by the target user terminal for acquiring the live video data, and determine the language matched with the target user as the target language according to the characteristic parameters of the target user terminal.
In addition, the cloud server can also query the database according to the identification information of the target user object, and determine the target language matched with the target user object according to the user data about the target user object recorded in the database.
In some embodiments, before implementation, corresponding preset processing rules may be configured in advance for each of the interactive data of different data types. Each preset processing rule may include a corresponding algorithm rule and an algorithm model. Further, a corresponding preset processing rule set may be obtained according to combining the plurality of preset processing rules. And the preset processing rule set also stores a matching relation between the preset processing rule and the data type of the interactive data.
In some embodiments, in the implementation, a matched preset processing rule may be determined from a preset processing rule set according to a data type of the interactive data, and the matched preset processing rule is used as the target processing rule.
In some embodiments, referring to fig. 3, in the case where the data type of the interactive data includes voice data, the processing the interactive data according to the target processing rule based on the target language may include the following steps when implementing:
s1: performing voice recognition on the voice data by using a voice recognition model to obtain corresponding text data; extracting voice characteristics of a first audience user from the voice data;
s2: processing the text data by using a translation model matched with the target language to obtain text data based on the target language;
s3: and processing the text data based on the target language by utilizing a preset voice synthesis model according to the voice characteristics of the first audience user to obtain corresponding synthesized voice data serving as the processed interaction data.
Based on the embodiment, the voice data input by the first audience user can be efficiently converted into the voice data based on the target language, and the voice data accords with the voice characteristics of the first audience user when speaking, so that the synthesized voice data of emotion information such as mood and attitude of the first audience user when speaking can be truly and comprehensively reflected, and then the synthesized voice data is used as the processed interaction data to reach the target user object, so that the target user object can understand related semantic content in the interaction data conveniently and efficiently based on the target language used by the target user object, and meanwhile, the real emotion information of the first audience user can be intuitively felt through the synthesized voice data, and relatively good interaction effect can be obtained.
Wherein the speech features include at least one of: tone, loudness, frequency, timbre, etc.
In the implementation, voice recognition is carried out on voice data by utilizing a voice recognition model, so that corresponding text data is obtained; the speech data may also be processed using a speech feature extraction model to extract speech features of the first audience user when the speech data is input. Determining a translation model matched with the target language from a plurality of translation models; and processing the text data by using the translation model matched with the target language to obtain translated text data based on the target language. Further, a preset voice synthesis model can be utilized, voice synthesis can be performed based on voice characteristics of a target user by using text data based on a target language, so that the processed interactive data containing semantic content based on the target language and simultaneously capable of conveying real emotion information of a first audience user can be obtained.
In some embodiments, referring to fig. 4, in the case where the data type of the interactive data includes an expression image, the processing the interactive data according to the target processing rule based on the target language may include the following steps when implementing:
S1: performing character detection on the expression image to determine whether meaningful text characters exist in the expression image;
s2: under the condition that meaningful text characters exist in the expression image, the image recognition model is utilized to recognize the expression image, so that the text characters in the expression image are extracted as text data;
s3: processing the text data by using a translation model matched with the target language to obtain text data based on the target language;
s4: constructing annotation data for the expression image according to the text data based on the target language;
s5: and combining the expression image and the annotation data to obtain the expression image carrying the annotation data, and taking the expression image carrying the annotation data as the processed interaction data.
Based on the embodiment, the expression image input by the first audience user can be efficiently converted into the processed interactive data carrying the labeling data based on the target language, so that the target user object can conveniently and efficiently understand the true meaning expressed by the expression image sent by the first audience user.
In particular, the emoticons may be left unprocessed in the event that it is determined that there are no meaningful text characters in the emoticons. And directly determining the expression image as the processed interactive data.
When determining whether a significant text character exists in the expression image, the text character detection can be performed on the expression image, and when determining that the text character exists in the expression image, the text character detection can be performed; and judging whether meaningful text characters exist in the text characters in the expression image according to the nonsensical character reference template matched with the first language.
When the expression image and the annotation data are specifically combined, an annotation frame containing the annotation data can be generated; and then splicing the annotation frame and the expression image, so that the expression image carrying the annotation data can be obtained.
In some embodiments, when the data type of the interactive data includes text data, the processing the interactive data according to the target processing rule based on the target language to obtain the processed interactive data may further include the following when implemented:
s1: processing the text data by using a translation model matched with the target language to obtain text data based on the target language;
s2: and constructing corresponding barrage data according to the text data of the target language, and taking the barrage data as the processed interactive data.
Based on the embodiment, the text data input by the first audience user can be efficiently and accurately converted into the text data based on the target language, so that the target user object can conveniently and rapidly understand the semantic content which is expected to be expressed by the interactive data sent by the first audience user.
In some embodiments, it is contemplated that the timeliness requirements are often high in the actual interaction process, but the precision requirements for semantic content are not very stringent. For example, it is common for different user objects to know the approximate meaning of each other in the communication interaction. Furthermore, in consideration of the fact that a plurality of communication interactions possibly exist at the same time in the same live broadcast room, if the translation terminal with higher calling precision performs related translation processing on text data, on one hand, the data processing capacity of the translation terminal is overlarge, and even the normal processing of live broadcast video stream data is affected; on the other hand, the timeliness of the alternating current interaction cannot be well met.
Based on the above consideration, in the present specification, text data is processed separately mainly by calling a translation model independent of a relatively simplified version of a translation terminal, so that user objects have higher precision compared with key information of interest during translation processing of the text data, timeliness of interaction is effectively ensured on the premise that basic interaction between the user objects is not affected, data processing burden is reduced, and meanwhile, influence on processing of live video stream data is avoided.
In specific implementation, the above processing text data by using the translation model matched with the target language may further include the following: determining the receiving time of the interaction request; determining theme information in live video when a first audience user initiates an interaction request through a first user terminal according to the initiation time; and then, using a translation model matched with the target language, and processing the text data based on the subject information as a reference, so that the text data is translated more efficiently on the premise of ensuring that the key information related to the subject information has higher precision, and the text data based on the target language, which meets the basic communication interaction requirement, is obtained rapidly.
In some embodiments, referring to fig. 5, when the method is implemented, the following may further be included:
s1: receiving a private message request initiated by a first user terminal; wherein, the private letter request at least carries an object identifier of a target user object;
s2: responding to the private message request, and establishing a private data communication channel between the first user terminal and the target user terminal based on a live broadcasting interface according to a related encryption communication protocol; and the cloud server displays the processed interactive data in the current live video displayed to the target user object through a live broadcasting room interface of the target user terminal based on the privacy data communication channel. Wherein the processed interactive data is perceivable only by the first viewer user and the target user object.
The privacy data communication channel is different from a common live broadcasting room barrage public screen. In particular, a privacy dialog box visible only to the first viewer user and the target user object may additionally be presented in the live room interface of the two user terminals establishing the privacy data communication channel. The privacy dialog box is only used for displaying and inputting interactive data which are mutually communicated between two user terminals with privacy data communication channels.
It should be noted that, based on the above-mentioned privacy data communication channel, the processed interactive data displayed in the current live video displayed to the target user object through the live broadcast room interface of the target user terminal is only perceptible to the first viewer by the target user object, but is not perceptible to other user objects in the live broadcast room.
Based on the above embodiment, in some live broadcast scenarios such as a business exhibition, the first audience user can also independently establish an independent and private data communication channel with the target user object in the live broadcast room, so that based on the private data communication channel, under the condition that other user objects cannot perceive, and meanwhile, the first audience user and the target user object are not influenced to watch the current live broadcast video, communication interaction with higher security is performed, and private data involved in the communication process is prevented from being revealed.
In some embodiments, after establishing the private data communication channel based on the live-room interface between the first user terminal and the target user terminal, the method may further include, when embodied, the following:
s1: encrypting the processed interaction data by utilizing public key data corresponding to the target user terminal to obtain ciphertext data of the interaction data;
s2: sending ciphertext data of the interactive data to a target user terminal through a privacy data communication channel; the target user terminal decrypts ciphertext data of the interaction data by using the private key data to obtain the processed interaction data; and displaying the processed interactive data in the current live video displayed to the target user object through a live broadcasting room interface.
In the implementation, after establishing a private data communication channel based on a live broadcasting interface between a first user terminal and a target user terminal, the cloud server can interact with the target user terminal according to a related encryption communication protocol, and generates public key data and private key data corresponding to the target user terminal by utilizing a terminal identifier of the target user terminal and a random number generator; and, the public key data is kept by the cloud server, and the private key data is kept by the target user terminal. Similarly, the cloud server may interact with the first user terminal according to a related encrypted communication protocol, and generate public key data and private key data corresponding to the first user terminal by using a terminal identifier of the first user terminal and a random number generator; and, the public key data is kept by the cloud server, and the private key data is kept by the first user terminal.
Based on the embodiment, the privacy data related to the interaction between the first user terminal and the target user terminal through the privacy data communication channel can be more effectively prevented from being revealed, and the data security during the interaction is better protected.
In some embodiments, before receiving the interaction request initiated by the first user terminal, the method may further include the following when implemented:
s1: receiving a connection request initiated by a first user terminal;
s2: responding to the connection request, and establishing a first data connection with a first user terminal;
s3: transmitting live video stream data to a first user terminal through the first data connection; and the first user terminal displays the current live video to the first user through a live broadcast room interface according to the live video stream data.
Based on the embodiment, the cloud server can quickly respond to the connection request initiated by the first user terminal, and the first user terminal can efficiently and stably acquire and display the real-time live video by establishing and using the first data connection.
In some embodiments, while transmitting live video streaming data to the first user terminal through the first data connection, the method may further include the following when implemented:
S1: collecting characteristic parameters of a first user terminal;
s2: determining a first language matched with a first audience user according to the characteristic parameters of the first user terminal;
s3: detecting whether live video stream data currently transmitted through a first data connection are live video stream data based on a first language;
s4: under the condition that live video stream data transmitted through the first data connection at present is not based on live video stream data of a first language, determining a cloud CDN cached with the live video stream data of the first language as a first target CDN;
s5: and switching the first data connection to a first target CDN so as to transmit the live video streaming data based on the first language to a first user terminal through the first data connection.
Based on the embodiment, the cloud server can automatically detect and identify the first language matched with the first audience user, and further can automatically switch to the live video matched with the first audience user based on the first language in time under the condition that the live video currently provided for the first audience user is actively judged to be not matched with the first user, so that the first audience user can obtain better interaction experience in a live broadcast room.
In some embodiments, the feature parameter package of the first user terminal may specifically include at least one of the following: the IP address of the first user terminal, default language parameters of the browser of the first user terminal, cookie data of the first user terminal, etc.
Based on the above embodiment, the cloud server may accurately and automatically determine the language type matched with the first user by collecting and according to the feature parameters of the first user terminal.
Specifically, the cloud server may collect, through the first data connection, a characteristic parameter of the first user terminal.
In some embodiments, the method may further include the following when implemented:
s1: acquiring live video stream data acquired in real time;
s2: calling a corresponding translation terminal to convert the live video stream data into video stream data based on a plurality of different languages; and video stream data based on a plurality of different languages are respectively cached in the corresponding cloud CDNs.
In some embodiments, after transmitting the live video streaming data to the first user terminal through the first data connection, the method may further include the following when implemented:
s1: receiving a language switching request initiated by a first user terminal;
S2: determining a second language customized by the first audience user according to the language switching request;
s3: determining a cloud CDN cached with live video stream data based on a second language as a second target CDN;
s4: and switching the first data connection to a second target CDN so as to transmit the live video stream data based on the second language to the first user terminal through the first data connection.
From the above, according to the data processing method of the live broadcasting room provided by the embodiment of the present application, after receiving the interaction request initiated by the first user terminal, the cloud server may determine, according to the object identifier of the target user object, the target language matched with the target user and the target user terminal held by the target user object; according to the data type of the interactive data, determining a matched target processing rule from a preset processing rule set; processing the interaction data according to the target processing rule based on the target language to obtain processed interaction data; and displaying the processed interactive data in the current live video displayed to the target user object through a live broadcasting room interface of the target user terminal. Therefore, the audience users can efficiently and conveniently interact with target user objects using different languages in a diversified manner in the live broadcasting room, the interaction experience of the audience users in the live broadcasting room is improved, and the diversified interaction requirements of the audience users are met. Furthermore, the cloud server can also establish an exclusive privacy data communication channel with the target user object in the living broadcast room for the audience user according to the needs of the audience user, and the audience user can further directly perform safer and more private communication interaction with the target user object in the living broadcast room interface based on the privacy data communication channel, so that the privacy data related to the communication interaction of the audience user is prevented from being revealed.
Referring to fig. 6, another method for processing data in a live broadcast room is further provided in the embodiment of the present application, and is applied to a first user terminal. The method can be implemented by the following steps:
s601: displaying a current live video based on a first language to a first audience user through a live room interface;
s602: receiving interaction data input by a first user through a live broadcasting room interface; determining a target user object aimed by the interactive data;
s603: generating a corresponding interaction request according to the interaction data; the interaction request also carries an object identifier of the target user object;
s604: the interaction request is sent to a cloud server; the cloud server determines a target language matched with the target user object according to the object identification of the target user object; processing the interaction data based on the target language to obtain processed interaction data; and the cloud server also displays the processed interaction data in the current live video displayed to the target user object through a live broadcasting room interface of the target user terminal.
Therefore, the data processing method of the live broadcasting room can enable audience users to efficiently and conveniently interact with target user objects using different languages in a diversified manner in the live broadcasting room, and improves interaction experience of the audience users. Furthermore, a private data communication channel exclusive to the target user object can be established in the living broadcast room for the audience user according to the needs of the audience user, and the audience user can perform safer and private communication interaction with the target user object in the living broadcast room interface directly based on the private data communication channel, so that the private data of the audience user is prevented from being revealed.
The embodiment of the application also provides a server, which comprises a processor and a memory for storing executable instructions of the processor, wherein the processor can execute the following steps according to the instructions when being implemented: receiving an interaction request initiated by a first user terminal; the interaction request at least carries interaction data and object identifiers of target user objects aimed at by the interaction data; according to the object identification of the target user object, determining a target language matched with the target user and a target user terminal held by the target user object; according to the data type of the interactive data, determining a matched target processing rule from a preset processing rule set; processing the interaction data according to a target processing rule based on a target language to obtain processed interaction data; and displaying the processed interactive data in the current live video displayed to the target user object through a live broadcasting room interface of the target user terminal.
In order to more accurately complete the above instructions, referring to fig. 7, another specific server is further provided in this embodiment of the present application, where the server includes a network communication port 701, a processor 702, and a memory 703, and the above structures are connected by an internal cable, so that each structure may perform specific data interaction.
The network communication port 701 may be specifically configured to receive an interaction request initiated by the first user terminal; the interaction request at least carries interaction data and object identification of a target user object aimed at by the interaction data.
The processor 702 may be specifically configured to determine, according to an object identifier of a target user object, a target language matched with the target user and a target user terminal held by the target user object; according to the data type of the interactive data, determining a matched target processing rule from a preset processing rule set; processing the interaction data according to a target processing rule based on a target language to obtain processed interaction data; and displaying the processed interactive data in the current live video displayed to the target user object through a live broadcasting room interface of the target user terminal.
The memory 703 may be used for storing a corresponding program of instructions.
In this embodiment, the network communication port 701 may be a virtual port that binds with different communication protocols, so that different data may be sent or received. For example, the network communication port may be a port responsible for performing web data communication, a port responsible for performing FTP data communication, or a port responsible for performing mail data communication. The network communication port may also be an entity's communication interface or a communication chip. For example, it may be a wireless mobile network communication chip, such as GSM, CDMA, etc.; it may also be a Wifi chip; it may also be a bluetooth chip.
In this embodiment, the processor 702 may be implemented in any suitable manner. For example, the processor may take the form of, for example, a microprocessor or processor, and a computer-readable medium storing computer-readable program code (e.g., software or firmware) executable by the (micro) processor, logic gates, switches, an application specific integrated circuit (Application Specific Integrated Circuit, ASIC), a programmable logic controller, and an embedded microcontroller, among others. The description is not intended to be limiting.
In this embodiment, the memory 703 may include a plurality of layers, and in a digital system, the memory may be any memory as long as it can hold binary data; in an integrated circuit, a circuit with a memory function without a physical form is also called a memory, such as a RAM, a FIFO, etc.; in the system, the storage device in physical form is also called a memory, such as a memory bank, a TF card, and the like.
The embodiment of the application also provides a user terminal, which comprises a processor and a memory for storing executable instructions of the processor, wherein the processor can execute the following steps according to the instructions when being implemented: displaying a current live video based on a first language to a first audience user through a live room interface; receiving interaction data input by a first user through a live broadcasting room interface; determining a target user object aimed by the interactive data; generating a corresponding interaction request according to the interaction data; the interaction request also carries an object identifier of the target user object; the interaction request is sent to a cloud server; the cloud server determines a target language matched with the target user object according to the object identification of the target user object; processing the interaction data based on the target language to obtain processed interaction data; and the cloud server also displays the processed interaction data in the current live video displayed to the target user object through a live broadcasting room interface of the target user terminal.
The embodiment of the application also provides a computer readable storage medium based on the data processing method of the live broadcast room, wherein the computer readable storage medium stores computer program instructions, and the following steps are realized when the computer program instructions are executed: receiving an interaction request initiated by a first user terminal; the interaction request at least carries interaction data and object identifiers of target user objects aimed at by the interaction data; according to the object identification of the target user object, determining a target language matched with the target user and a target user terminal held by the target user object; according to the data type of the interactive data, determining a matched target processing rule from a preset processing rule set; processing the interaction data according to a target processing rule based on a target language to obtain processed interaction data; and displaying the processed interactive data in the current live video displayed to the target user object through a live broadcasting room interface of the target user terminal.
In the present embodiment, the storage medium includes, but is not limited to, a random access Memory (Random Access Memory, RAM), a Read-Only Memory (ROM), a Cache (Cache), a Hard Disk (HDD), or a Memory Card (Memory Card). The memory may be used to store computer program instructions. The network communication unit may be an interface for performing network connection communication, which is set in accordance with a standard prescribed by a communication protocol.
In this embodiment, the functions and effects of the program instructions stored in the computer readable storage medium may be explained in comparison with other embodiments, and are not described herein.
The embodiment of the application also provides a computer program product, which comprises a computer program, wherein the computer program realizes the following steps when being executed by a processor: receiving an interaction request initiated by a first user terminal; the interaction request at least carries interaction data and object identifiers of target user objects aimed at by the interaction data; according to the object identification of the target user object, determining a target language matched with the target user and a target user terminal held by the target user object; according to the data type of the interactive data, determining a matched target processing rule from a preset processing rule set; processing the interaction data according to a target processing rule based on a target language to obtain processed interaction data; and displaying the processed interactive data in the current live video displayed to the target user object through a live broadcasting room interface of the target user terminal.
Referring to fig. 8, on a software level, the embodiment of the present application further provides a data processing device of a live broadcast room, which is applied to a cloud server side, and the device may specifically include the following structural modules:
The receiving module 801 may be specifically configured to receive an interaction request initiated by the first user terminal; the interaction request at least carries interaction data and object identifiers of target user objects aimed at by the interaction data;
the first determining module 802 may be specifically configured to determine, according to an object identifier of a target user object, a target language matched with the target user and a target user terminal held by the target user object;
the second determining module 803 may be specifically configured to determine, according to a data type of the interaction data, a matched target processing rule from a preset processing rule set;
the processing module 804 may be specifically configured to process the interaction data according to a target processing rule based on a target language, to obtain processed interaction data;
the touch module 805 may be specifically configured to display the processed interaction data in a current live video displayed to the target user object through a live room interface of the target user terminal.
In some embodiments, the data type of the interaction data may specifically include at least one of the following: text data, voice data, emoticons, etc.
In some embodiments, the processing module 804 may be specifically configured to process, when the data type of the interactive data includes voice data, the interactive data according to a target processing rule based on a target language, so as to obtain processed interactive data according to the following manner: performing voice recognition on the voice data by using a voice recognition model to obtain corresponding text data; extracting voice characteristics of a first audience user from the voice data; processing the text data by using a translation model matched with the target language to obtain text data based on the target language; and processing the text data based on the target language by utilizing a preset voice synthesis model according to the voice characteristics of the first audience user to obtain corresponding synthesized voice data serving as the processed interaction data.
In some embodiments, the processing module 804 may be specifically configured to process the interactive data according to the target processing rule based on the target language when the data type of the interactive data includes an expression image, so as to obtain the processed interactive data: performing character detection on the expression image to determine whether meaningful text characters exist in the expression image; under the condition that meaningful text characters exist in the expression image, the image recognition model is utilized to recognize the expression image, so that the text characters in the expression image are extracted as text data; processing the text data by using a translation model matched with the target language to obtain text data based on the target language; constructing annotation data for the expression image according to the text data based on the target language; and combining the expression image and the annotation data to obtain the expression image carrying the annotation data, and taking the expression image carrying the annotation data as the processed interaction data.
In some embodiments, the processing module 804 may be specifically configured to process, when the data type of the interactive data includes text data, the interactive data according to a target processing rule based on a target language, so as to obtain processed interactive data according to the following manner: processing the text data by using a translation model matched with the target language to obtain text data based on the target language; and constructing corresponding barrage data according to the text data of the target language, and taking the barrage data as the processed interactive data.
In some embodiments, the apparatus may be further configured to receive a private letter request initiated by the first user terminal when the apparatus is embodied; wherein, the private letter request at least carries an object identifier of a target user object; responding to the private message request, and establishing a private data communication channel between the first user terminal and the target user terminal based on a live broadcasting interface according to a related encryption communication protocol; and the cloud server displays the processed interactive data in the current live video displayed to the target user object through a live broadcasting room interface of the target user terminal based on the privacy data communication channel.
In some embodiments, after establishing a private data communication channel based on a live broadcasting interface between the first user terminal and the target user terminal, the device may be further configured to encrypt the processed interaction data by using public key data corresponding to the target user terminal to obtain ciphertext data of the interaction data when the device is implemented; sending ciphertext data of the interactive data to a target user terminal through a privacy data communication channel; the target user terminal decrypts ciphertext data of the interaction data by using the private key data to obtain the processed interaction data; and displaying the processed interactive data in the current live video displayed to the target user object through a live broadcasting room interface.
In some embodiments, before receiving the interaction request initiated by the first user terminal, the apparatus may be further configured to receive a connection request initiated by the first user terminal when the apparatus is implemented; responding to the connection request, and establishing a first data connection with a first user terminal; transmitting live video stream data to a first user terminal through the first data connection; and the first user terminal displays the current live video to the first user through a live broadcast room interface according to the live video stream data.
In some embodiments, the method may further include, while transmitting live video streaming data to the first user terminal via the first data connection, collecting a characteristic parameter of the first user terminal when the apparatus is implemented; determining a first language matched with a first audience user according to the characteristic parameters of the first user terminal; detecting whether live video stream data currently transmitted through a first data connection are live video stream data based on a first language; under the condition that live video stream data transmitted through the first data connection at present is not based on live video stream data of a first language, determining a cloud CDN cached with the live video stream data of the first language as a first target CDN; and switching the first data connection to a first target CDN so as to transmit the live video streaming data based on the first language to a first user terminal through the first data connection.
In some embodiments, the characteristic parameter of the first user terminal may specifically include at least one of the following: the IP address of the first user terminal, default language parameters of the browser of the first user terminal, cookie data of the first user terminal, etc.
Referring to fig. 9, the embodiment of the present application further provides another apparatus for processing data in a live broadcast room, which is applied to a side of a first user terminal, and may specifically include the following structural modules:
the display module 901 may be specifically configured to display a current live video based on a first language to a first viewer user through a live room interface;
the receiving module 902 may be specifically configured to receive interaction data input by the first user through the live broadcast room interface; determining a target user object aimed by the interactive data;
the generating module 903 may be specifically configured to generate a corresponding interaction request according to the interaction data; the interaction request also carries an object identifier of the target user object;
the sending module 904 may be specifically configured to send the interaction request to a cloud server; the cloud server determines a target language matched with the target user object according to the object identification of the target user object; processing the interaction data based on the target language to obtain processed interaction data; and the cloud server also displays the processed interaction data in the current live video displayed to the target user object through a live broadcasting room interface of the target user terminal.
It should be noted that, the units, devices, or modules described in the above embodiments may be implemented by a computer chip or entity, or may be implemented by a product having a certain function. For convenience of description, the above devices are described as being functionally divided into various modules, respectively. Of course, when the present description is implemented, the functions of each module may be implemented in the same piece or pieces of software and/or hardware, or a module that implements the same function may be implemented by a plurality of sub-modules or a combination of sub-units, or the like. The above-described apparatus embodiments are merely illustrative, for example, the division of the units is merely a logical function division, and there may be additional divisions when actually implemented, for example, multiple units or components may be combined or integrated into another system, or some features may be omitted or not performed. Alternatively, the coupling or direct coupling or communication connection shown or discussed with each other may be an indirect coupling or communication connection via some interfaces, devices or units, which may be in electrical, mechanical or other form.
From the above, based on the data processing device of the live broadcasting room provided by the embodiment of the application, the audience user can efficiently and conveniently interact with the target user objects using different languages in a diversified communication manner in the live broadcasting room, and the interaction experience of the audience user is improved. Furthermore, a private data communication channel exclusive to the target user object can be established in the living broadcast room for the audience user according to the needs of the audience user, and the audience user can perform safer and private communication interaction with the target user object in the living broadcast room interface directly based on the private data communication channel, so that the private data of the audience user is prevented from being revealed.
Although the present description provides method operational steps as described in the examples or flowcharts, more or fewer operational steps may be included based on conventional or non-inventive means. The order of steps recited in the embodiments is merely one way of performing the order of steps and does not represent a unique order of execution. When implemented by an apparatus or client product in practice, the methods illustrated in the embodiments or figures may be performed sequentially or in parallel (e.g., in a parallel processor or multi-threaded processing environment, or even in a distributed data processing environment). The terms "comprises," "comprising," or any other variation thereof, are intended to cover a non-exclusive inclusion, such that a process, method, article, or apparatus that comprises a list of elements does not include only those elements but may include other elements not expressly listed or inherent to such process, method, article, or apparatus. Without further limitation, it is not excluded that additional identical or equivalent elements may be present in a process, method, article, or apparatus that comprises a described element. The terms first, second, etc. are used to denote a name, but not any particular order.
Those skilled in the art will also appreciate that, in addition to implementing the controller in a pure computer readable program code, it is well possible to implement the same functionality by logically programming the method steps such that the controller is in the form of logic gates, switches, application specific integrated circuits, programmable logic controllers, embedded microcontrollers, etc. Such a controller can be regarded as a hardware component, and means for implementing various functions included therein can also be regarded as a structure within the hardware component. Or even means for achieving the various functions may be regarded as either software modules implementing the methods or structures within hardware components.
The description may be described in the general context of computer-executable instructions, such as program modules, being executed by a computer. Generally, program modules include routines, programs, objects, components, data structures, classes, etc. that perform particular tasks or implement particular abstract data types. The specification may also be practiced in distributed computing environments where tasks are performed by remote processing devices that are linked through a communications network. In a distributed computing environment, program modules may be located in both local and remote computer-readable storage media including memory storage devices.
From the above description of embodiments, it will be apparent to those skilled in the art that the present description may be implemented in software plus a necessary general hardware platform. Based on such understanding, the technical solutions of the present specification may be embodied essentially in the form of a software product, which may be stored in a storage medium, such as a ROM/RAM, a magnetic disk, an optical disk, etc., and include several instructions to cause a computer device (which may be a personal computer, a mobile terminal, a server, or a network device, etc.) to perform the methods described in the various embodiments or portions of the embodiments of the present specification.
Various embodiments in this specification are described in a progressive manner, and identical or similar parts are all provided for each embodiment, each embodiment focusing on differences from other embodiments. The specification is operational with numerous general purpose or special purpose computer system environments or configurations. For example: personal computers, server computers, hand-held or portable devices, tablet devices, multiprocessor systems, microprocessor-based systems, set top boxes, programmable electronic devices, network PCs, minicomputers, mainframe computers, distributed computing environments that include any of the above systems or devices, and the like.
Although the present specification has been described by way of example, it will be appreciated by those skilled in the art that there are many variations and modifications to the specification without departing from the spirit of the specification, and it is intended that the appended claims encompass such variations and modifications as do not depart from the spirit of the specification.

Claims (15)

1. The data processing method of the live broadcasting room is characterized by being applied to a cloud server and comprising the following steps of:
receiving an interaction request initiated by a first user terminal; the interaction request at least carries interaction data and object identifiers of target user objects aimed at by the interaction data;
according to the object identification of the target user object, determining a target language matched with the target user and a target user terminal held by the target user object;
according to the data type of the interactive data, determining a matched target processing rule from a preset processing rule set;
processing the interaction data according to a target processing rule based on a target language to obtain processed interaction data;
and displaying the processed interactive data in the current live video displayed to the target user object through a live broadcasting room interface of the target user terminal.
2. The method of claim 1, wherein the data type of the interaction data comprises at least one of: text data, voice data, emoticons.
3. The method of claim 2, wherein, in the case where the data type of the interactive data includes voice data, processing the interactive data according to a target processing rule based on a target language, to obtain processed interactive data, comprises:
performing voice recognition on the voice data by using a voice recognition model to obtain corresponding text data; extracting voice characteristics of a first audience user from the voice data;
processing the text data by using a translation model matched with the target language to obtain text data based on the target language;
and processing the text data based on the target language by utilizing a preset voice synthesis model according to the voice characteristics of the first audience user to obtain corresponding synthesized voice data serving as the processed interaction data.
4. The method according to claim 2, wherein, in case the data type of the interactive data includes an expression image, processing the interactive data according to a target processing rule based on a target language, to obtain processed interactive data, comprises:
Performing character detection on the expression image to determine whether meaningful text characters exist in the expression image;
under the condition that meaningful text characters exist in the expression image, the image recognition model is utilized to recognize the expression image, so that the text characters in the expression image are extracted as text data;
processing the text data by using a translation model matched with the target language to obtain text data based on the target language;
constructing annotation data for the expression image according to the text data based on the target language;
and combining the expression image and the annotation data to obtain the expression image carrying the annotation data, and taking the expression image carrying the annotation data as the processed interaction data.
5. The method of claim 2, wherein, in the case where the data type of the interactive data includes text data, processing the interactive data according to a target processing rule based on a target language, to obtain processed interactive data, comprises:
processing the text data by using a translation model matched with the target language to obtain text data based on the target language;
and constructing corresponding barrage data according to the text data of the target language, and taking the barrage data as the processed interactive data.
6. The method according to claim 1, wherein the method further comprises:
receiving a private message request initiated by a first user terminal; wherein, the private letter request at least carries an object identifier of a target user object;
responding to the private message request, and establishing a private data communication channel between the first user terminal and the target user terminal based on a live broadcasting interface according to a related encryption communication protocol; and the cloud server displays the processed interactive data in the current live video displayed to the target user object through a live broadcasting room interface of the target user terminal based on the privacy data communication channel.
7. The method of claim 6, wherein after establishing the private data communication channel based on the live room interface between the first user terminal and the target user terminal, the method further comprises:
encrypting the processed interaction data by utilizing public key data corresponding to the target user terminal to obtain ciphertext data of the interaction data;
sending ciphertext data of the interactive data to a target user terminal through a privacy data communication channel; the target user terminal decrypts ciphertext data of the interaction data by using the private key data to obtain the processed interaction data; and displaying the processed interactive data in the current live video displayed to the target user object through a live broadcasting room interface.
8. The method of claim 1, wherein prior to receiving the first user terminal initiated request for interaction, the method further comprises:
receiving a connection request initiated by a first user terminal;
responding to the connection request, and establishing a first data connection with a first user terminal;
transmitting live video stream data to a first user terminal through the first data connection; and the first user terminal displays the current live video to the first user through a live broadcast room interface according to the live video stream data.
9. The method of claim 8, wherein the method further comprises, while transmitting live video streaming data to the first user terminal over the first data connection:
collecting characteristic parameters of a first user terminal;
determining a first language matched with a first audience user according to the characteristic parameters of the first user terminal;
detecting whether live video stream data currently transmitted through a first data connection are live video stream data based on a first language;
under the condition that live video stream data transmitted through the first data connection at present is not based on live video stream data of a first language, determining a cloud CDN cached with the live video stream data of the first language as a first target CDN;
And switching the first data connection to a first target CDN so as to transmit the live video streaming data based on the first language to a first user terminal through the first data connection.
10. The method of claim 9, wherein the characteristic parameters of the first user terminal include at least one of: the IP address of the first user terminal, default language parameters of a browser of the first user terminal, cookie data of the first user terminal.
11. A method for processing data in a live broadcast room, applied to a first user terminal, comprising:
displaying a current live video based on a first language to a first audience user through a live room interface;
receiving interaction data input by a first user through a live broadcasting room interface; determining a target user object aimed by the interactive data;
generating a corresponding interaction request according to the interaction data; the interaction request also carries an object identifier of the target user object;
the interaction request is sent to a cloud server; the cloud server determines a target language matched with the target user object according to the object identification of the target user object; processing the interaction data based on the target language to obtain processed interaction data; and the cloud server also displays the processed interaction data in the current live video displayed to the target user object through a live broadcasting room interface of the target user terminal.
12. A data processing device of a live broadcast room, which is applied to a cloud server, and comprises:
the receiving module is used for receiving an interaction request initiated by the first user terminal; the interaction request at least carries interaction data and object identifiers of target user objects aimed at by the interaction data;
the first determining module is used for determining a target language matched with the target user and a target user terminal held by the target user object according to the object identification of the target user object;
the second determining module is used for determining a matched target processing rule from a preset processing rule set according to the data type of the interactive data;
the processing module is used for processing the interaction data according to target processing rules based on target language to obtain processed interaction data;
and the touch module is used for displaying the processed interaction data in the current live video displayed to the target user object through a live broadcasting room interface of the target user terminal.
13. A server comprising a processor and a memory for storing processor-executable instructions, which when executed by the processor implement the steps of the method of any one of claims 1 to 10.
14. A computer readable storage medium having stored thereon computer instructions which, when executed by a processor, implement the steps of the method of any of claims 1 to 10, or 11.
15. A computer program product comprising a computer program which, when executed by a processor, implements the steps of the method of any one of claims 1 to 10, or 11.
CN202211706177.4A 2022-12-26 2022-12-26 Data processing method, device and server of live broadcasting room Pending CN116055756A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202211706177.4A CN116055756A (en) 2022-12-26 2022-12-26 Data processing method, device and server of live broadcasting room

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202211706177.4A CN116055756A (en) 2022-12-26 2022-12-26 Data processing method, device and server of live broadcasting room

Publications (1)

Publication Number Publication Date
CN116055756A true CN116055756A (en) 2023-05-02

Family

ID=86124786

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202211706177.4A Pending CN116055756A (en) 2022-12-26 2022-12-26 Data processing method, device and server of live broadcasting room

Country Status (1)

Country Link
CN (1) CN116055756A (en)

Similar Documents

Publication Publication Date Title
CN105721462B (en) Information pushing method and device
US9984408B1 (en) Method, medium, and system for live video cooperative shopping
JP6442441B2 (en) Translation system and translation method
US20160337290A1 (en) Message Push Method and Apparatus
WO2015043547A1 (en) A method, device and system for message response cross-reference to related applications
CN112929253B (en) Virtual image interaction method and device
CN113568699B (en) Content display method, device, equipment and storage medium
CN111625308B (en) Information display method and device and electronic equipment
WO2020078050A1 (en) Comment information processing method and apparatus, and server, terminal and readable medium
CN112073307A (en) Mail processing method and device, electronic equipment and computer readable medium
CN110647827A (en) Comment information processing method and device, electronic equipment and storage medium
CN112055252A (en) Multi-screen interaction method and device, computer readable medium and electronic equipment
CN104811485A (en) Resource sharing method
CN108574878B (en) Data interaction method and device
CN110109594B (en) Drawing data sharing method and device, storage medium and equipment
KR20220061926A (en) Method and apparatus for switching skin of mini-program page, and electronic device
CN104268611A (en) Webpage synchronizing method and device
US9332206B2 (en) Frame sharing
CN113139377A (en) Method, device, equipment and computer readable medium for pushing information
CN113038192A (en) Video processing method and device, electronic equipment and storage medium
CN104980807A (en) Method and terminal for multimedia interaction
CN111147894A (en) Sign language video generation method, device and system
CN109948082A (en) Live information processing method and processing device, electronic equipment, storage medium
CN116055756A (en) Data processing method, device and server of live broadcasting room
CN111857901A (en) Data processing method, method for generating session background, electronic device and medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination