CN112055252A - Multi-screen interaction method and device, computer readable medium and electronic equipment - Google Patents
Multi-screen interaction method and device, computer readable medium and electronic equipment Download PDFInfo
- Publication number
- CN112055252A CN112055252A CN202010831939.8A CN202010831939A CN112055252A CN 112055252 A CN112055252 A CN 112055252A CN 202010831939 A CN202010831939 A CN 202010831939A CN 112055252 A CN112055252 A CN 112055252A
- Authority
- CN
- China
- Prior art keywords
- interactive
- interaction
- streaming media
- media file
- display
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
- 230000003993 interaction Effects 0.000 title claims abstract description 371
- 238000000034 method Methods 0.000 title claims abstract description 71
- 230000002452 interceptive effect Effects 0.000 claims abstract description 395
- 230000004044 response Effects 0.000 claims description 61
- 238000004590 computer program Methods 0.000 claims description 12
- 238000012545 processing Methods 0.000 claims description 7
- 238000010586 diagram Methods 0.000 description 30
- 230000005540 biological transmission Effects 0.000 description 26
- 238000004891 communication Methods 0.000 description 10
- 230000006870 function Effects 0.000 description 9
- 239000003795 chemical substances by application Substances 0.000 description 5
- 238000005516 engineering process Methods 0.000 description 5
- 230000008447 perception Effects 0.000 description 5
- 230000003287 optical effect Effects 0.000 description 4
- 230000008569 process Effects 0.000 description 4
- 230000009471 action Effects 0.000 description 3
- 230000014509 gene expression Effects 0.000 description 2
- 239000004973 liquid crystal related substance Substances 0.000 description 2
- 239000013307 optical fiber Substances 0.000 description 2
- 230000000644 propagated effect Effects 0.000 description 2
- 239000004065 semiconductor Substances 0.000 description 2
- 230000001960 triggered effect Effects 0.000 description 2
- VYZAMTAEIAYCRO-UHFFFAOYSA-N Chromium Chemical compound [Cr] VYZAMTAEIAYCRO-UHFFFAOYSA-N 0.000 description 1
- 230000006978 adaptation Effects 0.000 description 1
- 238000013473 artificial intelligence Methods 0.000 description 1
- 230000003190 augmentative effect Effects 0.000 description 1
- 230000008859 change Effects 0.000 description 1
- 230000006835 compression Effects 0.000 description 1
- 238000007906 compression Methods 0.000 description 1
- 230000006837 decompression Effects 0.000 description 1
- 230000007547 defect Effects 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 238000004880 explosion Methods 0.000 description 1
- 238000013507 mapping Methods 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 230000016776 visual perception Effects 0.000 description 1
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/43—Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
- H04N21/436—Interfacing a local distribution network, e.g. communicating with another STB or one or more peripheral devices inside the home
- H04N21/4363—Adapting the video stream to a specific local network, e.g. a Bluetooth® network
- H04N21/43637—Adapting the video stream to a specific local network, e.g. a Bluetooth® network involving a wireless protocol, e.g. Bluetooth, RF or wireless LAN [IEEE 802.11]
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/47—End-user applications
- H04N21/478—Supplemental services, e.g. displaying phone caller identification, shopping application
- H04N21/4781—Games
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/47—End-user applications
- H04N21/478—Supplemental services, e.g. displaying phone caller identification, shopping application
- H04N21/4782—Web browsing, e.g. WebTV
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/47—End-user applications
- H04N21/478—Supplemental services, e.g. displaying phone caller identification, shopping application
- H04N21/4788—Supplemental services, e.g. displaying phone caller identification, shopping application communicating with other users, e.g. chatting
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Computer Networks & Wireless Communication (AREA)
- General Engineering & Computer Science (AREA)
- Information Transfer Between Computers (AREA)
Abstract
The embodiment of the application provides a multi-screen interaction method and device, a computer readable medium and electronic equipment. The multi-screen interaction method comprises the following steps: receiving a joining request which is sent by an interactive end and used for requesting to join an interactive scene, wherein the joining request comprises a user identifier corresponding to the interactive end; after the interactive end joins the interactive scene, receiving an interactive message sent by the interactive end, wherein the interactive message comprises real-time interactive information; selecting a target streaming media file matched with the real-time interaction information from a plurality of streaming media files associated with the user identification; and controlling a display end of the interactive scene to display the target streaming media file according to the target streaming media file. According to the technical scheme, the streaming media file can be seamlessly displayed in multi-screen interaction, and user experience of the multi-screen interaction is enhanced.
Description
Technical Field
The application relates to the technical field of multi-screen interaction, in particular to a multi-screen interaction method, a multi-screen interaction device, a computer readable medium and electronic equipment.
Background
Multi-screen interaction refers to a series of operations such as transmission and display of multimedia contents between different terminal devices, such as common intelligent terminals, smart phones, televisions, and the like. The multi-screen interaction can realize the simultaneous sharing of the display contents on different terminals, and the multimedia life of the user is enriched.
At present, in a scene of implementing multi-screen interaction, hardware devices with the same communication protocol are mainly implemented through a short-distance communication mode, for example, bluetooth, infrared, Wi-Fi signal and other modes.
Disclosure of Invention
Embodiments of the present application provide a multi-screen interaction method, apparatus, computer-readable medium, and electronic device, so that seamless display of a streaming media file in multi-screen interaction can be achieved at least to a certain extent, and user experience of multi-screen interaction is enhanced.
Other features and advantages of the present application will be apparent from the following detailed description, or may be learned by practice of the application.
According to an aspect of an embodiment of the present application, a multi-screen interaction method is provided, including: receiving a joining request which is sent by an interactive end and used for requesting to join an interactive scene, wherein the joining request comprises a user identifier corresponding to the interactive end; after the interactive end joins the interactive scene, receiving an interactive message sent by the interactive end, wherein the interactive message comprises real-time interactive information; selecting a target streaming media file matched with the real-time interaction information from a plurality of streaming media files associated with the user identification; and controlling a display end of the interactive scene to display the target streaming media file according to the target streaming media file.
According to an aspect of an embodiment of the present application, a multi-screen interaction method is provided, including: sending an access request for requesting to join an interactive scene to a cloud server, wherein the interactive scene comprises a plurality of interactive ends; after the interaction scene is added, acquiring a target streaming media file corresponding to each interaction terminal, wherein the target streaming media file is selected from a plurality of streaming media files associated with user identifications corresponding to each interaction terminal by the cloud server according to interaction messages sent by each interaction terminal, and the target streaming media file corresponding to each interaction terminal is matched with real-time interaction information contained in the interaction messages sent by each interaction terminal; and displaying the target streaming media files corresponding to the interactive ends.
According to an aspect of an embodiment of the present application, there is provided a multi-screen interaction apparatus, including: the first receiving unit is configured to receive a join request which is sent by an interactive end and used for requesting to join an interactive scene, wherein the join request comprises a user identifier corresponding to the interactive end; the second receiving unit is configured to receive an interactive message sent by the interactive end after the interactive end joins the interactive scene, wherein the interactive message comprises real-time interactive information; the selecting unit is configured to select a target streaming media file matched with the real-time interaction information from a plurality of streaming media files associated with the user identifier; and the control unit is configured to control a display end of the interactive scene to display the target streaming media file according to the target streaming media file.
In some embodiments of the present application, based on the foregoing scheme, the first receiving unit is configured to: receiving an access request sent by the display end, and feeding back response information aiming at the access request to the display end, wherein the response information is used for indicating the display end to display a set prompt interface; and receiving the joining request sent by the interactive terminal through the prompt interface.
In some embodiments of the present application, based on the foregoing scheme, the second receiving unit is configured to: and if an interaction starting instruction is received, sending a starting message to the interaction end, wherein the starting message is used for indicating the interaction end to generate an interaction interface corresponding to the interaction scene and detecting interaction information aiming at the interaction scene.
In some embodiments of the present application, based on the foregoing solution, the selecting unit is configured to: and according to the user identification, determining identification information of a plurality of streaming media files related to the user identification from a relation table between the interactive scene and the identification information of the streaming media files, and selecting the identification information of a target streaming media file matched with the real-time interactive information from the identification information of the plurality of streaming media files.
In some embodiments of the present application, based on the foregoing solution, the control unit is configured to: generating an interactive response message containing the identification information of the target streaming media file, and sending the interactive response message to a streaming media server so that the streaming media server sends the target streaming media file to the display terminal for displaying according to the identification information of the target streaming media file; or sending the target streaming media file to a display end of the interactive scene so that the display end displays the target streaming media file; or generating an interactive response message containing the identification information of the target streaming media file, and sending the interactive response message to the display end, so that the display end acquires the target streaming media file according to the identification information of the target streaming media file and displays the acquired target streaming media file.
In some embodiments of the present application, based on the foregoing scheme, the real-time interaction information includes an interaction score, and the second receiving unit is configured to: receiving an interaction score sent by the interaction end, sending the interaction score to the display end, and sending interaction reward information to the interaction end if an interaction completion message sent by the display end according to the interaction score is received; or receiving the interaction score sent by the interaction end, if the interaction end is determined to finish the interaction according to the interaction score, sending interaction reward information to the interaction end, and sending interaction finish information to the display end, so that the display end displays the interaction reward information to the interaction end.
In some embodiments of the present application, based on the foregoing solution, the control unit is configured to: and counting the number of the interactive ends added into the interactive scene, and sending the counted number to the display end for displaying.
According to an aspect of an embodiment of the present application, there is provided a multi-screen interaction apparatus, including: the system comprises a sending unit, a receiving unit and a processing unit, wherein the sending unit is configured to send an access request for requesting to join an interactive scene to a cloud server, and the interactive scene comprises a plurality of interactive ends; an obtaining unit, configured to obtain, after joining the interactive scene, a target streaming media file corresponding to each interactive end, where the target streaming media file is selected by the cloud server from a plurality of streaming media files associated with the user identifier corresponding to each interactive end according to the interactive message sent by each interactive end, and the target streaming media file corresponding to each interactive end matches with real-time interactive information included in the interactive message sent by each interactive end; and the display unit is configured to display the target streaming media files corresponding to the interaction ends.
In some embodiments of the present application, based on the foregoing scheme, the obtaining unit is configured to: receiving the interaction scores of the interaction ends sent by the cloud server, and displaying interaction reward information of the target interaction end for the target interaction end with the interaction score reaching a preset threshold value; or receiving information of a target interaction end, sent by the cloud server, of which the interaction score reaches a preset threshold value, and displaying interaction reward information of the target interaction end.
In some embodiments of the present application, based on the foregoing scheme, the obtaining unit is configured to: receiving response information fed back by the cloud server for the access request; and generating a set prompt interface according to the response information, wherein the prompt interface is used for the interaction terminal to join the interaction scene.
In some embodiments of the present application, based on the foregoing solution, the display unit is configured to: if the target streaming media files corresponding to the interactive ends are obtained at the same time, determining a display sequence according to the user identification corresponding to each interactive end; and displaying the target streaming media files corresponding to the interactive ends according to the display sequence.
According to an aspect of the embodiments of the present application, there is provided a computer readable medium, on which a computer program is stored, which when executed by a processor, implements the multi-screen interaction method as described in the above embodiments.
According to an aspect of an embodiment of the present application, there is provided an electronic device including: one or more processors; a storage device for storing one or more programs, which when executed by the one or more processors, cause the one or more processors to implement the multi-screen interaction method as described in the above embodiments.
In the technical solutions provided in some embodiments of the present application, after an interactive scene is added at an interactive end, an interactive message sent by the interactive end is received, a target streaming media file matched with real-time interactive information included in the interactive message is selected from a plurality of streaming media files associated with a user identifier, and then a display end of the interactive scene is controlled to display the target streaming media file according to the target streaming media file, so that the problem of compatibility of multi-screen interaction between different devices is effectively solved, flexibility is higher, seamless display of the streaming media file can be realized through information transmission between the devices, and interactive experience of a user is improved.
It is to be understood that both the foregoing general description and the following detailed description are exemplary and explanatory only and are not restrictive of the application.
Drawings
The accompanying drawings, which are incorporated in and constitute a part of this specification, illustrate embodiments consistent with the present application and together with the description, serve to explain the principles of the application. It is obvious that the drawings in the following description are only some embodiments of the application, and that for a person skilled in the art, other drawings can be derived from them without inventive effort. In the drawings:
FIG. 1 is a diagram illustrating an implementation environment of a multi-screen interaction method according to an embodiment of the present application;
FIG. 2 is a diagram illustrating an environment for implementing a multi-screen interaction method according to another embodiment of the present application;
FIG. 3 illustrates a flow diagram of a multi-screen interaction method according to an embodiment of the present application;
FIG. 4 illustrates a flow diagram of a multi-screen interaction method according to an embodiment of the present application;
FIG. 5 illustrates a flow diagram of a multi-screen interaction method according to an embodiment of the present application;
FIG. 6 illustrates a flow diagram of a multi-screen interaction method according to an embodiment of the present application;
FIG. 7 illustrates a flow diagram of a multi-screen interaction method according to an embodiment of the present application;
FIG. 8 illustrates an interaction flow diagram of a multi-screen interaction method according to an embodiment of the present application;
FIG. 9 illustrates a logical decision diagram of a multi-screen interaction method according to an embodiment of the present application;
FIG. 10 is a schematic interface diagram of a multi-screen interaction method at a display end according to an embodiment of the application;
FIG. 11 is a schematic interface diagram of a multi-screen interaction method at an interaction end according to an embodiment of the application;
FIG. 12 illustrates a block diagram of a multi-screen interaction device, according to an embodiment of the present application;
FIG. 13 illustrates a block diagram of a multi-screen interaction device, according to an embodiment of the present application;
FIG. 14 illustrates a schematic structural diagram of a computer system suitable for use in implementing the electronic device of an embodiment of the present application.
Detailed Description
Example embodiments will now be described more fully with reference to the accompanying drawings. Example embodiments may, however, be embodied in many different forms and should not be construed as limited to the examples set forth herein; rather, these embodiments are provided so that this disclosure will be thorough and complete, and will fully convey the concept of example embodiments to those skilled in the art.
Furthermore, the described features, structures, or characteristics may be combined in any suitable manner in one or more embodiments. In the following description, numerous specific details are provided to give a thorough understanding of embodiments of the application. One skilled in the relevant art will recognize, however, that the subject matter of the present application can be practiced without one or more of the specific details, or with other methods, components, devices, steps, and so forth. In other instances, well-known methods, devices, implementations, or operations have not been shown or described in detail to avoid obscuring aspects of the application.
It is to be noted that the terms used in the specification and claims of the present application and the above-described drawings are only for describing the embodiments and are not intended to limit the scope of the present application. It will be understood that the terms "comprises," "comprising," "includes," "including," "has," "having," and the like, when used herein, specify the presence of stated features, integers, steps, operations, elements, components, and/or groups thereof, but do not preclude the presence or addition of one or more other features, integers, steps, operations, elements, components, and/or groups thereof.
It will be further understood that, although the terms first, second, third, etc. may be used herein to describe various elements, these elements should not be limited by these terms. These terms are only used to distinguish one element from another. For example, a first element could be termed a second element without departing from the scope of the present invention. Similarly, a second element may be termed a first element. As used herein, the term "and/or" includes any and all combinations of one or more of the associated listed items.
The block diagrams shown in the figures are functional entities only and do not necessarily correspond to physically separate entities. I.e. these functional entities may be implemented in the form of software, or in one or more hardware modules or integrated circuits, or in different networks and/or processor means and/or microcontroller means.
The flow charts shown in the drawings are merely illustrative and do not necessarily include all of the contents and operations/steps, nor do they necessarily have to be performed in the order described. For example, some operations/steps may be decomposed, and some operations/steps may be combined or partially combined, so that the actual execution sequence may be changed according to the actual situation.
It should be understood that in the present application, "at least one" means one or more, "a plurality" means two or more.
Before further detailed description of the embodiments of the present application, terms and expressions referred to in the embodiments of the present application will be described, and the terms and expressions referred to in the embodiments of the present application will be used for the following explanation.
Cloud computing: the increase, use and interaction mode of internet-based related services generally relates to the provision of dynamic and easily-expandable and often-virtualized resources through the internet, and cloud is a metaphor of networks and the internet, and in the past, a telecommunication network is often represented by cloud in a figure, and the abstraction of the internet and underlying infrastructure is also represented by the cloud, so that cloud computing can enable a user to experience computing power of 10 trillion times per second, the user can simulate nuclear explosion, forecast climate change and market development trend with the powerful computing power, and the user can access a data center through computers, notebooks, mobile phones and the like to operate according to own needs.
An interactive scene: the virtual scene output by the equipment is different from the real world and can support user interaction, such as a game scene, and visual perception of the interactive scene can be formed by naked eyes or the assistance of the equipment, for example, a two-dimensional image output by a display screen, and a three-dimensional image output by a stereoscopic display technology such as a stereoscopic projection technology, a virtual reality technology and an augmented reality technology; in addition, various real-world-simulated perceptions such as auditory perception, tactile perception, olfactory perception, motion perception and the like can be formed through various possible hardware.
Multi-screen interaction: the method applies the flash internet protocol, the Miracast protocol and the like, through Wi-Fi network connection, a series of operations such as transmission, analysis, display, control and the like of multimedia (audio, video and picture) contents can be carried out on different multimedia terminals, the displayed contents can be shared on different platform devices at the same time, and the multimedia life of users is enriched.
Websocket protocol: the browser and the server only need to perform a handshake action in a Websocket Application Programming Interface (API), then a fast channel is formed between the browser and the server, and data interaction can be directly performed between the browser and the server.
At present, in the application of a multi-screen interaction technology, not only are device hardware drivers required to be compatible with each other, but also a series of operations such as transmission, analysis, display, control and the like of multimedia (such as audio, video, pictures and the like) contents on different devices are mainly realized through short-distance communication, such as modes of bluetooth, infrared, Wi-Fi signals and the like, so that the existing multi-screen interaction has the defects of low flexibility, large limitation, low multi-screen interaction efficiency influenced by network transmission rate and the like.
In contrast, a multi-screen interaction method is provided in an embodiment of the present application, and includes receiving a join request sent by an interaction end for requesting to join an interaction scene, where the join request includes a user identifier corresponding to the interaction end, receiving an interaction message sent by the interaction end after the interaction end joins the interaction scene, where the interaction message includes real-time interaction information, selecting a target streaming media file matched with the real-time interaction information from a plurality of streaming media files associated with the user identifier, and controlling a display end of the interaction scene to display the target streaming media file according to the target streaming media file. The multi-screen interaction method provided by the embodiment of the application can effectively solve the problem of compatibility of multi-screen interaction among different devices, is higher in flexibility, can realize seamless display of the streaming media file through information transmission among the devices, and improves the interaction experience of users.
Fig. 1 is a diagram of an implementation environment of a multi-screen interaction method according to an embodiment of the present application, where the implementation environment includes: the interactive terminal 101, the cloud server 102 and the display terminal 103 are used for information transmission between the interactive terminal 101, the cloud server 102 and the display terminal 103 through a network, and the network may include various connection types, such as a wired communication link, a wireless communication link or an optical fiber cable.
The interactive terminal 101 is an electronic device for a user to participate in an interactive scene for interaction, and the electronic device may be a notebook computer, a mobile phone, a tablet computer, a vehicle-mounted terminal, an intelligent wearable device, an intelligent home appliance, or the like. The interactive terminal 101 may perform data interaction with the cloud server 102 through a network, for example, the interactive terminal 101 may send a join request requesting to join in an interactive scene to the cloud server 102, and after receiving a response to the join request returned by the cloud server 102, the interactive terminal 101 joins in the interactive scene.
The cloud server 102 may be an independent physical server, a server cluster or a distributed system formed by a plurality of physical servers, or a cloud server providing basic cloud computing services such as a cloud service, a cloud database, cloud computing, a cloud function, cloud storage, a network service, cloud communication, a middleware service, a domain name service, a security service, a CDN, a big data and artificial intelligence platform, and the like. The cloud server 102 has powerful expansion capability, huge computing capability and mass storage capability, and can provide services for mass users at the same time.
The display terminal 103 may be an electronic device with internet connection and a display screen, such as a notebook computer, a mobile phone, a tablet computer, a smart television, a general television, or other large-screen display devices, such as a projector, or a liquid crystal display array. For an online user, the corresponding interactive end 101 and display end 103 may be provided, for an offline user, the corresponding interactive end 101 and display end 103 may be provided, and for an offline user, all offline users may also share one display end 103, for example, a large screen is displayed. The display terminal 103 may perform data interaction with the cloud server 102 through a network.
In an embodiment of the application, the interactive end 101 may send a join request for requesting to join in an interactive scene to the cloud server 102, the join request sent by the interactive end 101 includes a user identifier corresponding to the interactive end 101, after the interactive end 101 joins in the interactive scene, the cloud server 102 may receive an interactive message sent by the interactive end 101, and select a target streaming media file matched with real-time interaction information included in the interactive message from a plurality of streaming media files associated with the user identifier, and then the cloud server 102 may control the display end 103 to display the target streaming media file according to the target streaming media file.
Fig. 2 is a diagram of an implementation environment of a multi-screen interaction method according to another embodiment of the present application, where the implementation environment includes: the interactive system comprises an interactive terminal 101, a cloud server 102, a display terminal 103 and a streaming media server 104, wherein the streaming media server 104 is a server for storing streaming media files corresponding to interactive scenes, and the streaming media server 104 can perform data interaction with the cloud server 102 and the display terminal 103 through a network.
In an embodiment of the present application, the cloud server 102 may generate an interactive response message including identification information of a target streaming media file, and then send the interactive response message to the streaming media server 104, and after receiving the identification information of the target streaming media file, the streaming media server 104 may determine the target streaming media file according to the identification information of the target streaming media file, and send the target streaming media file to the display terminal 103 for display.
In an embodiment of the present application, the cloud server 102 may send an interactive response message including identification information of a target streaming media file to the display terminal 103, and the display terminal 103 may obtain the target streaming media file from the streaming media server 104 according to the identification information of the target streaming media file and then display the obtained target streaming media file.
It should be understood that the numbers of the interactive terminal 101, the cloud server 102, the display terminal 103 and the streaming media server 104 in fig. 1 and 2 are only illustrative. There may be any number of interaction terminals 101, cloud servers 102, and display terminals 103, as desired for implementation. For example, the cloud server 102 may be a server cluster composed of a plurality of servers.
The multi-screen interaction method provided by the embodiment of the application is generally executed by the cloud server 102, and accordingly, the multi-screen interaction device is generally disposed in the cloud server 102. However, it is easily understood by those skilled in the art that the multi-screen interaction method provided in the present embodiment of the application can also be executed by the display terminal 103, and accordingly, the multi-screen interaction device can also be disposed in the display terminal 103, which is not particularly limited in the present exemplary embodiment. For example, in an exemplary embodiment, the display terminal 103 may obtain a target streaming media file corresponding to each interactive terminal 101, and then display the target streaming media file corresponding to each interactive terminal 101, where the target streaming media file is selected by the cloud server 102 from a plurality of streaming media files associated with the user identifier corresponding to each interactive terminal 101 according to an interactive message sent by each interactive terminal 101, and the target streaming media file corresponding to each interactive terminal 101 is matched with real-time interactive information included in the interactive message sent by each interactive terminal 101, and sends the obtained target streaming media file to the display terminal 103.
The implementation details of the technical solution of the embodiment of the present application are set forth in detail below:
fig. 3 illustrates a flowchart of a multi-screen interaction method according to an embodiment of the present application, which may be performed by a cloud server, which may be the cloud server 102 illustrated in fig. 1 or fig. 2. Referring to fig. 3, the multi-screen interaction method includes:
step S310, receiving a joining request which is sent by an interactive terminal and used for requesting to join an interactive scene, wherein the joining request comprises a user identification corresponding to the interactive terminal;
step S320, after the interactive end joins the interactive scene, receiving an interactive message sent by the interactive end, wherein the interactive message comprises real-time interactive information;
step S330, selecting a target streaming media file matched with the real-time interaction information from a plurality of streaming media files associated with the user identification;
and step S340, controlling a display end of the interactive scene to display the target streaming media file according to the target streaming media file.
These steps are described in detail below.
In step S310, a join request for requesting to join in an interactive scene sent by an interactive end is received, where the join request includes a user identifier corresponding to the interactive end.
Specifically, the interactive end may be a mobile phone, a tablet computer, and the like, the join request may be information sent by the interactive end to a cloud server according to a gesture that detects that a user slides out of a screen of the interactive end, the cloud server receives the join request sent by the interactive end to request to join in an interactive scene, and the join request carries a user identifier corresponding to the interactive end.
In the multi-screen interaction, multiple devices interact with each other, when the multiple devices interact with each other in multiple screens, a data interaction process is very complex, in order to ensure accuracy of data synchronization, a user identifier of a requester is carried in a join request of an interaction end, specifically, the user identifier is used for identifying which interaction end the join request is sent by, the user identifier may be a user identifier for logging in at the interaction end, for example, a nickname of a user, an account number of the user, and the like, and the user identifier may also be a physical address of the interaction end.
It should be noted that the interactive end in the embodiment of the present application may be any one of interactive ends requesting to join in an interactive scene, and the interactive end joining in the interactive scene may include an interactive end of an offline user or an interactive end of an online user, where the offline user means that the user is in an interactive site, and the online user means that the user is not in the interactive site.
In an embodiment of the present application, the interaction side may send a join request for requesting to join the interaction scenario to the cloud server through the display side, and in this embodiment, as shown in fig. 4, the method may specifically include steps S410 to S420, which are described in detail as follows:
and step S410, receiving an access request sent by the display end, and feeding back response information aiming at the access request to the display end, wherein the response information is used for indicating the display end to display a set prompt interface.
Specifically, the display terminal may join the interactive scene prior to the interactive terminal, the cloud server may receive an access request sent by the display terminal before receiving a join request sent by the interactive terminal to request to join the interactive scene, where the access request is a request of the display terminal to join the interactive scene, the cloud server receives the access request of the display terminal and feeds back response information for the access request to the display terminal, and after receiving the response information, the display terminal may generate a set prompt interface according to the response information.
The set prompt interface can be a two-dimensional code interface of a set interactive scene, the interactive end can join the interactive scene through the two-dimensional code interface on the scanning display end, the set prompt interface can also be an interface displaying a specified user name and a password, and the interactive end can join the interactive scene through the specified user name and the password.
In one embodiment of the present application, the display end may run a browser, and the browser is a browser supporting HTML5, such as a Chrome browser. Before the display end sends the access request, the display end can establish a Websocket data transmission channel with the cloud server through the browser, and then sends the access request to the cloud server through the Websocket data transmission channel.
The Websocket data transmission channel established between the display end and the cloud server can be achieved by the display end sending a handshake request and the cloud server responding to the handshake request. Specifically, the display end sends a Websocket handshake request to the cloud server through the browser, where the Websocket handshake request may include an IP address, a host name, and the like of the display end. The cloud server receives a Websocket handshake request sent by the display end, and acquires an IP address, a host name and the like of the display end. The cloud server parses the Websocket handshake request and generates a handshake response. And the cloud server sends the handshake response to the display end to realize successful handshake and establish a Websocket data transmission channel.
In the embodiment provided by the application, the display end sends the access request by establishing the Websocket data transmission channel with the cloud server, receives the response information aiming at the access request through the browser, and further generates the set prompt interface on the browser, so that the display end does not need to install client software any more, and the efficiency of adding each device into an interactive scene is improved. Especially in a home network, the display end and the cloud server establish Websocket connection, network flow is reduced, and network resources are saved.
And step S420, receiving a joining request sent by the interactive terminal through a prompt interface.
After the display end displays the set prompt interface, if the user wants to join the interactive scene, a joining request can be sent to the cloud server through the prompt interface so as to join the interactive scene.
In an implementation manner of the application, the prompt interface may be a two-dimensional code interface of an interactive scene, and therefore, a user may open a code scanning tool on the interactive end, such as a scanning function of a client program or a browser, scan a two-dimensional code generated by the display end and used for implementing multi-screen interaction, and send a join request requesting to join the interactive scene to the cloud server.
Similarly, the interactive end can also establish a Websocket data transmission channel with the cloud server, and then send an adding request for requesting to add to the interactive scene to the cloud server through the Websocket data transmission channel, and transmit related data through the Websocket data transmission channel, so that network resources can be saved.
With reference to fig. 3, in step S320, after the interactive end joins the interactive scene, an interactive message sent by the interactive end is received, where the interactive message includes real-time interactive information.
Specifically, the interaction end can send an interaction message to the cloud server when the interaction scene is opened, and the cloud server receives the interaction message sent by the interaction end. The interactive information comprises real-time interactive information, and it can be understood that the interactive end can detect the operation of the user on the interactive scene in real time under the condition that the interactive scene is opened, so as to obtain the real-time interactive information.
In step S330, a target streaming media file matching the real-time interaction information is selected from the plurality of streaming media files associated with the user identifier.
It should be noted that the streaming media file is a file in a streaming media format, and the streaming media format refers to a media format played on the internet in a streaming transmission manner. The streaming transmission mode is to divide the whole multimedia file such as audio or video into individual compressed packets by a special compression mode, and the packets are continuously transmitted to the user computer by the server in real time. In the system adopting the streaming transmission mode, a user does not need to wait until the whole file is completely downloaded like a downloading mode, but only needs to use a decompression device (hardware or software) to decompress a compressed multimedia file on a computer of the user and then play and watch the compressed multimedia file after the start delay of a few seconds or a few tens of seconds. The remainder of the multimedia file will now continue to be downloaded in the background server.
Since the user opens the interactive scene, and the user and the interactive scene have an association relationship, the multiple streaming media files associated with the user identifier can be understood as streaming media files corresponding to the interactive scene, and the streaming media files in the embodiment of the present application can be video files, audio files, text files, animation files, image files, and the like. The type of the streaming media file in the embodiment of the present application is not specifically limited, for example, the streaming media file may be various video files or various audio files.
After determining a plurality of streaming media files associated with the user identifier, a target streaming media file matched with the real-time interaction information can be selected from the streaming media files. The cloud server maps the real-time interaction information to a streaming media file according to the mapping relation, and acquires a target streaming media file.
In the foregoing process, the cloud server may pre-store a relationship table between the interactive scene and the identification information of the streaming media file in the database, so as to determine a plurality of streaming media files associated with the user identification, in this embodiment, the method may specifically include:
and according to the user identification, determining identification information of a plurality of streaming media files related to the user identification from a relation table between the interactive scene and the identification information of the streaming media files, and selecting the identification information of the target streaming media file matched with the real-time interactive information from the identification information of the plurality of streaming media files.
In this embodiment, the identification information of the streaming media file is a unique identification identifier of the streaming media file, and is used to distinguish different contents of the streaming media file, and the cloud server may pre-store a relationship table between an interactive scene and the identification information of the streaming media file, and then determine the identification information of a plurality of streaming media files associated with the user identifier from the relationship table according to the user identifier added to the interactive scene.
After the identification information of a plurality of streaming media files associated with the user identification is determined, the identification information of the target streaming media file matched with the real-time interaction information is selected from the identification information, the identification information of the target streaming media file is the only identification of the target streaming media file, and the target streaming media file can be determined through the identification information of the target streaming media file.
In step S340, the display end of the interactive scene is controlled to display the target streaming media file according to the target streaming media file.
Specifically, after determining the target streaming media file, the cloud server may control a display end of the interactive scene to display according to the target streaming media file. The control of the display terminal to display may include various implementation manners, specifically:
in a first implementation manner, after the cloud server selects the target streaming media file, an interactive response message including identification information of the target streaming media file may be generated, and the interactive response message is sent to the streaming media server, so that the streaming media server sends the target streaming media file to the display terminal for displaying according to the identification information of the target streaming media file.
In this implementation manner, the streaming media server is a server for caching a streaming media file, the streaming media server is connected to the cloud server and the display terminal, respectively, after the cloud server determines a target streaming media file matched with the real-time interaction information, the cloud server may generate an interaction response message including identification information of the target streaming media file, and send the interaction response message to the streaming media server, and the streaming media server sends the cached target streaming media file to the display terminal for display according to the interaction response message.
In a second implementation manner, after the target streaming media file is selected, the cloud server can directly send the target streaming media file to the display end of the interactive scene, so that the display end displays the target streaming media file.
The cloud server directly sends the target streaming media file to the display end to be displayed after selecting the target streaming media file, and does not send the target streaming media file according to the display request of the display end any more, so that the service pressure of the server can be reduced to a certain extent, and the network flow can be saved to a certain extent due to the fact that the display end is not required to send the display request at regular time.
In a third implementation manner, after the target streaming media file is selected, the cloud server generates an interactive response message containing identification information of the target streaming media file, and sends the interactive response message to the display end, so that the display end acquires the target streaming media file according to the identification information of the target streaming media file and displays the acquired target streaming media file.
In this implementation manner, after the cloud server determines the target streaming media file matched with the real-time interaction information, an interaction response message may be generated, and the interaction response message is sent to the display terminal, the display terminal may obtain the target streaming media file according to identification information of the target streaming media file included in the interaction response message, specifically, the display terminal may obtain the target streaming media file from the streaming media file cached in the local database according to the identification information of the target streaming media file, may also obtain the target streaming media file from the streaming media server, and then the display terminal displays the obtained target streaming media file.
Based on the technical scheme of the embodiment, after the interactive end joins the interactive scene, the interactive message sent by the interactive end is received, the target streaming media file matched with the real-time interactive information contained in the interactive message is selected from the streaming media files associated with the user identifier, and then the display end of the interactive scene is controlled to display the target streaming media file according to the target streaming media file, so that the problem of compatibility of multi-screen interaction among different devices is effectively solved, flexibility is higher, seamless display of the streaming media file can be realized through information transmission among the devices, and the interactive experience of the user is improved.
In an embodiment of the application, if the cloud server receives an interactive start instruction, the cloud server may send a start message to the interactive end, where the interactive start instruction may be triggered by the display end or the interactive end.
After the interactive end receives the opening message, an interactive interface corresponding to the interactive scene can be generated, the user can start interaction through the interactive interface, and after the user opens the interactive interface, the interactive end can detect the operation of the user on the interactive interface so as to obtain the interactive information of the user in the interactive scene.
In an embodiment of the application, the real-time interaction information of the interaction end may be an interaction score of the interaction end for performing interaction in an interaction scene, wherein the interaction score may be understood as an interaction completion condition of the interaction end in the interaction scene, and the interaction end may send the interaction score to the cloud server.
In some embodiments, the cloud server may send the received interaction score of the interaction end to the display end, so that the display end determines whether the interaction end completes the interaction according to the interaction score, and if the cloud server receives an interaction completion message sent by the display end, the cloud server may send interaction reward information to the interaction end.
In other embodiments, after receiving the interaction score sent by the interaction end, the cloud server may determine whether the interaction end completes the interaction according to the interaction score, if the cloud server determines that the interaction end completes the interaction, the cloud server may send the interaction reward information to the interaction end, meanwhile, the cloud server may also send the interaction completion message to the display end, and after receiving the interaction completion message, the display end may display the interaction reward information to the interaction end.
In an embodiment of the application, for the interactive terminals joining the interactive scene, the cloud server may also perform quantity statistics on the interactive terminals joining the interactive scene in real time or periodically, and send the quantity obtained through the statistics to the display terminal for displaying.
Fig. 5 is a flowchart illustrating a multi-screen interaction method according to an embodiment of the present application, where the multi-screen interaction method may be performed by a display terminal, which may be the display terminal 103 shown in fig. 1 or fig. 2. Referring to fig. 5, the multi-screen interaction method includes:
step S510, sending an access request for requesting to join an interactive scene to a cloud server, wherein the interactive scene comprises a plurality of interactive ends;
step S520, after the interactive scene is added, a target streaming media file corresponding to each interactive terminal is obtained, wherein the target streaming media file is selected from a plurality of streaming media files associated with the user identifier corresponding to each interactive terminal by the cloud server according to the interactive message sent by each interactive terminal, and the target streaming media file corresponding to each interactive terminal is matched with the real-time interactive information contained in the interactive message sent by each interactive terminal;
and step S530, displaying the target streaming media files corresponding to the interactive ends.
These steps are described in detail below.
Step S510, sending an access request for requesting to join in an interactive scene to the cloud server, where the interactive scene includes a plurality of interactive terminals.
In this embodiment, the execution main body is a display end, and the display end is configured to display the interactive scene, so that, in order to enable display of the interactive scene, the display end may send an access request to the cloud server, where the access request is used to request to join the interactive scene, and of course, in addition to the display end, the display end also includes a plurality of interactive ends, and the interactive ends are used to perform interaction of the interactive scene.
It should be noted that, no matter the online user or the offline user, the interaction of the interaction scene may be performed through the interaction end, for the online user, the online user may have a respective display end, for the offline user, the offline user may also have a respective display end, and of course, all the offline users may also share one display end, for example, a large display screen.
Step S520, after the interactive scene is added, obtaining a target streaming media file corresponding to each interactive end, where the target streaming media file is selected by the cloud server from a plurality of streaming media files associated with the user identifier corresponding to each interactive end according to the interactive message sent by each interactive end, and the target streaming media file corresponding to each interactive end matches with the real-time interactive information included in the interactive message sent by each interactive end.
After the display end and the interactive end join in an interactive scene, the display end can acquire target streaming media files corresponding to each interactive end, wherein the target streaming media files corresponding to each interactive end are selected from a plurality of streaming media files associated with user identifications corresponding to each interactive end by the cloud server according to interactive messages sent by each interactive end.
When the interactive end interacts with an interactive scene, the interactive end can send an interactive message to the cloud server, and the interactive message contains real-time interactive information of the interactive end, so that the cloud server can select a target streaming media file matched with the real-time interactive information from a plurality of streaming media files associated with the user identifier.
In an embodiment, the cloud server may further generate an interactive response message including identification information of the target streaming media file, and send the interactive response message to the display terminal, so that the display terminal may acquire the target streaming media file from a locally cached file according to the identification information of the target streaming media file and display the target streaming media file, or the display terminal acquires the target streaming media file from the streaming media server according to the identification information of the target streaming media file and displays the target streaming media file. In another embodiment, the cloud server may further send the generated interactive response message including the identification information of the target streaming media file to the streaming media server, so that the streaming media server obtains the target streaming media file according to the identification information of the target streaming media file, and then the streaming media server sends the obtained target streaming media file to the display terminal for display.
And step S530, displaying the target streaming media files corresponding to the interaction ends.
After the display terminal obtains the target streaming media files corresponding to each interactive terminal, each target streaming media file can be displayed on a display screen of the display terminal.
In an embodiment of the present application, as shown in fig. 7, step S530 may specifically include steps S710 to S720:
step S710, if the target streaming media files corresponding to each interactive end are obtained at the same time, determining a display sequence according to the user identification corresponding to each interactive end;
and S720, displaying the target streaming media files corresponding to the interactive ends according to the display sequence.
In this embodiment, if the display end obtains the target streaming media files corresponding to each interactive end at the same time, in order to better display the target streaming media files, the display end may determine a display order of the multiple streaming media files, specifically, the display end may determine the display order according to the user identifier corresponding to each interactive end, and then display the target streaming media files corresponding to each interactive end according to the display order.
In an embodiment of the application, the interaction message sent by the interaction end includes real-time interaction information, the real-time interaction information may be an interaction score, the display end may determine whether there is a target interaction end with an interaction score reaching a preset threshold value after receiving the interaction scores of the interaction ends sent by the cloud server, and if so, the display end displays interaction reward information for the target interaction end.
In an embodiment of the application, after the interaction end sends the interaction message containing the interaction scores to the cloud server, the cloud server can determine whether a target interaction end with an interaction score reaching a preset threshold value exists according to the interaction scores of the interaction ends, if yes, information of the target interaction end is sent to the display end, and after the display end receives the information of the target interaction end, the display end can display interaction reward information of the target interaction end.
In an embodiment of the present application, after the display terminal sends the access request to the cloud server, the cloud server may feed back response information to the display terminal, so that the display terminal may generate a set prompt interface according to the response information, in this embodiment, as shown in fig. 6, specifically including step S610 to step S620, which is described in detail below:
and step S610, receiving response information aiming at the access request fed back by the cloud server.
Specifically, after receiving an access request sent by a display terminal, the cloud server may feed back response information for the access request to the display terminal, and the display terminal receives the response information fed back by the cloud server and joins in an interactive scene.
And S620, generating a set prompt interface according to the response information, wherein the prompt interface is used for the interactive terminal to join in the interactive scene.
After receiving the response message, the display terminal may generate a set prompt interface according to the response message. The set prompt interface can be a two-dimensional code interface of an interactive scene, the interactive terminal can be added into the interactive scene through the two-dimensional code interface on the scanning display terminal, the set prompt interface can also display a designated user name and a password, and the interactive terminal is added into the interactive scene through a designated user and the password.
Fig. 8 illustrates an interaction flow diagram of a multi-screen interaction method according to an embodiment of the present application. As shown in fig. 8, the method includes steps S810 to S890:
and step S810, connecting the display end with the cloud server, and joining the interactive scene.
In an embodiment of the application, the display end can operate a browser, and the display end can establish a Websocket data transmission channel with the cloud server through the browser to achieve data transmission.
After the connection is established, the display end can send an access request for requesting to join the interactive scene to the cloud server through the Websocket data transmission channel, and the cloud server can feed back response information aiming at the access request to the display end, so that the interactive scene is joined.
And S820, generating a set prompt interface by the display end.
After receiving response information for the join request fed back by the cloud server, the display terminal may generate a set prompt interface according to the response information. The set prompt interface can be a two-dimensional code interface of an interactive scene, the interactive terminal can join the interactive scene through the two-dimensional code interface on the scanning display terminal, the set prompt interface can also display a designated user name and a password, and the interactive terminal joins the interactive scene through inputting the designated user and the password on the interactive terminal interface.
And S830, connecting the interactive end with the cloud server, and adding the interactive scene.
In an embodiment of the application, the interactive end may also establish a Websocket data transmission channel with the cloud server through the browser to implement data transmission. After the connection is established, the interactive end can send a joining request for requesting to join the interactive scene to the cloud server through the Websocket data transmission channel, and the cloud server can feed back response information aiming at the joining request to the interactive end, so that the joining of the interactive scene is completed. Specifically, the interaction end can send a joining request for requesting joining of the interaction scene to the cloud server through a prompt interface of the display end, for example, when the prompt interface is a two-dimensional code, the interaction end can start a scanning function, scan the two-dimensional code, and join the interaction scene.
Step 840, the cloud server receives an interaction starting instruction triggered by the display terminal.
And step S850, the cloud server sends an opening message to the interactive terminal.
And S860, the interaction end starts interaction and sends an interaction message to the cloud server.
Step S870, the cloud server sends the interactive message to a display terminal.
And step S880, the display end judges whether the interaction end completes the interaction according to the interaction message, if so, the display end sends an interaction completion message to the cloud server, and simultaneously displays the interaction reward information of the interaction end.
Step 890, the cloud server sends the interactive reward information to the interactive terminal.
It should be noted that, step S870-step S890 are an embodiment of processing the interactive message, and in other embodiments of the present application, the processing of the interactive message may be that the cloud server may determine whether the interactive end completes the interaction after receiving the interactive message sent by the interactive end, and if so, send the interactive reward information to the interactive end, and meanwhile, the cloud server may also send the interactive completion message to the display end, and after receiving the interactive completion message, the display end may display the interactive reward information to the interactive end.
It should be further noted that, after the interaction is started at step S860 and the interaction message is sent to the cloud server, in addition to executing step S870, the cloud server may further select a target streaming media file matched with the real-time interaction information included in the interaction message from the multiple streaming media files associated with the user identifier, and after selecting the target streaming media file, the cloud server may send the target streaming media file to the display terminal for display.
Fig. 9 is a logic decision diagram of a multi-screen interaction method according to an embodiment of the present application, and as shown in fig. 9, the logic decision diagram specifically includes:
step S910, whether the interactive end joins the interactive scene, if yes, step S920 is executed, and if not, the interaction is ended.
Step S920, the interactive terminal waits for the start of the interaction, and if so, performs step S930.
Step S930, the interactive end starts to interact.
Step S940, whether the interaction score of the interaction end reaches a preset threshold, if yes, step S950 is executed, and if not, the interaction end may restart to perform interaction.
And S950, the interactive end draws the interactive reward and the interaction is finished.
Fig. 10 illustrates an interface schematic diagram of a multi-screen interaction method at a display end according to an embodiment of the present application, in an example shown in fig. 10, an interaction scene is a game "fire line 666", after each interaction end starts to perform a game "fire line 666" interaction, the display end may display a target streaming media file corresponding to the interaction end in a text stream format, where the target streaming media file is matched with real-time interaction information of the interaction end, for example, when the interaction end where "AG" is located completes the interaction, the display end may display "AG completes the fire line 666", "AG is not enemy", "AG 666", "AG is too 6", "KZ" is located at the interaction end, and the display end may display "KZ does not exist".
Fig. 11 illustrates an interface schematic diagram of a multi-screen interaction method at an interaction end according to an embodiment of the present application, in an example illustrated in fig. 11, an interaction scene is a game "fire line 666", a page of the interaction scene provides views for a user, which can be divided into a main interface, an interaction interface and an information interface, where the main interface is an interface for the user to perform game operations, the interaction interface is an interface for displaying interaction situations between the users, such as chatting, cheering, praise, and the like, and the information interface is used for displaying information such as a highlight photo, an interaction ranking, and the like.
The user can switch between different interfaces through the controls below the page, for example, the main interface corresponds to the control "fire wire 666", the interactive interface corresponds to the control "chat room", and the information interface corresponds to the control "highlight photograph".
The following describes embodiments of the apparatus of the present application, which can be used to perform the multi-screen interaction method in the foregoing embodiments of the present application. For details that are not disclosed in the embodiments of the apparatus of the present application, please refer to the embodiments of the multi-screen interaction method described above in the present application.
Fig. 12 is a block diagram illustrating a multi-screen interaction apparatus according to an embodiment of the present application, and referring to fig. 12, a multi-screen interaction apparatus 1200 according to an embodiment of the present application includes: a first receiving unit 1202, a second receiving unit 1204, a selecting unit 1206 and a control unit 1208.
The first receiving unit 1202 is configured to receive a join request sent by an interactive end and used for requesting to join in an interactive scene, where the join request includes a user identifier corresponding to the interactive end; a second receiving unit 1204, configured to receive an interactive message sent by the interactive end after the interactive end joins the interactive scene, where the interactive message includes real-time interactive information; a selecting unit 1206, configured to select a target streaming media file matching the real-time interaction information from a plurality of streaming media files associated with the user identifier; the control unit 1208 is configured to control the display end of the interactive scene to display the target streaming media file according to the target streaming media file.
In some embodiments of the present application, the first receiving unit 1202 is configured to: receiving an access request sent by the display end, and feeding back response information aiming at the access request to the display end, wherein the response information is used for indicating the display end to display a set prompt interface; and receiving the joining request sent by the interactive terminal through the prompt interface.
In some embodiments of the present application, the second receiving unit 1204 is configured to: and if an interaction starting instruction is received, sending a starting message to the interaction end, wherein the starting message is used for indicating the interaction end to generate an interaction interface corresponding to the interaction scene and detecting interaction information aiming at the interaction scene.
In some embodiments of the present application, the selection unit 1206 is configured to: and according to the user identification, determining identification information of a plurality of streaming media files related to the user identification from a relation table between the interactive scene and the identification information of the streaming media files, and selecting the identification information of a target streaming media file matched with the real-time interactive information from the identification information of the plurality of streaming media files.
In some embodiments of the present application, the control unit 1208 is configured to: generating an interactive response message containing the identification information of the target streaming media file, and sending the interactive response message to a streaming media server so that the streaming media server sends the target streaming media file to the display terminal for displaying according to the identification information of the target streaming media file; or sending the target streaming media file to a display end of the interactive scene so that the display end displays the target streaming media file; or generating an interactive response message containing the identification information of the target streaming media file, and sending the interactive response message to the display end, so that the display end acquires the target streaming media file according to the identification information of the target streaming media file and displays the acquired target streaming media file.
In some embodiments of the present application, the real-time interaction information includes an interaction score, and the second receiving unit 1204 is configured to: receiving an interaction score sent by the interaction end, sending the interaction score to the display end, and sending interaction reward information to the interaction end if an interaction completion message sent by the display end according to the interaction score is received; or receiving the interaction score sent by the interaction end, if the interaction end is determined to finish the interaction according to the interaction score, sending interaction reward information to the interaction end, and sending interaction finish information to the display end, so that the display end displays the interaction reward information to the interaction end.
In some embodiments of the present application, the control unit 1208 is configured to: and counting the number of the interactive ends added into the interactive scene, and sending the counted number to the display end for displaying.
FIG. 13 illustrates a block diagram of a multi-screen interaction device, according to an embodiment of the present application.
Referring to fig. 13, a multi-screen interaction device 1300 according to an embodiment of the present application includes: a transmission unit 1302, an acquisition unit 1304, and a display unit 1306.
The sending unit 1302 is configured to send an access request for requesting to join an interactive scene to the cloud server, where the interactive scene includes multiple interactive terminals; an obtaining unit 1304, configured to obtain, after joining the interactive scene, a target streaming media file corresponding to each interactive end, where the target streaming media file is selected by the cloud server from multiple streaming media files associated with the user identifier corresponding to each interactive end according to the interactive message sent by each interactive end, and the target streaming media file corresponding to each interactive end matches with real-time interaction information included in the interactive message sent by each interactive end; a display unit 1306, configured to display target streaming media files corresponding to the interaction ends.
In some embodiments of the present application, the obtaining unit 1304 is configured to: receiving the interaction scores of the interaction ends sent by the cloud server, and displaying interaction reward information of the target interaction end for the target interaction end with the interaction score reaching a preset threshold value; or receiving information of a target interaction end, sent by the cloud server, of which the interaction score reaches a preset threshold value, and displaying interaction reward information of the target interaction end.
In some embodiments of the present application, the obtaining unit 1304 is configured to: receiving response information fed back by the cloud server for the access request; and generating a set prompt interface according to the response information, wherein the prompt interface is used for the interaction terminal to join the interaction scene.
In some embodiments of the present application, based on the foregoing solution, the display unit 1306 is configured to: if the target streaming media files corresponding to the interactive ends are obtained at the same time, determining a display sequence according to the user identification corresponding to each interactive end; and displaying the target streaming media files corresponding to the interactive ends according to the display sequence.
FIG. 14 illustrates a schematic structural diagram of a computer system suitable for use in implementing the electronic device of an embodiment of the present application.
It should be noted that the computer system 1400 of the electronic device shown in fig. 14 is only an example, and should not bring any limitation to the functions and the scope of use of the embodiments of the present application.
As shown in fig. 14, a computer system 1400 includes a Central Processing Unit (CPU)1401, which can perform various appropriate actions and processes, such as executing the methods described in the above embodiments, according to a program stored in a Read-Only Memory (ROM) 1402 or a program loaded from a storage portion 1408 into a Random Access Memory (RAM) 1403. In the RAM 1403, various programs and data necessary for system operation are also stored. The CPU 1401, ROM 1402, and RAM 1403 are connected to each other via a bus 1404. An Input/Output (I/O) interface 1405 is also connected to the bus 1404.
The following components are connected to the I/O interface 1405: an input portion 1406 including a keyboard, a mouse, and the like; an output portion 1407 including a Cathode Ray Tube (CRT), a Liquid Crystal Display (LCD), and the like, a speaker, and the like; a storage portion 1408 including a hard disk and the like; and a communication section 1409 including a Network interface card such as a LAN (Local Area Network) card, a modem, and the like. The communication section 1409 performs communication processing via a network such as the internet. The driver 1410 is also connected to the I/O interface 1405 as necessary. A removable medium 1411 such as a magnetic disk, an optical disk, a magneto-optical disk, a semiconductor memory, or the like is mounted on the drive 1410 as necessary, so that a computer program read out therefrom is installed into the storage section 1408 as necessary.
In particular, according to embodiments of the application, the processes described above with reference to the flow diagrams may be implemented as computer software programs. For example, embodiments of the present application include a computer program product comprising a computer program embodied on a computer readable medium, the computer program comprising a computer program for performing the method illustrated by the flow chart. In such an embodiment, the computer program may be downloaded and installed from a network via the communication portion 1409 and/or installed from the removable medium 1411. When the computer program is executed by a Central Processing Unit (CPU)1401, various functions defined in the system of the present application are executed.
It should be noted that the computer readable medium shown in the embodiments of the present application may be a computer readable signal medium or a computer readable storage medium or any combination of the two. A computer readable storage medium may be, for example, but not limited to, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, or device, or any combination of the foregoing. More specific examples of the computer readable storage medium may include, but are not limited to: an electrical connection having one or more wires, a portable computer diskette, a hard disk, a Random Access Memory (RAM), a Read-Only Memory (ROM), an Erasable Programmable Read-Only Memory (EPROM), a flash Memory, an optical fiber, a portable Compact Disc Read-Only Memory (CD-ROM), an optical storage device, a magnetic storage device, or any suitable combination of the foregoing. In the present application, a computer readable storage medium may be any tangible medium that can contain, or store a program for use by or in connection with an instruction execution system, apparatus, or device. In this application, however, a computer readable signal medium may include a propagated data signal with a computer program embodied therein, for example, in baseband or as part of a carrier wave. Such a propagated data signal may take many forms, including, but not limited to, electro-magnetic, optical, or any suitable combination thereof. A computer readable signal medium may also be any computer readable medium that is not a computer readable storage medium and that can communicate, propagate, or transport a program for use by or in connection with an instruction execution system, apparatus, or device. The computer program embodied on the computer readable medium may be transmitted using any appropriate medium, including but not limited to: wireless, wired, etc., or any suitable combination of the foregoing.
The flowchart and block diagrams in the figures illustrate the architecture, functionality, and operation of possible implementations of systems, methods and computer program products according to various embodiments of the present application. Each block in the flowchart or block diagrams may represent a module, segment, or portion of code, which comprises one or more executable instructions for implementing the specified logical function(s). It should also be noted that, in some alternative implementations, the functions noted in the block may occur out of the order noted in the figures. For example, two blocks shown in succession may, in fact, be executed substantially concurrently, or the blocks may sometimes be executed in the reverse order, depending upon the functionality involved. It will also be noted that each block of the block diagrams or flowchart illustration, and combinations of blocks in the block diagrams or flowchart illustration, can be implemented by special purpose hardware-based systems which perform the specified functions or acts, or combinations of special purpose hardware and computer instructions.
The units described in the embodiments of the present application may be implemented by software, or may be implemented by hardware, and the described units may also be disposed in a processor. Wherein the names of the elements do not in some way constitute a limitation on the elements themselves.
As another aspect, the present application also provides a computer-readable medium, which may be contained in the electronic device described in the above embodiments; or may exist separately without being assembled into the electronic device. The computer readable medium carries one or more programs which, when executed by an electronic device, cause the electronic device to implement the method described in the above embodiments.
It should be noted that although in the above detailed description several modules or units of the device for action execution are mentioned, such a division is not mandatory. Indeed, the features and functionality of two or more modules or units described above may be embodied in one module or unit, according to embodiments of the application. Conversely, the features and functions of one module or unit described above may be further divided into embodiments by a plurality of modules or units.
Through the above description of the embodiments, those skilled in the art will readily understand that the exemplary embodiments described herein may be implemented by software, or by software in combination with necessary hardware. Therefore, the technical solution according to the embodiments of the present application can be embodied in the form of a software product, which can be stored in a non-volatile storage medium (which can be a CD-ROM, a usb disk, a removable hard disk, etc.) or on a network, and includes several instructions to enable a computing device (which can be a personal computer, a server, a touch terminal, or a network device, etc.) to execute the method according to the embodiments of the present application.
Other embodiments of the present application will be apparent to those skilled in the art from consideration of the specification and practice of the embodiments disclosed herein. This application is intended to cover any variations, uses, or adaptations of the invention following, in general, the principles of the application and including such departures from the present disclosure as come within known or customary practice within the art to which the invention pertains.
It will be understood that the present application is not limited to the precise arrangements described above and shown in the drawings and that various modifications and changes may be made without departing from the scope thereof. The scope of the application is limited only by the appended claims.
Claims (15)
1. A multi-screen interaction method is characterized by comprising the following steps:
receiving a joining request which is sent by an interactive end and used for requesting to join an interactive scene, wherein the joining request comprises a user identifier corresponding to the interactive end;
after the interactive end joins the interactive scene, receiving an interactive message sent by the interactive end, wherein the interactive message comprises real-time interactive information;
selecting a target streaming media file matched with the real-time interaction information from a plurality of streaming media files associated with the user identification;
and controlling a display end of the interactive scene to display the target streaming media file according to the target streaming media file.
2. The method according to claim 1, wherein before receiving a join request for requesting to join an interactive scene sent by an interactive end, the method further comprises: receiving an access request sent by the display end, and feeding back response information aiming at the access request to the display end, wherein the response information is used for indicating the display end to display a set prompt interface;
receiving a joining request sent by an interactive terminal and used for requesting to join an interactive scene, comprising the following steps: and receiving the joining request sent by the interactive terminal through the prompt interface.
3. The method of claim 1, further comprising:
and if an interaction starting instruction is received, sending a starting message to the interaction end, wherein the starting message is used for indicating the interaction end to generate an interaction interface corresponding to the interaction scene and detecting interaction information aiming at the interaction scene.
4. The method of claim 1, wherein selecting a target streaming media file matching the real-time interaction information from a plurality of streaming media files associated with the user identifier comprises:
and according to the user identification, determining identification information of a plurality of streaming media files related to the user identification from a relation table between the interactive scene and the identification information of the streaming media files, and selecting the identification information of a target streaming media file matched with the real-time interactive information from the identification information of the plurality of streaming media files.
5. The method of claim 1, wherein controlling a display end of the interactive scene to display the target streaming media file according to the target streaming media file comprises:
generating an interactive response message containing the identification information of the target streaming media file, and sending the interactive response message to a streaming media server so that the streaming media server sends the target streaming media file to the display terminal for displaying according to the identification information of the target streaming media file; or
Sending the target streaming media file to a display end of the interactive scene so that the display end displays the target streaming media file; or
And generating an interactive response message containing the identification information of the target streaming media file, and sending the interactive response message to the display end, so that the display end acquires the target streaming media file according to the identification information of the target streaming media file and displays the acquired target streaming media file.
6. The method of claim 1, wherein the real-time interaction information comprises an interaction score, and the method further comprises:
receiving an interaction score sent by the interaction end, sending the interaction score to the display end, and sending interaction reward information to the interaction end if an interaction completion message sent by the display end according to the interaction score is received; or
And receiving the interaction score sent by the interaction end, if the interaction end completes the interaction according to the interaction score, sending interaction reward information to the interaction end, and sending interaction completion information to the display end, so that the display end displays the interaction reward information to the interaction end.
7. The method according to any one of claims 1 to 6, further comprising:
and counting the number of the interactive ends added into the interactive scene, and sending the counted number to the display end for displaying.
8. A multi-screen interaction method is characterized by comprising the following steps:
sending an access request for requesting to join an interactive scene to a cloud server, wherein the interactive scene comprises a plurality of interactive ends;
after the interaction scene is added, acquiring a target streaming media file corresponding to each interaction terminal, wherein the target streaming media file is selected from a plurality of streaming media files associated with user identifications corresponding to each interaction terminal by the cloud server according to interaction messages sent by each interaction terminal, and the target streaming media file corresponding to each interaction terminal is matched with real-time interaction information contained in the interaction messages sent by each interaction terminal;
and displaying the target streaming media files corresponding to the interactive ends.
9. The method of claim 8, wherein the real-time interaction information comprises an interaction score, and the method further comprises:
receiving the interaction scores of the interaction ends sent by the cloud server, and displaying interaction reward information of the target interaction end for the target interaction end with the interaction score reaching a preset threshold value; or
And receiving information of a target interaction end, which is sent by the cloud server and has an interaction score reaching a preset threshold value, and displaying interaction reward information of the target interaction end.
10. The method of claim 8, further comprising:
receiving response information fed back by the cloud server for the access request;
and generating a set prompt interface according to the response information, wherein the prompt interface is used for the interaction terminal to join the interaction scene.
11. The method according to claim 8, wherein displaying the target streaming media file corresponding to each interactive end comprises:
if the target streaming media files corresponding to the interactive ends are obtained at the same time, determining a display sequence according to the user identification corresponding to each interactive end;
and displaying the target streaming media files corresponding to the interactive ends according to the display sequence.
12. A multi-screen interaction device, the device comprising:
the first receiving unit is configured to receive a join request which is sent by an interactive end and used for requesting to join an interactive scene, wherein the join request comprises a user identifier corresponding to the interactive end;
the second receiving unit is configured to receive an interactive message sent by the interactive end after the interactive end joins the interactive scene, wherein the interactive message comprises real-time interactive information;
the selecting unit is configured to select a target streaming media file matched with the real-time interaction information from a plurality of streaming media files associated with the user identifier;
and the control unit is configured to control a display end of the interactive scene to display the target streaming media file according to the target streaming media file.
13. A multi-screen interaction device, the device comprising:
the system comprises a sending unit, a receiving unit and a processing unit, wherein the sending unit is configured to send a joining request for requesting to join an interactive scene to a cloud server, and the interactive scene comprises a plurality of interactive ends;
an obtaining unit, configured to obtain, after joining the interactive scene, a target streaming media file corresponding to each interactive end, where the target streaming media file is selected by the cloud server from a plurality of streaming media files associated with the user identifier corresponding to each interactive end according to the interactive message sent by each interactive end, and the target streaming media file corresponding to each interactive end matches with real-time interactive information included in the interactive message sent by each interactive end;
and the display unit is configured to display the target streaming media files corresponding to the interaction ends.
14. A computer-readable medium, on which a computer program is stored, which, when being executed by a processor, implements the multi-screen interaction method according to any one of claims 1 to 7, or implements the multi-screen interaction method according to any one of claims 8 to 11.
15. An electronic device, comprising:
one or more processors;
a storage device for storing one or more programs that, when executed by the one or more processors, cause the one or more processors to implement the multi-screen interaction method of any one of claims 1 to 7, or the multi-screen interaction method of any one of claims 8 to 11.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202010831939.8A CN112055252A (en) | 2020-08-18 | 2020-08-18 | Multi-screen interaction method and device, computer readable medium and electronic equipment |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202010831939.8A CN112055252A (en) | 2020-08-18 | 2020-08-18 | Multi-screen interaction method and device, computer readable medium and electronic equipment |
Publications (1)
Publication Number | Publication Date |
---|---|
CN112055252A true CN112055252A (en) | 2020-12-08 |
Family
ID=73600071
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202010831939.8A Pending CN112055252A (en) | 2020-08-18 | 2020-08-18 | Multi-screen interaction method and device, computer readable medium and electronic equipment |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN112055252A (en) |
Cited By (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN114422965A (en) * | 2022-01-14 | 2022-04-29 | Tcl通讯科技(成都)有限公司 | Information interaction method and device, storage medium and electronic equipment |
CN114785848A (en) * | 2022-03-02 | 2022-07-22 | 阿里巴巴(中国)有限公司 | Collaborative interaction and collaboration method, device and system between electronic devices |
WO2022252909A1 (en) * | 2021-06-03 | 2022-12-08 | Oppo广东移动通信有限公司 | Screen casting method and apparatus in wireless terminal, wireless terminal, and storage medium |
CN116627370A (en) * | 2023-04-12 | 2023-08-22 | 广东敏行数据有限公司 | Display screen updating content processing method, storage medium, device and system |
Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20130097635A1 (en) * | 2011-10-13 | 2013-04-18 | Gface Gmbh | Interactive remote participation in live entertainment |
CN104645614A (en) * | 2015-03-02 | 2015-05-27 | 郑州三生石科技有限公司 | Multi-player video on-line game method |
CN110225412A (en) * | 2019-07-05 | 2019-09-10 | 腾讯科技(深圳)有限公司 | Video interaction method, device and storage medium |
-
2020
- 2020-08-18 CN CN202010831939.8A patent/CN112055252A/en active Pending
Patent Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20130097635A1 (en) * | 2011-10-13 | 2013-04-18 | Gface Gmbh | Interactive remote participation in live entertainment |
CN104645614A (en) * | 2015-03-02 | 2015-05-27 | 郑州三生石科技有限公司 | Multi-player video on-line game method |
CN110225412A (en) * | 2019-07-05 | 2019-09-10 | 腾讯科技(深圳)有限公司 | Video interaction method, device and storage medium |
Cited By (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2022252909A1 (en) * | 2021-06-03 | 2022-12-08 | Oppo广东移动通信有限公司 | Screen casting method and apparatus in wireless terminal, wireless terminal, and storage medium |
CN114422965A (en) * | 2022-01-14 | 2022-04-29 | Tcl通讯科技(成都)有限公司 | Information interaction method and device, storage medium and electronic equipment |
CN114785848A (en) * | 2022-03-02 | 2022-07-22 | 阿里巴巴(中国)有限公司 | Collaborative interaction and collaboration method, device and system between electronic devices |
CN116627370A (en) * | 2023-04-12 | 2023-08-22 | 广东敏行数据有限公司 | Display screen updating content processing method, storage medium, device and system |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US11962861B2 (en) | Live streaming room red packet processing method and apparatus, and medium and electronic device | |
WO2022087920A1 (en) | Video playing method and apparatus, and terminal and storage medium | |
CN112055252A (en) | Multi-screen interaction method and device, computer readable medium and electronic equipment | |
WO2022188595A1 (en) | Method and apparatus for displaying application picture, and terminal, screen projection system and medium | |
KR20210135683A (en) | Method, system, and computer program for displaying reaction during voip-based call | |
CN112073754B (en) | Cloud game screen projection method and device, computer equipment, computer readable storage medium and cloud game screen projection interaction system | |
CN111586452A (en) | Cross-device interaction method and device and playing device | |
CN109819268B (en) | Live broadcast room play control method, device, medium and equipment in video live broadcast | |
KR20180027565A (en) | METHOD AND APPARATUS FOR PERFORMING SERVICE OPERATION BASED ON CHAT GROUP, AND METHOD AND APPARATUS FOR ACQUIRING GROUP MEMBER INFORMATION | |
WO2021204139A1 (en) | Video displaying method, device, equipment, and storage medium | |
CN111131850A (en) | Method and device for displaying special effect of virtual gift and electronic equipment | |
CN113840154A (en) | Live broadcast interaction method and system based on virtual gift and computer equipment | |
CN112511849A (en) | Game display method, device, equipment, system and storage medium | |
CN114727146A (en) | Information processing method, device, equipment and storage medium | |
CN111277890A (en) | Method for acquiring virtual gift and method for generating three-dimensional panoramic live broadcast room | |
CN112153457A (en) | Wireless screen projection connection method and device, computer storage medium and electronic equipment | |
CN109819341B (en) | Video playing method and device, computing equipment and storage medium | |
CN111147885B (en) | Live broadcast room interaction method and device, readable medium and electronic equipment | |
CN114125480B (en) | Live chorus interaction method, system, device and computer equipment | |
CN115052169A (en) | Interaction method, device, electronic equipment and computer readable medium | |
CN114489891A (en) | Control method, system, device, readable medium and equipment of cloud application program | |
CN105744338B (en) | A kind of method for processing video frequency and its equipment | |
CN112295217B (en) | Device joining method and device, electronic device and computer readable medium | |
CN114257572B (en) | Data processing method, device, computer readable medium and electronic equipment | |
WO2024002162A1 (en) | Method and apparatus for interaction in live-streaming room, and device and medium |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
REG | Reference to a national code |
Ref country code: HK Ref legal event code: DE Ref document number: 40035389 Country of ref document: HK |
|
RJ01 | Rejection of invention patent application after publication |
Application publication date: 20201208 |
|
RJ01 | Rejection of invention patent application after publication |