CN112104887B - Video source checking method, device, equipment and storage medium - Google Patents

Video source checking method, device, equipment and storage medium Download PDF

Info

Publication number
CN112104887B
CN112104887B CN202010972503.0A CN202010972503A CN112104887B CN 112104887 B CN112104887 B CN 112104887B CN 202010972503 A CN202010972503 A CN 202010972503A CN 112104887 B CN112104887 B CN 112104887B
Authority
CN
China
Prior art keywords
video
video source
audio
video stream
playing
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202010972503.0A
Other languages
Chinese (zh)
Other versions
CN112104887A (en
Inventor
许英俊
曾衍
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Guangzhou Cubesili Information Technology Co Ltd
Original Assignee
Guangzhou Cubesili Information Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Guangzhou Cubesili Information Technology Co Ltd filed Critical Guangzhou Cubesili Information Technology Co Ltd
Priority to CN202010972503.0A priority Critical patent/CN112104887B/en
Publication of CN112104887A publication Critical patent/CN112104887A/en
Application granted granted Critical
Publication of CN112104887B publication Critical patent/CN112104887B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/21Server components or server architectures
    • H04N21/218Source of audio or video content, e.g. local disk arrays
    • H04N21/2187Live feed
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/431Generation of visual interfaces for content selection or interaction; Content or additional data rendering
    • H04N21/4312Generation of visual interfaces for content selection or interaction; Content or additional data rendering involving specific graphical features, e.g. screen layout, special fonts or colors, blinking icons, highlights or animations
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/45Management operations performed by the client for facilitating the reception of or the interaction with the content or administrating data related to the end-user or to the client device itself, e.g. learning user preferences for recommending movies, resolving scheduling conflicts
    • H04N21/454Content or additional data filtering, e.g. blocking advertisements
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/45Management operations performed by the client for facilitating the reception of or the interaction with the content or administrating data related to the end-user or to the client device itself, e.g. learning user preferences for recommending movies, resolving scheduling conflicts
    • H04N21/462Content or additional data management, e.g. creating a master electronic program guide from data received from the Internet and a Head-end, controlling the complexity of a video stream by scaling the resolution or bit-rate based on the client capabilities
    • H04N21/4622Retrieving content or additional data from different sources, e.g. from a broadcast channel and the Internet
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/47End-user applications
    • H04N21/478Supplemental services, e.g. displaying phone caller identification, shopping application
    • H04N21/4784Supplemental services, e.g. displaying phone caller identification, shopping application receiving rewards

Abstract

The application discloses a video source checking method, a device, equipment and a storage medium, wherein the method comprises the following steps: correspondingly playing a plurality of video streams from different remote video sources in a plurality of playing windows, wherein the image and the audio of each video stream are in a suppressed state, so that the video sources of the video streams cannot be identified; sequentially removing the suppressed state of the audio of each video stream, and continuously maintaining the image in the suppressed state; receiving user input of selecting video source identification for each playing window, and checking whether the video source identification is correctly matched with the video source; and outputting the verification result information. The method and the device construct a verification mechanism for verifying the identification capability of the user video source technically, and are beneficial to the deep development and development of the network live broadcast application technology.

Description

Video source checking method, device, equipment and storage medium
Technical Field
The present application relates to the field of live webcasting technologies, and in particular, to a method, an apparatus, a device, and a storage medium for verifying a video source.
Background
The network live broadcast technology is usually in the form of a live broadcast room, a user in the live broadcast room can be an anchor user through applying for setting up the live broadcast room, and further, the anchor user can attract other users to join in the live broadcast room to watch video contents made by the anchor user through hosting the video contents in the live broadcast room, so that the purpose of transmitting audio and video information through network live broadcast is achieved.
One type of network live broadcast activity based on the live broadcast room is that a plurality of anchor users can jointly hold a connection activity, information is cooperatively spread and a common task is completed through the connection activity between the anchor users, and under the mode of the connection activity, the two live broadcast rooms are generally in a state that information can be intercommunicated. The interaction can be further carried out among the live broadcasting rooms, and activities are cooperatively executed together, for example, two anchor users can sing the same song for audience users of the two live broadcasting rooms to enjoy, so that the purpose of utilizing the live broadcasting rooms to realize support on-line chorus is achieved.
The live network connection activity needs a plurality of service logics for support, and the existing various connection activity modes cannot meet the increasing mental requirements of audience users, so that various novel service logic innovations are urgently needed for realizing the support. By providing various novel business logics, the user activity of the live broadcast room can be further improved, the operation of an economic transaction system of the live broadcast room is stimulated, and other related network activities are brought. Indeed, these new business logics all require support from the underlying technology.
The online activities are often temporarily initiated, and the combinations of the anchor users may also be random, so that the audience users may not be familiar with the anchor users themselves and their styles, but in some business logics, it is necessary to check whether the audience users have the ability to identify the anchor users, so that the parties can collect data or even make a judgment based on the ability. For computer devices, the authentication of this capability needs to be facilitated by technical means.
In the prior art, no application example with the authentication capability is seen in the field of live webcasting, so that further development and enrichment of live webcasting connection activities are limited to some extent.
Disclosure of Invention
The primary objective of the present application is to provide a video source verification method, so as to provide technical support for identifying video sources for audience users in live network connection activities.
As another object of the present application, a video source verification apparatus, an electronic device, and a non-volatile storage medium, which are compatible with the foregoing method, are provided.
In order to meet various purposes of the application, the following technical scheme is adopted in the application:
a video source verification method adapted to the first objective of the present application includes the following steps:
correspondingly playing a plurality of video streams from different remote video sources in a plurality of playing windows, wherein the image and the audio of each video stream are in a suppressed state, so that the video sources of the video streams cannot be identified;
sequentially removing the suppressed state of the audio of each video stream, and continuously maintaining the image in the suppressed state;
receiving user input of selecting video source identification for each playing window, and checking whether the video source identification is correctly matched with the video source;
and outputting the verification result information.
In one embodiment, in the step of making the video source unrecognizable because the image and the audio of each video stream are in the suppressed state, the video stream image is made to be in the suppressed state by performing image synthesis processing on the video stream or by overlapping a mask layer on the playing window to interfere with the video stream image.
In another class of embodiments, in the step of suppressing both the image and the audio of each video stream to make the video source thereof unrecognizable, the audio of the video stream is made to be in the suppressed state by performing audio synthesis processing on the video stream/by setting the volume of the video stream to mute.
In an optional embodiment, in the step of sequentially removing the suppressed state of the audio of each video stream and continuously maintaining the image in the suppressed state, after the suppressed state of the audio of one video stream is removed and maintained for a predetermined time, the suppressed state of the audio of the other video stream is restored, and then the same processing is performed on the audio of the other video stream until the same processing is performed on the audio of all the video streams.
In a preferred embodiment, the step of receiving user input selecting a video source identifier for each playback window is performed after removing the suppressed state of audio of the first video stream until said user input is completed.
In one class of embodiments, in the step of receiving a user input selecting a video source identifier for each of the playback windows, the user input is a correspondence between a video source indication picture provided in the graphical user interface and confirmed by the user and the playback window.
In one embodiment, in the step of checking whether the video source identifier and the video source are correctly matched, the video source identifiers are compared one by one according to the video source information of the video streams provided by the server, when the video source identifiers correspond to the video source information, the video source identifier and the video source are determined to be matched, otherwise, the video source identifier and the video source identifier are determined to be not matched.
In one embodiment, in the step of outputting the verification result information, whether the video source identifiers are correctly matched is displayed one by one to output the corresponding verification result information, and the verification result information is submitted to a server to be output.
In some embodiments, the method comprises the following pre-steps:
and receiving a verification starting instruction, and rebuilding the corresponding playing relation between the plurality of playing windows and the plurality of video streams in the graphical user interface.
In some embodiments, the method further comprises the following post-steps:
and removing the video and audio suppressed states of the video streams of all the playing windows to restore the normal playing.
In a preferred embodiment, the method further comprises the following post-steps:
and in the step of sequentially removing the suppressed state of the audio of each path of video stream, in the process of removing the suppressed state of the audio of one path of video stream, the playing window corresponding to the video stream is marked with a highlight.
In a further embodiment, the method comprises the steps of:
receiving electronic gift receiving entry information pushed by the server according to the inspection result information, and visually displaying the receiving entry on a graphical user interface;
and responding to the instruction of the user for accessing the receiving inlet and submitting corresponding instruction to the server to realize the acquisition of the electronic gift for the user.
Another object of the present application is to provide a video source verification apparatus, which includes:
the video playing module is used for correspondingly playing a plurality of video streams from different remote video sources in a plurality of playing windows, and the image and the audio of each video stream are in a suppressed state so that the video source of each video stream can not be identified;
the polling suppression module is used for sequentially removing the suppressed state of the audio of each path of video stream and continuously maintaining the image in the suppressed state;
the identification checking module is used for receiving user input of selecting a video source identification for each playing window and checking whether the video source identification is correctly matched with the video source or not;
and the result output module is used for outputting the verification result information.
Another object of the present application is to provide an electronic device, which includes a central processing unit and a memory, wherein the central processing unit is used to invoke and run a computer program stored in the memory to execute the steps of the video source verification method described in the present application.
A non-volatile storage medium storing a computer program according to the video source verification method is provided, and when the computer program is called by a computer, the computer program performs the steps included in the method.
Compared with the prior art, the application has the following advantages:
firstly, the video streams of a plurality of playing windows are technically inhibited, then the inhibited state of the images in the video streams is maintained by sequentially removing the inhibited state of the audio in the video streams, the user input is received to obtain the judgment of the video source identification of each video stream by the user, and on the basis, whether the user input is matched with the fact of the video source is checked, so that a technical support for checking whether the user can accurately identify each video source is provided.
Secondly, the method fully considers the characteristics of the two contents of the image and the audio contained in the video stream, the image is always kept in the restrained state, the restrained state is properly removed from the audio, so that the anchor user is allowed to prompt the user by providing the sound information including the original sound of the anchor user, the user can obtain a proper amount of materials to be used as an identification basis by combining the characteristics of the sound and the information content transmitted by the sound, the correct judgment of the user is facilitated, meanwhile, a proper fuzzy space is kept for the user identification due to the fact that the corresponding relation between the playing window and the video stream is unknown, and therefore the corresponding verification result information can be further obtained through computer verification, the thinking duration of the user is prolonged, and the online survival rate of the user in the live network broadcast is ensured.
In addition, the application provides a video source identifier, and a user selects a video source identifier for a certain playing window to indicate that a video stream played by the playing window is a video source of the video source identifier, so that the video source identifier can be represented in advance to have a certain indication meaning, so that the user can establish a relation between information conveyed by sound of one video stream according to the indication meaning, according to the relation, the user can relatively more reliably select a proper video source identifier for the playing window, and the background can also improve or reduce the user identification difficulty by adjusting the indication meaning expressed by the video source identifier, thereby improving the controllability of the technical support framework provided by the application.
Additional aspects and advantages of the present application will be set forth in part in the description which follows and, in part, will be obvious from the description, or may be learned by practice of the present application.
Drawings
The foregoing and/or additional aspects and advantages of the present application will become apparent and readily appreciated from the following description of the embodiments, taken in conjunction with the accompanying drawings of which:
fig. 1 is a schematic diagram of a typical network deployment architecture related to implementing the technical solution of the present application;
FIG. 2 is a flowchart illustrating an exemplary embodiment of a video source verification method according to the present application;
fig. 3 is a schematic diagram of a display state formed by a gui representation of a terminal device when the application program of the present invention is running in a live broadcast room;
FIG. 4 is a partial schematic view of a display state shown and formed on a graphical user interface after a verification result signal is obtained according to an activity rule evolution based on FIG. 3;
fig. 5 is a partial schematic view of a display state formed by a graphical user interface in the running process of an application program in the live broadcast room after the application program is applied to the application program, and is mainly used for explaining an interface prompt when a program background reconstructs a corresponding relationship between a play window and a video stream;
FIG. 6 is a flowchart illustrating a video source verification method according to another embodiment of the present application;
fig. 7 is a schematic block diagram of an exemplary embodiment of a video source verification apparatus according to the present application.
Detailed Description
Reference will now be made in detail to embodiments of the present application, examples of which are illustrated in the accompanying drawings, wherein like or similar reference numerals refer to the same or similar elements or elements having the same or similar function throughout. The embodiments described below with reference to the drawings are exemplary only for the purpose of explaining the present application and are not to be construed as limiting the present application.
As used herein, the singular forms "a", "an", "the" and "the" are intended to include the plural forms as well, unless the context clearly indicates otherwise. It will be further understood that the terms "comprises" and/or "comprising," when used in this specification, specify the presence of stated features, integers, steps, operations, elements, and/or components, but do not preclude the presence or addition of one or more other features, integers, steps, operations, elements, components, and/or groups thereof. It will be understood that when an element is referred to as being "connected" or "coupled" to another element, it can be directly connected or coupled to the other element or intervening elements may also be present. Further, "connected" or "coupled" as used herein may include wirelessly connected or wirelessly coupled. As used herein, the term "and/or" includes all or any element and all combinations of one or more of the associated listed items.
It will be understood by those within the art that, unless otherwise defined, all terms (including technical and scientific terms) used herein have the same meaning as commonly understood by one of ordinary skill in the art to which this application belongs. It will be further understood that terms, such as those defined in commonly used dictionaries, should be interpreted as having a meaning that is consistent with their meaning in the context of the prior art and will not be interpreted in an idealized or overly formal sense unless expressly so defined herein.
As will be appreciated by those skilled in the art, "client," "terminal," and "terminal device" as used herein include both devices that are wireless signal receivers, which are devices having only wireless signal receivers without transmit capability, and devices that are receive and transmit hardware, which have receive and transmit hardware capable of two-way communication over a two-way communication link. Such a device may include: cellular or other communication devices such as personal computers, tablets, etc. having single or multi-line displays or cellular or other communication devices without multi-line displays; PCS (Personal Communications Service), which may combine voice, data processing, facsimile and/or data communication capabilities; a PDA (Personal Digital Assistant) that may include a radio frequency receiver, a pager, internet/intranet access, web browser, notepad, calendar, and/or GPS (Global Positioning System) receiver; a conventional laptop and/or palmtop computer or other device having and/or including a radio frequency receiver. As used herein, a "client," "terminal device" can be portable, transportable, installed in a vehicle (aeronautical, maritime, and/or land-based), or situated and/or configured to operate locally and/or in a distributed fashion at any other location(s) on earth and/or in space. The "client", "terminal Device" used herein may also be a communication terminal, a web terminal, a music/video playing terminal, such as a PDA, an MID (Mobile Internet Device) and/or a Mobile phone with music/video playing function, and may also be a smart tv, a set-top box, and the like.
The hardware referred to by the names "server", "client", "service node", etc. is essentially an electronic device with the performance of a personal computer, and is a hardware device having necessary components disclosed by the von neumann principle such as a central processing unit (including an arithmetic unit and a controller), a memory, an input device, an output device, etc., a computer program is stored in the memory, and the central processing unit calls a program stored in an external memory into the internal memory to run, executes instructions in the program, and interacts with the input and output devices, thereby completing a specific function.
It should be noted that the concept of "server" as referred to in this application can be extended to the case of a server cluster. According to the network deployment principle understood by those skilled in the art, the servers should be logically divided, and in physical space, the servers may be independent from each other but can be called through an interface, or may be integrated into one physical computer or a set of computer clusters. Those skilled in the art will appreciate this variation and should not be so limited as to restrict the implementation of the network deployment of the present application.
Referring to fig. 1, the hardware basis required for implementing the related art embodiments of the present application may be deployed according to the architecture shown in the figure. The server 80 is deployed at the cloud end, and serves as a front-end application server, and is responsible for further connecting a related data server, a video streaming server, and other servers providing related support, so as to form a logically associated server cluster to provide services for related terminal devices, such as a smart phone 81 and a personal computer 82 shown in the figure. Both the smart phone and the personal computer can access the internet through a known network access mode, and establish a data communication link with the cloud server 80 so as to run a terminal application program related to the service provided by the server. In the related technical solution of the present application, the smart phone 81 or the personal computer (collectively referred to as a terminal device or a local device) may receive the message sent by the server 80, and the terminal device runs a corresponding application program to perform corresponding processing on the message.
It should be noted that, in a server cluster supporting the live broadcast room operation, in some scenarios, a server supporting the live broadcast room message service and a server supporting the live broadcast room video stream synthesis are merged into the same server or the same network address, and sometimes, the same application server may be used independently to establish the mutual association of the whole cluster, so that the server finally responsible for the service can be pointed to by using the same network address. In this regard, those skilled in the art will appreciate.
In order to support the operation of the application program, the terminal device is equipped with a related operating system, such as an IOS (operating system), an HMS (grand Mongolian), an Android and other operating systems providing equivalent functions, and with the support of such an operating system, the application program developed adaptively can be normally operated, so that human-computer interaction and remote interaction are realized.
The method is programmed and built in a terminal device application program providing the network live broadcast as a basic service function. The live webcast refers to a live webcast room network service realized based on the network deployment architecture.
The live broadcast room is a video chat room realized by means of an internet technology, generally has an audio and video broadcast control function and comprises a main broadcast user and audience users, wherein the audience users can comprise registered users registered in a platform or unregistered tourist users; either registered users who are interested in the anchor user or registered or unregistered users who are not interested in the anchor user. The interaction between the anchor user and the audience user can be realized through known online interaction modes such as voice, video, characters and the like, generally, the anchor user performs programs for the audience user in the form of audio and video streams, and economic transaction behaviors can also be generated in the interaction process. Of course, the application form of the live broadcast room is not limited to online entertainment, and can be popularized to other relevant scenes, such as an educational training scene, a video conference scene, a product recommendation and sale scene, and any other scene needing similar interaction.
The application program running on the terminal device can trigger and display the interface notification message according to the requirement designed by the platform side, and the currently common interface notification message is mainly used for playing dynamic effect information or broadcast information, and is certainly not limited to the method. It will thus be appreciated that any action supported by the live room service that is intended to output some notification message to the graphical user interface of the live room constructed when the application is running can be understood as an interface notification message of the present application.
The dynamic effect information is also called dynamic special effect information, interface special effect information and the like, and is mainly used as a basis for playing animation in a graphical user interface of a live broadcast room so as to play a corresponding dynamic special effect and further realize a message notification effect.
The broadcast information, like the dynamic effect information, is a notification message sent to the terminal user for implementing notification of a certain message, and after reaching the terminal device, the broadcast information finally causes the related content to be displayed in the graphical user interface of the live broadcast room of the terminal device.
The person skilled in the art will know this: although the various methods of the present application are described based on the same concept so as to be common to each other, they may be independently performed unless otherwise specified. In the same way, for each embodiment disclosed in the present application, it is proposed based on the same inventive concept, and therefore, concepts of the same expression and concepts of which expressions are different but are appropriately changed only for convenience should be equally understood.
Referring to fig. 2, a video source verification method of the present application, which is suitable for being implemented as a function of a live broadcast application program and is executed on a terminal device side, in an exemplary embodiment of the present application, includes the following steps:
step S11, correspondingly playing multiple video streams from different remote video sources in multiple playing windows, where the images and audio of each video stream are suppressed so that the video source is not recognizable:
for convenience of description, please refer to fig. 3, in a graphical user interface displayed after the terminal device operates, the live broadcast room application shown in fig. 3 displays that the current live broadcast room is in the online activity mode, where a horizontal area 112 displays three playing windows (the number of the playing windows is not limited thereto, and theoretically may be two or more than two arbitrary numbers), each playing window correspondingly plays one video stream, and each video stream belongs to an anchor user who is participating in the online activity, that is, three anchor users are participating in the activity. For audience users who are using the live room application to participate in the same online activity, the video streams in these playback windows are all from different anchor users, and are all video streams received from remote servers that belong to different video sources.
The horizontal area 114 displays three indication pictures which essentially belong to different video source identifiers, the contents of the pictures can be defined by the anchor user, for example, a face image screenshot can be acquired from each anchor user by means of a screenshot tool provided by a live broadcast room application program to embody, or a personal head portrait of the anchor user corresponding to each playing window is directly called, or a personal nickname of the anchor user is directly displayed, and the like. It can be understood that the video source identifier like the indication picture has established a corresponding relationship between the video source identifier and the video source in the background, so that the corresponding relationship is called for verification in the subsequent process. The indication pictures are used as video source identifiers in the method, and the positions of the three indication pictures can be exchanged pairwise in the interaction technology, so that a user can conveniently adjust the positions of the indication pictures through touch input. Since one indication picture corresponds to the right lower side of one playing window in the interface effect, it can be intuitively understood that, for each playing window and the video stream correspondingly processed in the background thereof, the corresponding relationship between each playing window and the indication picture right below the playing window is the user input of the video source identifier specified by the user for the playing windows (including the confirmation of the default corresponding relationship provided by the user for the system under the condition that the user does not actually perform touch operation).
As can be seen from the area 112 shown in fig. 3, the images in the video streams in the playing windows corresponding to the video streams of the respective anchor users are all in a suppressed state, although it cannot be illustrated, actually, due to the activity rule, the audio in the video streams is often set to be in a suppressed state at the beginning of the activity. As shown in the figure, the images of the video streams in the suppressed state are processed in various ways to present a distortion effect, and the audio of the video streams cannot be clearly heard, so that the video streams themselves cannot be identified due to being suppressed, and a user cannot receive information conveyed by the video streams of various anchor users temporarily, and cannot determine which anchor user's video stream is being played by the play windows at all, and cannot identify the video source of the video stream being played by the play windows.
The technical suppression of the video stream can be implemented on the server side for running the live broadcast room service and also can be implemented on the terminal equipment side where the user is located.
For images in video streams, on the side of the terminal equipment, technical suppression can be implemented on each video stream by multiple modes to achieve the aim of unidentifiable video sources. For example:
in one type of mode, the method is mainly implemented by adopting an image processing technology, after a live broadcast room application program of a terminal device receives video streams of each anchor user transmitted by a server, in the process of decoding and outputting the video streams to a corresponding playing window for playing, the video stream images are subjected to the suppression processing by using a known image blurring algorithm, an image distortion algorithm, even live broadcast overlay layers and the like, and the requirements of implementation of the interference means are met on the basis that the video sources of the video stream images cannot be identified visually, so that generally speaking, the video stream images are usually not identifiable relative to human vision. In the case where the image cannot be recognized, the owner of the video stream, that is, the video source cannot be recognized by the image.
In another mode, the method is mainly realized by adopting an image interference technology, when a live broadcast room application program of the terminal equipment decodes a plurality of paths of video streams and outputs the video streams to each corresponding playing window for playing, a mask layer is directly superposed corresponding to each playing window in a graphical user interface, the mask layer can have certain transparency and can also present a mosaic effect, the mask layer can be preset and provided for the live broadcast room application program to be called as required, and finally, the interference of video stream images is realized, so that the video sources of the video streams cannot be identified through the video images.
For the audio in the video stream, on the side of the terminal device, the technical suppression can be realized by multiple modes in the same way, so that the purpose that the video source can not be identified through the audio is achieved. For example:
in one type of mode, the audio processing technology is mainly adopted for implementation, after a live broadcast room application program of a terminal device receives video streams of all anchor users transmitted by a server, before audio is played, any available means is adopted for processing audio signals, for example, a large amount of noise is added to the audio signals, or the audio signals are directly removed, and finally the audio is distorted or silenced, so that the users can effectively hear the sound content transmitted by the anchor users corresponding to the audio signals, the sound content cannot be identified, and the purpose that video sources of the video streams cannot be identified through the audio is achieved.
In another type of mode, the audio control technology is mainly adopted for implementation, for this reason, the live broadcast application program of the terminal device prohibits the audio of the video stream to be processed from being output to the speaker or sets the audio to be mute, or reduces the output volume of the audio, so that the content of the audio of the corresponding video stream cannot be effectively identified by the user, and thus the video source of the video stream cannot be identified by the hearing of the user. Thus, the video source cannot be identified because its audio is disturbed.
It can be seen that there are obviously many technical means, although not exhaustive, for implementing technical suppression of images and audio of a video stream. Considering that a user can generally identify a video source through an image or an audio, in order to make the video source of a video stream really unrecognizable, the image and the audio of each video stream should be technically suppressed, so that the entire video stream is in a suppressed state, so as to ensure that the user cannot identify which anchor user the video source belongs to through any part of information of the video stream.
When the video streams of all the playing windows are simultaneously suppressed by the above-described technique, and the interface effect shown in the area 112 in fig. 3 is presented, it can be understood that the viewer user in the live room does not know to which anchor user a playing window corresponds, and in this case, the subsequent steps can be continued.
Step S12, sequentially removing the suppressed state of the audio of each video stream, and continuously maintaining the image in the suppressed state:
the graphical user interface shown in fig. 3 may be used for an on-line entertainment activity known colloquially as a "listener" in which the images and audio of the video streams of each anchor user are rendered unrecognizable by technical suppression, but the suppressed state may be removed by opening the audio of each video stream in turn for a predetermined period of time to allow the user to identify the video source of the video stream with a small amount of information conveyed by the audio. For such activities, it is necessary to exercise control over the individual video streams through certain technical steps.
In one embodiment, a preset activity rule may be adopted, each video stream is opened for a preset time period of, for example, 60 seconds in turn, when an anchor user comes, the live broadcast room application program removes the audio of the video stream from the suppressed state, that is, various technical suppression means applied to the audio before are cancelled, so that the terminal device normally plays the sound of the video stream, and the user can hear the audio content in the video stream of the anchor user within the range of the preset time period; when the predetermined time period is over, the original technical suppression means can be reapplied to restore the suppressed state of the audio (or not, although the prior technical suppression means may have an influence on the fairness of the activity rules, the prior technical suppression means has no substantial influence on the inventive spirit of the application). When one anchor user finishes the normal speech of the preset time, the next preset time can be entered, and the other anchor user continues to speak, and so on until the video streams of all the anchor users are displayed in the same way.
It should be noted that, because the playing windows of the respective anchor users are regularly arranged in the graphical user interface as shown in fig. 3, the operations of removing the suppressed state of the audio of the respective video streams are also performed sequentially and are usually opened for the same predetermined time period, so that theoretically, even if there is no obvious indication, the audience users can know to which playing window the audio of the current normal speaking speech belongs. Knowing this correspondence does not naturally affect the operation of the activity rules and the fair and normal participation of the user in the activity. However, on the other hand, considering that the whole process of removing the suppressed state from the audio of one anchor user is long, in order to avoid the memory confusion or other situations of the user, the playback window corresponding to the video stream in which the suppressed state of the current audio is removed and the normal loudspeaking is performed can be highlighted, for example, a frame with obvious flickers is added to the playback window, and a character guide is given.
It can be understood that, in this step, the images of the video streams are continuously in the suppressed state all the time, so as to prevent the user from identifying the video source through the images, that is, the user can only analyze to which anchor user the video source of the corresponding video stream belongs through the short-time open audio of each anchor user.
Through the process of the step, the user can master some materials of the video source according to which the video stream can be analyzed through the information of sound, content and the like contained in the audio, and can further submit the analysis result.
Step S13, receiving user input of selecting video source identification for each playing window, and checking whether the video source identification is correctly matched with the video source;
as shown in the area 114 of the graphical user interface shown in fig. 3, where the indication pictures represent a video source identifier, the user may drag each indication picture to the lower side of a different playing window to represent that the user "listens to the listener" and then select the video source identifier corresponding to each playing window, so as to give the analysis result. The user may drag these indicator pictures at any time throughout the activity, or may be limited to dragging the pictures within a certain period of time, with the live-air application process being responsible for receiving the user input during this process. In any case, there will eventually be a trigger condition to validate the user's analysis results in order to perform a verification of the user's analysis results based on the user input.
The condition for triggering the implementation of the verification can be flexibly set, for example, to meet the requirement of a preset activity rule, when the activity generates a timing end event, or when the user manually triggers a result submission event, it can be considered that the user analysis result is determined, and thus the verification is implemented.
The basis for checking whether the analysis result of the user is correct comprises information of two aspects, namely, the corresponding relation between the indication picture input by the user and the playing window on one hand, and the corresponding relation between the playing window and the video source of the video stream of the playing window on the other hand. The former is known from user input and the latter is also known from the video source information contained in the video stream pushed by the server, so that, in theory, the check can be performed by the live-room application at the terminal device. In some changed embodiments, for the purpose of preventing program cracking, before the verification is performed, the corresponding relationship between the playing window and the video source of the video stream thereof can be obtained from the server again for verification; similarly, the local user analysis result can also be submitted to the server to implement the verification. And so on, can be flexibly transformed by those skilled in the art.
During specific verification, for a playing window, if the video source represented by the corresponding indication picture input by the user is matched with the video source of the video stream actually corresponding to the playing window, it can be determined that the analysis result given by the user for the playing window is correct. And respectively carrying out the verification one by one aiming at each playing window, and finally obtaining the verification results of the analysis results of the user aiming at all the playing windows to form corresponding verification result information.
It should be noted that, it is also possible to change the matching condition according to different requirements by checking whether the user input matches with the actual situation composition of the background. For example, in a customized activity rule, it may only need to individually examine whether the video source identifier input by the user matches with the real video source of the playing window, and individually give the verification result of whether the matching is performed or not; in other customized activity rules, it may be required that a matching verification result is formed only when all correspondence relationships between video source identifiers input by a user and video source information corresponding to each playing window are correct, or a mismatch is formed only when one of the correspondence relationships is incorrect. It can be understood that no matter what matching mechanism is adopted, the embodiment of the inventive spirit of the present application is not affected.
Step S14, outputting verification result information:
after the verification is completed, the verification result information is formed, and according to the result information, the verification result information is available for further use by various active business logics, so that the verification result information needs to be output. It can be understood that the check result can be output to the memory variable, the local file, the remote server, and the current graphical user interface, which will be further described in combination with several cases as follows, so as to explain how the check result information can be output through different outputs to achieve various development purposes.
In one aspect, the inspection result information is output to a graphical user interface of a user to complete human-computer interaction. In one expression, as shown in fig. 4, on the basis of the corresponding relationship between the indication picture input by the user and the playing window, an icon representing correct (matching) or incorrect (not matching) is additionally added to each indication picture according to whether the corresponding relationship input by the user realizes matching, as shown in the lower right corner of each indication picture, so as to feed back the details of whether the analysis result is correct or not to the user.
On the other hand, the verification result information is also uploaded to the server to be output, so that the server records personal history data for the user according to the verification result information of the user or is used for distributing rewards such as electronic gifts according to the personal history data. For example, in the online entertainment activities of "listening to a person" shown in fig. 3 and 4, a reward can be applied to the user whose analysis result is verified to be correct according to the verification result information.
Of course, the test result information may be outputted and developed for other ways, for example, to provide the server with diagnosis, classification, etc. of the personal audio recognition capability of the user, so that the server subsequently provides the user with an adapted audio service or other relevant services. Such techniques are all helpful for further development and utilization of the technical framework provided by the present application.
The exemplary embodiments of the video source verification method and its variations disclosed above strongly illustrate the implementation of the method and some of the technical effects achieved thereby. To more fully demonstrate the advantages of the present application, the following description continues with further disclosure of various embodiments to enhance the flexibility of implementation of the present application:
in one embodiment, the step S13 of receiving the user input for selecting the video source identifier for each playing window may be executed after all three playing windows are speaking loudspeaked by the open audio, or may be executed after the audio suppressed state of the video stream of the first playing window is removed, and the user input is deemed to be completed after the user input for selecting all three playing windows is executed by the user. Through the embodiment, the application has a flexible implementation space in design aspect on the man-machine interaction logic.
In another embodiment, it is allowed to include a trigger mechanism for running the method in consideration of the needs of some online entertainment activities, and therefore, the method comprises a pre-step of receiving a verification start instruction and reconstructing the corresponding playing relationship between the plurality of playing windows and the plurality of video streams in the graphical user interface.
For example, at any time during the on-line activity in the live broadcast, the user may send a guess check activity such as "listening acquaintance" by touching an activity control provided in a graphical user interface, so that the live broadcast application program responds to the user' S touch and starts the "listening acquaintance" guess check activity, at this time, readjusting the correspondence between the video stream and the playing window in the background so that a certain playing window originally used for playing the video of a certain anchor user may be played by the user as the video stream of another anchor user, and after such background adjustment is completed, continuing to execute the other steps of the present application, since step S11 will realize technical suppression of the video stream, it is actually equivalent to starting the on-line activity such as "acquaintance" by the trigger mechanism of this embodiment, so that the user can be triggered to participate in such an activity at any time, the method becomes an online entertainment mechanism which can be called by a user at any time, and enriches the service capability of the application program in the live broadcast room. During the period of implementing the corresponding relationship between the reconstructed play window and the video stream in the background, the interface representation shown in fig. 5 can be referred in the graphical user interface to indicate that the user perceives background processing in the exchange of the anchor positions.
In another embodiment, since the user completes the entire verification process of the foregoing exemplary embodiment, a post-step may be added, in which the suppressed states of video and audio of the video streams of all the playing windows are removed, so that the normal playing is resumed. It can be understood that, in the online activity, the user may trigger and participate in the activity such as "listening to a listener" at any time, and then execute the steps of the present application, and after the execution of the activity is completed, the normal playing state may be recovered by removing various additionally applied technical inhibiting means, that is, the activity is exited.
Referring to fig. 6, in an embodiment adapted to online entertainment activities, the video source verification method of the present application flexibly includes various steps such as steps S11-S13, and further includes the following steps:
step S15, receiving the information of the electronic gift access entrance which is pushed by the server according to the checking result information, and visually displaying the access entrance on a graphical user interface:
if the verification result information is outputted to the server, the server calculates the number of electronic gifts available to the corresponding user according to the preset activity rule based on the verification result information, and the server pushes the electronic gift pick-up entrance information to the user who is qualified to obtain the electronic gift at step S13.
After receiving the information of the electronic gift pick-up entrance, the live broadcast application program at the user side can pop up an interface notification message on the current graphical user interface, play a corresponding dynamic special effect and provide a corresponding pick-up entrance, and the pick-up entrance can respond to the touch of the user to execute a process similar to contending for the electronic gift.
Step S16, in response to the instruction of the user to access the pick-up entrance, submitting corresponding instruction to the server to realize obtaining the electronic gift for the user:
after seeing the dynamic special effect and the receiving entrance on the graphical user interface of the live broadcast room, the user touches the receiving entrance, correspondingly sends a control instruction for contending for the electronic gift to the server, and after the control instruction is analyzed by the server, the server sends the electronic gifts with the corresponding number to the personal account of the user, so that the user obtains the electronic gift and informs the user as required.
It can be seen that through further enriching reward punishment mechanism, more help being used for realizing various online recreational activities with the technical scheme of this application to can further strengthen the interest, promote user interaction, lively live room flow.
Further, a video source verification apparatus of the present application can be constructed by functionalizing the steps in the method disclosed in the above embodiments, and according to this idea, please refer to fig. 7, wherein in an exemplary embodiment, the apparatus includes:
the video playing module 11 is configured to correspondingly play a plurality of video streams from different remote video sources in a plurality of playing windows, where images and audio of each video stream are in a suppressed state, so that the video source is not identifiable;
the polling suppression module 12 is used for sequentially removing the suppressed state of the audio of each path of video stream and continuously maintaining the image in the suppressed state;
an identifier checking module 13, configured to receive a user input for selecting a video source identifier for each playing window, and check whether the video source identifier is correctly matched with the video source;
and a result output module 14, configured to output the verification result information.
Further, to facilitate the implementation of the present application, the present application provides an electronic device, which includes a central processing unit and a memory, where the central processing unit is configured to invoke and run a computer program stored in the memory to perform the steps of the video source verification method in the foregoing embodiments.
It can be seen that the memory is suitable for adopting a nonvolatile storage medium, the aforementioned method is implemented as a computer program and installed in an electronic device such as a mobile phone or a computer, the related program code and data are stored in the nonvolatile storage medium of the electronic device, and further the program is executed by a central processing unit of the electronic device and is called from the nonvolatile storage medium to a memory for execution, so as to achieve the desired purpose of the present application. Therefore, it is understood that in an embodiment of the present application, a non-volatile storage medium may also be provided, in which a computer program implemented according to various embodiments of the video source verification method is stored, and when the computer program is called by a computer, the computer program executes the steps included in the method.
In summary, the verification mechanism for verifying the identification capability of the user video source is technically constructed, and the deep development and development of the network live broadcast application technology are facilitated.
As will be appreciated by one skilled in the art, the present application includes apparatus that are directed to performing one or more of the operations, methods described herein. These devices may be specially designed and manufactured for the required purposes, or they may comprise known devices in general-purpose computers. These devices have stored within their memory computer programs that are selectively activated or reconfigured. Such a computer program may be stored in a device (e.g., computer) readable medium, including, but not limited to, any type of disk including floppy disks, hard disks, optical disks, CD-ROMs, and magnetic-optical disks, ROMs (Read-Only memories), RAMs (Random Access memories), EPROMs (Erasable Programmable Read-Only memories), EEPROMs (Electrically Erasable Programmable Read-Only memories), flash memories, magnetic cards, or optical cards, or any type of media suitable for storing electronic instructions, and each coupled to a bus. That is, a readable medium includes any medium that stores or transmits information in a form readable by a device (e.g., a computer).
It will be understood by those within the art that each block of the block diagrams and/or flowchart illustrations, and combinations of blocks in the block diagrams and/or flowchart illustrations, can be implemented by computer program instructions. Those skilled in the art will appreciate that the computer program instructions may be implemented by a processor of a general purpose computer, special purpose computer, or other programmable data processing apparatus to produce a machine, such that the instructions, which execute via the processor of the computer or other programmable data processing apparatus, implement the aspects specified in the block or blocks of the block diagrams and/or flowchart illustrations disclosed herein.
Those of skill in the art will appreciate that the various operations, methods, steps in the processes, acts, or solutions discussed in this application can be interchanged, modified, combined, or eliminated. Further, other steps, measures, or schemes in various operations, methods, or flows that have been discussed in this application can be alternated, altered, rearranged, broken down, combined, or deleted. Further, steps, measures, schemes in the prior art having various operations, methods, procedures disclosed in the present application may also be alternated, modified, rearranged, decomposed, combined, or deleted.
The foregoing is only a partial embodiment of the present application, and it should be noted that, for those skilled in the art, several modifications and decorations can be made without departing from the principle of the present application, and these modifications and decorations should also be regarded as the protection scope of the present application.

Claims (15)

1. A video source verification method is characterized by comprising the following steps:
correspondingly playing a plurality of video streams from different remote video sources in a plurality of playing windows in a connection activity mode displayed in a current live broadcast room, wherein the images and the audios of the video streams are in a suppressed state, so that the video sources of the video streams cannot be identified;
sequentially removing the suppressed state of the audio of each video stream, and continuously maintaining the image in the suppressed state;
receiving user input of selecting video source identification for each playing window, and verifying whether the video source identification is correctly matched with the video source;
and outputting the verification result information.
2. The method of claim 1, wherein: in the step that the image and the audio of each video stream are in the suppressed state and the video source of the video stream is not recognizable, the video stream image is in the suppressed state by carrying out image synthesis processing on the video stream or by overlapping a masking layer on a playing window to interfere the video stream image.
3. The method of claim 1, wherein: in the step of making the image and the audio of each video stream in the suppressed state and making the video source unrecognizable, the audio of the video stream is made in the suppressed state by performing audio synthesizing processing on the video stream/by setting the volume of the video stream to mute.
4. The method of claim 1, wherein: and in the step of sequentially removing the suppressed state of the audio of each path of video stream and continuously maintaining the image in the suppressed state, after the suppressed state of the audio of one path of video stream is removed and the image is maintained for a preset time, the suppressed state of the audio of the other path of video stream is restored, and then the same processing is performed on the audio of the other path of video stream until the same processing is performed on the audio of all the paths of video streams.
5. The method of claim 1, wherein the step of receiving user input selecting a video source identification for each playback window is performed after removing the suppressed state of audio of the first video stream until said user input is completed.
6. The method of claim 1, wherein the step of receiving a user input selecting a video source identifier for each of the plurality of playback windows comprises providing a user-determined correspondence between the video source indication pictures and the playback windows in the graphical user interface as the user input.
7. The method according to claim 1, wherein in the step of checking whether the video source identifier is correctly matched with the video source, the video source identifiers are compared one by one according to the video source information of the video streams provided by the server, when the video source identifiers are respectively corresponding to the video source information, it is determined that the video source identifiers are matched with the video source, otherwise, it is determined that the video source identifiers are not matched with the video source.
8. The method as claimed in claim 1, wherein in the step of outputting the verification result information, displaying whether the video source identifiers are correctly matched one by one to output the corresponding verification result information, and submitting the verification result information to a server for output.
9. Method according to any of claims 1 to 8, characterized in that it comprises the following pre-steps:
and receiving a verification starting instruction, and rebuilding the corresponding playing relation between the plurality of playing windows and the plurality of video streams in the graphical user interface.
10. The method according to any one of claims 1 to 8, further comprising the following post-step:
and removing the restrained states of the video and the audio of the video streams of all the playing windows to restore the normal playing.
11. The method according to any one of claims 1 to 8, further comprising the following post-step:
and in the step of sequentially removing the suppressed state of the audio of each path of video stream, in the process of removing the suppressed state of the audio of one path of video stream, the playing window corresponding to the video stream is marked with a highlight.
12. Method according to any one of claims 1 to 8, characterized in that it comprises the following steps:
receiving electronic gift receiving entry information pushed by the server according to the inspection result information, and visually displaying the receiving entry on a graphical user interface;
and responding to the instruction of the user for accessing the receiving inlet and submitting corresponding instruction to the server to realize the acquisition of the electronic gift for the user.
13. A video source verification apparatus, comprising:
the video playing module is used for correspondingly playing a plurality of video streams from different remote video sources in a plurality of playing windows which are displayed in a current live broadcast room and are in a connection activity mode, and the images and the audios of the video streams are in a suppressed state so that the video sources can not be identified;
the polling suppression module is used for sequentially removing the suppressed state of the audio of each path of video stream and continuously maintaining the image in the suppressed state;
the identification checking module is used for receiving user input of selecting a video source identification for each playing window and checking whether the video source identification is correctly matched with the video source or not;
and the result output module is used for outputting the verification result information.
14. An electronic device comprising a central processing unit and a memory, wherein the central processing unit is configured to invoke execution of a computer program stored in the memory to perform the steps of the video source verification method according to any one of claims 1 to 9.
15. A non-volatile storage medium, characterized in that it stores a computer program implemented by the video source verification method according to any one of claims 1 to 9, which, when invoked by a computer, performs the steps comprised by the method.
CN202010972503.0A 2020-09-16 2020-09-16 Video source checking method, device, equipment and storage medium Active CN112104887B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202010972503.0A CN112104887B (en) 2020-09-16 2020-09-16 Video source checking method, device, equipment and storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202010972503.0A CN112104887B (en) 2020-09-16 2020-09-16 Video source checking method, device, equipment and storage medium

Publications (2)

Publication Number Publication Date
CN112104887A CN112104887A (en) 2020-12-18
CN112104887B true CN112104887B (en) 2022-05-17

Family

ID=73759699

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202010972503.0A Active CN112104887B (en) 2020-09-16 2020-09-16 Video source checking method, device, equipment and storage medium

Country Status (1)

Country Link
CN (1) CN112104887B (en)

Family Cites Families (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9049496B2 (en) * 2011-09-01 2015-06-02 Gracenote, Inc. Media source identification
CN105979379B (en) * 2016-06-01 2019-12-10 腾讯科技(深圳)有限公司 method and device for playing audition content
CN106791906B (en) * 2016-12-31 2020-06-23 北京星辰美豆文化传播有限公司 Multi-user network live broadcast method and device and electronic equipment thereof
CN107645682B (en) * 2017-10-20 2019-11-26 广州酷狗计算机科技有限公司 The method and system being broadcast live
CN110267064B (en) * 2019-06-12 2021-11-12 百度在线网络技术(北京)有限公司 Audio playing state processing method, device, equipment and storage medium
CN110910860B (en) * 2019-11-29 2022-07-08 北京达佳互联信息技术有限公司 Online KTV implementation method and device, electronic equipment and storage medium

Also Published As

Publication number Publication date
CN112104887A (en) 2020-12-18

Similar Documents

Publication Publication Date Title
US11386903B2 (en) Methods and systems for speech presentation based on simulated binaural audio signals
US11575531B2 (en) Dynamic virtual environment
US9621851B2 (en) Augmenting web conferences via text extracted from audio content
US11650790B2 (en) Centrally controlling communication at a venue
CN113453029B (en) Live broadcast interaction method, server and storage medium
US11916984B2 (en) System and methods for running conference applications before, during, and after a network conference
EP4229835A1 (en) System and methods for recording conference application activity associated with a network conference
US20220116434A1 (en) System and methods for sharing a screen shot of a conference application during a network conference
CN104394437A (en) Live broadcasting method and system
US11425176B2 (en) Transmitting conference application content during a network conference
CN112104887B (en) Video source checking method, device, equipment and storage medium
EP4229831A1 (en) System and methods for associating conference application content with an instance of a network conference
US20230091856A1 (en) System for Managing Remote Presentations
US20220201370A1 (en) Simulating audience reactions for performers on camera
Aguilera et al. Spatial audio for audioconferencing in mobile devices: Investigating the importance of virtual mobility and private communication and optimizations
JP2022129627A (en) Electronic conference management device, electronic conference recording system, electronic conference reproduction system, electronic conference management method and program
Davaz Production and editing practices in The Sweet Roll gaming podcast
CN113096674A (en) Audio processing method and device and electronic equipment
CN116828222A (en) Live broadcast flow control method and device, equipment and medium thereof

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
TA01 Transfer of patent application right

Effective date of registration: 20210115

Address after: 511442 3108, 79 Wanbo 2nd Road, Nancun Town, Panyu District, Guangzhou City, Guangdong Province

Applicant after: GUANGZHOU CUBESILI INFORMATION TECHNOLOGY Co.,Ltd.

Address before: 29th floor, building B-1, Wanda Plaza, Nancun Town, Panyu District, Guangzhou City, Guangdong Province

Applicant before: GUANGZHOU HUADUO NETWORK TECHNOLOGY Co.,Ltd.

TA01 Transfer of patent application right
GR01 Patent grant
GR01 Patent grant