CN115484467A - Live video processing method and device, computer readable medium and electronic equipment - Google Patents

Live video processing method and device, computer readable medium and electronic equipment Download PDF

Info

Publication number
CN115484467A
CN115484467A CN202110603559.3A CN202110603559A CN115484467A CN 115484467 A CN115484467 A CN 115484467A CN 202110603559 A CN202110603559 A CN 202110603559A CN 115484467 A CN115484467 A CN 115484467A
Authority
CN
China
Prior art keywords
video
user
live
live broadcast
segment
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202110603559.3A
Other languages
Chinese (zh)
Inventor
黄其亮
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Tencent Technology Shenzhen Co Ltd
Original Assignee
Tencent Technology Shenzhen Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Tencent Technology Shenzhen Co Ltd filed Critical Tencent Technology Shenzhen Co Ltd
Priority to CN202110603559.3A priority Critical patent/CN115484467A/en
Publication of CN115484467A publication Critical patent/CN115484467A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/21Server components or server architectures
    • H04N21/218Source of audio or video content, e.g. local disk arrays
    • H04N21/2187Live feed
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/44Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/442Monitoring of processes or resources, e.g. detecting the failure of a recording device, monitoring the downstream bandwidth, the number of times a movie has been viewed, the storage space available from the internal hard disk
    • H04N21/44204Monitoring of content usage, e.g. the number of times a movie has been viewed, copied or the amount which has been watched
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/83Generation or processing of protective or descriptive data associated with content; Content structuring
    • H04N21/845Structuring of content, e.g. decomposing content into time segments
    • H04N21/8456Structuring of content, e.g. decomposing content into time segments by decomposing the content in the time domain, e.g. in time segments

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Databases & Information Systems (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Two-Way Televisions, Distribution Of Moving Picture Or The Like (AREA)

Abstract

The embodiment of the application provides a live video processing method and device, a computer readable medium and electronic equipment. The live video processing method comprises the following steps: receiving a joining request aiming at a target live broadcast room, wherein the joining request comprises identification information of a user requesting to join the target live broadcast room; determining live broadcast content missed by the user according to the played time of the target live broadcast room and the identification information of the user; selecting at least one video clip from live content missed by the user; and generating a video collection segment according to the selected video segment, and presenting the video collection segment to the user. According to the technical scheme, the user can efficiently know the played content in the live broadcast room according to the video highlight segments when entering the live broadcast room, and the acceptability of the live broadcast content is improved.

Description

Live video processing method and device, computer readable medium and electronic equipment
Technical Field
The present application relates to the field of computer and communication technologies, and in particular, to a live video processing method and apparatus, a computer-readable medium, and an electronic device.
Background
The live video broadcast is a mode of live broadcast by utilizing the internet and a streaming media technology, in the related technology, after a user enters a virtual live broadcast room for live video broadcast, real-time live broadcast content can be directly played, and because the user misses the previous live broadcast content, the user is difficult to keep up with the current live broadcast progress, so that the acceptability of the live broadcast content is influenced, and the user viscosity of a live broadcast platform can be reduced.
Disclosure of Invention
Embodiments of the present application provide a method and an apparatus for processing a live video, a computer-readable medium, and an electronic device, so that a user can efficiently know a played content in a live broadcast room according to a video highlight segment when entering the live broadcast room at least to a certain extent, and the acceptability of the live broadcast content is improved.
Other features and advantages of the present application will be apparent from the following detailed description, or may be learned by practice of the application.
According to an aspect of an embodiment of the present application, a method for processing a live video is provided, including: receiving a joining request aiming at a target live broadcast room, wherein the joining request comprises identification information of a user requesting to join the target live broadcast room; determining live broadcast content missed by the user according to the broadcast duration of the target live broadcast room and the identification information of the user; selecting at least one video segment from the live content missed by the user; and generating a video collection fragment according to the selected video fragment, and presenting the video collection fragment to the user.
According to an aspect of an embodiment of the present application, a method for processing a live video is provided, including: displaying live broadcasting room information which is being live broadcasted; if the fact that the user enters the target live broadcast room is detected, sending a joining request aiming at the target live broadcast room to a server side, wherein the joining request comprises identification information of the user; receiving a video collection segment transmitted by the server, wherein the video collection segment is generated according to a video segment selected from live broadcast contents missed by the user; and playing the video collection segment.
According to an aspect of an embodiment of the present application, there is provided a live video processing apparatus, including: the device comprises a first receiving unit, a second receiving unit and a third receiving unit, wherein the first receiving unit is configured to receive a joining request aiming at a target live broadcast room, and the joining request comprises identification information of a user requesting to join the target live broadcast room; the determining unit is configured to determine live broadcast contents missed by the user according to the played time length of the target live broadcast room and the identification information of the user; a selection unit configured to select at least one video segment from the live content missed by the user; a first processing unit configured to generate a video highlight clip according to the selected video clip, and present the video highlight clip to the user.
In some embodiments of the present application, based on the foregoing scheme, the determining unit is configured to: acquiring a time point when the user joins the target live broadcast room according to the identification information of the user; and determining the live broadcast time period missed by the user according to the broadcast time length of the target live broadcast time period and the time point of the user joining the target live broadcast time period, so as to determine the live broadcast content missed by the user based on the live broadcast time period missed by the user.
In some embodiments of the present application, based on the foregoing solution, the first processing unit is further configured to: before at least one video segment is selected from the live content missed by the user, determining the live time missed by the user according to the live time period missed by the user; if the live broadcast duration missed by the user is less than or equal to a first duration, presenting the real-time live broadcast content of the target live broadcast room to the user; the selection unit is configured to: and if the live broadcast time length missed by the user is greater than the first time length, executing a process of selecting at least one video clip from the live broadcast content missed by the user.
In some embodiments of the present application, based on the foregoing solution, the apparatus for processing live video further includes: the second processing unit is configured to segment the played content of the target live broadcast room by taking a second time length as a segmentation reference to obtain a plurality of video time periods; and splitting the video content in the video time interval to obtain a plurality of video segments corresponding to the played content.
In some embodiments of the present application, based on the foregoing solution, the second processing unit is configured to split the video content in the video period by at least one of:
performing semantic analysis on the video content in the video time period to split the video content in the video time period based on a semantic analysis result;
performing content identification processing on the video content in the video time interval so as to split the video content in the video time interval based on a content identification result;
performing heat analysis on the video content in the video time interval so as to split the video content in the video time interval based on the heat analysis result.
In some embodiments of the present application, based on the foregoing solution, the selecting unit is configured to: determining at least one video time interval to which the live content missed by the user belongs based on a plurality of video segments corresponding to the played content; at least one video segment is selected from the at least one video period, respectively.
In some embodiments of the present application, based on the foregoing solution, the selecting unit is configured to: according to the user characteristics of the user, selecting video clips matched with the user characteristics from the at least one video time interval respectively; and if the video clip matched with the user characteristic is not selected from the target video time interval in the at least one video time interval according to the user characteristic, selecting the at least one video clip from the video clips obtained by splitting the target video time interval according to the heat value of the video clip.
In some embodiments of the present application, based on the foregoing solution, the processing apparatus of live video further includes: the third processing unit is configured to acquire interaction data corresponding to each video clip; determining the heat factor of each video clip according to the interactive data corresponding to each video clip; calculating the interaction change trend corresponding to each video clip according to the heat factor of each video clip; and calculating the heat value of each video clip based on the heat factor of each video clip and the interaction change trend corresponding to each video clip.
In some embodiments of the present application, based on the foregoing solution, the first processing unit is configured to: synthesizing the selected video clips according to the time sequence of the video to obtain synthesized video clips; and generating a video collection segment with a set duration range based on the synthesized video segment.
In some embodiments of the present application, based on the foregoing solution, the first processing unit is configured to: if the time length of the synthesized video clips exceeds the set time length range and at least two video clips are selected from the same video time period, removing the video clips from the at least two video clips according to the heat value or the matching degree with the user characteristics until the time length of the synthesized video clips is within the set time length range; and taking the video segment within the set duration range obtained by synthesis as the video collection segment.
In some embodiments of the present application, based on the foregoing solution, the first processing unit is configured to: and if the duration of the synthesized video clip exceeds the set duration range, intercepting the video clip which accords with the set duration range from the synthesized video clip to be used as the video collection clip.
In some embodiments of the present application, based on the foregoing solution, the first processing unit is configured to: if the duration of the synthesized video clip does not reach the set duration range, selecting a video clip from other video clips of the at least one video time interval according to the heat value until the duration of the synthesized video clip is within the set duration range; and taking the video segment within the set duration range obtained by synthesis as the video collection segment.
In some embodiments of the present application, based on the foregoing solution, the first processing unit is further configured to: in the process of presenting the video highlight clip to the user, if a skip playing instruction for the video highlight clip is detected, identifying a target video clip presented when the skip playing instruction is received; and adjusting the user characteristics of the user according to the attribute label of the target video clip.
In some embodiments of the present application, based on the foregoing solution, the first processing unit is further configured to: and performing semantic analysis and/or content identification processing on the video clip of the played content in the target live broadcast room to obtain an attribute label of the video clip, wherein the attribute label is used for calculating the matching degree with the user characteristics.
In some embodiments of the present application, based on the foregoing solution, the first processing unit is further configured to: and adjusting the attribute tags of the video clips according to the interactive data corresponding to the video clips.
According to an aspect of an embodiment of the present application, there is provided a processing apparatus for live video, including: a display unit configured to display live-cast room information being live-cast; the system comprises a sending unit, a receiving unit and a processing unit, wherein the sending unit is configured to send a joining request aiming at a target live broadcast room to a server side if a user is detected to enter the target live broadcast room, and the joining request comprises identification information of the user; a second receiving unit, configured to receive a video highlight segment transmitted by the server, where the video highlight segment is generated according to a video segment selected from live content missed by the user; and the playing unit is configured to play the video collection segment.
In some embodiments of the present application, based on the foregoing solution, the playing unit is further configured to: and if a skip playing instruction for the video collection segment is received or the playing of the video collection segment is completed, playing the real-time live broadcast content of the target live broadcast room.
According to an aspect of embodiments of the present application, there is provided a computer-readable medium on which a computer program is stored, the computer program, when executed by a processor, implementing a method for processing live video as described in the above embodiments.
According to an aspect of an embodiment of the present application, there is provided an electronic device including: one or more processors; a storage device for storing one or more programs which, when executed by the one or more processors, cause the one or more processors to implement a method of processing live video as described in the above embodiments.
According to an aspect of embodiments herein, there is provided a computer program product or computer program comprising computer instructions stored in a computer readable storage medium. The processor of the computer device reads the computer instructions from the computer-readable storage medium, and the processor executes the computer instructions to cause the computer device to execute the processing method of the live video provided in the above-mentioned various optional embodiments.
In the technical scheme provided by some embodiments of the application, live broadcast content missed by a user is determined according to the broadcast duration of a target live broadcast room and identification information of the user, then a video clip is selected from the live broadcast content missed by the user, and a video highlight clip is generated according to the selected video clip and is presented to the user, so that the user can efficiently know the broadcast content of the live broadcast room according to the video highlight clip when entering the live broadcast room, and then the progress of the video live broadcast can be quickly followed up, not only is the acceptability of the live broadcast content improved, but also the live broadcast viewing experience of the user is improved, and meanwhile, the user viscosity of a live broadcast platform can also be improved.
It is to be understood that both the foregoing general description and the following detailed description are exemplary and explanatory only and are not restrictive of the application.
Drawings
The accompanying drawings, which are incorporated in and constitute a part of this specification, illustrate embodiments consistent with the present application and together with the description, serve to explain the principles of the application. It is obvious that the drawings in the following description are only some embodiments of the application, and that for a person skilled in the art, other drawings can be derived from them without inventive effort. In the drawings:
FIG. 1 shows a schematic diagram of an exemplary system architecture to which aspects of embodiments of the present application may be applied;
FIG. 2 shows a flow diagram of a method of processing live video according to an embodiment of the present application;
FIG. 3 shows a flow diagram of a method of processing live video according to an embodiment of the present application;
FIG. 4 shows a flow diagram for calculating a video heat value according to an embodiment of the present application;
FIG. 5 shows a flow diagram of a method of processing live video according to an embodiment of the application;
FIG. 6 shows a flow diagram of a method of processing live video according to an embodiment of the application;
FIG. 7 illustrates a diagram of splitting played content by time according to one embodiment of the application;
FIG. 8 shows a schematic diagram of the splitting of a large fragment according to an embodiment of the present application;
FIG. 9 illustrates a schematic diagram of splitting a large fragment according to heat value, according to an embodiment of the present application;
FIG. 10 shows a schematic diagram of labeling small segments according to an embodiment of the present application;
FIG. 11 illustrates a diagram of small segments selected according to a user's interest tag, according to one embodiment of the present application;
fig. 12-14 show interface diagrams of a live client according to an embodiment of the application;
FIG. 15 shows a block diagram of a processing device for live video according to an embodiment of the present application;
FIG. 16 shows a block diagram of a processing device for live video according to an embodiment of the application;
FIG. 17 illustrates a schematic structural diagram of a computer system suitable for use to implement the electronic device of the embodiments of the subject application.
Detailed Description
Example embodiments will now be described more fully with reference to the accompanying drawings. Example embodiments may, however, be embodied in many different forms and should not be construed as limited to the examples set forth herein; rather, these embodiments are provided so that this disclosure will be thorough and complete, and will fully convey the concept of example embodiments to those skilled in the art.
Furthermore, the described features, structures, or characteristics may be combined in any suitable manner in one or more embodiments. In the following description, numerous specific details are provided to give a thorough understanding of embodiments of the application. One skilled in the relevant art will recognize, however, that the embodiments of the present application can be practiced without one or more of the specific details, or with other methods, components, devices, steps, and so forth. In other instances, well-known methods, devices, implementations, or operations have not been shown or described in detail to avoid obscuring aspects of the application.
The block diagrams shown in the figures are functional entities only and do not necessarily correspond to physically separate entities. I.e. these functional entities may be implemented in the form of software, or in one or more hardware modules or integrated circuits, or in different networks and/or processor means and/or microcontroller means.
The flowcharts shown in the figures are illustrative only and do not necessarily include all of the contents and operations/steps, nor do they necessarily have to be performed in the order described. For example, some operations/steps may be decomposed, and some operations/steps may be combined or partially combined, so that the actual execution sequence may be changed according to the actual situation.
It should be noted that: reference herein to "a plurality" means two or more. "and/or" describe the association relationship of the associated objects, meaning that there may be three relationships, e.g., A and/or B may mean: a exists alone, A and B exist simultaneously, and B exists alone. The character "/" generally indicates that the former and latter associated objects are in an "or" relationship.
Fig. 1 shows a schematic diagram of an exemplary system architecture to which the technical solution of the embodiments of the present application can be applied.
As shown in fig. 1, system architecture 100 may include terminal device 101, game server 102, live server 103, terminal device 104, and network 105. The terminal device 101 and the terminal device 104 are used for users to use, and may be electronic devices having a live video function or a live video watching function. Specifically, the terminal device 101 and the terminal device 104 may be electronic devices such as a mobile phone, a tablet Computer, a wearable device, a PC (Personal Computer), and the like.
Optionally, the terminal device 101 and the terminal device 104 may be installed and operated with a live client, where the live client refers to a client used by a user to explain a live video or watch the live video, and the live client may have functions of collecting, publishing, downloading, and playing the live video. In the exemplary embodiment, a display screen and/or an audio capture device are disposed in the terminal device 101 and the terminal device 104, the display screen is used for displaying live video, and the audio capture device is used for capturing audio data of the anchor.
In a possible implementation manner, it is assumed that the terminal device 101 is a client for performing live game playing, and the terminal device 104 is a client for watching live game playing, and then the terminal device 101 may be equipped with a game client besides the live game client, where the game client may be a Shooting game, a First-Person Shooting game (FPS), a third-Person Shooting game, a multi-player gunplay survival game, a multi-player Online tactical game (MOBA), a role playing game, an instant strategy game, a racing game, a music game, and other game clients. Besides, if the game server 102 can provide other services, such as a virtual reality service, a three-dimensional map service, a military simulation service, and the like, the terminal device 101 may also install a corresponding application program to perform live video.
The game server 102 is a server that provides game services, and the game server 102 may include one server or a plurality of servers. Alternatively, the game server 102 may be a server that provides a cloud game.
The live broadcast server 103 is a server that provides a background service for a live broadcast client, the live broadcast server 103 may include one server or multiple servers, and optionally, the live broadcast server 103 may be a cloud server.
As shown in fig. 1, terminal apparatuses 101, game servers 102, live servers 103, and terminal apparatuses 104 can communicate with each other via a network 105. For example, the terminal device 101 and the game server 102 may communicate with each other through a wired network or a wireless network; the terminal device 101 and the live broadcast server 103 can communicate with each other through a wired network or a wireless network; the terminal device 104 and the live broadcast server 103 can communicate with each other through a wired network or a wireless network; the game server 102 and the live broadcast server 103 can communicate with each other through a wired network or a wireless network.
In an embodiment of the present application, an example is described in which the terminal device 101 is a client for playing a live game, and the terminal device 104 is a client for watching a live game. The terminal device 101 can log in the game server 102 through the first account to play a game, and can log in the live broadcast server 103 through the second account to live broadcast a game being played, and the terminal device 104 can acquire a game video live broadcast by the terminal device 101 from the live broadcast server 103 and provide the game video to a user for watching.
Specifically, the terminal apparatus 101 can create a live broadcast room through the live broadcast server 103 and play a game live broadcast based on the live broadcast room. The terminal device 104 can selectively enter the live broadcast room corresponding to the terminal device 101 through the live broadcast room list provided by the live broadcast server 103, and watch the corresponding game live broadcast video. Because the terminal device 104 may have been live in the live broadcast room for a period of time when entering the live broadcast room corresponding to the terminal device 101, when the user needs to know the played content, a live broadcast scheme is to find the playback video released by the anchor to select to watch live playback when the current live broadcast is over and the anchor releases the playback video, and this scheme results in that the user cannot check the played content in time and the watching experience is poor. Another live broadcasting scheme is that when the user enters the online live broadcasting mode, the user selects to watch the broadcasted content by dragging the progress bar and switches the current live broadcasting progress back through operation.
In order to prevent the user of the terminal device 104 from keeping up with the progress of the current live broadcast, in an embodiment of the present application, after detecting a request sent by the terminal device 104 to join the live broadcast room, the live broadcast server 103 may determine live broadcast content missed by the user of the terminal device 104 according to the played time of the live broadcast room and identification information of the user corresponding to the terminal device 104 (the identification information of the user may be user identification information of a live broadcast client installed on the terminal device 104), then select at least one video segment (for example, a video segment matching with a user characteristic of the user) from the live broadcast content missed by the user of the terminal device 104, then generate a video highlight segment according to the selected video segment, and send the video highlight segment to the terminal device 104 for playing, so that the user of the terminal device 104 can efficiently understand the played content of the live broadcast room by watching the video highlight segment, and can quickly keep up with the progress of the live broadcast, which not only improves the acceptability of the live broadcast content, but also improves the viewing experience of the user, and also improves the viscosity of a live broadcast platform.
Illustratively, when the terminal device 104 finishes playing the video highlight clip or the user chooses to skip playing the video highlight clip, the real-time live content of the live broadcast room may be played.
The implementation details of the technical solution of the embodiment of the present application are set forth in detail below:
fig. 2 shows a flowchart of a live video processing method according to an embodiment of the present application, where the live video processing method may be performed by a live server, which may be the live server 103 shown in fig. 1. Referring to fig. 2, the processing method of the live video at least includes steps S210 to S240, and the following is described in detail:
in step S210, a join request for a target live broadcast room is received, where the join request includes identification information of a user requesting to join the target live broadcast room.
In an embodiment of the application, when a user needs to enter a live broadcast room to watch a live broadcast, the user can log in a live broadcast client, select a target live broadcast room through a live broadcast room list displayed by the live broadcast client, and then the live broadcast client can send an adding request to a live broadcast server based on the target live broadcast room selected by the user, and add identification information (such as account information of the user in a live broadcast application program) of the user and information of the target live broadcast room in the adding request.
Alternatively, the target live broadcast room may be a live broadcast room, and the live broadcast content may be game play, video commentary, exhibition activities, and the like. The join request can be a join request triggered when the user enters the target live broadcast time for the first time, or a join request triggered when the user enters the live broadcast time again after exiting midway.
In step S220, the live content missed by the user is determined according to the played time length of the target live broadcast room and the identification information of the user.
In an embodiment of the application, because the user has the identification information, and the information of the joining time point exists when the user joins the target live broadcasting room, the time point of the user joining the target live broadcasting room can be obtained according to the identification information of the user, then the live broadcasting time period missed by the user is determined according to the played time length of the target live broadcasting room and the time point of the user joining the target live broadcasting room, and further the live broadcasting content missed by the user can be determined based on the live broadcasting time period missed by the user.
For example, a certain live broadcast is broadcast from 9. Assuming that the user has exited the live room at 10 after joining the live room at 10.
In step S230, at least one video clip is selected from the live content missed by the user.
In an embodiment of the application, the selection of at least one video segment from the live content missed by the user may be performed randomly, or one video segment may be selected in each live time period, or may be selected according to a user characteristic, or may be selected according to a heat value of the video segment.
Specifically, if the video clip is selected according to the user characteristics, an attribute tag may be added to the video clip, where the attribute tag is used to indicate the type of the video clip, for example, for a game video, a group war tag, a tag of a certain game hero, a tag of a certain hero type (such as a french hero, an auxiliary hero, and the like), and then a matching degree may be calculated according to the attribute tag and the user characteristics, where the user characteristics are used to indicate that the interest of the user is good, for example, the interest of the user is in a certain game hero or a certain hero type, and then the video clip with the tag of the game hero or the tag of the hero type is a video clip matching the user characteristics of the user. The technical scheme of the embodiment can improve the accuracy of at least one selected video segment, and is favorable for realizing personalized selection.
If the selection is based on the heat value of the video segments, at least one video segment may be selected in order of high to low heat value.
Optionally, semantic analysis and/or content recognition processing may be performed on the video segment of the played content in the target live broadcast room to obtain the attribute tag of the video segment. The video speech analysis and the content recognition process are both based on artificial intelligence technology, specifically, based on computer vision technology to analyze and process video segments.
Specifically, the Artificial Intelligence (AI) technology is a theory, method, technique, and application system that simulates, extends, and expands human Intelligence using a digital computer or a machine controlled by a digital computer, senses the environment, acquires knowledge, and uses the knowledge to obtain the best results. In other words, artificial intelligence is a comprehensive technique of computer science that attempts to understand the essence of intelligence and produce a new intelligent machine that can react in a manner similar to human intelligence. Artificial intelligence is the research of the design principle and the realization method of various intelligent machines, so that the machines have the functions of perception, reasoning and decision making. The artificial intelligence technology is a comprehensive subject and relates to the field of extensive technology, namely the technology of a hardware level and the technology of a software level. The artificial intelligence base technologies generally include technologies such as sensors, dedicated artificial intelligence chips, cloud computing, distributed storage, big data processing technologies, operation/interaction systems, mechatronics, and the like. The artificial intelligence software technology mainly comprises a computer vision technology, a voice processing technology, a natural language processing technology, machine learning/deep learning and the like.
Computer Vision (CV) technology is a science for researching how to make a machine look, and further refers to replacing human eyes with a camera and a Computer to perform machine Vision such as identification, tracking and measurement on a target, and further performing graphic processing, so that the Computer processing becomes an image more suitable for human eyes to observe or to transmit to an instrument to detect. As a scientific discipline, computer vision research-related theories and techniques attempt to build artificial intelligence systems that can capture information from images or multidimensional data. The computer vision technology generally includes technologies such as image processing, image Recognition, image semantic understanding, image retrieval, OCR (Optical Character Recognition), video processing, video semantic understanding, video content/behavior Recognition, three-dimensional object reconstruction, 3D technology, virtual reality, augmented reality, synchronous positioning, map construction, and the like, and also includes common biometric technologies such as face Recognition, fingerprint Recognition, and the like.
In an embodiment of the application, after the attribute tag of the video segment is determined, the attribute tag of the video segment may be further adjusted according to the interaction data corresponding to the video segment. For example, if the viewing amount of a certain video segment is large, or the favorite amount is large, the attribute tag of the video segment may be adjusted, for example, adjusted to be a hot video.
In an embodiment of the application, when a video segment matched with a user characteristic is selected from live content missed by a user according to the user characteristic of the user, the similarity between the user characteristic and an attribute label of the video segment can be calculated, and if the similarity is greater than a set threshold value, the user characteristic is matched with the attribute label of the video segment. Alternatively, the similarity may be determined by calculating a distance between the user feature and the attribute label, such as a Euclidean distance, a Minkowski distance, a Manhattan distance, or the like, where a smaller distance indicates a greater similarity.
Continuing to refer to fig. 2, in step S240, a video highlight segment is generated according to the selected video segment, and the video highlight segment is presented to the user.
In an embodiment of the present application, the selected video segments may be subjected to synthesis processing, and the synthesized video segments are used as the generated video highlight segments. Alternatively, the process of presenting the video highlight clip to the user may be sending the generated video highlight clip to a live client of the viewer, and then playing by the live client.
For example, in the process of presenting the video highlight segment to the user, if a skip play instruction for the video highlight segment is detected, the target video segment presented when the skip play instruction is received may be identified, and then the user characteristic of the user is adjusted according to the attribute tag of the target video segment. Specifically, since the video highlight clip is generated from the selected video clip, when a skip play instruction for the video highlight clip is detected, the user may not be interested in the currently played video clip, so that the user characteristics of the user can be adjusted based on the skip play instruction, and the accuracy of subsequent recommendation can be optimized.
Based on the technical solution of the embodiment shown in fig. 2, in an embodiment of the present application, as shown in fig. 3, after step S220, step S310 may further be included, namely, it is determined whether the live broadcast time length missed by the user is greater than a first time length, for example, the live broadcast time length missed by the user is determined according to the live broadcast time period missed by the user.
If the duration of the live broadcast missed by the user is longer than the first duration, step S230 and step S240 are executed again, that is, the process of selecting a video segment and generating a video highlight segment is performed. And if the live broadcast duration missed by the user is less than or equal to the first duration, the real-time live broadcast content of the target live broadcast room can be directly presented to the user. According to the technical scheme of the embodiment, when the live broadcast duration missed by the user is short, the real-time live broadcast content can be directly presented to the user. Alternatively, the first duration may be, for example, half an hour, 1 hour, or the like.
Based on the technical solutions of the embodiments shown in fig. 2 and fig. 3, in an embodiment of the present application, the played content of the target live broadcast room may also be segmented to generate video segments, so as to select a video segment from the generated video segments. Specifically, the second duration may be used as a reference for segmenting the played content in the target live broadcasting room to obtain a plurality of video time periods, and then the video content in the video time periods is split to obtain a plurality of video segments corresponding to the played content. For example, the played content in the target live broadcast room may be divided into a plurality of video time periods of 10 minutes on the basis of 10 minutes as a segmentation reference, and then the video time periods are split to obtain a plurality of video segments corresponding to the played content. Alternatively, the second time period may be set according to actual requirements, and may be, for example, 5 minutes, 15 minutes, 20 minutes, or the like, in addition to 10 minutes.
In an embodiment of the application, when splitting video content in a video period, semantic analysis can be performed on the video content in the video period, and then the video content in the video period is split based on a semantic analysis result; or the content identification processing can be carried out on the video content in the video time interval, and then the video content in the video time interval is split based on the content identification result; or the video content in the video time interval can be subjected to heat analysis, and then the video content in the video time interval is split based on the heat analysis result. Of course, the video content in the video period may also be split by performing a plurality of processes of semantic analysis, content recognition process, and heat analysis on the video content. For example, the game hero may be divided into video segments, video segments including a peak of heat, or video segments for a certain game hero battle.
After segmenting the played content of the target live broadcast room, in an embodiment of the present application, the process of selecting a video segment matching the user characteristic from the live broadcast content missed by the user may be: the method comprises the steps of determining at least one video time interval to which live content missed by a user belongs based on a plurality of video segments corresponding to played content, and then respectively selecting at least one video segment from the at least one video time interval. That is, in order to ensure the content continuity of the generated video highlight segments, at least one video segment may be selected from each video period missed by the user for the composition processing.
Optionally, when the video segments are respectively selected from the video periods missed by the user, the video segments matching with the user characteristics may be selected from the video periods only depending on the user characteristics, the video segments with higher heat values may be selected from the video periods only depending on the heat values, or the video segments may be selected by combining the user characteristics and the heat values. Specifically, if a video segment matching the user feature is not selected from a target video period missed by the user according to the user feature, at least one video segment may be selected from the video segments split from the target video period according to the heat value of the video segment. That is, 1 or more video clips can be selected in the order of high to low heat value.
In an embodiment of the present application, a scheme for calculating a heat value of a video segment is further provided, and specifically referring to fig. 4, the scheme includes the following steps:
step S410, obtaining interactive data corresponding to each video clip.
In an embodiment of the present application, the interactive data corresponding to the video clip may include: watching people number data, commenting people number data, barrage data, connecting wheat interaction data and the like.
Step S420, determining a heat factor of each video segment according to the interactive data corresponding to each video segment.
In one embodiment of the present application, the popularity factor may be, for example, the number of people watching, the number of people commenting, the number of barrage, the duration of the interactive session, the amount of received gift, etc. If the video clip has multiple heat factors, the multiple heat factors can be counted (e.g., averaged, or weighted averaged, etc.) to obtain the final heat factor of the video clip.
Step S430, calculating the interactive change trend corresponding to each video clip according to the heat factor of each video clip.
In one embodiment of the present application, the interactive trend is used for feeding back the interactive change condition of the live content, for example, the interactive trend can be measured by the trend of the heat factor over time. Alternatively, the interaction trend may be represented by a natural index of the derivative of the heat factor.
Step S440, calculating the heat value of each video clip based on the heat factor of each video clip and the interaction change trend corresponding to each video clip.
In an embodiment of the application, the heat factor of the video clip and the interaction change trend corresponding to the video clip can be integrated to obtain the heat value of the video clip.
Specifically, the heat value of the video clip can be calculated by the following formula:
H=Base(P)+e f'(P)
wherein H represents a heat value of the video clip; p represents a heat factor; base (P) represents the cardinality of the heat factor, such as the number of viewers, etc.; f' (P) represents the derivative of the heat factor dynamics, which reflects the interaction trend.
In an embodiment of the present application, the process of generating the video highlight segment according to the selected video segment in step S240 shown in fig. 2 and 3 may be: and synthesizing the selected video clips according to the time sequence of the video to obtain synthesized video clips, and then generating a video collection clip with a set duration range based on the synthesized video clips.
Specifically, if the duration of the synthesized video segment is within the set duration range, the synthesized video segment may be directly used as the video highlight segment.
If the time length of the synthesized video clip exceeds the set time length range and at least two video clips are selected from the same video time interval, the video clips can be removed from the at least two video clips according to the heat value or the matching degree with the user characteristics until the time length of the synthesized video clip is within the set time length range, and then the synthesized video clip within the set time length range is used as a video collection clip. Alternatively, video segments with lower heat values or video segments with lower matching degrees with the user features may be removed.
Of course, if the duration of the synthesized video segment exceeds the set duration range, the video segment conforming to the set duration range can be directly intercepted from the synthesized video segment to be used as the video collection segment. For example, the second half of the synthesized video clip may be intercepted, that is, from back to front, so as to ensure that the generated video highlight clip has a higher correlation with the real-time live content. Or a video highlight clip can be generated by respectively cutting out a part of the first half, the middle part and the second half of the synthesized video clip.
If the duration of the synthesized video clip does not reach the set duration range, the video clip can be selected from other video clips of at least one video time interval according to the heat value until the duration of the synthesized video clip is within the set duration range, and then the synthesized video clip within the set duration range is used as the video collection clip.
Fig. 2 to fig. 4 are illustrations from the perspective of a live broadcast server, and the following describes in detail the technical solution of the embodiment of the present application from the perspective of a live broadcast client in conjunction with fig. 5:
fig. 5 shows a flowchart of a processing method of a live video according to an embodiment of the present application, which may be performed by a live client that may run in the terminal device 104 shown in fig. 1. Referring to fig. 5, the processing method of the live video at least includes steps S510 to S540, which are described in detail as follows:
in step S510, live room information that is being live is displayed.
In an embodiment of the present application, a live client (specifically, a client watching a live broadcast) may obtain live room information being live broadcast from a live broadcast server, and then display the live room information being live broadcast on an interface on the live broadcast client (for example, the live room information may be displayed in a list form or a thumbnail form), where the live room information may be a name displaying a live room, profile information, live content, a live broadcast time period, and the like.
In step S520, if it is detected that the user enters the target live broadcast room, a join request for the target live broadcast room is sent to the server, where the join request includes identification information of the user.
In an embodiment of the application, when a user needs to enter a live broadcast room to watch live broadcast, the user can log in a live broadcast client, select a target live broadcast room through live broadcast room information displayed by the live broadcast client, and then the live broadcast client can send an adding request to a live broadcast server based on the target live broadcast room selected by the user, and add identification information (such as account information of the user in a live broadcast application program) of the user and information of the target live broadcast room in the adding request.
In step S530, a video highlight segment transmitted by the server is received, where the video highlight segment is generated according to a video segment selected from live content missed by the user.
The process of generating the video collection segment by the server may refer to the technical solution of the foregoing embodiment, and is not described again.
In step S540, the video highlight segment is played.
Optionally, when the user clicks to enter the target live broadcast room and the video highlight clip is obtained, the video highlight clip may be directly played on the interface of the live broadcast client. Or displaying a playing control of the video highlight clip on an interface of the live broadcast client, and playing the video highlight clip after detecting a triggering instruction of the playing control.
In an embodiment of the present application, during the process of playing the video highlight reel, if a skip play instruction for the video highlight reel is received, or if the video highlight reel is completely played, the real-time live content in the target live broadcast room may be played. Optionally, in this embodiment, the playing window of the video highlight segment and the playing window of the live broadcast content in the live broadcast room may be the same.
In addition, in an embodiment of the application, a playing window of the video highlight clip and a playing window of the real-time live content can be displayed on an interface of the live broadcast client, and then whether the video highlight clip or the real-time live content is played is selected according to a triggering action of a user.
The technical solutions of the embodiments of the present application are described above from the perspective of the live broadcast server and the live broadcast client, and details of implementation of the technical solutions of the embodiments of the present application are described below with reference to fig. 6 to 14:
as shown in fig. 6, a method for processing live video according to an embodiment of the present application includes the following steps:
step S601, after the user enters the live broadcast room, the live broadcast client sends the time node and the user ID to the live broadcast server.
Specifically, when a user enters a live broadcast room to start watching an online live broadcast, a live broadcast client sends a request to a live broadcast server, where the request may include time node information and a user ID. Optionally, the time node information may be a time point when the user enters a live broadcast room, and the live broadcast server may determine a live broadcast period missed by the user accordingly; or the time node information may also be a live broadcast period missed by the user determined by the live broadcast client.
And step S602, matching the interest tag and the video clip popularity under the user ID by the live broadcast server, generating a video collection and sending the video collection to the client.
In an embodiment of the application, after a main broadcast of a live broadcast room is started, a live broadcast server can start to acquire the broadcasted content of the live broadcast room in real time, then split the content into large segments (namely Video time slots in the above embodiment) according to time, and store the large segments as Video1, video2 and Video3 \8230insequence. As shown in fig. 7, the time lengths of the split large segments may be the same or different, and the time length of each large segment may be set according to actual requirements.
For the video content in each large segment, the video content can be split by combining an AI algorithm of semantic analysis and picture content identification, or according to the heat distribution condition (such as a waveband peak value) of the video content, so as to obtain small segments, and the small segments are named correspondingly in sequence. Specifically, as shown in fig. 8, video1 can be split into small fragments v1.1, v1.2, v1.3, v1.4; resolving Video2 into small fragments v2.1, v2.2, v2.3 and v2.4; resolving Video3 into small fragments v3.1, v3.2 and the like.
Assuming that the Video1 is split according to the heat distribution, as shown in fig. 9, the heat curve of the Video1 includes the peaks and valleys of the heat values, and the splitting can be performed with the peak of each heat value as the center to obtain small segments v1.1, v1.2, v1.3, v1.4. When the AI algorithm combining semantic analysis and picture content recognition is used for splitting, a large segment can be divided into a plurality of small segments related to content comparison, for example, small segments of hero in each game.
After the large segments are split to obtain the small segments, semantic analysis and picture content recognition can be performed on each small segment to mark corresponding attribute tags on each small segment, and the small segments and the attribute tags are stored in the live broadcast server together. Specifically, as shown in fig. 10, for example, the small segment v1.1 is labeled with an attribute tag a; labeling the small segment v1.2 with an attribute label b; labeling the small fragment v1.3 with an attribute label a; labeling the small fragment v1.4 with an attribute label c; labeling the small fragment v2.1 with an attribute label d; labeling the small fragment v2.2 with an attribute label f; labeling the small fragment v2.3 with an attribute label g; labeling the small fragment v2.4 with an attribute label a; labeling the small fragment v3.1 with an attribute label b; the small fragment v3.2 is labeled with an attribute label f.
Optionally, the attribute tags may be tags of game hero, for example, to indicate that the small segment is a small segment corresponding to a game hero; or the attribute tag can also be a hero type tag to indicate that the small segment is a small segment corresponding to a certain hero type; or the attribute tag may be of a combat type to indicate whether the small segment is a team video segment or a singleton video segment, etc.
In an embodiment of the present application, the heat value score of each small segment can be obtained according to the amount of interaction of each small segment in unit time. The mutual amount can include factors such as the number of people watching small segments, the number of people commenting on small segments, the number of bullet screens and the like, and meanwhile, because live broadcast and real-time data are related, variables with time-varying trends can be considered when the heat value score of each small segment is calculated. Alternatively, the trend may be represented by a natural index of the derivative of the heat factor P.
For example, in one embodiment, the heat value score for each snippet may be calculated by the following formula: h = Base (P) + e f'(P)
Wherein, P represents a popularity factor, which can be the number of people watching live broadcast, or the amount of money sending gifts, the number of comments on live broadcast per capita, the number of barrages, the duration of interaction with wheat, and the like. When a plurality of heat factors exist, the weighted average of the plurality of heat factors can be taken, so that the purposes of operation such as control, balance, screen refreshing and the like are achieved. Base (P) represents the Base of the heat factor, e.g., the number of people watching live in real time; and f' (P) represents the derivative of the dynamic change of people number, and the trend of the change of the live interaction is improved. Alternatively, the above expression may be understood as: when the number of people who live and watch is large and the number of interactions is large, and the number of people in the past unit time increases rapidly or the number of interactions increases suddenly, the popularity is high.
In an embodiment of the application, an interest tag of a user is a user feature, an attribute tag of a small video segment is a feature of the small video segment, and the user feature and the feature of the small video segment are two related concepts. The initial characteristics of the small video segments can be artificially divided, for example, the anchor can select a corresponding classification during playing, or the platform performs manual marking according to the content, and the like. The characteristics of the small video segments are not constant, and can be determined by weighting the user who views the video and the interaction behavior as a whole, for example, the higher the interaction and the longer the viewing time of the user with the corresponding interest tag to a certain small video segment, the larger the influence of the interest tag of the user on the attribute tag of the small video segment, and the attribute tag of the small video segment can be adjusted accordingly.
The user features are feature vectors formed according to favorite categories actively selected by the user and behavior data (such as attention behaviors in a live broadcast room, comment praise behaviors and the like) of the user.
When the user features are associated with the features of the video snippets, the user features may be used as a feature vector x, and the features of the video snippets may be used as a feature vector y, such as the feature vector x = (x =) (x) 1 ,x 2 ,...,x n ) Eigenvector y = (y) 1 ,y 2 ,...,y n ) The two feature vectors can be represented as two points in Euclidean space, and then the Euclidean distance d between the point x and the point y Euclidean (x, y) or d Euclidean (y, x) can be expressed as:
Figure BDA0003093405660000191
the euclidean distance between two feature vectors may represent the magnitude of the correlation between the feature vectors. The smaller the value is, the higher the matching degree between the two feature vectors is, so that one or more small segments with higher matching degree can be selected from each large segment to synthesize a video collection and display the video collection to a user, thereby achieving the purpose of thousands of people.
Specifically, as shown in fig. 11, assuming that the interest tags of the user are a, d, e, and g, the matching video snippets obtained by matching with the attribute tags of the video snippets may be v1.1, v1.3, v2.1, v2.3, and v2.4.
Since there are no small segments in the large segment Video3 that match the interest tag of the user, and it is necessary to ensure at least one small segment is selected from each large segment in order to ensure the continuity of the Video highlight content, a small segment can be selected from the Video3 according to the heat value, assuming that the small segments that are finally screened are v1.1, v1.3, v2.1, v2.3, v2.4, v3.2.
After the small segments are screened out, the small segments can be edited and combined into a video collection in the live broadcast server in a time sequence. If the duration of the synthesized video highlights is longer than the preset time of 1 minute (and 5 seconds), the small segments with low heat value can be removed according to the heat value sequence so as to meet the requirement that the duration of the video highlights is equal to 1 minute (and 5 seconds). After the small segments are removed, it is also satisfied that at least 1 small segment is selected in each large segment to ensure the continuity of the content. Of course, in one embodiment of the present application, if the duration of the synthesized video highlights is greater than the preset 1 minute time (and ± 5 seconds), then a duration of 1 minute (and ± 5 seconds) may also be truncated therefrom as the final video highlight.
If the duration of the composite video compilation is less than the preset 1 minute time (and 5 seconds), then other small video segments can be added in sequence according to the heat value until the composite video compilation meets the duration requirement. The time length requirement in this embodiment is only an example, and can be flexibly set according to the requirement in practical application.
After the live client generates the video highlights, the video highlights can be sent to the live client.
With continued reference to FIG. 6, the method further includes the steps of:
step S603, the live broadcast client starts playing the video album. Such as playing the video highlights on a live room interface of a live client.
Step S604, the live broadcast client judges whether to skip the video highlights, if not, the live broadcast progress of the current live broadcast room is played after the video highlights are played; and if the video collection is skipped, directly playing the live broadcast progress of the current live broadcast room.
Step S605, if the live broadcast client skips the video highlights, the live broadcast server identifies the attribute tag a of the video clip played during skipping, and then reduces the weight of the tag a in the interest tag under the current user ID, so as to reduce recommendation of such video clip.
Based on the technical scheme of the embodiment shown in fig. 6, in an application scenario of the present application, as shown in fig. 12, a plurality of live broadcast rooms in live broadcast are displayed in a live broadcast list interface of a live broadcast client, and a user can enter a certain live broadcast room through a click operation, for example, after clicking a region 1201 in the interface, enter a live broadcast room named "national service fielding i returns to dash and dash". After the user enters the live broadcast room, the live broadcast server extracts a plurality of segments from the broadcast content according to the interest tag of the current user, automatically generates a video album for 1 minute (for example only) in real time (optionally, the video album is generated when the broadcast content not watched by the user exceeds a certain threshold, for example, 1 hour), and plays the video album through the live broadcast client. As shown in fig. 13, after entering the live broadcasting room, the video album is played in a window 1301 of the live broadcasting interface, and a play progress bar 1303 may be displayed in a lower area of the window 1301. Meanwhile, a guidance button 1302 may be displayed in the window 1301 to prompt the user to skip the video highlights by triggering the guidance button 1302.
After the live client has finished playing the video highlights, the current live progress can be automatically played in a window 1401, as shown in fig. 14. Of course, the user may also click on the guide button 1302 to skip the video highlights and directly enter the interface shown in fig. 14 to view the current live schedule.
It should be noted that: if the user exits the live broadcast room while watching the live broadcast, if the user enters the live broadcast room again, if the unviewed time interval exceeds the threshold time for generating the video highlights by 1 hour (for example only), a new video highlight in the unviewed period is also generated in real time.
According to the technical scheme of the embodiment, the user can efficiently know the played content of the live broadcast room according to the video collection segment when entering the live broadcast room, so that the user can quickly follow up the progress of the live broadcast of the video, the acceptability of the live broadcast content is improved, the live broadcast watching experience of the user is also improved, and meanwhile the user viscosity of a live broadcast platform can also be improved.
The following describes an embodiment of an apparatus of the present application, which may be used to execute a live video processing method in the foregoing embodiment of the present application. For details that are not disclosed in the embodiments of the apparatus of the present application, please refer to the embodiments of the live video processing method described above in the present application.
Fig. 15 shows a block diagram of a processing apparatus of a live video according to an embodiment of the present application, which may be disposed in a live server, which may be the live server 103 shown in fig. 1.
Referring to fig. 15, a device 1500 for processing live video according to an embodiment of the present application includes: a first receiving unit 1502, a determining unit 1504, a selecting unit 1506, and a first processing unit 1508.
The first receiving unit 1502 is configured to receive a join request for a target live broadcast room, where the join request includes identification information of a user requesting to join the target live broadcast room; the determining unit 1504 is configured to determine live content missed by the user according to the played time length of the target live broadcast room and the identification information of the user; the selection unit 1506 is configured to select at least one video clip from the live content missed by the user; the first processing unit 1508 is configured to generate a video highlight segment according to the selected video clip and present the video highlight segment to the user.
In some embodiments of the present application, based on the foregoing scheme, the determining unit 1504 is configured to: acquiring a time point when the user joins the target live broadcast room according to the identification information of the user; and determining the live broadcast time period missed by the user according to the broadcast time length of the target live broadcast time period and the time point of the user joining the target live broadcast time period, so as to determine the live broadcast content missed by the user based on the live broadcast time period missed by the user.
In some embodiments of the present application, based on the foregoing solution, the first processing unit 1508 is further configured to: before at least one video segment is selected from the live content missed by the user, determining the live time missed by the user according to the live time period missed by the user; if the live broadcasting time length missed by the user is less than or equal to a first time length, presenting the real-time live broadcasting content of the target live broadcasting room to the user; the selection unit is configured to: and if the live broadcast time length missed by the user is greater than the first time length, executing a process of selecting at least one video clip from the live broadcast content missed by the user.
In some embodiments of the present application, based on the foregoing solution, the processing apparatus 1500 of the live video further includes: the second processing unit is configured to segment the played content of the target live broadcast room by taking a second time length as a segmentation reference to obtain a plurality of video time periods; and splitting the video content in the video time interval to obtain a plurality of video segments corresponding to the played content.
In some embodiments of the present application, based on the foregoing solution, the second processing unit is configured to split the video content in the video period by at least one of:
performing semantic analysis on the video content in the video time interval to split the video content in the video time interval based on a semantic analysis result;
performing content identification processing on the video content in the video time interval so as to split the video content in the video time interval based on a content identification result;
performing heat analysis on the video content in the video time interval so as to split the video content in the video time interval based on the heat analysis result.
In some embodiments of the present application, based on the foregoing scheme, the selecting unit 1506 is configured to: determining at least one video time interval to which the live content missed by the user belongs based on a plurality of video segments corresponding to the played content; at least one video segment is selected from the at least one video period, respectively.
In some embodiments of the present application, based on the foregoing scheme, the selecting unit 1506 is configured to: according to the user characteristics of the user, video clips matched with the user characteristics are selected from the at least one video time interval respectively; and if a video clip matched with the user characteristic is not selected from a target video time interval in the at least one video time interval according to the user characteristic, selecting at least one video clip from the video clips obtained by splitting the target video time interval according to the heat value of the video clip.
In some embodiments of the present application, based on the foregoing solution, the processing apparatus 1500 of the live video further includes: the third processing unit is configured to acquire interaction data corresponding to each video clip; determining the heat factor of each video clip according to the interactive data corresponding to each video clip; calculating the interactive change trend corresponding to each video clip according to the heat factor of each video clip; and calculating the heat value of each video clip based on the heat factor of each video clip and the interaction change trend corresponding to each video clip.
In some embodiments of the present application, based on the foregoing solution, the first processing unit 1508 is configured to: synthesizing the selected video clips according to the time sequence of the video to obtain synthesized video clips; and generating a video collection segment with a set duration range based on the synthesized video segment.
In some embodiments of the present application, based on the foregoing solution, the first processing unit 1508 is configured to: if the time length of the synthesized video clip exceeds the set time length range and at least two video clips are selected from the same video time interval, removing the video clips from the at least two video clips according to the heat value or the matching degree with the user characteristics until the time length of the synthesized video clip is within the set time length range; and taking the video segment within the set duration range obtained by synthesis as the video collection segment.
In some embodiments of the present application, based on the foregoing solution, the first processing unit 1508 is configured to: and if the duration of the synthesized video clip exceeds the set duration range, intercepting the video clip which accords with the set duration range from the synthesized video clip to be used as the video collection clip.
In some embodiments of the present application, based on the foregoing solution, the first processing unit 1508 is configured to: if the duration of the synthesized video clip does not reach the set duration range, selecting a video clip from other video clips of the at least one video time interval according to the heat value until the duration of the synthesized video clip is within the set duration range; and taking the video segment within the set duration range obtained by synthesis as the video collection segment.
In some embodiments of the present application, based on the foregoing solution, the first processing unit 1508 is further configured to: in the process of presenting the video highlight segment to the user, if a skip playing instruction for the video highlight segment is detected, identifying a target video segment presented when the skip playing instruction is received; and adjusting the user characteristics of the user according to the attribute label of the target video clip.
In some embodiments of the present application, based on the foregoing solution, the first processing unit 1508 is further configured to: and performing semantic analysis and/or content recognition processing on the video clip of the played content of the target live broadcasting room to obtain an attribute tag of the video clip, wherein the attribute tag is used for calculating the matching degree with the user characteristics.
In some embodiments of the present application, based on the foregoing solution, the first processing unit 1508 is further configured to: and adjusting the attribute tags of the video clips according to the interactive data corresponding to the video clips.
Fig. 16 shows a block diagram of a processing apparatus for live video according to an embodiment of the present application, which may be disposed in a live client, which may be the terminal device 104 shown in fig. 1.
Referring to fig. 16, a device 1600 for processing live video according to an embodiment of the present application includes: a display unit 1602, a sending unit 1604, a second receiving unit 1606, and a playing unit 1608.
Wherein the display unit 1602 is configured to display live room information that is being live; the sending unit 1604 is configured to send, to a server, an affiliation request for a target live broadcast room if it is detected that a user enters the target live broadcast room, where the affiliation request includes identification information of the user; the second receiving unit 1606 is configured to receive a video highlight segment transmitted by the server, where the video highlight segment is generated according to a video segment selected from live content missed by the user; the playing unit 1608 is configured to play the video highlight clip.
In some embodiments of the present application, based on the foregoing scheme, the playing unit 1608 is further configured to: and if a skip playing instruction for the video collection segment is received or the video collection segment is played completely, playing the real-time live broadcast content of the target live broadcast room.
FIG. 17 illustrates a schematic structural diagram of a computer system suitable for use to implement the electronic device of the embodiments of the subject application.
It should be noted that the computer system 1700 of the electronic device shown in fig. 17 is only an example, and should not bring any limitation to the functions and the scope of the application of the embodiments.
As shown in fig. 17, a computer system 1700 includes a Central Processing Unit (CPU) 1701 that can perform various appropriate actions and processes, such as executing the methods described in the above embodiments, according to a program stored in a Read-Only Memory (ROM) 1702 or a program loaded from a storage portion 1708 into a Random Access Memory (RAM) 1703. The RAM 1703 also stores various programs and data necessary for system operation. The CPU 1701, ROM 1702, and RAM 1703 are connected to one another by a bus 1704. An Input/Output (I/O) interface 1705 is also connected to the bus 1704.
The following components are connected to the I/O interface 1705: an input portion 1706 including a keyboard, a mouse, and the like; an output section 1707 including a Display such as a Cathode Ray Tube (CRT), a Liquid Crystal Display (LCD), and a speaker; a storage portion 1708 including a hard disk and the like; and a communication section 1709 including a Network interface card such as a Local Area Network (LAN) card, a modem, or the like. The communication section 1709 performs communication processing via a network such as the internet. A driver 1710 is also connected to the I/O interface 1705 as necessary. A removable medium 1711 such as a magnetic disk, an optical disk, a magneto-optical disk, a semiconductor memory, or the like is mounted on the drive 1710 as necessary, so that a computer program read out therefrom is mounted into the storage portion 1708 as necessary.
In particular, according to embodiments of the application, the processes described above with reference to the flow diagrams may be implemented as computer software programs. For example, embodiments of the present application include a computer program product comprising a computer program embodied on a computer readable medium, the computer program comprising a computer program for performing the method illustrated by the flow chart. In such an embodiment, the computer program may be downloaded and installed from a network through the communication section 1709 and/or installed from the removable medium 1711. When the computer program is executed by a Central Processing Unit (CPU) 1701, various functions defined in the system of the present application are executed.
It should be noted that the computer readable medium shown in the embodiments of the present application may be a computer readable signal medium or a computer readable storage medium or any combination of the two. A computer readable storage medium may be, for example, but not limited to, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, or device, or any combination of the foregoing. More specific examples of the computer readable storage medium may include, but are not limited to: an electrical connection having one or more wires, a portable computer diskette, a hard disk, a Random Access Memory (RAM), a Read-Only Memory (ROM), an Erasable Programmable Read-Only Memory (EPROM), a flash Memory, an optical fiber, a portable Compact Disc Read-Only Memory (CD-ROM), an optical storage device, a magnetic storage device, or any suitable combination of the foregoing. In the context of this application, a computer readable storage medium may be any tangible medium that can contain, or store a program for use by or in connection with an instruction execution system, apparatus, or device. In this application, however, a computer readable signal medium may include a propagated data signal with a computer program embodied therein, for example, in baseband or as part of a carrier wave. Such a propagated data signal may take any of a variety of forms, including, but not limited to, electro-magnetic, optical, or any suitable combination thereof. A computer readable signal medium may also be any computer readable medium that is not a computer readable storage medium and that can communicate, propagate, or transport a program for use by or in connection with an instruction execution system, apparatus, or device. The computer program embodied on the computer readable medium may be transmitted using any appropriate medium, including but not limited to: wireless, wired, etc., or any suitable combination of the foregoing.
The flowchart and block diagrams in the figures illustrate the architecture, functionality, and operation of possible implementations of systems, methods and computer program products according to various embodiments of the present application. Each block in the flowchart or block diagrams may represent a module, segment, or portion of code, which comprises one or more executable instructions for implementing the specified logical function(s). It should also be noted that, in some alternative implementations, the functions noted in the block may occur out of the order noted in the figures. For example, two blocks shown in succession may, in fact, be executed substantially concurrently, or the blocks may sometimes be executed in the reverse order, depending upon the functionality involved. It will also be noted that each block of the block diagrams or flowchart illustration, and combinations of blocks in the block diagrams or flowchart illustration, can be implemented by special purpose hardware-based systems which perform the specified functions or acts, or combinations of special purpose hardware and computer instructions.
The units described in the embodiments of the present application may be implemented by software, or may be implemented by hardware, and the described units may also be disposed in a processor. Wherein the names of the elements do not in some way constitute a limitation on the elements themselves.
As another aspect, the present application also provides a computer-readable medium, which may be contained in the electronic device described in the above embodiment; or may exist separately without being assembled into the electronic device. The computer readable medium carries one or more programs, which when executed by one of the electronic devices, cause the electronic device to implement the method described in the above embodiments.
It should be noted that although in the above detailed description several modules or units of the device for action execution are mentioned, such a division is not mandatory. Indeed, the features and functions of two or more modules or units described above may be embodied in one module or unit according to embodiments of the application. Conversely, the features and functions of one module or unit described above may be further divided into embodiments by a plurality of modules or units.
Through the above description of the embodiments, those skilled in the art will readily understand that the exemplary embodiments described herein may be implemented by software, or by software in combination with necessary hardware. Therefore, the technical solution according to the embodiments of the present application may be embodied in the form of a software product, which may be stored in a non-volatile storage medium (which may be a CD-ROM, a usb disk, a removable hard disk, etc.) or on a network, and includes several instructions to enable a computing device (which may be a personal computer, a server, a touch terminal, or a network device, etc.) to execute the method according to the embodiments of the present application.
Other embodiments of the present application will be apparent to those skilled in the art from consideration of the specification and practice of the embodiments disclosed herein. This application is intended to cover any variations, uses, or adaptations of the invention following, in general, the principles of the application and including such departures from the present disclosure as come within known or customary practice within the art to which the invention pertains.
It will be understood that the present application is not limited to the precise arrangements described above and shown in the drawings and that various modifications and changes may be made without departing from the scope thereof. The scope of the application is limited only by the appended claims.

Claims (20)

1. A method for processing live video is characterized by comprising the following steps:
receiving a joining request aiming at a target live broadcasting room, wherein the joining request comprises identification information of a user requesting to join the target live broadcasting room;
determining live broadcast content missed by the user according to the played time of the target live broadcast room and the identification information of the user;
selecting at least one video segment from the live content missed by the user;
and generating a video collection segment according to the selected video segment, and presenting the video collection segment to the user.
2. The method for processing the live video according to claim 1, wherein determining the live content missed by the user according to the played time length of the target live room and the identification information of the user includes:
acquiring a time point when the user joins the target live broadcast room according to the identification information of the user;
and determining the live broadcast time period missed by the user according to the played time length of the target live broadcast time period and the time point of the user joining the target live broadcast time period, so as to determine the live broadcast content missed by the user based on the live broadcast time period missed by the user.
3. The method of processing live video according to claim 2, wherein before selecting at least one video segment from the live content missed by the user, the method of processing live video further comprises:
determining the live broadcast duration missed by the user according to the live broadcast period missed by the user;
if the live broadcast duration missed by the user is less than or equal to a first duration, presenting the real-time live broadcast content of the target live broadcast room to the user;
and if the live broadcast time length missed by the user is greater than the first time length, executing a process of selecting at least one video clip from the live broadcast content missed by the user.
4. The method for processing live video according to claim 1, wherein before selecting at least one video segment from live content missed by the user, the method for processing live video further comprises:
segmenting the played content of the target live broadcast room by taking a second time length as a segmentation reference to obtain a plurality of video time periods;
and splitting the video content in the video time interval to obtain a plurality of video segments corresponding to the played content.
5. The method for processing the live video according to claim 4, wherein splitting the video content in the video period includes at least one of:
performing semantic analysis on the video content in the video time interval to split the video content in the video time interval based on a semantic analysis result;
performing content identification processing on the video content in the video time interval so as to split the video content in the video time interval based on a content identification result;
performing heat analysis on the video content in the video time interval so as to split the video content in the video time interval based on the heat analysis result.
6. The method for processing live video according to claim 4, wherein selecting at least one video segment from live content missed by the user comprises:
determining at least one video time interval to which the live content missed by the user belongs based on a plurality of video segments corresponding to the played content;
at least one video segment is selected from the at least one video period, respectively.
7. The method of processing live video according to claim 6, wherein selecting at least one video clip from the at least one video period respectively comprises:
according to the user characteristics of the user, video clips matched with the user characteristics are selected from the at least one video time interval respectively;
and if the video clip matched with the user characteristic is not selected from the target video time interval in the at least one video time interval according to the user characteristic, selecting the at least one video clip from the video clips obtained by splitting the target video time interval according to the heat value of the video clip.
8. The method for processing live video according to claim 7, further comprising:
acquiring interactive data corresponding to each video clip;
determining the heat factor of each video clip according to the interactive data corresponding to each video clip;
calculating the interactive change trend corresponding to each video clip according to the heat factor of each video clip;
and calculating the heat value of each video clip based on the heat factor of each video clip and the interaction change trend corresponding to each video clip.
9. The method for processing the live video according to any one of claims 6 to 8, wherein generating a video highlight segment according to the selected video segment comprises:
synthesizing the selected video clips according to the time sequence of the video to obtain synthesized video clips;
and generating a video collection segment with a set duration range based on the synthesized video segment.
10. The method for processing the live video according to claim 9, wherein generating a video highlight segment with a set duration range based on the synthesized video segment comprises:
if the time length of the synthesized video clip exceeds the set time length range and at least two video clips are selected from the same video time interval, removing the video clips from the at least two video clips according to the heat value or the matching degree with the user characteristics until the time length of the synthesized video clip is within the set time length range;
and taking the synthesized video segment within the set duration range as the video collection segment.
11. The method for processing the live video according to claim 9, wherein generating a video highlight clip with a set duration range based on the synthesized video clip comprises:
and if the duration of the synthesized video clip exceeds the set duration range, intercepting the video clip which accords with the set duration range from the synthesized video clip to be used as the video collection clip.
12. The method for processing the live video according to claim 9, wherein generating a video highlight clip with a set duration range based on the synthesized video clip comprises:
if the duration of the synthesized video clip does not reach the set duration range, selecting a video clip from other video clips of the at least one video time interval according to the heat value until the duration of the synthesized video clip is within the set duration range;
and taking the video segment within the set duration range obtained by synthesis as the video collection segment.
13. The method for processing live video according to any one of claims 1 to 8, further comprising:
in the process of presenting the video highlight segment to the user, if a skip playing instruction for the video highlight segment is detected, identifying a target video segment presented when the skip playing instruction is received;
and adjusting the user characteristics of the user according to the attribute tag of the target video clip.
14. The method for processing live video according to any one of claims 1 to 8, further comprising:
and performing semantic analysis and/or content identification processing on the video clip of the played content in the target live broadcast room to obtain an attribute label of the video clip, wherein the attribute label is used for calculating the matching degree with the user characteristics.
15. The method for processing live video according to claim 14, further comprising:
and adjusting the attribute tags of the video clips according to the interactive data corresponding to the video clips.
16. A method for processing live video is characterized by comprising the following steps:
displaying live broadcasting room information which is being live broadcasted;
if the fact that the user enters the target live broadcast room is detected, sending a joining request aiming at the target live broadcast room to a server side, wherein the joining request comprises identification information of the user;
receiving a video highlight segment transmitted by the server, wherein the video highlight segment is generated according to a video segment selected from live broadcast contents missed by the user;
and playing the video collection segment.
17. The method for processing live video according to claim 16, further comprising:
and if a skip playing instruction for the video collection segment is received or the playing of the video collection segment is completed, playing the real-time live broadcast content of the target live broadcast room.
18. A live video processing apparatus, comprising:
the device comprises a first receiving unit, a second receiving unit and a third receiving unit, wherein the first receiving unit is configured to receive a joining request aiming at a target live broadcast room, and the joining request comprises identification information of a user requesting to join the target live broadcast room;
the determining unit is configured to determine live broadcast contents missed by the user according to the played time length of the target live broadcast room and the identification information of the user;
a selection unit configured to select at least one video segment from the live content missed by the user;
a first processing unit configured to generate a video highlight clip according to the selected video clip, and present the video highlight clip to the user.
19. A computer-readable medium, on which a computer program is stored, which, when being executed by a processor, carries out a method of processing live video according to any one of claims 1 to 15; or implementing a method of processing live video as claimed in any of claims 16 to 17.
20. An electronic device, comprising:
one or more processors;
storage means for storing one or more programs which, when executed by the one or more processors, cause the one or more processors to implement a method of processing live video as claimed in any one of claims 1 to 15; or implementing a method of processing live video as claimed in any of claims 16 to 17.
CN202110603559.3A 2021-05-31 2021-05-31 Live video processing method and device, computer readable medium and electronic equipment Pending CN115484467A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202110603559.3A CN115484467A (en) 2021-05-31 2021-05-31 Live video processing method and device, computer readable medium and electronic equipment

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202110603559.3A CN115484467A (en) 2021-05-31 2021-05-31 Live video processing method and device, computer readable medium and electronic equipment

Publications (1)

Publication Number Publication Date
CN115484467A true CN115484467A (en) 2022-12-16

Family

ID=84420195

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202110603559.3A Pending CN115484467A (en) 2021-05-31 2021-05-31 Live video processing method and device, computer readable medium and electronic equipment

Country Status (1)

Country Link
CN (1) CN115484467A (en)

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109862388A (en) * 2019-04-02 2019-06-07 网宿科技股份有限公司 Generation method, device, server and the storage medium of the live video collection of choice specimens
CN110225411A (en) * 2019-05-31 2019-09-10 北京奇艺世纪科技有限公司 The segment of programme televised live reviews method, system, computer equipment and medium
CN111918085A (en) * 2020-08-06 2020-11-10 腾讯科技(深圳)有限公司 Live broadcast processing method and device, electronic equipment and computer readable storage medium

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109862388A (en) * 2019-04-02 2019-06-07 网宿科技股份有限公司 Generation method, device, server and the storage medium of the live video collection of choice specimens
CN110225411A (en) * 2019-05-31 2019-09-10 北京奇艺世纪科技有限公司 The segment of programme televised live reviews method, system, computer equipment and medium
CN111918085A (en) * 2020-08-06 2020-11-10 腾讯科技(深圳)有限公司 Live broadcast processing method and device, electronic equipment and computer readable storage medium

Similar Documents

Publication Publication Date Title
US12094209B2 (en) Video data processing method and apparatus, device, and medium
CN108184144B (en) Live broadcast method and device, storage medium and electronic equipment
US10345897B2 (en) Spectator interactions with games in a specatating system
US10484439B2 (en) Spectating data service for a spectating system
US10390064B2 (en) Participant rewards in a spectating system
US20120072936A1 (en) Automatic Customized Advertisement Generation System
CN107633441A (en) Commodity in track identification video image and the method and apparatus for showing merchandise news
US20170001112A1 (en) Game content interface in a spectating system
US20170001122A1 (en) Integrating games systems with a spectating system
US20170003784A1 (en) Game effects from spectating community inputs
KR20180022866A (en) Integration of the specification and game systems
CN111491173A (en) Live broadcast cover determining method and device, computer equipment and storage medium
CN106658200A (en) Live video sharing and obtaining methods and devices, and terminal equipment thereof
US10363488B1 (en) Determining highlights in a game spectating system
US10864447B1 (en) Highlight presentation interface in a game spectating system
CN114025188B (en) Live advertisement display method, system, device, terminal and readable storage medium
CN113490004B (en) Live broadcast interaction method and related device
CN112287848B (en) Live broadcast-based image processing method and device, electronic equipment and storage medium
JP2007129531A (en) Program presentation system
CN113766299A (en) Video data playing method, device, equipment and medium
CN111970525B (en) Live broadcast room searching method and device, server and storage medium
CN111615002A (en) Video background playing control method, device and system and electronic equipment
CN111970521B (en) Live broadcast method and device of virtual anchor, computer equipment and storage medium
CN111479119A (en) Method, device and system for collecting feedback information in live broadcast and storage medium
CN114139491A (en) Data processing method, device and storage medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination