CN111050190A - Encoding method, device and equipment of live video stream and storage medium - Google Patents

Encoding method, device and equipment of live video stream and storage medium Download PDF

Info

Publication number
CN111050190A
CN111050190A CN201911422688.1A CN201911422688A CN111050190A CN 111050190 A CN111050190 A CN 111050190A CN 201911422688 A CN201911422688 A CN 201911422688A CN 111050190 A CN111050190 A CN 111050190A
Authority
CN
China
Prior art keywords
video stream
region
live
integrated
live video
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201911422688.1A
Other languages
Chinese (zh)
Other versions
CN111050190B (en
Inventor
陈文琼
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Guangzhou Kugou Computer Technology Co Ltd
Original Assignee
Guangzhou Kugou Computer Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Guangzhou Kugou Computer Technology Co Ltd filed Critical Guangzhou Kugou Computer Technology Co Ltd
Priority to CN201911422688.1A priority Critical patent/CN111050190B/en
Publication of CN111050190A publication Critical patent/CN111050190A/en
Application granted granted Critical
Publication of CN111050190B publication Critical patent/CN111050190B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/21Server components or server architectures
    • H04N21/218Source of audio or video content, e.g. local disk arrays
    • H04N21/2187Live feed
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/102Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
    • H04N19/124Quantisation
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/134Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
    • H04N19/146Data rate or code amount at the encoder output
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/134Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
    • H04N19/167Position within a video image, e.g. region of interest [ROI]
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/23Processing of content or additional data; Elementary server operations; Server middleware
    • H04N21/234Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs
    • H04N21/23418Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs involving operations for analysing video streams, e.g. detecting features or characteristics
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/23Processing of content or additional data; Elementary server operations; Server middleware
    • H04N21/234Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs
    • H04N21/23424Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs involving splicing one content stream with another content stream, e.g. for inserting or substituting an advertisement
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/23Processing of content or additional data; Elementary server operations; Server middleware
    • H04N21/234Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs
    • H04N21/2343Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs involving reformatting operations of video signals for distribution or compliance with end-user requests or end-user device requirements
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/47End-user applications
    • H04N21/478Supplemental services, e.g. displaying phone caller identification, shopping application
    • H04N21/4788Supplemental services, e.g. displaying phone caller identification, shopping application communicating with other users, e.g. chatting

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Databases & Information Systems (AREA)
  • Business, Economics & Management (AREA)
  • Marketing (AREA)
  • General Engineering & Computer Science (AREA)
  • Compression Or Coding Systems Of Tv Signals (AREA)

Abstract

The embodiment of the disclosure provides a method, a device, equipment and a storage medium for encoding a live video stream, which relate to the field of multimedia processing, and the method comprises the following steps: acquiring a first direct-playing video stream; carrying out target area identification on the first direct-playing video stream to obtain first area information; receiving a second live video stream which is sent by a second terminal and corresponds to second area information; integrating the first live video stream and the second live video stream and obtaining integrated area information according to an integration mode; and carrying out partition coding by combining the integrated area information to obtain a live broadcast coding stream. According to the method and the device, when the integrated video is coded, the integrated video is coded in a partitioned mode according to the information of the integrated area, different coding quantization parameters are adopted in different areas, so that coding with high code rate is carried out on important areas, coding with low code rate is carried out on non-important areas, and the bandwidth occupied by the coded video of the integrated video is reduced.

Description

Encoding method, device and equipment of live video stream and storage medium
Technical Field
The present disclosure relates to the field of multimedia processing, and in particular, to a method, an apparatus, a device, and a storage medium for encoding a live video stream.
Background
The webcast is a multimedia form for performing webcast on video content on a webcast platform by using a communication network and a multimedia communication technology, optionally, the webcast platform can perform webcast on video content of a single anchor, and also can perform webcast on a plurality of anchor content, wherein the webcast means that video interaction is performed between at least two anchors, and a live broadcast mode for performing live broadcast on integrated video of the video interaction is performed.
In the related technology, the example that the anchor A and the anchor B start live broadcast with live broadcast is taken as an example, the example that audiences watch live broadcast videos in a live broadcast room of the anchor A from the view angle of the anchor A is taken as an example, the live broadcast videos comprise a live broadcast video stream a of the anchor A and a live broadcast video stream B of the anchor B, the anchor B codes the live broadcast video stream B and sends the coded live broadcast video stream B to a terminal of the anchor A, and the live broadcast video stream a and the live broadcast video stream B are integrated by the terminal of the anchor A and then are pushed to a live broadcast server for live broadcast.
However, when the live broadcast is performed through the foregoing manner, since the terminal of the anchor broadcast a needs to integrate the live video stream a and the live video stream b and then push the integrated live video stream a and live video stream b to the live broadcast server, the bandwidth occupied by the live video stream is large, the storage space occupied by the live video stream is large, and the video processing efficiency is reduced.
Disclosure of Invention
The embodiment of the disclosure provides a method, a device, equipment and a storage medium for encoding a live video stream, which can reduce the bandwidth occupation of the live video stream. The technical scheme is as follows:
according to an aspect of the present disclosure, there is provided a method of encoding a live video stream, the method including:
acquiring a first direct-playing video stream, wherein the first direct-playing video stream is a video stream generated by a first terminal in a process of carrying out live microphone connection with a second terminal;
identifying a target area of the first direct-playing video stream to obtain first area information, wherein the target area is an area where the content of a target type is located;
receiving a second live video stream sent by the second terminal, wherein the second live video stream corresponds to second area information, and the second area information is used for indicating an area where the target type content in the second live video stream is located;
integrating the first live video stream and the second live video stream to obtain an integrated video, and combining the first region information and the second region information according to an integration mode to obtain integrated region information;
and carrying out partition coding on the integrated video in combination with the integrated area information to obtain a live broadcast coding stream.
In an optional embodiment, the performing partition coding on the integrated video in combination with the integrated region information to obtain a live broadcast coded stream includes:
coding a first region corresponding to the integrated region information in the integrated video at a first code rate;
encoding a second region in the integrated video at a second code rate, the second region being a region other than the first region;
wherein the first code rate is higher than the second code rate.
In an optional embodiment, a target video frame in the integrated video corresponds to frame region information in the integrated region information;
the encoding a first region corresponding to the integrated region information in the integrated video at a first code rate includes:
encoding the first region corresponding to the frame region information in the target video frame at the first code rate;
the encoding a second region in the integrated video at a second code rate, comprising:
encoding the second region in the target video frame at the second code rate.
In an optional embodiment, after performing target area identification on the first live video stream to obtain first area information, the method further includes:
coding the first direct-playing video stream by combining the first area information to obtain a first coded stream;
and sending the first coding stream to the second terminal.
In an optional embodiment, the encoding the first direct current video stream with the first region information to obtain a first encoded stream includes:
encoding a third region corresponding to the first region information in the first live video stream at a third code rate;
encoding a fourth region in the first direct-playing video stream at a fourth code rate, wherein the fourth region is a region except the third region;
wherein the third code rate is higher than the fourth code rate.
In an optional embodiment, the receiving a second live video stream sent by the second terminal further includes:
receiving a second coding stream sent by the second terminal;
decoding the second encoded stream to obtain the second live video stream;
and acquiring the second area information corresponding to the second live video stream.
In an optional embodiment, the obtaining the second region information corresponding to the second live video stream includes:
and acquiring the second area information from the supplementary enhancement information attached to the image frame in the second live video stream.
In an optional embodiment, the performing target area identification on the first live video stream to obtain first area information includes:
and performing face recognition on the first live video stream to obtain first face region information as the first region information, wherein the second region information is used for indicating second face region information in the second live video stream.
In another aspect, an apparatus for encoding a live video stream is provided, the apparatus including:
the system comprises an acquisition module, a processing module and a processing module, wherein the acquisition module is used for acquiring a first direct-playing video stream, and the first direct-playing video stream is a video stream generated by a first terminal in the process of carrying out live wheat connection with a second terminal;
the identification module is used for identifying a target area of the first direct-playing video stream to obtain first area information, wherein the target area is an area where the content of a target type is located;
a receiving module, configured to receive a second live video stream sent by the second terminal, where the second live video stream corresponds to second area information, and the second area information is used to indicate an area in which the target type content in the second live video stream is located;
the processing module is used for integrating the first live video stream and the second live video stream to obtain an integrated video and combining the first area information and the second area information according to an integration mode to obtain integrated area information;
and the coding module is used for carrying out partition coding on the integrated video combined with the integrated area information to obtain a live broadcast coding stream.
In an optional embodiment, the encoding module is further configured to encode a first region corresponding to the integrated region information in the integrated video at a first code rate; encoding a second region in the integrated video at a second code rate, the second region being a region other than the first region;
wherein the first code rate is higher than the second code rate.
In an optional embodiment, a target video frame in the integrated video corresponds to frame region information in the integrated region information;
the encoding module is further configured to encode the first region corresponding to the frame region information in the target video frame at the first code rate; encoding the second region in the target video frame at the second code rate.
In an optional embodiment, the encoding module is further configured to encode the first direct-play video stream in combination with the first region information to obtain a first encoded stream;
the device further comprises:
and the sending module is used for sending the first coding stream to the second terminal.
In an optional embodiment, the encoding module is further configured to encode a third region corresponding to the first region information in the first direct-view video stream at a third code rate; encoding a fourth region in the first direct-playing video stream at a fourth code rate, wherein the fourth region is a region except the third region;
wherein the third code rate is higher than the fourth code rate.
In an optional embodiment, the receiving module is further configured to receive a second encoded stream sent by the second terminal;
the device, still include:
a decoding module, configured to decode the second encoded stream to obtain the second live video stream;
the obtaining module is further configured to obtain the second region information corresponding to the second live video stream.
In an optional embodiment, the obtaining module is further configured to obtain the second area information from supplemental enhancement information attached to an image frame in the second live video stream.
In an optional embodiment, the identification module is further configured to perform face identification on the first live video stream to obtain first face region information as the first region information, and the second region information is used to indicate second face region information in the second live video stream.
In another aspect, a computer device is provided, which includes a processor and a memory, where at least one instruction is stored in the memory, and the instruction is loaded and executed by the processor to implement the encoding method of a live video stream as provided in the embodiments of the present disclosure.
In another aspect, a computer-readable storage medium is provided, wherein the storage medium stores at least one instruction, and the instruction is loaded and executed by a processor to implement the encoding method of a live video stream as provided in the embodiments of the present disclosure.
In another aspect, a computer program product is provided, which when run on a computer causes the computer to perform the method of encoding a live video stream as provided in the embodiments of the present disclosure described above.
The technical scheme provided by the embodiment of the disclosure has the beneficial effects that:
the method comprises the steps of determining first region information in a first live video stream and second region information in a second live video stream, integrating the first region information and the second region information after integrating the first live video stream and the second live video stream to obtain integrated region information, carrying out partition coding on the integrated video according to the integrated region information when the integrated video is coded, and adopting different coding quantization parameters in different regions to carry out coding with higher code rate on important regions and coding with lower code rate on non-important regions, so that the bandwidth occupied by the coded video of the integrated video is reduced, and the coding efficiency is improved.
Drawings
The accompanying drawings, which are incorporated in and constitute a part of this specification, illustrate embodiments consistent with the present disclosure and together with the description, serve to explain the principles of the disclosure.
Fig. 1 is a schematic diagram illustrating an encoding manner of a video image frame according to an exemplary embodiment of the present disclosure;
FIG. 2 is a schematic illustration of an implementation environment provided by an exemplary embodiment of the present disclosure;
fig. 3 is a flowchart of an encoding method of a live video stream provided by an exemplary embodiment of the present disclosure;
fig. 4 is a schematic arrangement of a first live video stream and a second live video stream provided based on the embodiment shown in fig. 3;
fig. 5 is a flowchart of an encoding method of a live video stream provided by another exemplary embodiment of the present disclosure;
fig. 6 is a flowchart of an encoding method of a live video stream provided by another exemplary embodiment of the present disclosure;
fig. 7 is a block diagram of an apparatus for encoding a live video stream according to an exemplary embodiment of the present disclosure;
fig. 8 is a block diagram of an apparatus for encoding a live video stream according to another exemplary embodiment of the present disclosure;
fig. 9 is a block diagram of a terminal according to an exemplary embodiment of the present disclosure.
Detailed Description
Reference will now be made in detail to the exemplary embodiments, examples of which are illustrated in the accompanying drawings. When the following description refers to the accompanying drawings, like numbers in different drawings represent the same or similar elements unless otherwise indicated. The implementations described in the exemplary embodiments below are not intended to represent all implementations consistent with the present disclosure. Rather, they are merely examples of apparatus and methods consistent with certain aspects of the present disclosure, as detailed in the appended claims.
First, terms related to embodiments of the present disclosure are briefly described:
region of Interest (ROI): the area to be processed is selected from the processed image in a frame selection manner in the image processing, and optionally, the frame selection manner may be a manner sketched by a square frame, a circular frame, an elliptical frame or an irregular frame. Optionally, the ROI belongs to one of Intelligent Video Encoding (IVE), the IVE technology can intelligently encode the video according to requirements, and under the ROI function, the high-quality Encoding is performed on the framed region in the video image frame, and the bitrate and the image quality are reduced for the non-framed region.
Optionally, in the embodiment of the present disclosure, a face region is taken as an ROI for example, that is, after a live video stream is acquired, a face recognition is performed on an image frame in the live video stream to obtain the face region, when the live video stream is encoded, the face region is encoded at a first code rate, and other regions except the face region are encoded at a second code rate, where the first code rate is higher than the second code rate.
Referring to fig. 1, schematically, a face region 110 is obtained by performing face recognition on a video image frame 100, when the video image frame 100 in a video stream is encoded, the face region 110 is encoded at a first code rate, and other regions 120 outside the face region 110 are encoded at a second code rate, so as to obtain an encoding result 130, where the first code rate is higher than the second code rate.
Next, an implementation environment related to the embodiment of the present disclosure is described, please refer to fig. 2, which shows a schematic diagram of an implementation environment provided by an exemplary embodiment of the present disclosure, as shown in fig. 2, the implementation environment includes an anchor terminal 210, an anchor terminal 220, a microphone connecting server 230, a streaming server 240, and a viewing terminal 250;
a connected-to-live connection relationship is established between the anchor terminal 210 and the anchor terminal 220 through the connected-to-live server 230, that is, a connected-to-live video live broadcast is performed between the anchor terminal 210 and the anchor terminal 220, the anchor terminal 210 pushes a live video of an anchor a to the anchor terminal 220, the anchor terminal 220 pushes a live video of an anchor B to the anchor terminal 210, optionally, the anchor terminal 220 pushes a live video of a live video B to the anchor terminal 210 for illustration, after acquiring a live video stream a, the anchor terminal 220 sends the live video stream a to the anchor terminal 210 through the connected-to-live server 230, after receiving the live video stream a, the anchor terminal 210 integrates the received live video stream a and the acquired live video stream B to obtain a connected-to-live video stream, and sends the connected-to-stream server 240, and the pushed-to-viewing terminal 250 through the pushed-stream server 240, therefore, the watching terminal can watch the connecting microphone video.
Optionally, when a connection relationship between the anchor terminal 210 and the anchor terminal 220 is established, the connection server 230 stores the connection relationship between the anchor terminal 210 and the anchor terminal 220, and after the anchor terminal 220 sends the live video stream a to the connection server 230, the connection server 230 sends the live video stream a to the anchor terminal 210 according to the connection relationship.
Optionally, the viewing terminal 250 needs to first select to view the live broadcast of the anchor terminal 210, the streaming server 240 records the on-demand relationship between the viewing terminal 250 and the anchor terminal 210, and when receiving the live broadcast of the anchor terminal 210, pushes the live broadcast of the anchor terminal 210 to the viewing terminal 250 that requests the live broadcast of the anchor terminal 250 according to the recorded on-demand relationship.
Optionally, a live video stream a pushed by the anchor terminal 220 to the anchor terminal 210 includes first ROI information, a live video stream b pushed by the anchor terminal 210 to the anchor terminal 220 includes second ROI information, when the anchor terminal 210 broadcasts a live video stream directly, first ROI information in the live video stream a is obtained, the live video stream a and the live video stream b are integrated, the first ROI information and the second ROI information are combined according to an integration manner to obtain integrated ROI information, and the integrated live video stream is encoded by combining the integrated ROI information with the integrated ROI information. And coding a target region corresponding to the integrated ROI information in the continuous microphone video stream at a first code rate, and coding other regions except the target region in the continuous microphone video stream at a second code rate.
The method for encoding a live video stream provided by the present disclosure is described in conjunction with the noun introduction and the implementation environment. Fig. 3 is a flowchart of an encoding method of a live video stream according to an exemplary embodiment of the present disclosure, which is described by way of example as being applied to a first terminal, and as shown in fig. 3, the method includes:
step 301, a first direct-playing video stream is obtained, where the first direct-playing video stream is a video stream generated in a process of performing a live microphone connection between a first terminal and a second terminal.
Optionally, a live connection relationship between the first terminal and the second terminal is established, and optionally, the live connection relationship may be a live video live connection, or a live game live connection, a live song hall live connection, or the like.
The method comprises the steps that video live broadcast is a mode that a first terminal collects a first video stream through a camera and sends the first video stream to a second terminal, the second terminal collects a second video stream through the camera and sends the second video stream to the first terminal, and therefore live broadcast is conducted on a video process between the first terminal and the second terminal, wherein when a target terminal conducts on-demand broadcast on a live broadcast video of the first terminal, the first terminal integrates the first video stream and the second video stream, then pushes a first integrated video to the target terminal to be played, and the integration mode of the first video stream and the second video stream is determined by the first terminal; when the target terminal requests the connected-to-microphone video of the second terminal, the second terminal integrates the first video stream and the second video stream and then pushes the second integrated video to the target terminal to be played, and the integration mode of the first video stream and the second video stream is determined by the second terminal;
the method comprises the steps that when a first terminal operates a target game program to fight or cooperate with a second terminal, the first terminal records a first game video stream operated by the target game program and sends the first game video stream to the second terminal, and the second terminal records a second game video stream operated by the target game program and sends the second game video stream to the first terminal, so that a game interaction process between the first terminal and the second terminal is carried out in a live broadcast mode; when the target terminal clicks the microphone-connected video of the second terminal, the second terminal integrates the first game video stream and the second game video stream, then pushes the second integrated video to the target terminal for playing, and the integration mode of the first game video stream and the second game video stream is determined by the second terminal;
the live broadcasting of the song hall and the wheat refers to a live broadcasting mode in which the first terminal performs song interaction with the second terminal in a mode of acquiring images or playing Music Videos (MVs) through a camera, the song interaction can be chorus interaction or alternative singing interaction, and the method is not limited in the embodiment of the disclosure. Optionally, in this embodiment, a first terminal is used to play a music short, a second terminal captures a video through a camera for example, the first terminal sends an MV video stream of the music short to the second terminal according to a playing progress, and the second terminal sends a live video stream captured by the camera to the first terminal, so that a song interaction between the first terminal and the second terminal is live. When a target terminal requests a wheat-connected video of a first terminal, the first terminal integrates an MV video stream and a live video stream, then pushes a first integrated video to the target terminal to play, and the integration mode of the MV video stream and the live video stream is determined by the first terminal; when the target terminal clicks the wheat-connected video of the second terminal, the second terminal integrates the MV video stream and the live video stream, then pushes the second integrated video to the target terminal to play, and the integration mode of the MV video stream and the live video stream is determined by the second terminal.
Step 302, performing target area identification on the first direct-playing video stream to obtain first area information.
Optionally, the target area is an area where the target type of content is located. Optionally, the target type content may be implemented as a face content, a character content, an animation character content, a plant type content, an animal type content, or the like, and in this disclosure, for example, a target area is taken as an area where a face is located, that is, a face is identified for the first live video stream, and first face area information is obtained as the first area information.
Optionally, when performing face recognition on the first live video stream, face recognition may be performed on each frame of video frame in the first live video stream, and first region information corresponding to each frame of video frame is recorded; or, performing face recognition on key frames in the first live video stream, recording first region information corresponding to each key frame, and applying the first region information to the corresponding key frames, which are the same as the corresponding key frames, for non-key frames; or, performing face recognition on the specified frame in the first direct-playing video stream, and recording first region information corresponding to the specified frame of each frame.
Optionally, when the target area of the first live video stream is identified, the first live video stream is input into the preset neural network model, and the target area of the video image frame in the first live video stream is identified through the preset neural network model. Optionally, the preset neural network model is obtained by performing supervised training on a sample image marked with a region division result, that is, obtaining a target region where the content of the target type in the sample image is located, inputting the sample image into the preset neural network model to be trained, outputting the sample image to obtain a recognition result region, and performing reverse adjustment on the preset neural network model according to a difference between the target region and the recognition result region, so as to train the preset neural network model.
Step 303, receiving a second live video stream sent by the second terminal, where the second live video stream corresponds to the second region information.
Optionally, the second area information is used to indicate an area in which the target type content in the second live video stream is located.
Optionally, after acquiring the second live video stream, the second terminal performs target area identification on the second live video stream to obtain the second area information, encodes the second live video stream in combination with the second area information, and sends the encoded stream to the first terminal, where the encoded stream is accompanied by the second area information.
Optionally, the second area information is used to indicate an area where a face is located in the second live video stream, and optionally, for the face recognition process of the second live video stream, reference is made to the face recognition process of the first live video stream in step 302, which is not described herein again.
Optionally, the second region information is used to indicate second face region information in a second live video stream.
It should be noted that the steps 301 to 302 and the step 303 are in parallel, that is, the steps 301 to 302 may be executed first, the step 303 may be executed first, and the steps 301 to 302 and the step 303 may be executed simultaneously.
And step 304, integrating the first live video stream and the second live video stream to obtain an integrated video.
Optionally, the manner of integrating the first live video stream and the second live video stream may be an integration manner default by the system, or an integration manner set by the first terminal. Optionally, the first live video stream and the second live video stream may be integrated in a parallel arrangement manner, may also be integrated in an overlapping manner, and may also be integrated in a partial overlapping manner, which is not limited in this disclosure. Optionally, the display size of the first live video stream and the display size of the second live video stream may be different or the same, which is schematic, please refer to fig. 4, a first live video stream 410 acquired by a camera connected to the first terminal is displayed in a live interface 400 of the first terminal, a second live video stream 420 is displayed on the first live video stream 410 in an overlapping manner, the first live video stream 410 is displayed in the live interface 400 in the maximum size displayed by the terminal, and according to the setting of the first terminal, the second live video stream 420 is displayed in the lower right corner of the live interface 400 in a small size.
Step 305, merging the first region information and the second region information according to the integration mode to obtain the integrated region information.
Optionally, the first region information corresponds to a first live video stream, the second region information corresponds to a second live video stream, and the region position of the first region information and the region position of the second region information are converted according to the integration manner of the first live video stream and the second live video stream to obtain integrated region information including a region expressed by the first region information and a region expressed by the second region information.
Optionally, the integration manner of the first region information and the second region information includes at least one of the following manners:
firstly, when a first live video stream and a second live video stream are arranged in parallel, an intersection does not exist in an integrated video in a region expressed by first region information and a region expressed by second region information, the first region information is converted, and after the second region information is converted, conversion results are combined to obtain integrated region information;
secondly, when the first live broadcast video stream and the second live broadcast video stream are superposed and the region expressed by the first region information and the region expressed by the second region information do not have intersection in the integrated video, converting the first region information, and after converting the second region information, combining the conversion results to obtain integrated region information;
thirdly, when the first live broadcast video stream and the second live broadcast video stream are superposed and the region expressed by the first region information and the region expressed by the second region information have intersection, converting the first region information, converting the second region information, and merging the conversion results to obtain integrated region information.
And step 306, performing partition coding on the integrated video combined with the integrated area information to obtain a live broadcast coding stream.
Optionally, a first region corresponding to the integrated region information in the integrated video is encoded at a first code rate, and a second region in the integrated video is encoded at a second code rate, where the second region is a region other than the first region, and the first code rate is higher than the second code rate.
In summary, in the encoding method for the live video stream provided in this embodiment, the first region information in the first live video stream and the second region information in the second live video stream are determined, and after the first live video stream and the second live video stream are integrated, the first region information and the second region information are integrated, so as to obtain the integrated region information.
In an optional embodiment, the first region corresponding to the integrated region information is encoded at a first code rate, and the other regions are encoded at a second code rate, fig. 5 is a flowchart of a method for encoding a live video stream according to another exemplary embodiment of the present disclosure, which is described by taking the method as an example when the method is applied to a first terminal, and as shown in fig. 5, the method includes:
step 501, a first direct-playing video stream is obtained, wherein the first direct-playing video stream is a video stream generated in the process of performing live broadcast of a first terminal and a second terminal.
Optionally, a live connection relationship between the first terminal and the second terminal is established, and optionally, the live connection relationship may be a live video live connection, or a live game live connection, a live song hall live connection, or the like.
Step 502, performing target area identification on the first direct-playing video stream to obtain first area information.
Optionally, the target area is an area where the target type of content is located. Optionally, the target type content may be implemented as a human face content, a character content, an animation character content, a plant type content, an animal type content, and the like, and in this disclosure, for example, a target area is taken as an area where a human face is located, that is, a human face is identified for the first live video stream, so as to obtain the first area information.
Optionally, when performing face recognition on the first live video stream, face recognition may be performed on each frame of video frame in the first live video stream, and first region information corresponding to each frame of video frame is recorded; or, performing face recognition on key frames in the first live video stream, recording first region information corresponding to each key frame, and applying the first region information to the corresponding key frames, which are the same as the corresponding key frames, for non-key frames; or, performing face recognition on the specified frame in the first direct-playing video stream, and recording first region information corresponding to the specified frame of each frame.
Step 503, receiving a second live video stream sent by the second terminal, where the second live video stream corresponds to the second region information.
Optionally, the second area information is used to indicate an area in which the target type content in the second live video stream is located.
Optionally, after acquiring the second live video stream, the second terminal performs target area identification on the second live video stream to obtain the second area information, encodes the second live video stream in combination with the second area information, and sends the encoded stream to the first terminal, where the encoded stream is accompanied by the second area information.
Optionally, after the second terminal and the first terminal establish a session connection, when the second terminal invokes a session Software Development Kit (SDK) interface to send a video frame of a second live video stream, the second terminal sets the local-end ROI information (i.e., the second region information) to a stream pushing encoder of the session SDK, and then the second terminal encodes the second live video stream by combining the ROI information.
Optionally, when receiving the second live video stream, first receiving a second encoded stream sent by the second terminal, decoding the second encoded stream to obtain the second live video stream, and acquiring second region information corresponding to the second live video stream. Optionally, the second area Information is obtained from Supplemental Enhancement Information (SEI) attached to an image frame in the second live video stream.
And step 504, integrating the first live video stream and the second live video stream to obtain an integrated video.
Optionally, the manner of integrating the first live video stream and the second live video stream may be an integration manner default by the system, or an integration manner set by the first terminal. Optionally, the first live video stream and the second live video stream may be integrated in a parallel arrangement manner, may also be integrated in an overlapping manner, and may also be integrated in a partial overlapping manner, which is not limited in this disclosure.
And 505, merging the first region information and the second region information according to the integration mode to obtain integrated region information.
Optionally, the first region information corresponds to a first live video stream, the second region information corresponds to a second live video stream, and the region position of the first region information and the region position of the second region information are converted according to the integration manner of the first live video stream and the second live video stream to obtain integrated region information including a region expressed by the first region information and a region expressed by the second region information.
Step 506, a first region corresponding to the integrated region information in the integrated video is encoded at a first code rate.
Optionally, the target video frame in the integrated video corresponds to frame region information in the integrated region information, and when the target video frame is encoded, a first region corresponding to the frame region information in the target video frame is encoded at a first code rate, and a second region in the target video frame is encoded at a second code rate.
Step 507, encoding a second region in the integrated video at a second code rate, wherein the second region is a region except the first region.
Optionally, a second region in the target video frame is encoded at a second code rate. Optionally, the first code rate is higher than the second code rate.
In summary, in the encoding method for the live video stream provided in this embodiment, the first region information in the first live video stream and the second region information in the second live video stream are determined, and after the first live video stream and the second live video stream are integrated, the first region information and the second region information are integrated, so as to obtain the integrated region information.
In the method provided by the embodiment, the first region is encoded at the first code rate, and the second region is encoded at the second code rate, so that the important region is encoded at a higher code rate, and the non-important region is encoded at a lower code rate, thereby reducing the bandwidth occupied by the video for integrating the video and improving the encoding efficiency.
In an optional embodiment, the first terminal further needs to encode the first live video stream and then transmit the encoded first live video stream to the second terminal, fig. 6 is a flowchart of a method for encoding a live video stream according to another exemplary embodiment of the present disclosure, which is described by taking as an example that the method is applied to the first terminal, and as shown in fig. 6, the method includes:
step 601, obtaining a first direct playing video stream, where the first direct playing video stream is a video stream generated in a process of performing live wheat broadcast by a first terminal and a second terminal.
Optionally, a live connection relationship between the first terminal and the second terminal is established, and optionally, the live connection relationship may be a live video live connection, or a live game live connection, a live song hall live connection, or the like.
Step 602, performing target area identification on the first direct-playing video stream to obtain first area information.
Optionally, the target area is an area where the target type of content is located. Optionally, the target type content may be implemented as a human face content, a character content, an animation character content, a plant type content, an animal type content, and the like, and in this disclosure, for example, a target area is taken as an area where a human face is located, that is, a human face is identified for the first live video stream, so as to obtain the first area information.
Optionally, when performing face recognition on the first live video stream, face recognition may be performed on each frame of video frame in the first live video stream, and first region information corresponding to each frame of video frame is recorded; or, performing face recognition on key frames in the first live video stream, recording first region information corresponding to each key frame, and applying the first region information to the corresponding key frames, which are the same as the corresponding key frames, for non-key frames; or, performing face recognition on the specified frame in the first direct-playing video stream, and recording first region information corresponding to the specified frame of each frame.
Step 603, the first direct-play video stream is encoded by combining the first region information, so as to obtain a first encoded stream.
Optionally, a third region corresponding to the first region information in the first live video stream is encoded at a third code rate, and a fourth region in the first live video stream is encoded at a fourth code rate, where the fourth region is a region other than the third region, and the third code rate is higher than the fourth code rate.
Alternatively, the third code rate may be identical to the first code rate, and the fourth code rate may be identical to the second code rate.
Step 604, the first encoded stream is sent to the second terminal.
Optionally, the second terminal is configured to generate a microphone connecting video stream by combining the second live video stream and the second region information according to the first region information in the first encoded stream, and push the microphone connecting video stream to a viewing terminal of the second terminal.
In summary, in the encoding method for the live video stream provided in this embodiment, the first region information in the first live video stream and the second region information in the second live video stream are determined, and after the first live video stream and the second live video stream are integrated, the first region information and the second region information are integrated, so as to obtain the integrated region information.
Fig. 7 is a block diagram of an encoding apparatus for a live video stream according to an exemplary embodiment of the present disclosure, which is exemplified or explained by applying the apparatus to a terminal, and includes: an acquisition module 710, an identification module 720, a receiving module 730, a processing module 740, and an encoding module 750;
an obtaining module 710, configured to obtain a first direct-play video stream, where the first direct-play video stream is a video stream generated by a first terminal in a process of performing a live connection with a second terminal;
an identifying module 720, configured to perform target area identification on the first direct-playing video stream to obtain first area information, where the target area is an area where content of a target type is located;
a receiving module 730, configured to receive a second live video stream sent by the second terminal, where the second live video stream corresponds to second area information, and the second area information is used to indicate an area where the target type content in the second live video stream is located;
a processing module 740, configured to integrate the first live video stream and the second live video stream to obtain an integrated video, and combine the first region information and the second region information according to an integration manner to obtain integrated region information;
and the encoding module 750 is configured to perform partition encoding on the integrated video in combination with the integrated region information to obtain a live broadcast encoded stream.
In an optional embodiment, the encoding module 750 is further configured to encode a first region corresponding to the integrated region information in the integrated video at a first code rate; encoding a second region in the integrated video at a second code rate, the second region being a region other than the first region;
wherein the first code rate is higher than the second code rate.
In an optional embodiment, a target video frame in the integrated video corresponds to frame region information in the integrated region information;
the encoding module 750 is further configured to encode the first region corresponding to the frame region information in the target video frame at the first code rate; encoding the second region in the target video frame at the second code rate.
In an optional embodiment, the encoding module 750 is further configured to encode the first direct-playing video stream with the first region information to obtain a first encoded stream;
as shown in fig. 8, the apparatus further includes:
a sending module 760, configured to send the first encoded stream to the second terminal.
In an optional embodiment, the encoding module 750 is further configured to encode a third region corresponding to the first region information in the first direct current video stream at a third code rate; encoding a fourth region in the first direct-playing video stream at a fourth code rate, wherein the fourth region is a region except the third region;
wherein the third code rate is higher than the fourth code rate.
In an optional embodiment, the receiving module 730 is further configured to receive a second encoded stream sent by the second terminal;
the device, still include:
a decoding module 770, configured to decode the second encoded stream to obtain the second live video stream;
the obtaining module 710 is further configured to obtain the second area information corresponding to the second live video stream.
In an optional embodiment, the obtaining module 710 is further configured to obtain the second area information from supplemental enhancement information attached to an image frame in the second live video stream.
In an optional embodiment, the recognition module 720 is further configured to perform face recognition on the first live video stream to obtain first face region information as the first region information, where the second region information is used to indicate second face region information in the second live video stream.
In summary, in the encoding apparatus for a live video stream provided in this embodiment, the first region information in the first live video stream and the second region information in the second live video stream are determined, and after the first live video stream and the second live video stream are integrated, the first region information and the second region information are integrated, so as to obtain the integrated region information.
It should be noted that: the encoding apparatus for live video stream provided in the foregoing embodiment is only illustrated by dividing the functional modules, and in practical application, the functions may be allocated by different functional modules according to needs, that is, the internal structure of the device is divided into different functional modules, so as to complete all or part of the functions described above. In addition, the encoding apparatus for a live video stream and the encoding method embodiment for a live video stream provided in the foregoing embodiments belong to the same concept, and specific implementation processes thereof are described in detail in the method embodiments and are not described herein again.
FIG. 9 is a block diagram illustrating a computer device 900 in accordance with an example embodiment. The computer device 900 may be, for example, the terminal introduced above. For example, the terminal may be an electronic device such as a mobile phone, a tablet Computer, an e-book reader, a multimedia playing device, a Personal Computer (PC), and a wearable device.
Referring to fig. 9, computer device 900 may include one or more of the following components: processing component 902, memory 904, power component 906, multimedia component 908, audio component 910, Input/Output (I/O) interface 912, sensor component 914, and communication component 916.
The processing component 902 generally controls the overall operation of the computer device 900, such as operations associated with display, telephone calls, data communications, camera operations, and recording operations. Processing component 902 may include one or more processors 920 to execute instructions to perform all or a portion of the steps of the methods described above. Further, processing component 902 can include one or more modules that facilitate interaction between processing component 902 and other components. For example, the processing component 902 can include a multimedia module to facilitate interaction between the multimedia component 908 and the processing component 902.
The memory 904 is configured to store various types of data to support operation at the computer device 900. Examples of such data include instructions for any application or method operating on computer device 900, contact data, phonebook data, messages, pictures, videos, and so forth. The Memory 904 may be implemented by any type of volatile or non-volatile Memory device or combination thereof, such as Static Random-Access Memory (SRAM), Electrically Erasable Programmable Read Only Memory (EEPROM), Erasable Programmable Read Only Memory (EPROM), Programmable Read Only Memory (PROM), Read Only Memory (ROM), magnetic Memory, flash Memory, magnetic or optical disk.
The power component 906 provides power to the various components of the computer device 900. The power components 906 may include a power management system, one or more power supplies, and other components associated with generating, managing, and distributing power for the computer device 900.
The multimedia components 908 include a screen that provides an output interface between the computer device 900 and a user. In some embodiments, the screen may include an Organic Light-Emitting Diode (OLED) display screen and a Touch Panel (TP). If the screen includes a touch panel, the screen may be implemented as a touch screen to receive an input signal from a user. The touch panel includes one or more touch sensors to sense touch, slide, and gestures on the touch panel. The touch sensor may not only sense the boundary of a touch or slide action, but also detect the duration and pressure associated with the touch or slide operation. In some embodiments, the multimedia component 908 includes a front facing camera and/or a rear facing camera. The front-facing camera and/or the rear-facing camera may receive external multimedia data when the computer device 900 is in an operating mode, such as a shooting mode or a video mode. Each front camera and rear camera may be a fixed optical lens system or have a focal length and optical zoom capability.
The audio component 910 is configured to output and/or input audio signals. For example, audio component 910 includes a Microphone (MIC) configured to receive external audio signals when computer device 900 is in an operational mode, such as a call mode, a recording mode, and a voice recognition mode. The received audio signals may further be stored in the memory 904 or transmitted via the communication component 916. In some embodiments, audio component 910 also includes a speaker for outputting audio signals.
I/O interface 912 provides an interface between processing component 902 and peripheral interface modules, which may be keyboards, click wheels, buttons, etc. These buttons may include, but are not limited to: a home button, a volume button, a start button, and a lock button.
The sensor component 914 includes one or more sensors for providing various aspects of state assessment for the computer device 900. For example, the sensor component 914 may detect an open/closed state of the computer device 900, the relative positioning of components, such as a display and keypad of the computer device 900, the sensor component 914 may also detect a change in the position of the computer device 900 or a component of the computer device 900, the presence or absence of user contact with the computer device 900, orientation or acceleration/deceleration of the computer device 900, and a change in the temperature of the computer device 900. The sensor assembly 914 may include a proximity sensor configured to detect the presence of a nearby object in the absence of any physical contact. The sensor assembly 914 may also include a photosensor, such as a Complementary Metal Oxide Semiconductor (CMOS) or Charge-coupled Device (CCD) image sensor, for use in imaging applications. In some embodiments, the sensor assembly 914 may also include an acceleration sensor, a gyroscope sensor, a magnetic sensor, a pressure sensor, or a temperature sensor.
The communication component 916 is configured to facilitate communications between the computer device 900 and other devices in a wired or wireless manner. Computer device 900 may access a wireless network based on a communication standard, such as Wi-Fi, 2G, or 3G, or a combination thereof. In an exemplary embodiment, the communication component 916 receives a broadcast signal or broadcast associated information from an external broadcast management system via a broadcast channel. In an exemplary embodiment, the communication component 916 further includes a Near Field Communication (NFC) module to facilitate short-range communications. For example, the NFC module may be implemented based on Radio Frequency Identification (RFID) technology, Infrared Data Association (IrDA) technology, Ultra Wide Band (UWB) technology, BlueTooth (BlueTooth, BT) technology, and other technologies.
In an exemplary embodiment, the computer Device 900 may be implemented by one or more Application Specific Integrated Circuits (ASICs), Digital Signal Processors (DSPs), Digital Signal Processing Devices (DSPDs), Programmable Logic Devices (PLDs), Field Programmable Gate Arrays (FPGAs), controllers, microcontrollers, microprocessors, or other electronic components for performing the above-described encoding method of a live video stream.
In an exemplary embodiment, there is also provided a non-transitory computer-readable storage medium having stored thereon a computer program which, when executed by a processor of the computer device 900, enables the computer device 900 to implement the above-described encoding method of a live video stream. For example, the non-transitory computer-readable storage medium may be a ROM, a Random-Access Memory (RAM), a CD-ROM, a magnetic tape, a floppy disk, an optical data storage device, and the like.
The embodiment of the present disclosure further provides a computer device, where the computer device includes a memory and a processor, where the memory stores at least one instruction, at least one program, a code set, or an instruction set, and the at least one instruction, the at least one program, the code set, or the instruction set is loaded by the processor and implements the encoding method of the live video stream.
The disclosed embodiment also provides a computer readable storage medium, in which at least one instruction, at least one program, a code set, or an instruction set is stored, and the at least one instruction, the at least one program, the code set, or the instruction set is loaded and executed by the processor to implement the above encoding method for a live video stream.
The present disclosure also provides a computer program product, which when running on a computer, causes the computer to execute the encoding method of the live video stream provided by the above method embodiments.
It should be understood that reference to "a plurality" herein means two or more. "and/or" describes the association relationship of the associated objects, meaning that there may be three relationships, e.g., a and/or B, which may mean: a exists alone, A and B exist simultaneously, and B exists alone. The character "/" generally indicates that the former and latter associated objects are in an "or" relationship.
Other embodiments of the disclosure will be apparent to those skilled in the art from consideration of the specification and practice of the disclosure disclosed herein. This disclosure is intended to cover any variations, uses, or adaptations of the disclosure following, in general, the principles of the disclosure and including such departures from the present disclosure as come within known or customary practice within the art to which the disclosure pertains. It is intended that the specification and examples be considered as exemplary only, with a true scope and spirit of the disclosure being indicated by the following claims.
It will be understood that the present disclosure is not limited to the precise arrangements described above and shown in the drawings and that various modifications and changes may be made without departing from the scope thereof. The scope of the present disclosure is limited only by the appended claims.

Claims (10)

1. A method of encoding a live video stream, the method comprising:
acquiring a first direct-playing video stream, wherein the first direct-playing video stream is a video stream generated by a first terminal in a process of carrying out live microphone connection with a second terminal;
identifying a target area of the first direct-playing video stream to obtain first area information, wherein the target area is an area where the content of a target type is located;
receiving a second live video stream sent by the second terminal, wherein the second live video stream corresponds to second area information, and the second area information is used for indicating an area where the target type content in the second live video stream is located;
integrating the first live video stream and the second live video stream to obtain an integrated video, and combining the first region information and the second region information according to an integration mode to obtain integrated region information;
and carrying out partition coding on the integrated video in combination with the integrated area information to obtain a live broadcast coding stream.
2. The method of claim 1, wherein the performing the partition coding on the integrated video and the integrated region information to obtain a live coded stream comprises:
coding a first region corresponding to the integrated region information in the integrated video at a first code rate;
encoding a second region in the integrated video at a second code rate, the second region being a region other than the first region;
wherein the first code rate is higher than the second code rate.
3. The method of claim 2, wherein the target video frame in the integrated video corresponds to the frame region information in the integrated region information;
the encoding a first region corresponding to the integrated region information in the integrated video at a first code rate includes:
encoding the first region corresponding to the frame region information in the target video frame at the first code rate;
the encoding a second region in the integrated video at a second code rate, comprising:
encoding the second region in the target video frame at the second code rate.
4. The method according to any one of claims 1 to 3, wherein after the performing the target area identification on the first live video stream to obtain the first area information, further comprises:
coding the first direct-playing video stream by combining the first area information to obtain a first coded stream;
and sending the first coding stream to the second terminal.
5. The method of claim 4, wherein said encoding the first direct video stream in combination with the first region information to obtain a first encoded stream comprises:
encoding a third region corresponding to the first region information in the first live video stream at a third code rate;
encoding a fourth region in the first direct-playing video stream at a fourth code rate, wherein the fourth region is a region except the third region;
wherein the third code rate is higher than the fourth code rate.
6. The method according to any one of claims 1 to 3, wherein the receiving a second live video stream transmitted by the second terminal further comprises:
receiving a second coding stream sent by the second terminal;
decoding the second encoded stream to obtain the second live video stream;
and acquiring the second area information corresponding to the second live video stream.
7. The method of claim 6, wherein the obtaining the second region information corresponding to the second live video stream comprises:
and acquiring the second area information from the supplementary enhancement information attached to the image frame in the second live video stream.
8. The method according to any one of claims 1 to 3, wherein the performing target area identification on the first live video stream to obtain first area information comprises:
and performing face recognition on the first live video stream to obtain first face region information as the first region information, wherein the second region information is used for indicating second face region information in the second live video stream.
9. An apparatus for encoding a live video stream, the apparatus comprising:
the system comprises an acquisition module, a processing module and a processing module, wherein the acquisition module is used for acquiring a first direct-playing video stream, and the first direct-playing video stream is a video stream generated by a first terminal in the process of carrying out live wheat connection with a second terminal;
the identification module is used for identifying a target area of the first direct-playing video stream to obtain first area information, wherein the target area is an area where the content of a target type is located;
a receiving module, configured to receive a second live video stream sent by the second terminal, where the second live video stream corresponds to second area information, and the second area information is used to indicate an area in which the target type content in the second live video stream is located;
the processing module is used for integrating the first live video stream and the second live video stream to obtain an integrated video and combining the first area information and the second area information according to an integration mode to obtain integrated area information;
and the coding module is used for carrying out partition coding on the integrated video combined with the integrated area information to obtain a live broadcast coding stream.
10. The apparatus of claim 9, wherein the encoding module is further configured to encode a first region of the integrated video corresponding to the integrated region information at a first code rate; encoding a second region in the integrated video at a second code rate, the second region being a region other than the first region;
wherein the first code rate is higher than the second code rate.
CN201911422688.1A 2019-12-31 2019-12-31 Encoding method, device and equipment of live video stream and storage medium Active CN111050190B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201911422688.1A CN111050190B (en) 2019-12-31 2019-12-31 Encoding method, device and equipment of live video stream and storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201911422688.1A CN111050190B (en) 2019-12-31 2019-12-31 Encoding method, device and equipment of live video stream and storage medium

Publications (2)

Publication Number Publication Date
CN111050190A true CN111050190A (en) 2020-04-21
CN111050190B CN111050190B (en) 2022-02-18

Family

ID=70243311

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201911422688.1A Active CN111050190B (en) 2019-12-31 2019-12-31 Encoding method, device and equipment of live video stream and storage medium

Country Status (1)

Country Link
CN (1) CN111050190B (en)

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106131583A (en) * 2016-06-30 2016-11-16 北京小米移动软件有限公司 A kind of live processing method, device, terminal unit and system
CN110049324A (en) * 2019-04-12 2019-07-23 深圳壹账通智能科技有限公司 Method for video coding, system, equipment and computer readable storage medium
CN110213587A (en) * 2019-07-08 2019-09-06 北京达佳互联信息技术有限公司 Method for video coding, device, electronic equipment and storage medium

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106131583A (en) * 2016-06-30 2016-11-16 北京小米移动软件有限公司 A kind of live processing method, device, terminal unit and system
CN110049324A (en) * 2019-04-12 2019-07-23 深圳壹账通智能科技有限公司 Method for video coding, system, equipment and computer readable storage medium
CN110213587A (en) * 2019-07-08 2019-09-06 北京达佳互联信息技术有限公司 Method for video coding, device, electronic equipment and storage medium

Also Published As

Publication number Publication date
CN111050190B (en) 2022-02-18

Similar Documents

Publication Publication Date Title
US20170304735A1 (en) Method and Apparatus for Performing Live Broadcast on Game
CN106911961B (en) Multimedia data playing method and device
CN102215373B (en) In conference display adjustments
CN105100829B (en) Video content intercept method and device
US20090096927A1 (en) System and method for video coding using variable compression and object motion tracking
US20170311004A1 (en) Video processing method and device
EP4007289A1 (en) Video uploading method and apparatus, electronic device, and storage medium
CN111583952B (en) Audio processing method, device, electronic equipment and storage medium
EP3264774B1 (en) Live broadcasting method and device for live broadcasting
CN109862380B (en) Video data processing method, device and server, electronic equipment and storage medium
CN103141085A (en) Information processing device and information processing method
CN114710640A (en) Video call method, device and terminal based on virtual image
CN110677734B (en) Video synthesis method and device, electronic equipment and storage medium
CN109922252B (en) Short video generation method and device and electronic equipment
KR102063463B1 (en) Multimedia information reproduction method and system, standardization server, live broadcasting terminal
CN106559631A (en) Method for processing video frequency and device
CN105120301A (en) Video processing method and apparatus, and intelligent equipment
CN112269554B (en) Display system and display method
CN111050190B (en) Encoding method, device and equipment of live video stream and storage medium
CN112312147A (en) Live broadcast control method and device and storage medium
US10616724B2 (en) Method, device, and non-transitory computer-readable recording medium for supporting relay broadcasting using mobile device
US20150371365A1 (en) Method and technical equipment for image capturing and viewing
CN113422997B (en) Method and device for playing audio data and readable storage medium
CN117813652A (en) Audio signal encoding method, device, electronic equipment and storage medium
CN115065835A (en) Live-broadcast expression display processing method, server, electronic equipment and storage medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant