CN115334310A - Video data encoding method, video data encoding device, electronic apparatus, and medium - Google Patents

Video data encoding method, video data encoding device, electronic apparatus, and medium Download PDF

Info

Publication number
CN115334310A
CN115334310A CN202110512362.9A CN202110512362A CN115334310A CN 115334310 A CN115334310 A CN 115334310A CN 202110512362 A CN202110512362 A CN 202110512362A CN 115334310 A CN115334310 A CN 115334310A
Authority
CN
China
Prior art keywords
image
target
image areas
image area
queue
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202110512362.9A
Other languages
Chinese (zh)
Inventor
宋嘉文
豆修鑫
樊鸿飞
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing Kingsoft Cloud Network Technology Co Ltd
Original Assignee
Beijing Kingsoft Cloud Network Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing Kingsoft Cloud Network Technology Co Ltd filed Critical Beijing Kingsoft Cloud Network Technology Co Ltd
Priority to CN202110512362.9A priority Critical patent/CN115334310A/en
Publication of CN115334310A publication Critical patent/CN115334310A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/134Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
    • H04N19/167Position within a video image, e.g. region of interest [ROI]
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/102Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
    • H04N19/103Selection of coding mode or of prediction mode
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/102Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
    • H04N19/124Quantisation
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/134Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
    • H04N19/136Incoming video signal characteristics or properties
    • H04N19/14Coding unit complexity, e.g. amount of activity or edge presence estimation
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/134Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
    • H04N19/156Availability of hardware or computational resources, e.g. encoding based on power-saving criteria

Abstract

The embodiment of the disclosure discloses a video data encoding method and device, electronic equipment and a computer readable medium. One embodiment of the method comprises: based on the current bandwidth condition, selecting a target number of image areas from an image area queue as target image areas, wherein the image areas in the image area queue are arranged according to the quantization parameter score corresponding to each image area, and the quantization parameter score is determined according to the spatial coding complexity and the temporal coding complexity corresponding to each image area; the target image region is encoded in a first encoding mode and the remaining image regions in the image region queue are encoded in a second encoding mode. The implementation mode realizes the purposes of reducing the decoding time delay to the maximum extent and optimizing the playing effect under the condition of not influencing data transmission.

Description

Video data encoding method, video data encoding device, electronic apparatus, and medium
Technical Field
Embodiments of the present disclosure relate to the field of computer technologies, and in particular, to a video data encoding method and apparatus, an electronic device, and a computer-readable medium.
Background
Related VR (virtual reality) video data transmission generally adopts a data transmission mode based on a visual angle. Specifically, encoded data of high definition is provided for a main view, and encoded data of low definition is provided for other views. This has the advantage that the decoding pressure and bandwidth requirements at the playing end can be reduced due to the relatively low complexity and code rate of the low definition data coding.
However, when the above transmission method is adopted, the following technical problems are often caused:
when the viewing angle is switched, the playing end needs to pull the high-definition encoded data again and decode the data again. Related video playback techniques include coding with reference to previous frames and key frames or coding with reference to key frames only. For the mode of referring to the previous frame and the key frame, decoding can be performed only after the previous frame is completely decoded during decoding, so that a larger decoding delay is generated. For the encoding mode only referring to the key frame, although the decoding delay is short, the code rate is greatly increased, the transmission cost is high, and the requirement on the network bandwidth is greatly increased.
Disclosure of Invention
This summary is provided to introduce a selection of concepts in a simplified form that are further described below in the detailed description. This summary is not intended to identify key features or essential features of the claimed subject matter, nor is it intended to be used to limit the scope of the claimed subject matter. Some embodiments of the present disclosure propose a video data encoding method, a data reading and writing method, an apparatus, an electronic device, and a computer readable medium to solve one or more of the technical problems mentioned in the background section above.
In a first aspect, some embodiments of the present disclosure provide a video data encoding method, including: based on the current bandwidth condition, selecting a target number of image areas from an image area queue as target image areas, wherein the image areas in the image area queue are arranged according to the quantization parameter score corresponding to each image area, and the quantization parameter score is determined according to the space coding complexity and the time coding complexity corresponding to each image area; the target image region is encoded in a first encoding mode and the remaining image regions in the image region queue are encoded in a second encoding mode.
In a second aspect, some embodiments of the present disclosure provide a video data encoding apparatus, including: the selecting unit is configured to select a target number of image areas from an image area queue as target image areas based on the current bandwidth condition, wherein the image areas in the image area queue are arranged according to the quantization parameter score corresponding to each image area, and the quantization parameter score is determined according to the spatial coding complexity and the temporal coding complexity corresponding to each image area; and the encoding unit is configured to encode the target image area in a first encoding mode and encode the rest image areas in the image area queue in a second encoding mode.
In a third aspect, some embodiments of the present disclosure provide an electronic device, comprising: one or more processors; a storage device having one or more programs stored thereon, which when executed by one or more processors, cause the one or more processors to implement the method described in any of the implementations of the first or second aspects.
In a fourth aspect, some embodiments of the disclosure provide a computer readable medium on which a computer program is stored, wherein the program when executed by a processor implements the method described in any implementation of the first or second aspect.
The above embodiments of the present disclosure have the following advantages: according to the network bandwidth condition, the coding mode is dynamically selected, the decoding time delay is reduced to the maximum extent under the condition of not influencing data transmission, and the playing effect is optimized.
Drawings
The above and other features, advantages and aspects of various embodiments of the present disclosure will become more apparent by referring to the following detailed description when taken in conjunction with the accompanying drawings. Throughout the drawings, the same or similar reference numbers refer to the same or similar elements. It should be understood that the drawings are schematic and that elements and elements are not necessarily drawn to scale.
Fig. 1 is a schematic diagram of coding with LDP reference mode in related video playing technology;
fig. 2 is a flow diagram of some embodiments of a video data encoding method according to the present disclosure;
fig. 3 is an exemplary diagram of a plurality of image regions corresponding to a video in a video data encoding method according to the present disclosure;
fig. 4 is a flow diagram of further embodiments of a video data encoding method according to the present disclosure;
fig. 5 is a schematic structural diagram of some embodiments of a video data encoding apparatus according to the present disclosure;
FIG. 6 is a schematic block diagram of an electronic device suitable for use in implementing some embodiments of the present disclosure.
Detailed Description
Embodiments of the present disclosure will be described in more detail below with reference to the accompanying drawings. While certain embodiments of the disclosure are shown in the drawings, it is to be understood that the disclosure may be embodied in various forms and should not be construed as limited to the embodiments set forth herein. Rather, these embodiments are provided so that this disclosure will be thorough and complete. It should be understood that the drawings and embodiments of the disclosure are for illustration purposes only and are not intended to limit the scope of the disclosure.
It should be noted that, for convenience of description, only the portions related to the related invention are shown in the drawings. The embodiments and features of the embodiments in the present disclosure may be combined with each other without conflict.
It should be noted that the terms "first", "second", and the like in the present disclosure are only used for distinguishing different devices, modules or units, and are not used for limiting the order or interdependence relationship of the functions performed by the devices, modules or units.
It is noted that references to "a", "an", and "the" modifications in this disclosure are intended to be illustrative rather than limiting, and that those skilled in the art will recognize that "one or more" may be used unless the context clearly dictates otherwise.
The names of messages or information exchanged between devices in the embodiments of the present disclosure are for illustrative purposes only, and are not intended to limit the scope of the messages or information.
The present disclosure will be described in detail below with reference to the accompanying drawings in conjunction with embodiments.
To better illustrate the present disclosure, the encoding method of the related video playing technology will be described first. Fig. 1 is a diagram illustrating a related video playing technology using LDP reference mode for encoding.
As shown in fig. 1, 5 frames of pictures with numbers (POC) 1-5 are included in total. POC1 is an I-frame, i.e., a key frame. POC2-5 are P frames, i.e., predicted frames. The arrows therein represent the encoded reference frames for each frame. As shown, POC2 refers to I-frames. POC3 refers to I-frames and POC2.POC4 refers to I-frames and POC3.POC5 refers to I-frames and PO4. That is, in LDP reference mode, each frame refers to I frame and previous frame except POC1 and POC2. Therefore, when the playing side decodes, it is necessary to wait for the previous frame to be decoded after all the previous frames are decoded. Taking POC5 as an example, the decoding delay reaches 5 frames, which easily causes the video buffering time to be too long.
With continued reference to fig. 2, a flow 200 of some embodiments of a video data encoding method according to the present disclosure is shown. The video data encoding method includes the steps of:
step 201, based on the current bandwidth situation, selecting a target number of image areas from the image area queue as target image areas.
In some embodiments, the subject of execution of the video data encoding method may select a target data amount of image regions from the image queue as target image regions based on the current bandwidth conditions. The image areas in the image queue are arranged according to the quantization parameter score corresponding to each image area. In practice, in the process of encoding a video, a video picture is often divided as needed, so as to obtain a plurality of image areas (tiles) corresponding to the video. The content in each frame image belonging to the same image area has certain relevance in time, which is beneficial to the realization of coding.
Fig. 3 shows an exemplary schematic diagram of a plurality of image regions corresponding to a video. Taking the example where the video includes the frame image 301 and the frame image 302, the image 301 may be divided into three rectangular regions 3011, 3012, and 3013, respectively, from the vertical direction. It is understood that the division into rectangular regions from the horizontal and/or vertical direction can be flexibly selected according to the needs. Similarly, the frame image 302 may be divided into rectangular regions 3021, 3022, and 3023. In practice, the same division method is generally adopted for division within a certain time range. On the basis, rectangular areas at the same position in each frame image belong to the same image area. For example, the rectangular region 3011 in the frame image 301 and the rectangular region 3021 in the frame image 302 belong to the same image region. The video corresponds to three image areas. In addition, the content (rectangular region) in each frame image corresponding to each image region is encoded, and encoded data corresponding to the image region can be obtained.
Within a certain time range, image areas corresponding to a plurality of video frames within the time range can be sorted according to the quantization parameter scores of the image areas, so that an image area queue is obtained. In practice, the time range may be flexibly determined, for example, the time interval corresponding to one or more video frame groups may be determined as the time range. It should be noted that, in practice, when the nth video frame group is to be encoded, the quantization parameter score of the nth video frame group cannot be calculated because the nth video frame group is not encoded temporarily. However, it was found that the N-1 st video frame group generally has a small variation because the time from the nth video frame group is short. Accordingly, a target image region may be selected based on the N-1 th video frame group, and then the nth video frame group may be encoded. Further, although one video frame Group (GOP) is transmitted at a time here as an example, it is understood that the present disclosure is equally applicable when transmission is performed in a packetized manner (a plurality of video frame groups are included in each packet).
In some embodiments, the quantization parameter score for each image region may be determined according to the spatial coding complexity and the temporal coding complexity corresponding to the image region. In practice, as shown in fig. 1, each frame image corresponding to an image area may be divided into a key frame (I frame), a prediction frame (P frame), and the like according to its function. On this basis, for a certain image region, the average quantization parameter (qp) of the key frame in each frame image corresponding to the image region can be determined first I ) And average quantization parameter (qp) of predicted frame P ). The average quantization parameter of the key frame characterizes temporal complexity, and the average quantization parameter of the predicted frame characterizes spatial complexity. In practice, when qp I Smaller and qp P The larger the bit rate is, the representing time complexity is lower, the space complexity is higher, and the code rate for reducing decoding delay cannot be greatly increased by using an encoding mode of only referring to key frames by all references. Therefore, the quantization parameter score (qpscore) of the image region may be calculated by the following formula: qpscore = qp P -qp I
On the basis, as an example, the image areas may be sorted in descending order according to the quantization parameter scores of the image areas, so as to obtain an image area queue. Then, a target number of image areas are selected from the head of the image area queue as target image areas. Since the queues are sorted down according to the quantization parameter score. Therefore, the selected target image areas have higher quantization parameter scores. That is, the target image regions all represent a lower temporal complexity and a higher spatial complexity. Therefore, by performing the coding method of referring only to the key frame for these target image regions, the code rate for reducing the decoding delay is not significantly increased. The determination of the target quantity can be obtained by specification or screening under certain conditions. For example, when the bandwidth is sufficient, a first value may be specified; and when the bandwidth is insufficient, the second data may be specified.
Step 202, encoding the target image area in a first encoding mode, and encoding the rest image areas in the image area queue in a second encoding mode.
In some embodiments, the execution body may encode the target image region in a first encoding manner, and encode the remaining image regions in the image region queue in a second encoding manner. For example, the first encoding method is an encoding method with a low decoding delay, for example, an encoding method that only refers to a key frame. The second encoding mode may be an encoding mode with a higher decoding delay but a lower code rate, such as an encoding mode with reference to a previous frame and a key frame.
In some embodiments, by using the video data encoding method provided by the present disclosure, an encoding mode is dynamically selected according to a network bandwidth condition, so that decoding delay is reduced to the maximum extent and a playing effect is optimized without affecting data transmission.
With further reference to fig. 4, a flow 400 of further embodiments of a video data encoding method is shown. The process 400 of the video data encoding method comprises the following steps:
step 401, in response to that the current bandwidth index meets a preset condition, incrementally increasing the historical number to obtain a target number, and selecting a target number of image areas from the image area queue as target image areas, where the historical number is the number of image areas coded in the first coding mode last time.
In some embodiments, the current bandwidth indicator may be a bandwidth delay duration. At this time, as an example, the preset condition may be less than the duration threshold.
In some optional implementations of some embodiments, the current bandwidth indicator is calculated by: acquiring current network delay and historical network delay; determining the change rate of the current network delay relative to the historical network delay; the rate of change is determined as the current bandwidth indicator. In the implementation modes, the change rate can accurately measure the change condition of the network so as to correspondingly adjust the coding mode, thereby better realizing adaptive coding. At this time, the preset condition may be that the preset condition is less than the change rate threshold. In practice, the current network delay may be a transmission delay of transmitting video data, and may be obtained by a difference between a transmission time and a reception time. The current network delay may be the most recently counted network delay. The historical network latency may be the next newly counted network latency. Optionally, the current bandwidth indicator may also be a difference between the current network delay and the historical network delay.
In some embodiments, in response to the current bandwidth indicator satisfying a predetermined condition indicating that the current bandwidth margin is sufficient, the executing entity of the video data encoding method may increment (e.g., add one) the historical amount to obtain the target amount. The history number is the number of image areas coded in the first coding mode at the last time. Therefore, dynamic adjustment can be realized on the basis of the historical quantity, and the current network bandwidth condition can be better adapted.
And 402, responding to the condition that the current bandwidth index does not meet the preset condition, progressively reducing the historical quantity to obtain a target quantity, and selecting image areas with the target quantity from the image area queue as target image areas.
In some embodiments, in response to the current bandwidth indicator not meeting the preset condition, indicating that the current bandwidth margin is insufficient, the executing subject of the video data encoding method may decrement (e.g., by one) the historical amount to obtain the target amount. Therefore, dynamic adjustment can be realized on the basis of the historical quantity, and the current network bandwidth condition can be better adapted.
Step 403, encoding the target image area in a first encoding mode, and encoding the rest image areas in the image area queue in a second encoding mode.
In some embodiments, the specific implementation of step 403 and the technical effect thereof may refer to step 202 in those embodiments corresponding to fig. 2, and are not described herein again.
In the embodiments corresponding to fig. 4, the target number of this time is obtained by dynamically adjusting the number of image regions that were encoded in the first encoding method last time. Therefore, the method can better adapt to the network condition and realize the self-adaptive adjustment of the target quantity.
With further reference to fig. 5, as an implementation of the methods shown in the above figures, the present disclosure provides some embodiments of a video data encoding apparatus, which correspond to those shown in fig. 2, and which may be applied in various electronic devices in particular.
As shown in fig. 5, the video data encoding apparatus 500 of some embodiments includes: a selecting unit 501 and an encoding unit 502. The selecting unit 501 is configured to select a target number of image regions from an image region queue as target image regions based on a current bandwidth condition, where the image regions in the image region queue are arranged according to a quantization parameter score corresponding to each image region, and the quantization parameter score is determined according to a spatial coding complexity and a temporal coding complexity corresponding to each image region. The encoding unit 502 is configured to encode the target image region in a first encoding manner and to encode the remaining image regions in the image region queue in a second encoding manner.
In an optional implementation of some embodiments, the selecting unit 501 is further configured to: and responding to the condition that the current bandwidth index meets the preset condition, increasing the historical number to obtain a target number, and selecting image areas with the target number from the image area queue as target image areas, wherein the historical number is the number of the image areas which are coded in the first coding mode at the last time.
In an optional implementation of some embodiments, the selecting unit 501 is further configured to: and in response to the fact that the current bandwidth index does not meet the preset condition, the historical number is decreased to obtain the target number, and the target number of image areas are selected from the image area queue to serve as the target image areas.
In an alternative implementation of some embodiments, the current bandwidth indicator is calculated by: acquiring current network delay and historical network delay; determining the change rate of the current network delay relative to the historical network delay; the rate of change is determined as the current bandwidth indicator.
It will be understood that the elements described in the apparatus 500 correspond to various steps in the method described with reference to fig. 2. Thus, the operations, features and advantages described above with respect to the method are also applicable to the apparatus 500 and the units included therein, and are not described herein again.
Referring now to fig. 6, a block diagram of an electronic device (e.g., server in fig. 1) 600 suitable for use in implementing some embodiments of the present disclosure is shown. The electronic device shown in fig. 6 is only an example, and should not bring any limitation to the functions and the scope of use of the embodiments of the present disclosure.
As shown in fig. 6, the electronic device 600 may include a processing means (e.g., central processing unit, graphics processor, etc.) 601 that may perform various appropriate actions and processes in accordance with a program stored in a Read Only Memory (ROM) 602 or a program loaded from a storage means 608 into a Random Access Memory (RAM) 603. In the RAM603, various programs and data necessary for the operation of the electronic apparatus 600 are also stored. The processing device 601, the ROM 602, and the RAM603 are connected to each other via a bus 604. An input/output (I/O) interface 605 is also connected to bus 604.
Generally, the following devices may be connected to the I/O interface 605: input devices 606 including, for example, a touch screen, touch pad, keyboard, mouse, camera, microphone, accelerometer, gyroscope, etc.; output devices 607 including, for example, a Liquid Crystal Display (LCD), a speaker, a vibrator, and the like; storage 608 including, for example, tape, hard disk, etc.; and a communication device 609. The communication means 609 may allow the electronic device 600 to communicate with other devices wirelessly or by wire to exchange data. While fig. 6 illustrates an electronic device 600 having various means, it is to be understood that not all illustrated means are required to be implemented or provided. More or fewer devices may alternatively be implemented or provided. Each block shown in fig. 6 may represent one device or may represent multiple devices as desired.
In particular, according to some embodiments of the present disclosure, the processes described above with reference to the flow diagrams may be implemented as computer software programs. For example, some embodiments of the present disclosure include a computer program product comprising a computer program embodied on a computer readable medium, the computer program comprising program code for performing the method illustrated in the flow chart. In some such embodiments, the computer program may be downloaded and installed from a network through the communication device 609, or installed from the storage device 608, or installed from the ROM 602. The computer program, when executed by the processing device 601, performs the above-described functions defined in the methods of some embodiments of the present disclosure.
It should be noted that the computer readable medium described in some embodiments of the present disclosure may be a computer readable signal medium or a computer readable storage medium or any combination of the two. A computer readable storage medium may be, for example, but not limited to, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, or device, or any combination of the foregoing. More specific examples of the computer readable storage medium may include, but are not limited to: an electrical connection having one or more wires, a portable computer diskette, a hard disk, a Random Access Memory (RAM), a read-only memory (ROM), an erasable programmable read-only memory (EPROM or flash memory), an optical fiber, a portable compact disc read-only memory (CD-ROM), an optical storage device, a magnetic storage device, or any suitable combination of the foregoing. In some embodiments of the disclosure, a computer readable storage medium may be any tangible medium that can contain, or store a program for use by or in connection with an instruction execution system, apparatus, or device. In some embodiments of the present disclosure, however, a computer readable signal medium may include a propagated data signal with computer readable program code embodied therein, for example, in baseband or as part of a carrier wave. Such a propagated data signal may take many forms, including, but not limited to, electro-magnetic, optical, or any suitable combination thereof. A computer readable signal medium may be any computer readable medium that is not a computer readable storage medium and that can communicate, propagate, or transport a program for use by or in connection with an instruction execution system, apparatus, or device. Program code embodied on a computer readable medium may be transmitted using any appropriate medium, including but not limited to: electrical wires, optical cables, RF (radio frequency), etc., or any suitable combination of the foregoing.
In some embodiments, the clients, servers may communicate using any currently known or future developed network Protocol, such as HTTP (HyperText Transfer Protocol), and may be interconnected with any form or medium of digital data communication (e.g., a communication network). Examples of communication networks include a local area network ("LAN"), a wide area network ("WAN"), the Internet (e.g., the Internet), and peer-to-peer networks (e.g., ad hoc peer-to-peer networks), as well as any currently known or future developed network.
The computer readable medium may be embodied in the electronic device; or may exist separately without being assembled into the electronic device. The computer readable medium carries one or more programs which, when executed by the electronic device, cause the electronic device to: based on the current bandwidth condition, selecting a target number of image areas from an image area queue as target image areas, wherein the image areas in the image area queue are arranged according to the quantization parameter score corresponding to each image area, and the quantization parameter score is determined according to the spatial coding complexity and the temporal coding complexity corresponding to each image area; the target image region is encoded in a first encoding mode and the remaining image regions in the image region queue are encoded in a second encoding mode.
Computer program code for carrying out operations for embodiments of the present disclosure may be written in any combination of one or more programming languages, including an object oriented programming language such as Java, smalltalk, C + +, and conventional procedural programming languages, such as the "C" programming language or similar programming languages. The program code may execute entirely on the user's computer, partly on the user's computer, as a stand-alone software package, partly on the user's computer and partly on a remote computer or entirely on the remote computer or server. In the case of a remote computer, the remote computer may be connected to the user's computer through any type of network, including a Local Area Network (LAN) or a Wide Area Network (WAN), or the connection may be made to an external computer (for example, through the Internet using an Internet service provider).
The flowchart and block diagrams in the figures illustrate the architecture, functionality, and operation of possible implementations of systems, methods and computer program products according to various embodiments of the present disclosure. In this regard, each block in the flowchart or block diagrams may represent a module, segment, or portion of code, which comprises one or more executable instructions for implementing the specified logical function(s). It should also be noted that, in some alternative implementations, the functions noted in the block may occur out of the order noted in the figures. For example, two blocks shown in succession may, in fact, be executed substantially concurrently, or the blocks may sometimes be executed in the reverse order, depending upon the functionality involved. It will also be noted that each block of the block diagrams and/or flowchart illustration, and combinations of blocks in the block diagrams and/or flowchart illustration, can be implemented by special purpose hardware-based systems which perform the specified functions or acts, or combinations of special purpose hardware and computer instructions.
The units described in some embodiments of the present disclosure may be implemented by software, and may also be implemented by hardware. The described units may also be provided in a processor, which may be described as: a processor includes a selecting unit and an encoding unit. Where the names of these units do not constitute a limitation on the unit itself in some cases, for example, the selecting unit may also be described as a "unit that selects a target number of image areas from the image area queue as target image areas".
The functions described herein above may be performed, at least in part, by one or more hardware logic components. For example, without limitation, exemplary types of hardware logic components that may be used include: field Programmable Gate Arrays (FPGAs), application Specific Integrated Circuits (ASICs), application Specific Standard Products (ASSPs), system on a chip (SOCs), complex Programmable Logic Devices (CPLDs), and the like.
The foregoing description is only exemplary of the preferred embodiments of the disclosure and is illustrative of the principles of the technology employed. It will be appreciated by those skilled in the art that the scope of the invention in the embodiments of the present disclosure is not limited to the specific combinations of the above-mentioned features, and other embodiments in which the above-mentioned features or their equivalents are combined arbitrarily without departing from the spirit of the invention are also encompassed. For example, the above features and (but not limited to) technical features with similar functions disclosed in the embodiments of the present disclosure are mutually replaced to form the technical solution.

Claims (10)

1. A method for encoding video data, comprising:
based on the current bandwidth condition, selecting a target number of image areas from an image area queue as target image areas, wherein the image areas in the image area queue are arranged according to the size of a quantization parameter score corresponding to each image area, and the quantization parameter score is determined according to the space coding complexity and the time coding complexity corresponding to each image area;
and coding the target image area in a first coding mode, and coding the rest image areas in the image area queue in a second coding mode.
2. The method of claim 1, wherein selecting a target number of image regions from the image region queue as target image regions based on the current bandwidth condition comprises:
and responding to the situation that the current bandwidth index meets a preset condition, increasing the historical number to obtain the target number, and selecting the image areas with the target number from the image area queue as the target image areas, wherein the historical number is the number of the image areas which are coded in the first coding mode at the last time.
3. The method of claim 2, wherein selecting a target number of image regions from the image region queue as target image regions based on the current bandwidth condition comprises:
and in response to the fact that the current bandwidth index does not meet the preset condition, the historical number is decreased to obtain the target number, and the target number of image areas are selected from the image area queue to serve as the target image areas.
4. The method of claim 2, wherein the current bandwidth indicator is calculated by:
acquiring current network delay and historical network delay;
determining a rate of change of the current network delay relative to the historical network delay;
determining the rate of change as the current bandwidth indicator.
5. The method of claim 1, wherein the first encoding mode is a reference-only key frame encoding mode, and the second encoding mode is a reference key frame and a previous frame encoding mode.
6. An apparatus for encoding video data, comprising:
the selecting unit is configured to select a target number of image areas from an image area queue as target image areas based on the current bandwidth condition, wherein the image areas in the image area queue are arranged according to the quantization parameter score corresponding to each image area, and the quantization parameter score is determined according to the space coding complexity and the time coding complexity corresponding to each image area;
an encoding unit configured to encode the target image region in a first encoding manner and encode remaining image regions in the image region queue in a second encoding manner.
7. The apparatus of claim 6, wherein the selecting unit is configured to:
and responding to the situation that the current bandwidth index meets a preset condition, increasing the historical number to obtain the target number, and selecting the image areas with the target number from the image area queue as the target image areas, wherein the historical number is the number of the image areas which are coded in the first coding mode at the last time.
8. The apparatus of claim 7, wherein the selecting unit is configured to:
and in response to the fact that the current bandwidth index does not meet the preset condition, the historical number is decreased to obtain the target number, and the target number of image areas are selected from the image area queue to serve as the target image areas.
9. An electronic device, comprising:
one or more processors;
a storage device having one or more programs stored thereon,
when executed by the one or more processors, cause the one or more processors to implement the method of any one of claims 1-5.
10. A computer-readable medium, on which a computer program is stored, wherein the program, when executed by a processor, implements the method of any one of claims 1-5.
CN202110512362.9A 2021-05-11 2021-05-11 Video data encoding method, video data encoding device, electronic apparatus, and medium Pending CN115334310A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202110512362.9A CN115334310A (en) 2021-05-11 2021-05-11 Video data encoding method, video data encoding device, electronic apparatus, and medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202110512362.9A CN115334310A (en) 2021-05-11 2021-05-11 Video data encoding method, video data encoding device, electronic apparatus, and medium

Publications (1)

Publication Number Publication Date
CN115334310A true CN115334310A (en) 2022-11-11

Family

ID=83912281

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202110512362.9A Pending CN115334310A (en) 2021-05-11 2021-05-11 Video data encoding method, video data encoding device, electronic apparatus, and medium

Country Status (1)

Country Link
CN (1) CN115334310A (en)

Similar Documents

Publication Publication Date Title
CN111225209B (en) Video data plug flow method, device, terminal and storage medium
CN113115067A (en) Live broadcast system, video processing method and related device
CN112954354B (en) Video transcoding method, device, equipment and medium
CN112312229A (en) Video transmission method and device, electronic equipment and storage medium
CN112312137A (en) Video transmission method and device, electronic equipment and storage medium
CN113490055B (en) Data processing method and device
CN112866746A (en) Multi-path streaming cloud game control method, device, equipment and storage medium
CN111183640A (en) Transmission device, transmission method, and program
CN111385576B (en) Video coding method and device, mobile terminal and storage medium
CN111385574B (en) Code rate control method and device in video coding, mobile terminal and storage medium
CN111787322B (en) Video coding method and device, electronic equipment and computer readable storage medium
CN111478916B (en) Data transmission method, device and storage medium based on video stream
CN112261408B (en) Image processing method and device for head-mounted display equipment and electronic equipment
US20130094571A1 (en) Low latency video compression
KR101399951B1 (en) Multi-view video steaming system and providing method thereof
CN113542856A (en) Reverse playing method, device, equipment and computer readable medium for online video
CN115334310A (en) Video data encoding method, video data encoding device, electronic apparatus, and medium
CN115767149A (en) Video data transmission method and device
US10356439B2 (en) Flexible frame referencing for display transport
US20160057431A1 (en) Method and device for transmitting a sequence of pictures, and corresponding method and device for reception, computer program and recording medium
US10148965B2 (en) Moving image coding apparatus and moving image coding method
CN110418134B (en) Video coding method and device based on video quality and electronic equipment
CN115334305A (en) Video data transmission method, video data transmission device, electronic equipment and medium
CN112995780B (en) Network state evaluation method, device, equipment and storage medium
EP4038892B1 (en) Methods, systems, and media for streaming video content using adaptive buffers

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination