WO2023241485A1 - 视频处理方法、发送端、接收端、存储介质和程序产品 - Google Patents

视频处理方法、发送端、接收端、存储介质和程序产品 Download PDF

Info

Publication number
WO2023241485A1
WO2023241485A1 PCT/CN2023/099449 CN2023099449W WO2023241485A1 WO 2023241485 A1 WO2023241485 A1 WO 2023241485A1 CN 2023099449 W CN2023099449 W CN 2023099449W WO 2023241485 A1 WO2023241485 A1 WO 2023241485A1
Authority
WO
WIPO (PCT)
Prior art keywords
video
target
frame rate
content type
identification information
Prior art date
Application number
PCT/CN2023/099449
Other languages
English (en)
French (fr)
Inventor
张殿凯
高思
刘少丽
Original Assignee
中兴通讯股份有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 中兴通讯股份有限公司 filed Critical 中兴通讯股份有限公司
Publication of WO2023241485A1 publication Critical patent/WO2023241485A1/zh

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L65/00Network arrangements, protocols or services for supporting real-time applications in data packet communication
    • H04L65/60Network streaming of media packets
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L65/00Network arrangements, protocols or services for supporting real-time applications in data packet communication
    • H04L65/80Responding to QoS
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N7/00Television systems
    • H04N7/14Systems for two-way working
    • H04N7/15Conference systems

Definitions

  • This application relates to the field of communication technology, in particular to a video processing method, sending end, receiving end, storage medium and program product.
  • video conferencing systems In order to better integrate work efficiency and safety issues, remote working and home working have gradually become a new normal.
  • many businesses have introduced video conferencing systems.
  • the construction of video conferencing systems has solved the communication problems between enterprises in different places to a certain extent, provided convenient and fast communication methods for branches located in different regions, reduced the travel costs of enterprises, and thus enabled people to communicate without leaving home.
  • Embodiments of the present application provide a video processing method, a video sending end, a video receiving end, a storage medium and a program product.
  • embodiments of the present application provide a video processing method, which is applied to a video sending end.
  • the method includes: obtaining a first video content type of a target video, and determining a corresponding first video content type according to the first video content type.
  • identification information and a first frame rate parameter and use the first identification information as target identification information and the first frame rate parameter as a target frame rate parameter; collect the target video according to the target frame rate parameter , and encode the target identification information and the collected target video to obtain a video code stream; send the video code stream to the video receiving end, so that the video receiving end performs coding on the video code stream After decoding to obtain the decoded video and the target identification information, the decoded video is displayed according to the target frame rate parameter corresponding to the target identification information.
  • embodiments of the present application also provide a video processing method, which is applied to a video receiving end.
  • the method includes: receiving a video code stream from a video sending end; decoding the video code stream to obtain a decoded video and target identification information; determining the corresponding target frame rate parameter according to the target identification information, and displaying the decoded video according to the target frame rate parameter.
  • embodiments of the present application also provide a video sending end, including: a memory, a processor and a A computer program is stored on the memory and can be run on the processor. When the processor runs the computer program, the video processing method of the first aspect is executed.
  • embodiments of the present application also provide a video receiving end, including: a memory, a processor, and a computer program stored in the memory and executable on the processor, and the processor runs the The computer program executes the video processing method of the above second aspect.
  • embodiments of the present application also provide a computer-readable storage medium that stores computer-executable instructions, and the computer-executable instructions are used to execute the video processing method as described above.
  • embodiments of the present application further provide a computer program product, including a computer program or computer instructions.
  • the computer program or computer instructions are stored in a computer-readable storage medium, and the processor of the computer device obtains the information from the computer program or computer instructions.
  • the computer readable storage medium reads the computer program or the computer instructions, and the processor runs the computer program or the computer instructions, so that the computer device performs the video processing method as described above.
  • Figure 1 is a schematic structural diagram of an implementation environment for executing a video processing method provided by an embodiment of the present application
  • Figure 2 is a flow chart of a video processing method on the video sending end side provided by an embodiment of the present application
  • Figure 3 is a flow chart of a video processing method on the video sending end side provided by another embodiment of the present application.
  • Figure 4 is a flow chart of a video processing method on the video sending end side provided by another embodiment of the present application.
  • Figure 5 is a flow chart of a video processing method on the video sending end side provided by another embodiment of the present application.
  • Figure 6 is a flow chart of a video processing method on the video sending end side provided by another embodiment of the present application.
  • Figure 7 is a flow chart of a video processing method on the video sending end side provided by another embodiment of the present application.
  • Figure 8 is a flow chart of a video processing method on the video sending end side provided by another embodiment of the present application.
  • Figure 9 is a flow chart of a video processing method on the video sending end side provided by another embodiment of the present application.
  • Figure 10 is a flow chart of a video processing method on the video sending end side provided by another embodiment of the present application.
  • Figure 11 is a flow chart of a video processing method on the video receiving end side provided by an embodiment of the present application.
  • Figure 12 is a schematic diagram of the principle of an implementation environment for executing a video processing method provided by an embodiment of the present application
  • Figure 13 is an overall flow chart of a video processing method provided by an embodiment of the present application.
  • Figure 14 is an overall flow chart of a video processing method provided by another embodiment of the present application.
  • Figure 15 is a schematic structural diagram of a video sending end provided by an embodiment of the present application.
  • Figure 16 is a schematic structural diagram of a video receiving end provided by an embodiment of the present application.
  • video conferencing system In order to better integrate work efficiency and safety issues, remote working and home working have gradually become a new normal. In order to enable businesses to operate normally, many businesses have introduced video conferencing systems.
  • the construction of video conferencing system in one To a certain extent, it solves the communication problems between enterprises in different places, provides convenient and fast communication methods for branches located in different regions, and reduces the travel costs of enterprises, so that people can achieve face-to-face audio and video without leaving home. comminicate.
  • video conferencing auxiliary streaming video In the process of using the video conferencing system, in order to better communicate, it is often necessary to use video conferencing auxiliary streaming video to share some materials, such as documents, videos, pictures, etc., with other participants.
  • embodiments of the present application provide a video processing method, a video transmitter, a video receiver, a computer-readable storage medium and a computer program product, which can not only save bandwidth resources, but also reduce playback lags and enhance the subjective experience.
  • Figure 1 is a schematic structural diagram of an implementation environment for executing a video processing method provided by an embodiment of the present application.
  • the implementation environment includes but is not limited to the video sending end 100 and the video receiving end 200 , where the video sending end 100 and the video receiving end 200 are connected by communication.
  • the relative positions and numbers of the video sending end 100 and the video receiving end 200 can be set accordingly in specific application scenarios.
  • the embodiments of this application do not make any changes to the relative positions and numbers of the video sending end 100 and the video receiving end 200. Specific limitations.
  • the implementation environment for executing the video processing method can be applied to 3G communication network systems, LTE communication network systems, 5G communication network systems, 6G communication network systems, and subsequently evolved mobile/fixed communication network systems. etc., the embodiments of the present application do not specifically limit this.
  • Figure 2 is a flow chart of a video processing method on the video sending end provided by an embodiment of the present application.
  • the video processing method can be applied to the video sending end in Figure 1, including but not limited to steps S100, Step S200 and step S300.
  • Step S100 Obtain the first video content type of the target video, determine the corresponding first identification information and the first frame rate parameter according to the first video content type, and use the first identification information as the target identification information and the first frame rate parameter.
  • the target frame rate parameter As the target frame rate parameter;
  • Step S200 Collect the target video according to the target frame rate parameter, and encode the target identification information and the collected target video to obtain a video code stream;
  • Step S300 Send the video code stream to the video receiving end, so that the video receiving end decodes the video code stream to obtain the decoded video and target identification information, and then decodes the decoded video according to the target frame rate parameter corresponding to the target identification information. display.
  • the video sending end obtains the first video content type of the target video, determines the corresponding first identification information and the first frame rate parameter according to the first video content type, and sends The first identification information is used as the target identification information and the first frame rate parameter is used as the target frame rate parameter; then, the video sending end will Collect the target video using rate parameters, encode the target identification information and the collected target video, obtain the video code stream, and send the video code stream to the video receiving end; finally, the video receiving end will decode the video code stream Obtain the decoded video and target identification information, obtain the corresponding target frame rate parameters according to the target identification information, and then display the decoded video according to the target frame rate parameters.
  • the above target video refers to the content to be displayed on the video sending end. In other words, it refers to the content that the video sending end needs to share with the video receiving end.
  • first video content type may be a document type, a non-document type such as a video image, or other content types.
  • the embodiment of the present application does not make any restrictions on the type of the first video content type. Specific limitations.
  • the first identification information and the first frame rate parameter have a one-to-one correspondence.
  • the first identification information corresponds to the value 1
  • the first frame rate parameter corresponds to f1
  • the first identification information corresponds to the value 0
  • the first frame rate parameter corresponds to f2.
  • the user can independently input the first video content type during initial sharing, and at the same time, the user independently inputs the first identification information and The first frame rate parameter; the first video content type may also be determined through the motion vector parameters of the encoded frame during the video encoding process, and the process is similar to the method steps in Figure 3 below; or the first video content type may be determined through the motion vector parameters in the target video
  • the image characteristics of the video image are used to determine the first video content type.
  • the process is similar to the method steps in Figure 7 below; other acquisition methods are also possible.
  • the embodiment of the present application does not specifically limit the acquisition method of the first video content type. .
  • the video sending end can determine the corresponding first identification information and the first frame rate parameter through local table lookup; it can also be that the first video content type Send to the cloud server, and receive the first identification information and the first frame rate parameter fed back by the cloud server; other acquisition methods may also be used, and the embodiment of the present application does not make any changes to the acquisition method of the first identification information and the first frame rate parameter. Specific limitations.
  • the embodiment of the present application can adaptively adjust the video collection and encoding frame rate parameters based on the first frame rate parameter; and, the video sending end also The first identification information and the target video will be encoded and written into a specific position in the code stream, and then transmitted to the video receiving end through the network for sharing; the video receiving end decodes the received video code stream and extracts the first identification information, according to The first identification information determines the first frame rate parameter, and adaptively adjusts the video display frame rate parameter based on the first frame rate parameter.
  • the embodiment of the present application can determine the corresponding target frame rate parameter according to the first video content type of the target video, and uses different frame rate parameters for different video content types to process the video, which not only can play low-frame When playing content that requires a high frame rate, such as sharing document content, it saves bandwidth resources; it can also reduce playback freezes and enhance the subjective experience when playing content that requires a high frame rate, such as sharing video content.
  • Figure 3 is a flow chart of a video processing method on the video sending end side provided by another embodiment of the present application.
  • the video processing method in the embodiment of the present application also includes but is not limited to step S410, step S420 and step S430.
  • Step S410 Obtain the encoded frame during the encoding process, and determine the motion vector parameters of the encoded frame;
  • Step S420 Determine the second video content type of the target video according to the motion vector parameter, and determine the corresponding second identification information and second frame rate parameter according to the second video content type;
  • Step S430 Update the target identification information to the second identification information, and update the target frame rate parameter to the second frame rate parameter.
  • Embodiments of the present application will continuously update the target identification information and target frame rate parameters.
  • the video sending end will use the motion vector parameters of the encoded frame during the encoding process. to determine the second video content type of the target video, and then determine the corresponding second identification information and the second frame rate parameter according to the second video content type, thereby achieving the update of the target identification information and the target frame rate parameter respectively, so that in the subsequent During the collection and encoding process, the second identification information and the second frame rate parameter are used.
  • the above-mentioned second video content type may be a document type, a non-document type such as a video image, or other content types.
  • This application implements
  • the type of the second video content type is not specifically limited.
  • the second identification information corresponds to the value 1
  • the second frame rate parameter corresponds to f1
  • the second identification information corresponds to the value 0
  • the second frame rate parameter corresponds to f2.
  • the video sending end can determine the corresponding second identification information and the second frame rate parameter through local table lookup; it can also be that the second video content type Send to the cloud server, and receive the second identification information and the second frame rate parameter fed back by the cloud server; other acquisition methods may also be used, and the embodiment of the present application does not make any changes to the acquisition method of the second identification information and the second frame rate parameter. Specific limitations.
  • Figure 4 is a flow chart of a video processing method on the video sending end side provided by another embodiment of the present application, regarding the determination of the second video content type of the target video according to the motion vector parameters in the above step S420. , including but not limited to step S510, step S520 and step S530.
  • Step S510 perform image division on the encoded frame to obtain multiple block images, and determine the total image number of all block images
  • Step S520 Calculate the differences between the motion vector parameters of all two adjacent block images, and determine the number of motion blocks of the motion block image in the block image based on all the differences;
  • Step S530 Determine the second video content type of the target video according to the total number of images and the number of motion blocks.
  • the above-mentioned process of determining the second video content type of the target video according to the motion vector parameters is as follows: first, the video sending end performs image division on the encoded frame, thereby obtaining multiple block images, and statistics The total image number of all block images; then, for each group of two adjacent block images, the video sender will calculate the difference in motion vector parameters of the two adjacent block images, and determine it based on the size of the difference Whether the second block image in the two adjacent block images is a motion block image, and count the number of motion blocks of the motion block images in the two adjacent block images of all groups; finally, according to the total number of images and the number of motion blocks to calculate the second video content type of the target video.
  • Figure 5 is a flow chart of a video processing method on the video sending end side provided by another embodiment of the present application, regarding the determination of motion blocks in the block image based on all differences in the above step S520.
  • the number of motion blocks of the image includes but is not limited to step S610 and step S620.
  • Step S610 For every two adjacent block images, if the difference between the motion vector parameter of the current block image and the motion vector parameter of the next block image is greater than the first preset threshold, the latter block is The image is recorded as a motion block image;
  • Step S620 Count the number of motion blocks in all motion block images.
  • the difference between the motion vector parameter of the previous block image and the motion vector parameter of the next block image is greater than the first preset threshold, It indicates that the front and rear images have changed significantly, and the video sending end will record the latter block image as a motion block image; in addition, if the difference between the motion vector parameters of the previous block image and the motion vector parameters of the next block image is If the difference is less than or equal to the first preset threshold, it means that the change between the front and back images is small, and the video sending end will not record the next block image as a motion block image.
  • the above-mentioned first preset threshold can be input by the user in advance, or can be determined according to the preset rules according to the current video transmission situation.
  • the embodiment of the present application does not specify the method for determining the first preset threshold. limited.
  • Figure 6 is a flow chart of a video processing method on the video sending end side provided by another embodiment of the present application.
  • step S530 determining the target video according to the total number of images and the number of motion blocks.
  • Two video content types including but not limited to step S710 and step S720.
  • Step S710 Calculate the ratio of the number of motion blocks to the number of total images
  • Step S720 Compare the ratio with the second preset threshold, and determine the second video content type of the target video according to the comparison result.
  • the video sending end after calculating the total number of images and the number of motion blocks, the video sending end will calculate the ratio between the number of motion blocks and the total number of images. If the ratio is greater than the second preset threshold, the target video can be determined.
  • the second video content type is a non-document type; if the ratio is less than or equal to the second preset threshold, it can be determined that the second video content type of the target video is a document type.
  • the above-mentioned second preset threshold can be input in advance by the user, or can be determined according to the preset rules according to the current video transmission situation.
  • the embodiment of the present application does not specify the method for determining the second preset threshold. limited.
  • Figure 7 is a flow chart of a video processing method on the video transmitting end side provided by another embodiment of the present application.
  • step S530 determining the target video according to the total number of images and the number of motion blocks.
  • Two video content types including but not limited to step S810 and step S820.
  • Step S810 For each encoded frame, determine a video content type based on the total number of images and the number of motion blocks;
  • Step S820 Select the video content type with the largest number as the second video content type of the target video.
  • the video sending end in order to make the second video content type more accurate, the video sending end will determine a video content type for each encoded frame. Therefore, for multiple encoded frames, the video sending end will determine to obtain multiple videos. Content type; finally, the video sending end will select the video content type with the largest number as the second video content type of the target video.
  • Figure 8 is a flow chart of a video processing method on the video sending end side provided by another embodiment of the present application. After collecting the target video according to the target frame rate parameter in the above step S200, this application The video processing method of the application embodiment also includes but is not limited to step S910, step S920 and step S930.
  • Step S910 Perform feature extraction on the video image of the target video to obtain image features
  • Step S920 Determine the second video content type of the target video according to the image characteristics, and determine the corresponding second identification information and second frame rate parameter according to the second video content type;
  • Step S930 Update the target identification information to the second identification information, and update the target frame rate parameter to the second frame rate parameter.
  • the embodiments of this application will continuously update the target identification information and target frame rate parameters.
  • the video sending end after collecting the target video according to the target frame rate parameters, the video sending end will perform feature extraction on the video image of the target video to obtain the image. characteristics, and then determine the second video content type according to the image characteristics, and then determine the second identification information and the second frame rate parameter corresponding to the second video content type, thereby achieving the update of the target identification information and the target frame rate parameter respectively, so that in back During the continuous collection and encoding process, the second identification information and the second frame rate parameter are used.
  • the above-mentioned second video content type may be a document type, a non-document type such as a video image, or other content types.
  • This application implements
  • the type of the second video content type is not specifically limited.
  • the second identification information corresponds to the value 1
  • the second frame rate parameter corresponds to f1
  • the second identification information corresponds to the value 0
  • the second frame rate parameter corresponds to f2.
  • the video sending end can determine the corresponding second identification information and the second frame rate parameter through local table lookup; it can also be that the second video content type Send to the cloud server, and receive the second identification information and the second frame rate parameter fed back by the cloud server; other acquisition methods may also be used, and the embodiment of the present application does not make any changes to the acquisition method of the second identification information and the second frame rate parameter. Specific limitations.
  • Figure 9 is a flow chart of a video processing method on the video sending end side provided by another embodiment of the present application.
  • the video processing method in the embodiment of the present application also includes but is not limited to step S1010 and step S1020.
  • Step S1010 Classify and identify the image features to obtain the probability value of the video content type of the target video
  • Step S1020 Compare the probability value with the third preset threshold, and determine the second video content type of the target video according to the comparison result.
  • the video sending end after extracting the image features, will classify and identify the image features to obtain the probability value of the video content type of the target video; then, the video sending end will combine the probability value and the third preset threshold Comparison is performed, and finally the second video content type of the target video is determined based on the comparison result.
  • the above probability value is a probability value that the category belongs to a non-document, such as a video
  • the probability value is greater than the third preset threshold
  • it is determined that the second video content type of the target video is a video type
  • the probability value is less than or equal to the third preset threshold, Three preset thresholds, then the second video content type of the target video is determined to be the document type.
  • the above probability value is a probability value that the category to which it belongs is a document
  • the probability value is greater than the third preset threshold
  • it is determined that the second video content type of the target video is a document type
  • the probability value is less than or equal to the third preset threshold threshold, it is determined that the second video content type of the target video is a non-document type.
  • Figure 10 is a flow chart of a video processing method on the video sending end side provided by another embodiment of the present application.
  • the video processing method in the embodiment of the present application also includes but is not limited to step S1110 and step S1120.
  • Step S1110 For each video image, determine a comparison result based on the probability value and the third preset threshold;
  • Step S1120 Determine the second video content type of the target video according to the largest number of comparison results.
  • the video sending end in order to make the second video content type more accurate, the video sending end will determine a video content type for each video image. Therefore, for multiple video images, the video sending end will determine to obtain multiple videos. Content type; finally, the video sending end will select the video content type with the largest number as the second video content type of the target video.
  • Figure 11 is a flow chart of a video processing method on the video receiving end provided by an embodiment of the present application.
  • the video processing method can be applied to the video receiving end in Figure 1, including but not limited to steps S1200, Step S1300 and step S1400.
  • Step S1200 Receive the video code stream from the video sending end
  • Step S1400 Determine the corresponding target frame rate parameter according to the target identification information, and display the decoded video according to the target frame rate parameter.
  • the video sending end obtains the first video content type of the target video, determines the corresponding first identification information and the first frame rate parameter according to the first video content type, and sends The first identification information is used as the target identification information and the first frame rate parameter is used as the target frame rate parameter; then, the video sending end collects the target video according to the target frame rate parameter, and compares the target identification information with the collected target video. Encoding, obtain the video code stream, and send the video code stream to the video receiving end; finally, the video receiving end will decode the video code stream to obtain the decoded video and target identification information, and obtain the corresponding target frame rate parameters based on the target identification information , and then display the decoded video according to the target frame rate parameters.
  • the video receiving end can determine the corresponding second identification information and the second frame rate parameter through local table lookup; it can also be that the second video content type Send to the cloud server, and receive the second identification information and the second frame rate parameter fed back by the cloud server; other acquisition methods may also be used, and the embodiment of the present application does not make any changes to the acquisition method of the second identification information and the second frame rate parameter. Specific limitations.
  • the video receiving end decodes the received video code stream and extracts the target identification information, determines the target frame rate parameters based on the target identification information, and adaptively adjusts the video display frame rate parameters based on the target frame rate parameters.
  • the video processing method on the video receiving end side in the embodiment of the present application corresponds to the video processing method on the video sending end side in the above embodiment
  • the video processing method on the video receiving end side in the embodiment of the present application is for specific implementation methods and technical effects, reference may be made to the specific implementation methods and technical effects of the video processing method on the video sending end side.
  • Figure 12 is a schematic diagram of the principle of an implementation environment for executing a video processing method provided by an embodiment of the present application. It includes the following: First, the video sending end sets relevant parameters, including video content type flag, frame rate, etc., then video encodes the video content type flag and the collected original video to output the video stream, and transmits the video stream to the At the receiving end, the video receiving end performs video decoding on the received code stream and extracts the video content type flag, and adjusts the frame rate parameters of the video display according to the video content type flag; at the same time, it uses the original video and the information generated by the encoding to detect and analyze the video content. , update the video content type flag and frame rate parameters based on the detection results for subsequent video processing.
  • relevant parameters including video content type flag, frame rate, etc.
  • Parameter settings Set parameters such as video content type flag, video frame rate, etc.
  • Video capture capture video according to the set frame rate.
  • Network transmission transmit the video stream from the sending end to the receiving end.
  • Video display Display and render the decoded video.
  • Video decoding Decode the video stream.
  • Content detection Use the video and its related information to analyze and judge the video content type, and output the video content type detection results.
  • Figure 13 is an overall flow chart of a video processing method provided by an embodiment of the present application. Includes the following steps:
  • Step S1-1 Set the auxiliary stream video related parameters of the video conferencing system.
  • the initial auxiliary stream video content type is document, its flag is set to 1, and the initial frame rate parameter is set to f1;
  • Step S1-2 Collect the original video according to the frame rate f1, use the H.265 video encoder to encode the original video and the video content type flag flag to obtain the video code stream, in which the video type flag information is written into the protocol reserved position, At the same time, the motion vector information of the encoded frame is saved for use by subsequent content detection modules;
  • Step S1-4 The video receiving end uses the corresponding H.265 video decoder to decode the received code stream and obtain the decoded video and content type flag;
  • Step S1-5 The video receiving end adjusts the video frame rate parameter according to the video type flag flag to display and play the decoded video. If the video content type flag is 0, the frame rate is adjusted to f2 (f2>>f1), otherwise the frame The rate is adjusted to f1;
  • Step S1-6 The content detection algorithm analyzes and processes the motion vector of the encoded frame generated in step S1-2, including the following: dividing the image into blocks, such as inputting an image with width W and height H, and using the preset patchsize size to divide the image into blocks. Divide into blocks to get the number of blocks in the entire image The motion vector in each block is compared with the motion vector value of the corresponding block in the previous frame.
  • Step S1-7 Based on the detection result of the previous step, modify the video content type flag flag of the video sender and the frame rate parameters of video collection and encoding, including the following: If the video content type detection result is a document, then the video content type flag flag Set to 1, the video frame rate parameter is set to f1; if the video content type detection result is non-document, the video content type flag flag is set to 0, and the video frame rate parameter is set to f2 (f2>>f1).
  • Figure 14 is an overall flow chart of a video processing method provided by another embodiment of the present application. Includes the following steps:
  • Step S2-1 Set the auxiliary stream video related parameters of the video conferencing system.
  • the initial auxiliary stream video content type is document, its flag is set to 1, and the initial frame rate parameter is set to f1;
  • Step S2-2 Collect the original video according to the frame rate f1, use the H.265 hardware video encoder to encode the original video and the video content type flag flag to obtain the video code stream, in which the video type flag information is written into the protocol reserved location. ;
  • Step S2-3 Transmit the encoded video stream in step S2-2 to the video receiving end through the UDP network;
  • Step S2-4 The video receiving end uses the corresponding H.265 video decoder to perform video decoding on the received code stream, and obtains the decoded video and video content type flag;
  • Step S2-5 The video receiving end adjusts the video frame rate parameter according to the video type flag flag to display and play the decoded video. If the video content type flag flag is 0, the frame rate is adjusted to f2 (f2>>f1), otherwise the frame The rate is adjusted to f1;
  • Step S2-6 The content detection module uses a convolutional neural network to detect video content.
  • the network input is the original video collected in step S2-2.
  • Image features are extracted through a deep learning network (such as mobilenet, resnet, etc.), and the classifier (such as softmax, sigmoid, etc.) to judge the category attributes of image features, including the following: using the annotation data set to train the neural network resnet50 to obtain network parameters for subsequent feature extraction; using the trained neural network resnet50 to perform the input image Feature extraction obtains image feature features; input the image feature features into the sigmoid classifier to obtain the probability value prob that the category of the current feature is video; if the video category probability value prob>0.5, the current frame type is judged to be a document, otherwise it is not document, save the current frame type flag; vote on the frame type flags that store N frames continuously, and the type with the most votes is the final video content type detection result of the current frame;
  • a deep learning network such as mobilenet, re
  • Step S2-7 Based on the detection result of the previous step, modify the video content type flag flag of the video sender and the frame rate parameters of video collection and encoding, including the following: If the video content type detection result is a document, then the video content type flag flag Set to 1, the video frame rate parameter is set to f1; if the video content type detection result is non-document, the video content type flag flag is set to 0, and the video frame rate parameter is set to f2 (f2>>f1).
  • dynamically adjusting the video frame rate through video content detection can ensure a good subjective experience of different video contents; in addition, the adaptive frame rate can control video collection, video encoding and other aspects, and can effectively utilize and save bandwidth. resources; in addition, the video receiving end shares the sending end detection results to ensure synchronization and reduce computational complexity.
  • Figure 15 is a schematic structural diagram of a video sending end provided by an embodiment of the present application; an embodiment of the present application also discloses a video sending end 100, including: a first memory 120, a first processor 110 and a computer program stored in the first memory 120 and executable on the first processor 110.
  • the first processor 110 runs the computer program, it executes the video processing method on the video sending end side as in any of the previous embodiments.
  • the first memory 120 can be used to store non-transitory software programs and non-transitory computer executable programs.
  • the first memory 120 may include high-speed random access memory, and may also include non-transitory memory, such as at least one magnetic disk storage device, flash memory device, or other non-transitory solid-state storage device.
  • the first memory 120 may include memory located remotely relative to the first processor 110, and these remote memories may be connected to the implementation environment through a network. Examples of the above-mentioned networks include but are not limited to the Internet, intranets, local area networks, mobile communication networks and combinations thereof.
  • the video sending end 100 in this embodiment can correspond to the video sending end in the implementation environment of the embodiment as shown in Figure 1.
  • the two belong to the same application concept, so they have the same implementation principles and beneficial effects. No further details will be given.
  • the non-transitory software programs and instructions required to implement the video processing method on the video sending end side of the above embodiment are stored in the first memory 120.
  • the video sending end side of the above embodiment is executed.
  • the video processing method for example, performs the method steps in FIGS. 2 to 10 described above.
  • the video sending end 100 of the embodiment of the present application can execute the video processing method on the video sending end side of the above embodiment, the specific implementation manner and technical effects of the video sending end 100 of the embodiment of the present application can be Correspondingly refer to the specific implementation and technical effects of the video processing method on the video sending end side.
  • Figure 16 is a schematic structural diagram of a video receiving end provided by an embodiment of the present application.
  • An embodiment of the invention also discloses a video receiving end 200, including: a second memory 220, a second processor 210, and a computer program stored in the second memory 220 and executable on the second processor 210.
  • the second processor 210 runs the computer program, it executes the video processing method on the video receiving end side as in any of the previous embodiments.
  • the second memory 220 can be used to store non-transitory software programs and non-transitory computer executable programs.
  • the second memory 220 may include high-speed random access memory, and may also include non-transitory memory, such as at least one magnetic disk storage device, flash memory device, or other non-transitory solid-state storage device.
  • the second memory 220 may include memory located remotely relative to the second processor 210, and these remote memories may be connected to the implementation environment through a network. Examples of the above-mentioned networks include but are not limited to the Internet, intranets, local area networks, mobile communication networks and combinations thereof.
  • the video receiving end 200 in this embodiment can correspond to the video receiving end in the implementation environment of the embodiment as shown in Figure 1.
  • the two belong to the same application concept, so they have the same implementation principles and beneficial effects. No further details will be given.
  • the non-transitory software programs and instructions required to implement the video processing method on the video receiving end side of the above embodiment are stored in the second memory 220.
  • the video receiving end side of the above embodiment is executed.
  • the video processing method for example, performs the method steps in Figure 11 described above.
  • the video receiving end 200 of the embodiment of the present application can perform the video processing method on the video receiving end side of the above embodiment, the specific implementation manner and technical effects of the video receiving end 200 of the embodiment of the present application can be Correspondingly refer to the specific implementation and technical effects of the video processing method on the video receiving end side.
  • an embodiment of the present application also discloses a computer-readable storage medium in which computer-executable instructions are stored, and the computer-executable instructions are used to execute the video processing method as in any of the previous embodiments.
  • the computer-readable storage medium in the embodiment of the present application can execute the video processing method on the video sending end side or the video receiving end side in the above embodiment
  • the specific details of the computer-readable storage medium in the embodiment of the present application are:
  • an embodiment of the present application also discloses a computer program product, which includes a computer program or computer instructions.
  • the computer program or computer instructions are stored in a computer-readable storage medium.
  • the processor of the computer device reads the computer program from the computer-readable storage medium.
  • the computer program or computer instructions are obtained, and the processor executes the computer program or computer instructions, so that the computer device performs the video processing method as in any of the previous embodiments.
  • the video sending end will obtain the first video content type of the target video, determine the corresponding first identification information and the first frame rate parameter according to the first video content type, and use the first identification information as the target identification information and the first frame rate parameter as the target frame rate parameter; then, the video sending end will collect the target video according to the target frame rate parameter, and encode the target identification information and the collected target video to obtain the video code stream , and sends the video code stream to the video receiving end; finally, the video receiving end decodes the video code stream to obtain the decoded video and target identification information, and then processes the decoded video according to the target frame rate parameter corresponding to the target identification information. show.
  • the corresponding target frame rate parameters can be determined according to the first video content type of the target video, and different frame rate parameters are used for video processing for different video content types.
  • the rate requirements include: When sharing document content, it saves bandwidth resources; it can also reduce playback freezes when playing content with high frame rate requirements, such as sharing video content, and enhances the subjective experience.
  • Computer storage media includes, but is not limited to, RAM, ROM, EEPROM, flash memory or other memory technology, CD-ROM, Digital Versatile Disk (DVD) or other optical disk storage, magnetic cassettes, tapes, disk storage or other magnetic storage devices, or may Any other medium used to store the desired information and that can be accessed by a computer.
  • communication media typically embodies computer readable instructions, data structures, program modules or other data in a modulated data signal such as a carrier wave or other transport mechanism, and may include any information delivery media .

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Two-Way Televisions, Distribution Of Moving Picture Or The Like (AREA)

Abstract

本申请公开了一种视频处理方法、发送端、接收端、存储介质和程序产品,包括:视频发送端获取目标视频的第一视频内容类型,根据第一视频内容类型确定第一标识信息和第一帧率参数,将第一标识信息作为目标标识信息和将第一帧率参数作为目标帧率参数(S100);根据目标帧率参数采集目标视频,并将目标标识信息和目标视频进行编码得到视频码流(S200);将视频码流发送至视频接收端,以使视频接收端对视频码流进行解码得到解码后视频和目标标识信息之后,再根据与目标标识信息对应的目标帧率参数对解码后视频进行显示(S300)。

Description

视频处理方法、发送端、接收端、存储介质和程序产品
相关申请的交叉引用
本申请基于申请号为202210678897.8、申请日为2022年06月16日的中国专利申请提出,并要求该中国专利申请的优先权,该中国专利申请的全部内容在此引入本申请作为参考。
技术领域
本申请涉及通信技术领域,尤其是一种视频处理方法、发送端、接收端、存储介质和程序产品。
背景技术
目前,为了更好的综合工作效率与安全的问题,远程工作和居家办公逐渐变成一种新常态。为了使企业能够正常运营,许多企业都引入了视频会议系统。视频会议系统的建设在一定程度上解决了异地企业之间的沟通问题,为分布在不同地区的分支机构提供了方便快捷的沟通方式,降低了企业的出行成本,从而使得人们足不出户就能实现面对面的音视频交流。其中,在使用视频会议系统过程中,为了更好的沟通交流,经常需要利用视频会议辅流视频分享一些资料,如文档、视频、图片等给其他参会人员。
但是,现有的视频会议往往存在带宽资源浪费或者播放卡顿等问题,例如:当分享文档内容时,可能经过几秒钟后才会翻页一次,内容变化缓慢,若此时采用较高的帧率进行播放则会导致带宽资源的浪费;另外,当分享视频内容时,若此时采用较低的帧率进行播放则会导致播放卡顿,主观体验感差。
发明内容
本申请实施例提供了一种视频处理方法、视频发送端、视频接收端、存储介质和程序产品。
第一方面,本申请实施例提供了一种视频处理方法,应用于视频发送端,所述方法包括:获取目标视频的第一视频内容类型,根据所述第一视频内容类型确定对应的第一标识信息和第一帧率参数,并将所述第一标识信息作为目标标识信息以及将所述第一帧率参数作为目标帧率参数;根据所述目标帧率参数对所述目标视频进行采集,并将所述目标标识信息和采集得到的所述目标视频进行编码,得到视频码流;将所述视频码流发送至视频接收端,以使所述视频接收端对所述视频码流进行解码得到解码后视频和所述目标标识信息之后,再根据与所述目标标识信息对应的所述目标帧率参数对所述解码后视频进行显示。
第二方面,本申请实施例还提供了一种视频处理方法,应用于视频接收端,所述方法包括:接收来自视频发送端的视频码流;对所述视频码流进行解码,得到解码后视频和目标标识信息;根据所述目标标识信息确定对应的目标帧率参数,并根据所述目标帧率参数对所述解码后视频进行显示。
第三方面,本申请实施例还提供了一种视频发送端,包括:存储器、处理器及存储在所 述存储器上并可在所述处理器上运行的计算机程序,所述处理器运行所述计算机程序时执行如上述第一方面的视频处理方法。
第四方面,本申请实施例还提供了一种视频接收端,包括:存储器、处理器及存储在所述存储器上并可在所述处理器上运行的计算机程序,所述处理器运行所述计算机程序时执行如上述第二方面的视频处理方法。
第五方面,本申请实施例还提供了一种计算机可读存储介质,存储有计算机可执行指令,所述计算机可执行指令用于执行如前面所述的视频处理方法。
第六方面,本申请实施例还提供了一种计算机程序产品,包括计算机程序或计算机指令,所述计算机程序或所述计算机指令存储在计算机可读存储介质中,计算机设备的处理器从所述计算机可读存储介质读取所述计算机程序或所述计算机指令,所述处理器运行所述计算机程序或所述计算机指令,使得所述计算机设备执行如前面所述的视频处理方法。
附图说明
图1是本申请一个实施例提供的用于执行视频处理方法的实施环境的结构示意图;
图2是本申请一个实施例提供的视频发送端侧的视频处理方法的流程图;
图3是本申请另一个实施例提供的视频发送端侧的视频处理方法的流程图;
图4是本申请另一个实施例提供的视频发送端侧的视频处理方法的流程图;
图5是本申请另一个实施例提供的视频发送端侧的视频处理方法的流程图;
图6是本申请另一个实施例提供的视频发送端侧的视频处理方法的流程图;
图7是本申请另一个实施例提供的视频发送端侧的视频处理方法的流程图;
图8是本申请另一个实施例提供的视频发送端侧的视频处理方法的流程图;
图9是本申请另一个实施例提供的视频发送端侧的视频处理方法的流程图;
图10是本申请另一个实施例提供的视频发送端侧的视频处理方法的流程图;
图11是本申请一个实施例提供的视频接收端侧的视频处理方法的流程图;
图12是本申请一个实施例提供的用于执行视频处理方法的实施环境的原理示意图;
图13是本申请一个实施例提供的视频处理方法的整体流程图;
图14是本申请另一个实施例提供的视频处理方法的整体流程图;
图15是本申请一个实施例提供的视频发送端的结构示意图;
图16是本申请一个实施例提供的视频接收端的结构示意图。
具体实施方式
为了使本申请的目的、技术方法及优点更加清楚明白,以下结合附图及实施例,对本申请进行进一步详细说明。应当理解,此处所描述的具体实施例仅用以解释本申请,并不用于限定本申请。
需要说明的是,虽然在流程图中示出了逻辑顺序,但是在某些情况下,可以以不同于流程图中的顺序执行所示出或描述的步骤。说明书和权利要求书及上述附图中的术语“第一”、“第二”等是用于区别类似的对象,而不必用于描述特定的顺序或先后次序。
目前,为了更好的综合工作效率与安全的问题,远程工作和居家办公逐渐变成一种新常态。为了使企业能够正常运营,许多企业都引入了视频会议系统。视频会议系统的建设在一 定程度上解决了异地企业之间的沟通问题,为分布在不同地区的分支机构提供了方便快捷的沟通方式,降低了企业的出行成本,从而使得人们足不出户就能实现面对面的音视频交流。其中,在使用视频会议系统过程中,为了更好的沟通交流,经常需要利用视频会议辅流视频分享一些资料,如文档、视频、图片等给其他参会人员。
但是,现有的视频会议往往存在带宽资源浪费或者播放卡顿等问题,例如:当分享文档内容时,可能经过几秒钟后才会翻页一次,内容变化缓慢,若此时采用较高的帧率进行播放则会导致带宽资源的浪费;另外,当分享视频内容时,若此时采用较低的帧率进行播放则会导致播放卡顿,主观体验感差。
基于此,本申请实施例提供了一种视频处理方法、视频发送端、视频接收端、计算机可读存储介质和计算机程序产品,不但能够节省带宽资源,而且还能够减少播放卡顿情况,增强了主观体验感。
下面结合附图,对本申请实施例作进一步阐述。
如图1所示,图1是本申请一个实施例提供的用于执行视频处理方法的实施环境的结构示意图。
在图1的示例中,该实施环境包括但不限于视频发送端100和视频接收端200,其中,视频发送端100和视频接收端200之间通信连接。
在一实施方式中,视频发送端100和视频接收端200的相对位置、数量等可以在具体应用场景中相应设置,本申请实施例对于视频发送端100和视频接收端200的相对位置和数量不作具体限定。
本领域技术人员可以理解的是,用于执行视频处理方法的实施环境可以应用于3G通信网络系统、LTE通信网络系统、5G通信网络系统、6G通信网络系统以及后续演进的移动/固定通信网络系统等,本申请实施例对此并不作具体限定。
本领域技术人员可以理解的是,图1中示出的实施环境并不构成对本申请实施例的限定,可以包括比图示更多或更少的部件,或者组合某些部件,或者不同的部件布置。
基于上述实施环境,下面提出本申请的视频发送端侧的视频处理方法的各个实施例。
如图2所示,图2是本申请一个实施例提供的视频发送端侧的视频处理方法的流程图,该视频处理方法可以应用于图1中的视频发送端,包括但不限于步骤S100、步骤S200和步骤S300。
步骤S100、获取目标视频的第一视频内容类型,根据第一视频内容类型确定对应的第一标识信息和第一帧率参数,并将第一标识信息作为目标标识信息以及将第一帧率参数作为目标帧率参数;
步骤S200、根据目标帧率参数对目标视频进行采集,并将目标标识信息和采集得到的目标视频进行编码,得到视频码流;
步骤S300、将视频码流发送至视频接收端,以使视频接收端对视频码流进行解码得到解码后视频和目标标识信息之后,再根据与目标标识信息对应的目标帧率参数对解码后视频进行显示。
在一实施例中,在视频处理过程中,首先,视频发送端会获取目标视频的第一视频内容类型,根据第一视频内容类型确定对应的第一标识信息和第一帧率参数,并将第一标识信息作为目标标识信息以及将第一帧率参数作为目标帧率参数;然后,视频发送端会根据目标帧 率参数对目标视频进行采集,并将目标标识信息和采集得到的目标视频进行编码,得到视频码流,并将视频码流发送至视频接收端;最后,视频接收端会对视频码流进行解码得到解码后视频和目标标识信息,根据目标标识信息得到对应的目标帧率参数,再根据目标帧率参数对解码后视频进行显示。
需要说明的是,关于上述的目标视频,是指要在视频发送端上显示的内容,换句话说,是指视频发送端需要分享给视频接收端的内容。
另外,需要说明的是,关于上述的第一视频内容类型,可以是文档类型,也可以是非文档类型如视频影像,也可以是其他内容类型,本申请实施例对第一视频内容类型的种类不作具体限定。
值得注意的是,关于第一视频内容类型、第一标识信息和第一帧率参数,上述三者为一一对应的关系。例如,当第一视频内容类型为文档类型,第一标识信息对应为数值1,第一帧率参数对应为f1;当第一视频内容类型为非文档类型,第一标识信息对应为数值0,第一帧率参数对应为f2。其中,由于文档类型的播放帧率需求小于非文档类型的播放帧率需求,因此可以设置f1小于f2。
需要说明的是,关于上述步骤S100中的目标视频的第一视频内容类型的获取方法,可以是在初始分享时由用户自主输入第一视频内容类型,并且同时由用户自主输入第一标识信息和第一帧率参数;也可以是在视频编码过程中通过编码帧的运动矢量参数来确定第一视频内容类型,其过程与下述图3中的方法步骤类似;也可以是通过目标视频中的视频图像的图像特征来确定第一视频内容类型,其过程与下述图7中的方法步骤类似;也可以是其他的获取方式,本申请实施例对第一视频内容类型的获取方式不作具体限定。
另外,值得注意的是,视频发送端在获取第一视频内容类型之后,可以通过本地查表的方式来确定对应的第一标识信息和第一帧率参数;也可以是将第一视频内容类型发送至云端服务器,接收云端服务器所反馈的第一标识信息和第一帧率参数;也可以是其他的获取方式,,本申请实施例对第一标识信息和第一帧率参数的获取方式不作具体限定。
另外,值得注意的是,当确定了目标帧率参数为第一帧率参数之后,本申请实施例能够基于第一帧率参数自适应调整视频采集和编码帧率参数;并且,视频发送端还会将第一标识信息和目标视频一同编码写入码流中特定的位置,然后通过网络传输到视频接收端共用;视频接收端对接收到的视频码流进行解码并提取第一标识信息,根据第一标识信息确定第一帧率参数,并基于第一帧率参数自适应调整视频显示帧率参数。
值得注意的是,本申请实施例能够根据目标视频的第一视频内容类型确定对应的目标帧率参数,对于不同的视频内容类型采用不同的帧率参数进行视频的处理,不但能够在播放低帧率需求的内容如分享文档内容的情况下,节省带宽资源;而且还能够在播放高帧率需求的内容如分享视频内容的情况下,减少播放卡顿的情况,增强了主观体验感。
另外,如图3所示,图3是本申请另一个实施例提供的视频发送端侧的视频处理方法的流程图,在上述步骤S200中的将目标标识信息和采集得到的目标视频进行编码之后,本申请实施例的视频处理方法还包括但不限于步骤S410、步骤S420和步骤S430。
步骤S410、获取编码过程中的编码帧,并确定编码帧的运动矢量参数;
步骤S420、根据运动矢量参数确定目标视频的第二视频内容类型,并根据第二视频内容类型确定对应的第二标识信息和第二帧率参数;
步骤S430、将目标标识信息更新为第二标识信息,以及将目标帧率参数更新为第二帧率参数。
本申请实施例会不断更新目标标识信息和目标帧率参数,在一实施例中,在对将目标标识信息和目标视频进行编码的期间,视频发送端会根据编码过程中的编码帧的运动矢量参数来确定目标视频的第二视频内容类型,接着再根据第二视频内容类型确定对应的第二标识信息和第二帧率参数,从而分别实现目标标识信息和目标帧率参数的更新,使得在后续采集和编码过程中切换为采用第二标识信息和第二帧率参数。
另外,需要说明的是,关于上述的第二视频内容类型,与上述的第一视频内容类型一样,可以是文档类型,也可以是非文档类型如视频影像,也可以是其他内容类型,本申请实施例对第二视频内容类型的种类不作具体限定。
值得注意的是,关于第二视频内容类型、第二标识信息和第二帧率参数,上述三者为一一对应的关系。例如,当第二视频内容类型为文档类型,第二标识信息对应为数值1,第二帧率参数对应为f1;当第二视频内容类型为非文档类型,第二标识信息对应为数值0,第二帧率参数对应为f2。其中,由于文档类型的播放帧率需求小于非文档类型的播放帧率需求,因此可以设置f1小于f2。
另外,值得注意的是,视频发送端在获取第二视频内容类型之后,可以通过本地查表的方式来确定对应的第二标识信息和第二帧率参数;也可以是将第二视频内容类型发送至云端服务器,接收云端服务器所反馈的第二标识信息和第二帧率参数;也可以是其他的获取方式,,本申请实施例对第二标识信息和第二帧率参数的获取方式不作具体限定。
另外,如图4所示,图4是本申请另一个实施例提供的视频发送端侧的视频处理方法的流程图,关于上述步骤S420中的根据运动矢量参数确定目标视频的第二视频内容类型,包括但不限于步骤S510、步骤S520和步骤S530。
步骤S510、对编码帧进行图像划分得到多个分块图像,并确定所有分块图像的总图像数量;
步骤S520、计算出所有的相邻两个分块图像的运动矢量参数的差值,并根据所有的差值确定分块图像中的运动块图像的运动块数量;
步骤S530、根据总图像数量和运动块数量确定目标视频的第二视频内容类型。
在一实施例中,关于上述的根据运动矢量参数确定目标视频的第二视频内容类型的过程,如下:首先,视频发送端会对编码帧进行图像划分,从而得到多个分块图像,并统计所有分块图像的总图像数量;接着,对于每组的相邻两个分块图像,视频发送端会计算出相邻两个分块图像的运动矢量参数的差值,根据差值的大小来确定相邻两个分块图像中的第二个分块图像是否为运动块图像,并统计所有组的相邻两个分块图像中的运动块图像的运动块数量;最后,再根据总图像数量和运动块数量来计算出目标视频的第二视频内容类型。
另外,如图5所示,图5是本申请另一个实施例提供的视频发送端侧的视频处理方法的流程图,关于上述步骤S520中的根据所有的差值确定分块图像中的运动块图像的运动块数量,包括但不限于步骤S610和步骤S620。
步骤S610、对于每两个相邻的分块图像,当前一个分块图像的运动矢量参数和后一个分块图像的运动矢量参数之间的差值大于第一预设阈值,将后一个分块图像记为运动块图像;
步骤S620、统计所有运动块图像的运动块数量。
在一实施例中,对于每组的相邻两个分块图像,如果前一个分块图像的运动矢量参数和后一个分块图像的运动矢量参数之间的差值大于第一预设阈值,则表明前后图像变化较大,则视频发送端会将后一个分块图像记为运动块图像;另外,如果前一个分块图像的运动矢量参数和后一个分块图像的运动矢量参数之间的差值小于或等于第一预设阈值,则表明前后图像变化较小,则视频发送端不会将后一个分块图像记为运动块图像。
可以理解的是,关于上述的第一预设阈值,可以是用户预先输入的,也可以根据当前视频传输情况按照预设规则确定得到,本申请实施例对第一预设阈值的确定方式不作具体限定。
另外,如图6所示,图6是本申请另一个实施例提供的视频发送端侧的视频处理方法的流程图,关于上述步骤S530中的根据总图像数量和运动块数量确定目标视频的第二视频内容类型,包括但不限于步骤S710和步骤S720。
步骤S710、计算出运动块数量和总图像数量的比值;
步骤S720、将比值和第二预设阈值进行比较,根据比较结果确定目标视频的第二视频内容类型。
在一实施例中,当计算得到总图像数量和运动块数量之后,视频发送端会计算出运动块数量和总图像数量之间的比值,如果比值大于第二预设阈值,则可以确定目标视频的第二视频内容类型为非文档类型;如果比值小于或等于第二预设阈值,则可以确定目标视频的第二视频内容类型为文档类型。
可以理解的是,关于上述的第二预设阈值,可以是用户预先输入的,也可以根据当前视频传输情况按照预设规则确定得到,本申请实施例对第二预设阈值的确定方式不作具体限定。
另外,如图7所示,图7是本申请另一个实施例提供的视频发送端侧的视频处理方法的流程图,关于上述步骤S530中的根据总图像数量和运动块数量确定目标视频的第二视频内容类型,包括但不限于步骤S810和步骤S820。
步骤S810、对于每个编码帧,根据总图像数量和运动块数量确定一个视频内容类型;
步骤S820、选择数量最多的视频内容类型作为目标视频的第二视频内容类型。
在一实施例中,为了使得第二视频内容类型更加准确,视频发送端会对每个编码帧均确定出一个视频内容类型,因此,对于多个编码帧,视频发送端会确定得到多个视频内容类型;最后,视频发送端会选择数量最多的视频内容类型作为目标视频的第二视频内容类型。
另外,如图8所示,图8是本申请另一个实施例提供的视频发送端侧的视频处理方法的流程图,在上述步骤S200中的根据目标帧率参数对目标视频进行采集之后,本申请实施例的视频处理方法还包括但不限于步骤S910、步骤S920和步骤S930。
步骤S910、对目标视频的视频图像进行特征提取,得到图像特征;
步骤S920、根据图像特征确定目标视频的第二视频内容类型,并根据第二视频内容类型确定对应的第二标识信息和第二帧率参数;
步骤S930、将目标标识信息更新为第二标识信息,以及将目标帧率参数更新为第二帧率参数。
本申请实施例会不断更新目标标识信息和目标帧率参数,在一实施例中,在根据目标帧率参数对目标视频进行采集之后,视频发送端会对目标视频的视频图像进行特征提取以得到图像特征,接着再根据图像特征确定第二视频内容类型,再确定第二视频内容类型对应的第二标识信息和第二帧率参数,从而分别实现目标标识信息和目标帧率参数的更新,使得在后 续采集和编码过程中切换为采用第二标识信息和第二帧率参数。
另外,需要说明的是,关于上述的第二视频内容类型,与上述的第一视频内容类型一样,可以是文档类型,也可以是非文档类型如视频影像,也可以是其他内容类型,本申请实施例对第二视频内容类型的种类不作具体限定。
值得注意的是,关于第二视频内容类型、第二标识信息和第二帧率参数,上述三者为一一对应的关系。例如,当第二视频内容类型为文档类型,第二标识信息对应为数值1,第二帧率参数对应为f1;当第二视频内容类型为非文档类型,第二标识信息对应为数值0,第二帧率参数对应为f2。其中,由于文档类型的播放帧率需求小于非文档类型的播放帧率需求,因此可以设置f1小于f2。
另外,值得注意的是,视频发送端在获取第二视频内容类型之后,可以通过本地查表的方式来确定对应的第二标识信息和第二帧率参数;也可以是将第二视频内容类型发送至云端服务器,接收云端服务器所反馈的第二标识信息和第二帧率参数;也可以是其他的获取方式,,本申请实施例对第二标识信息和第二帧率参数的获取方式不作具体限定。
另外,如图9所示,图9是本申请另一个实施例提供的视频发送端侧的视频处理方法的流程图,关于上述步骤S920中的根据图像特征确定目标视频的第二视频内容类型,本申请实施例的视频处理方法还包括但不限于步骤S1010和步骤S1020。
步骤S1010、对图像特征进行分类识别,得到目标视频的视频内容类型的概率值;
步骤S1020、将概率值和第三预设阈值进行比较,根据比较结果确定目标视频的第二视频内容类型。
在一实施例中,在提取得到图像特征之后,视频发送端会对图像特征进行分类识别,得到目标视频的视频内容类型的概率值;接着,视频发送端会将概率值和第三预设阈值进行比较,最后根据比较结果确定目标视频的第二视频内容类型。
其中,当上述的概率值为所属类别为非文档如视频的概率值,若概率值大于第三预设阈值,则确定目标视频的第二视频内容类型为视频类型;若概率值小于或等于第三预设阈值,则确定目标视频的第二视频内容类型为文档类型。
另外,当上述的概率值为所属类别为文档的概率值,若概率值大于第三预设阈值,则确定目标视频的第二视频内容类型为文档类型;若概率值小于或等于第三预设阈值,则确定目标视频的第二视频内容类型为非文档类型。
可以理解的是,关于上述的第三预设阈值,可以是用户预先输入的,也可以根据当前视频传输情况按照预设规则确定得到,本申请实施例对第三预设阈值的确定方式不作具体限定。
另外,如图10所示,图10是本申请另一个实施例提供的视频发送端侧的视频处理方法的流程图,关于上述步骤S1020中的根据比较结果确定目标视频的第二视频内容类型,本申请实施例的视频处理方法还包括但不限于步骤S1110和步骤S1120。
步骤S1110、对于每个视频图像,根据概率值和第三预设阈值确定一个比较结果;
步骤S1120、根据数量最多的比较结果确定目标视频的第二视频内容类型。
在一实施例中,为了使得第二视频内容类型更加准确,视频发送端会对每个视频图像均确定出一个视频内容类型,因此,对于多个视频图像,视频发送端会确定得到多个视频内容类型;最后,视频发送端会选择数量最多的视频内容类型作为目标视频的第二视频内容类型。
基于上述实施环境和视频发送端侧的视频处理方法,下面提出本申请的视频接收端侧的 视频处理方法的各个实施例。
如图11所示,图11是本申请一个实施例提供的视频接收端侧的视频处理方法的流程图,该视频处理方法可以应用于图1中的视频接收端,包括但不限于步骤S1200、步骤S1300和步骤S1400。
步骤S1200、接收来自视频发送端的视频码流;
步骤S1300、对视频码流进行解码,得到解码后视频和目标标识信息;
步骤S1400、根据目标标识信息确定对应的目标帧率参数,并根据目标帧率参数对解码后视频进行显示。
在一实施例中,在视频处理过程中,首先,视频发送端会获取目标视频的第一视频内容类型,根据第一视频内容类型确定对应的第一标识信息和第一帧率参数,并将第一标识信息作为目标标识信息以及将第一帧率参数作为目标帧率参数;然后,视频发送端会根据目标帧率参数对目标视频进行采集,并将目标标识信息和采集得到的目标视频进行编码,得到视频码流,并将视频码流发送至视频接收端;最后,视频接收端会对视频码流进行解码得到解码后视频和目标标识信息,根据目标标识信息得到对应的目标帧率参数,再根据目标帧率参数对解码后视频进行显示。
另外,值得注意的是,视频接收端在获取第二视频内容类型之后,可以通过本地查表的方式来确定对应的第二标识信息和第二帧率参数;也可以是将第二视频内容类型发送至云端服务器,接收云端服务器所反馈的第二标识信息和第二帧率参数;也可以是其他的获取方式,,本申请实施例对第二标识信息和第二帧率参数的获取方式不作具体限定。
视频接收端对接收到的视频码流进行解码并提取目标标识信息,根据目标标识信息确定目标帧率参数,并基于目标帧率参数自适应调整视频显示帧率参数。
值得注意的是,由于本申请实施例的视频接收端侧的视频处理方法对应于上述实施例的视频发送端侧的视频处理方法,因此,本申请实施例的视频接收端侧的视频处理方法的具体实施方式和技术效果,可对应参照上述视频发送端侧的视频处理方法的具体实施方式和技术效果。
基于上述实施环境、视频发送端侧的视频处理方法和视频接收端侧的视频处理方法,下面提出本申请的视频处理方法的整体实施例。
如图12所示,图12是本申请一个实施例提供的用于执行视频处理方法的实施环境的原理示意图。包括如下:首先视频发送端进行相关参数设置,包括视频内容类型标志、帧率等,然后把视频内容类型标志和采集的原始视频一起进行视频编码输出视频码流,通过网络把视频码流传输到接收端,视频接收端对接收到的码流进行视频解码并提取视频内容类型标志,根据视频内容类型标志调整视频显示的帧率参数;同时利用原始视频和编码产生的信息进行视频内容检测和分析,根据检测结果更新视频内容类型标志和帧率参数,用于后续视频处理。
其中,关于图12中的原理步骤,分别如下:
参数设置:设置视频内容类型标志、视频帧率等参数。
视频采集:根据设定的帧率采集视频。
视频编码:采用通用的视频编码器如H.26x,AV1,AVS等,对采集的视频和视频内容类型标志进行编码。
网络传输:把视频流从发送端传输到接收端。
视频显示:对解码后视频进行显示渲染。
视频解码:对视频码流解码。
内容检测:利用视频及其相关信息对视频内容类型进行分析和判断,输出视频内容类型检测结果。
另外,如图13所示,图13是本申请一个实施例提供的视频处理方法的整体流程图。包括如下步骤:
步骤S1-1:设置视频会议系统辅流视频相关参数,初始辅流视频内容类型为文档,其标志位flag设置为1,初始帧率参数设置为f1;
步骤S1-2:按照帧率f1采集原始视频,采用H.265视频编码器对原始视频和视频内容类型标志位flag进行编码得到视频码流,其中视频类型标志信息写入协议预留位置中,同时保存编码帧的运动矢量信息供后续内容检测模块使用;
步骤S1-3:将步骤S1-2已编码的视频码流通过TCP网络传输到视频接收端;
步骤S1-4:视频接收端采用对应的H.265视频解码器对接收到的码流进行视频解码,获取解码后视频和内容类型标志位flag;
步骤S1-5:视频接收端根据视频类型标志flag调整视频帧率参数对解码后视频进行显示播放,如果视频内容类型标志位为0,则帧率调整为f2(f2>>f1),否则帧率调整为f1;
步骤S1-6:内容检测算法对步骤S1-2产生的编码帧运动矢量进行分析处理,包括如下:对图像进行块划分,如输入图像宽度W,高度为H,采用预设patchsize大小对图像进行分块,得到整个图像的块数目每一个块中的运动矢量与前一帧对应块的运动矢量值进行比较,如果二者差值大于一定阈值则认为当前块为运动的,用motionblock记录;计算整个图像中运动块占比ratio=motionblock/num,如果ratio>T,可根据需求通过实验获得,则当前图像内容类型为视频,否则为文档类型,将当前帧类型标志位进行保存;对连续存储N帧的帧类型标志位进行投票,获得票数最多的类型即为当前帧最终的视频内容类型结果;
步骤S1-7:根据上一步检测结果,修改视频发送端的视频内容类型标志位flag和视频采集、编码的帧率参数,包括如下:如果视频内容类型检测结果为文档,则视频内容类型标志位flag设置为1,视频帧率参数设置为f1;如果视频内容类型检测结果为非文档,则视频内容类型标志位flag设置为0,视频帧率参数设置为f2(f2>>f1)。
如图14所示,图14是本申请另一个实施例提供的视频处理方法的整体流程图。包括如下步骤:
步骤S2-1:设置视频会议系统辅流视频相关参数,初始辅流视频内容类型为文档,其标志位flag设置为1,初始帧率参数设置为f1;
步骤S2-2:按照帧率f1采集原始视频,采用H.265硬件视频编码器对原始视频和视频内容类型标志位flag进行编码得到视频码流,其中视频类型标志信息写入协议预留位置中;
步骤S2-3:将步骤S2-2已编码的视频码流通过UDP网络传输到视频接收端;
步骤S2-4:视频接收端采用对应的H.265视频解码器对接收到的码流进行视频解码,获取解码后视频和视频内容类型标志位flag;
步骤S2-5:视频接收端根据视频类型标志flag调整视频帧率参数对解码后视频进行显示播放,如果视频内容类型标志flag为0,则帧率调整为f2(f2>>f1),否则帧率调整为f1;
步骤S2-6:内容检测模块采用基于卷积神经网络的进行视频内容检测,网络输入为步骤S2-2中采集的原始视频,通过深度学习网络(如mobilenet、resnet等)提取图像特征,分类器(如softmax、sigmoid等)对图像特征进行类别属性判断,包括如下:利用标注数据集对神经网络resnet50进行训练,获取网络参数,用于后续特征提取;采用已训练的神经网络resnet50对输入图像进行特征提取获取图像特征features;将图像特征features输入到sigmoid分类器中,获取当前特征所属类别为视频的概率值prob;如果视频类别概率值prob>0.5,则当前帧类型判断为文档,否则为非文档,将当前帧类型标志位进行保存;对连续存储N帧的帧类型标志位进行投票,获得票数最多的类型即为当前帧最终的视频内容类型检测结果;
步骤S2-7:根据上一步检测结果,修改视频发送端的视频内容类型标志位flag和视频采集、编码的帧率参数,包括如下:如果视频内容类型检测结果为文档,则视频内容类型标志位flag设置为1,视频帧率参数设置为f1;如果视频内容类型检测结果为非文档,则视频内容类型标志位flag设置为0,视频帧率参数设置为f2(f2>>f1)。
根据本申请实施例的技术方案,通过视频内容检测动态调整视频帧率可以保证不同视频内容的主观体验良好;另外,自适应帧率可以控制视频采集、视频编码等环节,可以有效利用和节省带宽资源;另外,视频接收端共用发送端检测结果,保证同步性同时降低计算复杂度。
基于上述实施环境、视频发送端侧的视频处理方法和视频接收端侧的视频处理方法,下面提出本申请的视频发送端、视频接收端、计算机可读存储介质和计算机程序产品的各个实施例。
如图15所示,图15是本申请一个实施例提供的视频发送端的结构示意图;本申请的一个实施例还公开了一种视频发送端100,包括:第一存储器120、第一处理器110及存储在第一存储器120上并可在第一处理器110上运行的计算机程序,第一处理器110运行计算机程序时执行如前面任意实施例中的视频发送端侧的视频处理方法。
第一存储器120作为一种非暂态计算机可读存储介质,可用于存储非暂态软件程序以及非暂态性计算机可执行程序。此外,第一存储器120可以包括高速随机存取存储器,还可以包括非暂态存储器,例如至少一个磁盘存储器件、闪存器件、或其他非暂态固态存储器件。在一些实施方式中,第一存储器120可包括相对于第一处理器110远程设置的存储器,这些远程存储器可以通过网络连接至该实施环境。上述网络的实例包括但不限于互联网、企业内部网、局域网、移动通信网及其组合。
本实施例中的视频发送端100,可以对应为如图1所示实施例的实施环境中的视频发送端,两者属于相同的申请构思,因此两者具有相同的实现原理以及有益效果,此处不再详述。
实现上述实施例的视频发送端侧的视频处理方法所需的非暂态软件程序以及指令存储在第一存储器120中,当被第一处理器110执行时,执行上述实施例的视频发送端侧的视频处理方法,例如,执行以上描述的图2至图10中的方法步骤。
值得注意的是,由于本申请实施例的视频发送端100能够执行上述实施例的视频发送端侧的视频处理方法,因此,本申请实施例的视频发送端100的具体实施方式和技术效果,可对应参照上述视频发送端侧的视频处理方法的具体实施方式和技术效果。
另外,如图16所示,图16是本申请一个实施例提供的视频接收端的结构示意图;本申 请的一个实施例还公开了一种视频接收端200,包括:第二存储器220、第二处理器210及存储在第二存储器220上并可在第二处理器210上运行的计算机程序,第二处理器210运行计算机程序时执行如前面任意实施例中的视频接收端侧的视频处理方法。
第二存储器220作为一种非暂态计算机可读存储介质,可用于存储非暂态软件程序以及非暂态性计算机可执行程序。此外,第二存储器220可以包括高速随机存取存储器,还可以包括非暂态存储器,例如至少一个磁盘存储器件、闪存器件、或其他非暂态固态存储器件。在一些实施方式中,第二存储器220可包括相对于第二处理器210远程设置的存储器,这些远程存储器可以通过网络连接至该实施环境。上述网络的实例包括但不限于互联网、企业内部网、局域网、移动通信网及其组合。
本实施例中的视频接收端200,可以对应为如图1所示实施例的实施环境中的视频接收端,两者属于相同的申请构思,因此两者具有相同的实现原理以及有益效果,此处不再详述。
实现上述实施例的视频接收端侧的视频处理方法所需的非暂态软件程序以及指令存储在第二存储器220中,当被第二处理器210执行时,执行上述实施例的视频接收端侧的视频处理方法,例如,执行以上描述的图11中的方法步骤。
值得注意的是,由于本申请实施例的视频接收端200能够执行上述实施例的视频接收端侧的视频处理方法,因此,本申请实施例的视频接收端200的具体实施方式和技术效果,可对应参照上述视频接收端侧的视频处理方法的具体实施方式和技术效果。
另外,本申请的一个实施例还公开了一种计算机可读存储介质,其中存储有计算机可执行指令,计算机可执行指令用于执行如前面任意实施例中的视频处理方法。
值得注意的是,由于本申请实施例的计算机可读存储介质能够执行上述实施例的视频发送端侧或者视频接收端侧的视频处理方法,因此,本申请实施例的计算机可读存储介质的具体实施方式和技术效果,可对应参照上述视频发送端侧或者视频接收端侧的视频处理方法的具体实施方式和技术效果。
此外,本申请的一个实施例还公开了一种计算机程序产品,包括计算机程序或计算机指令,计算机程序或计算机指令存储在计算机可读存储介质中,计算机设备的处理器从计算机可读存储介质读取计算机程序或计算机指令,处理器执行计算机程序或计算机指令,使得计算机设备执行如前面任意实施例中的视频处理方法。
值得注意的是,由于本申请实施例的计算机程序产品能够执行上述实施例的视频发送端侧或者视频接收端侧的视频处理方法,因此,本申请实施例的计算机程序产品的具体实施方式和技术效果,可对应参照上述视频发送端侧或者视频接收端侧的视频处理方法的具体实施方式和技术效果。
本申请实施例中,首先,视频发送端会获取目标视频的第一视频内容类型,根据第一视频内容类型确定对应的第一标识信息和第一帧率参数,并将第一标识信息作为目标标识信息以及将第一帧率参数作为目标帧率参数;然后,视频发送端会根据目标帧率参数对目标视频进行采集,并将目标标识信息和采集得到的目标视频进行编码,得到视频码流,并将视频码流发送至视频接收端;最后,视频接收端会对视频码流进行解码得到解码后视频和目标标识信息,再根据与目标标识信息对应的目标帧率参数对解码后视频进行显示。根据本申请实施例的技术方案,能够根据目标视频的第一视频内容类型确定对应的目标帧率参数,对于不同的视频内容类型采用不同的帧率参数进行视频的处理,不但能够在播放低帧率需求的内容如 分享文档内容的情况下,节省带宽资源;而且还能够在播放高帧率需求的内容如分享视频内容的情况下,减少播放卡顿的情况,增强了主观体验感。
本领域普通技术人员可以理解,上文中所公开方法中的全部或某些步骤、系统可以被实施为软件、固件、硬件及其适当的组合。某些物理组件或所有物理组件可以被实施为由处理器,如中央处理器、数字信号处理器或微处理器执行的软件,或者被实施为硬件,或者被实施为集成电路,如专用集成电路。这样的软件可以分布在计算机可读介质上,计算机可读介质可以包括计算机存储介质(或非暂时性介质)和通信介质(或暂时性介质)。如本领域普通技术人员公知的,术语计算机存储介质包括在用于存储信息(诸如计算机可读指令、数据结构、程序模块或其他数据)的任何方法或技术中实施的易失性和非易失性、可移除和不可移除介质。计算机存储介质包括但不限于RAM、ROM、EEPROM、闪存或其他存储器技术、CD-ROM、数字多功能盘(DVD)或其他光盘存储、磁盒、磁带、磁盘存储或其他磁存储装置、或者可以用于存储期望的信息并且可以被计算机访问的任何其他的介质。此外,本领域普通技术人员公知的是,通信介质通常包含计算机可读指令、数据结构、程序模块或者诸如载波或其他传输机制之类的调制数据信号中的其他数据,并且可包括任何信息递送介质。

Claims (14)

  1. 一种视频处理方法,应用于视频发送端,所述方法包括:
    获取目标视频的第一视频内容类型,根据所述第一视频内容类型确定对应的第一标识信息和第一帧率参数,并将所述第一标识信息作为目标标识信息以及将所述第一帧率参数作为目标帧率参数;
    根据所述目标帧率参数对所述目标视频进行采集,并将所述目标标识信息和采集得到的所述目标视频进行编码,得到视频码流;
    将所述视频码流发送至视频接收端,以使所述视频接收端对所述视频码流进行解码得到解码后视频和所述目标标识信息之后,再根据与所述目标标识信息对应的所述目标帧率参数对所述解码后视频进行显示。
  2. 根据权利要求1所述的视频处理方法,其中,在所述将所述目标标识信息和采集得到的所述目标视频进行编码之后,所述方法还包括:
    获取编码过程中的编码帧,并确定所述编码帧的运动矢量参数;
    根据所述运动矢量参数确定所述目标视频的第二视频内容类型,并根据所述第二视频内容类型确定对应的第二标识信息和第二帧率参数;
    将所述目标标识信息更新为所述第二标识信息,以及将所述目标帧率参数更新为所述第二帧率参数。
  3. 根据权利要求2所述的视频处理方法,其中,所述根据所述运动矢量参数确定所述目标视频的第二视频内容类型,包括:
    对所述编码帧进行图像划分得到多个分块图像,并确定所有所述分块图像的总图像数量;
    计算出所有的相邻两个所述分块图像的运动矢量参数的差值,并根据所有的所述差值确定所述分块图像中的运动块图像的运动块数量;
    根据所述总图像数量和所述运动块数量确定所述目标视频的第二视频内容类型。
  4. 根据权利要求3所述的视频处理方法,其中,所述根据所有的所述差值确定所述分块图像中的运动块图像的运动块数量,包括:
    对于每两个相邻的所述分块图像,当前一个分块图像的运动矢量参数和后一个分块图像的运动矢量参数之间的差值大于第一预设阈值,将所述后一个分块图像记为运动块图像;
    统计所有所述运动块图像的运动块数量。
  5. 根据权利要求3所述的视频处理方法,其中,所述根据所述总图像数量和所述运动块数量确定所述目标视频的第二视频内容类型,包括:
    计算出所述运动块数量和所述总图像数量的比值;
    将所述比值和第二预设阈值进行比较,根据比较结果确定所述目标视频的第二视频内容类型。
  6. 根据权利要求3至5中任意一项所述的视频处理方法,其中,所述根据所述总图像数量和所述运动块数量确定所述目标视频的第二视频内容类型,包括:
    对于每个所述编码帧,根据所述总图像数量和所述运动块数量确定一个视频内容类型;
    选择数量最多的视频内容类型作为所述目标视频的第二视频内容类型。
  7. 根据权利要求1所述的视频处理方法,其中,在所述根据所述目标帧率参数对所述目 标视频进行采集之后,所述方法还包括:
    对所述目标视频的视频图像进行特征提取,得到图像特征;
    根据所述图像特征确定所述目标视频的第二视频内容类型,并根据所述第二视频内容类型确定对应的第二标识信息和第二帧率参数;
    将所述目标标识信息更新为所述第二标识信息,以及将所述目标帧率参数更新为所述第二帧率参数。
  8. 根据权利要求7所述的视频处理方法,其中,所述根据所述图像特征确定所述目标视频的第二视频内容类型,包括:
    对所述图像特征进行分类识别,得到所述目标视频的视频内容类型的概率值;
    将所述概率值和第三预设阈值进行比较,根据比较结果确定所述目标视频的第二视频内容类型。
  9. 根据权利要求8所述的视频处理方法,其中,所述根据比较结果确定所述目标视频的第二视频内容类型,包括:
    对于每个所述视频图像,根据所述概率值和第三预设阈值确定一个比较结果;
    根据数量最多的所述比较结果确定所述目标视频的第二视频内容类型。
  10. 一种视频处理方法,应用于视频接收端,所述方法包括:
    接收来自视频发送端的视频码流;
    对所述视频码流进行解码,得到解码后视频和目标标识信息;
    根据所述目标标识信息确定对应的目标帧率参数,并根据所述目标帧率参数对所述解码后视频进行显示。
  11. 一种视频发送端,包括:存储器、处理器及存储在所述存储器上并可在所述处理器上运行的计算机程序,所述处理器运行所述计算机程序时执行权利要求1至9中任意一项所述的视频处理方法。
  12. 一种视频接收端,包括:存储器、处理器及存储在所述存储器上并可在所述处理器上运行的计算机程序,所述处理器运行所述计算机程序时执行权利要求10所述的视频处理方法。
  13. 一种计算机可读存储介质,存储有计算机可执行指令,所述计算机可执行指令用于执行权利要求1至9中任意一项所述的视频处理方法或者权利要求10所述的视频处理方法。
  14. 一种计算机程序产品,包括计算机程序或计算机指令,所述计算机程序或所述计算机指令存储在计算机可读存储介质中,计算机设备的处理器从所述计算机可读存储介质读取所述计算机程序或所述计算机指令,所述处理器运行所述计算机程序或所述计算机指令,使得所述计算机设备执行权利要求1至9中任意一项所述的视频处理方法或者权利要求10所述的视频处理方法。
PCT/CN2023/099449 2022-06-16 2023-06-09 视频处理方法、发送端、接收端、存储介质和程序产品 WO2023241485A1 (zh)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN202210678897.8A CN117294683A (zh) 2022-06-16 2022-06-16 视频处理方法、发送端、接收端、存储介质和程序产品
CN202210678897.8 2022-06-16

Publications (1)

Publication Number Publication Date
WO2023241485A1 true WO2023241485A1 (zh) 2023-12-21

Family

ID=89192202

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2023/099449 WO2023241485A1 (zh) 2022-06-16 2023-06-09 视频处理方法、发送端、接收端、存储介质和程序产品

Country Status (2)

Country Link
CN (1) CN117294683A (zh)
WO (1) WO2023241485A1 (zh)

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104811722A (zh) * 2015-04-16 2015-07-29 华为技术有限公司 一种视频数据的编解码方法及装置
CN108833923A (zh) * 2018-06-20 2018-11-16 腾讯科技(深圳)有限公司 视频编码、解码方法、装置、存储介质和计算机设备
CN112995776A (zh) * 2021-01-26 2021-06-18 北京字跳网络技术有限公司 共享屏幕内容抓屏帧率确定方法、装置、设备和存储介质
US20210400233A1 (en) * 2019-03-06 2021-12-23 Autel Robotics Co., Ltd. Encoding method, image encoder and image transmission system
CN113965751A (zh) * 2021-10-09 2022-01-21 腾讯科技(深圳)有限公司 屏幕内容编码方法、装置、设备及存储介质

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104811722A (zh) * 2015-04-16 2015-07-29 华为技术有限公司 一种视频数据的编解码方法及装置
CN108833923A (zh) * 2018-06-20 2018-11-16 腾讯科技(深圳)有限公司 视频编码、解码方法、装置、存储介质和计算机设备
US20210400233A1 (en) * 2019-03-06 2021-12-23 Autel Robotics Co., Ltd. Encoding method, image encoder and image transmission system
CN112995776A (zh) * 2021-01-26 2021-06-18 北京字跳网络技术有限公司 共享屏幕内容抓屏帧率确定方法、装置、设备和存储介质
CN113965751A (zh) * 2021-10-09 2022-01-21 腾讯科技(深圳)有限公司 屏幕内容编码方法、装置、设备及存储介质

Also Published As

Publication number Publication date
CN117294683A (zh) 2023-12-26

Similar Documents

Publication Publication Date Title
US10990812B2 (en) Video tagging for video communications
US11025959B2 (en) Probabilistic model to compress images for three-dimensional video
US9936208B1 (en) Adaptive power and quality control for video encoders on mobile devices
US8385425B2 (en) Temporal video filtering for real time communication systems
US11272094B2 (en) System and method providing object-oriented zoom in multimedia messaging
US20200106708A1 (en) Load Balancing Multimedia Conferencing System, Device, and Methods
US10237581B2 (en) Presentation of composite streams to users
US20120275511A1 (en) System and method for providing content aware video adaptation
US20130044192A1 (en) Converting 3d video into 2d video based on identification of format type of 3d video and providing either 2d or 3d video based on identification of display device type
CN110072119B (zh) 一种基于深度学习网络的内容感知视频自适应传输方法
CN102158690A (zh) 远程多路实时视频监控系统
Dou et al. Edge computing-enabled deep learning for real-time video optimization in IIoT
US20210211768A1 (en) Video Tagging For Video Communications
CN108353127B (zh) 基于深度相机的图像稳定
CN111093094A (zh) 视频转码方法、装置、系统及电子设备及可读存储介质
KR102084510B1 (ko) 컨텐츠 특성 기반 트리거 피처를 갖는 컴퓨팅 시스템
WO2021057477A1 (zh) 视频编解码方法和相关装置
WO2015085873A1 (zh) 视频码流的获取方法及装置
US20210352347A1 (en) Adaptive video streaming systems and methods
WO2023241485A1 (zh) 视频处理方法、发送端、接收端、存储介质和程序产品
US11095901B2 (en) Object manipulation video conference compression
CN110753243A (zh) 图像处理方法、图像处理服务器及图像处理系统
KR102574353B1 (ko) 모바일 단말에서 유해 동영상 차단을 위한 단말 자원 기반의 적응적 프레임 추출 및 스트리밍 제어 시스템 및 방법
US20140362178A1 (en) Novel Transcoder and 3D Video Editor
CN114827617A (zh) 一种基于感知模型的视频编解码方法及系统

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 23823053

Country of ref document: EP

Kind code of ref document: A1