WO2024078064A1 - Image processing method and apparatus, and terminal - Google Patents

Image processing method and apparatus, and terminal Download PDF

Info

Publication number
WO2024078064A1
WO2024078064A1 PCT/CN2023/105927 CN2023105927W WO2024078064A1 WO 2024078064 A1 WO2024078064 A1 WO 2024078064A1 CN 2023105927 W CN2023105927 W CN 2023105927W WO 2024078064 A1 WO2024078064 A1 WO 2024078064A1
Authority
WO
WIPO (PCT)
Prior art keywords
image
auxiliary stream
current frame
processed
information
Prior art date
Application number
PCT/CN2023/105927
Other languages
French (fr)
Chinese (zh)
Inventor
鄢玉民
宋晨
Original Assignee
中兴通讯股份有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 中兴通讯股份有限公司 filed Critical 中兴通讯股份有限公司
Publication of WO2024078064A1 publication Critical patent/WO2024078064A1/en

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/23Processing of content or additional data; Elementary server operations; Server middleware
    • H04N21/235Processing of additional data, e.g. scrambling of additional data or processing content descriptors
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/81Monomedia components thereof
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/83Generation or processing of protective or descriptive data associated with content; Content structuring
    • H04N21/845Structuring of content, e.g. decomposing content into time segments
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N5/00Details of television systems
    • H04N5/222Studio circuitry; Studio devices; Studio equipment
    • H04N5/262Studio circuits, e.g. for mixing, switching-over, change of character of image, other special effects ; Cameras specially adapted for the electronic generation of special effects
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N5/00Details of television systems
    • H04N5/222Studio circuitry; Studio devices; Studio equipment
    • H04N5/262Studio circuits, e.g. for mixing, switching-over, change of character of image, other special effects ; Cameras specially adapted for the electronic generation of special effects
    • H04N5/265Mixing
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N7/00Television systems
    • H04N7/14Systems for two-way working
    • H04N7/15Conference systems

Definitions

  • the present application relates to but is not limited to the field of image processing technology.
  • the terminal will interact with the user by displaying multiple frames of images.
  • users' demand for real-time interaction is becoming more and more prominent.
  • Traditional video conferencing only displays a certain frame of image acquired, and cannot synchronously update the difference between consecutive frames, which reduces the interactivity between the terminal and the user and cannot meet the user's usage needs.
  • the present application provides an image processing method, device, terminal, electronic device and storage medium.
  • the present application provides an image processing method, the method comprising: synthesizing an auxiliary stream image and its corresponding annotation information to generate a current frame synthesized image; detecting the current frame synthesized image and the previous frame synthesized image to determine difference information; encoding the current frame synthesized image based on the difference information to generate encoded data; sending the encoded data to a peer device so that the peer device processes the encoded data to obtain and display a decoded image including the annotation information corresponding to the auxiliary stream image.
  • the present application provides an image processing method, the method comprising: obtaining encoded data, which is the data sent by the image processing method in the first aspect; decoding the encoded data to obtain a decoded image, which is an image carrying an auxiliary stream image and its corresponding annotation information; and displaying the decoded image.
  • the present application provides an encoding device, which includes: a synthesis module, configured to synthesize an auxiliary stream image and its corresponding annotation information to generate a current frame synthesized image; a detection module, configured to detect the current frame synthesized image and the previous frame synthesized image, and determine the difference information encoding module, configured to encode the current frame synthesized image according to the difference information to generate encoded data; a sending module, configured to send the encoded data to a peer device, so that the peer device processes the encoded data, obtains and displays a decoded image including the annotation information corresponding to the auxiliary stream image.
  • the present application provides a decoding device, comprising: an acquisition module, configured to acquire encoded data, the encoded data being the data sent by the image processing method in the first aspect; a decoding module, configured to decode the encoded data to obtain a decoded image, the decoded image being an image carrying an auxiliary stream image and its corresponding annotation information; and a display module, configured to display the decoded image.
  • the present application provides a terminal, comprising: an encoding device and/or a decoding device; the encoding device is configured to execute the image processing method in the first aspect of the present application; and the decoding device is configured to execute the image processing method in the second aspect of the present application.
  • the present application provides an image processing system, the image processing system comprising: a plurality of terminals connected in communication, the terminals being configured to implement any one of the image processing methods in the present application.
  • the present application provides an electronic device, comprising: one or more processors; a memory on which one or more programs are stored, and when the one or more programs are executed by one or more processors, the one or more processors implement any image processing method in the present application.
  • the present application provides a readable storage medium, which stores a computer program, and when the computer program is executed by a processor, any one of the image processing methods in the present application is implemented.
  • FIG1 is a schematic flow chart of an image processing method provided in the present application.
  • FIG. 2 is a schematic diagram showing a flow chart of image processing by the image synthesis device provided in the present application.
  • FIG. 3 is a schematic diagram showing a flow chart of detecting auxiliary stream images provided by the present application.
  • FIG4 is a schematic flow chart showing the image processing method provided by the present application.
  • FIG5 is a block diagram showing the composition of the image processing system provided by the present application.
  • FIG6 shows a block diagram of the image processing system provided by the present application.
  • FIG. 7 shows a block diagram of the image processing system provided by the present application.
  • FIG8 shows a schematic diagram of a display interface for auxiliary stream images provided in the present application.
  • FIG9 shows a block diagram of the encoding device provided in the present application.
  • FIG10 is a block diagram showing the composition of the decoding device provided in the present application.
  • FIG. 11 shows a block diagram of the components of the terminal provided in the present application.
  • FIG. 12 is a block diagram showing the composition of the image processing system provided by the present application.
  • FIG. 13 is a block diagram showing an exemplary hardware architecture of a computing device capable of implementing the image processing method and apparatus according to the present application.
  • FIG1 is a flow chart of an image processing method provided by the present application. The method can be applied to an encoding device. As shown in FIG1 , the image processing method in the present application includes but is not limited to the following steps S101 to S104.
  • Step S101 synthesize the auxiliary stream image and its corresponding annotation information to generate a current frame synthesized image.
  • Step S102 Detect the current frame synthesized image and the previous frame synthesized image to determine difference information.
  • the previous frame composite image is an image generated by synthesizing the previous frame image of the auxiliary stream image and the annotation information corresponding to the previous frame image of the auxiliary stream image.
  • Step S103 encoding the current frame synthesized image according to the difference information to generate encoded data.
  • Step S104 Send the coded data to the peer device so that the peer device can The decoded image including the annotation information corresponding to the auxiliary stream image is obtained and displayed.
  • the counterpart device is a device that can process the encoded data, obtain and display the decoded image including the annotation information corresponding to the auxiliary stream image.
  • the counterpart device can be a decoding device, a receiving terminal and other devices.
  • the counterpart device can be set based on the actual application scenario. Other unspecified counterpart devices are also within the scope of protection of this application and will not be repeated here.
  • the information on the terminal's annotation of the auxiliary stream image can be clearly identified; the current frame synthesized image and the previous frame synthesized image are detected to determine the difference information, so that the user can synchronously obtain the difference information between two consecutive frames, thereby improving the interactivity between the terminal and the user; the current frame synthesized image is encoded based on the difference information to generate encoded data, which can speed up the encoding speed of the image to reduce the energy consumption of encoding; and, by sending the encoded data to the opposite device, so that the opposite device processes the encoded data, obtains and displays the decoded image including the annotation information corresponding to the auxiliary stream image, so that the opposite device can view the decoded image with the annotation information, so that the opposite device can display the annotation information more clearly.
  • the synthesis of the auxiliary stream image and its corresponding annotation information to generate a current frame synthesized image in step S101 can be implemented in the following manner: based on multiple frame rates, obtaining the annotation information corresponding to the auxiliary stream image; processing the annotation information corresponding to the auxiliary stream image according to a preset container and a preset image format to generate an annotated image; and integrating the auxiliary stream image and the annotated image to generate a current frame synthesized image.
  • the annotation information corresponding to the auxiliary stream image may be information based on multiple frame rates in the form of point set data.
  • Frame rate refers to the number of frames or images shown or displayed per second.
  • Frame rate is mainly used to refer to the number of frames of an image played per second in the synchronized audio and/or image of a movie, television or video.
  • the frame rate may be 120 frames per second, or 24 frames per second (or 25 frames per second, 30 frames per second), etc.
  • the real-time change of the auxiliary stream image can be clarified, and then the annotation information corresponding to the auxiliary stream image is processed according to a preset container (such as a bitmap container, etc.) and a preset image format (such as an image format of a red green blue alpha (RGBA) color space; a YUV image format, etc.), so that the obtained annotated image can better reflect the real-time change characteristics and meet the user's needs.
  • a preset container such as a bitmap container, etc.
  • a preset image format such as an image format of a red green blue alpha (RGBA) color space; a YUV image format, etc.
  • the "Y” in the YUV image format represents brightness (Luminance or Luma), which is the grayscale value; while “U” and “V” represent chrominance (Chrominance or Chroma), which describes the color and saturation of the image and is used to specify the color of the pixel.
  • auxiliary stream image and the annotated image are integrated (for example, it can be superimposed synthesis or differential synthesis, etc.) to generate a current frame composite image, which is convenient for subsequent processing and improves the image processing efficiency.
  • the auxiliary stream image and the annotated image are integrated to generate a current frame composite image, including: converting the image formats of the auxiliary stream image and the annotated image respectively to obtain a converted image set; scaling each image in the converted image set according to a preset image resolution to obtain a scaled image set; synchronizing each image in the scaled image set according to a preset frame rate to obtain a processed auxiliary stream image and a processed annotated image; and superimposing and synthesizing the processed auxiliary stream image and the processed annotated image to generate a current frame composite image.
  • the processed auxiliary stream images and the processed annotated images can be more conveniently superimposed and synthesized, thereby ensuring the accuracy of the superimposed images and improving the image processing efficiency.
  • FIG2 shows a schematic diagram of a process flow of an image synthesis device provided by the present application for processing an image.
  • the image synthesis device 200 includes but is not limited to the following modules: a label collector 201, a data conversion module 202, an auxiliary stream image acquisition module 203, an image format conversion module 204, an image scaling module 205, a frame rate synchronization module 206, and an image overlay module 207.
  • the annotation collector 201 is configured to collect annotation information and support the collection of annotation information at multiple frame rates, obtain annotation information presented in the form of point set data, or passively receive point set data pushed by an annotation source.
  • the auxiliary stream image acquisition module 203 is configured to acquire auxiliary stream images, support acquisition of auxiliary stream images of various frame rates, and support various image formats, and can actively acquire auxiliary stream images or passively receive auxiliary stream image data push.
  • the data conversion module 202 is configured to process the point set data, for example, to convert the point set data based on a preset container such as a bitmap, and convert it into an appropriate
  • the synthesized annotated image supports output in preset image formats such as the image format of the RGBA color space and the YUV image format.
  • the image format conversion module 204 is configured to convert the format of the auxiliary stream image and the format of the annotated image into the same type of image format to avoid image synthesis failure caused by different image formats.
  • the image scaling module 205 is configured to stretch the auxiliary stream image and the annotated image to the same image resolution according to a preset image resolution, and is applied to the scenario where the resolution of the auxiliary stream image, the resolution of the annotated image and the resolution of the target image are inconsistent.
  • the frame rate synchronization module 206 is configured to synchronize the acquisition frequencies of the auxiliary stream image and the annotated image according to a preset frame rate, and control the frequency of the synthesized current frame synthesized image by dropping frames and/or inserting frames, thereby reducing the data processing pressure of the image synthesis device 200 and improving the efficiency and stability of image synthesis.
  • the image synthesis device 200 may process the input auxiliary stream image and annotation information in the following manner.
  • the auxiliary stream image is collected by the auxiliary stream image collection module 203, and the annotation information is collected by the annotation collector 201. Then, the annotation information corresponding to the auxiliary stream image is processed by the data conversion module 202 according to the preset container and the preset image format to generate an annotated image.
  • the image format conversion module 204 performs image format conversion on the auxiliary stream image and the annotated image respectively to obtain a converted image set, wherein the converted image set includes: the auxiliary stream image after image format conversion and the annotated image after image format conversion.
  • the image scaling module 205 performs scaling processing on the auxiliary stream image after the image format conversion and the annotated image after the image format conversion, for example, according to the preset image resolution, the resolution of the auxiliary stream image after the image format conversion is adjusted to obtain the scaled auxiliary stream image; according to the preset image resolution, the annotated image after the image format conversion is adjusted to obtain the scaled annotated image. This ensures that the image resolutions of the scaled annotated image and the scaled auxiliary stream image are both the preset image resolutions, which facilitates the subsequent image processing.
  • the scaled annotated image and the scaled auxiliary stream image are synchronized by the frame rate synchronization module 206 to obtain a processed auxiliary stream image and a processed annotated image with the same frame rate (for example, both are preset frame rates).
  • each image in the scaled image set is synchronized according to a preset frame rate to obtain a processed auxiliary stream image and a processed annotated image, including: when it is determined that the actual frame rate of the images in the scaled image set is greater than the preset frame rate, frame dropping processing is performed on each image in the scaled image set based on a sampling method to obtain a processed auxiliary stream image and a processed annotated image; when it is determined that the actual frame rate of the images in the scaled image set is less than the preset frame rate, internal interpolation is used to process each image in the scaled image set to obtain a processed auxiliary stream image and a processed annotated image.
  • auxiliary stream images and annotated images By processing auxiliary stream images and annotated images in different frame rate synchronization modes, the success rate of images in the superposition synthesis process can be increased, and the image processing efficiency can be improved.
  • the image superposition module 207 is used to superimpose the processed auxiliary stream image and the processed annotated image to generate a current frame composite image.
  • the processed auxiliary stream image and the processed annotated image are superimposed and synthesized to generate a current frame synthesized image, including: using the processed auxiliary stream image as the background image, superimposing the annotation features in the processed annotated image onto the processed auxiliary stream image, and obtaining the current frame synthesized image.
  • the transparency of the annotated image is set to be fully transparent, thereby obtaining the annotation features in the processed annotated image.
  • the annotation features in the processed annotated image are superimposed on the processed auxiliary stream image to obtain the current frame composite image.
  • the current frame composite image can have both the annotation features in the processed annotated image and the image features of the processed auxiliary stream image, thereby enriching the content of the current frame composite image.
  • the processed auxiliary stream image and the processed annotated image are superimposed and synthesized to generate a current frame synthesized image, including: processing the processed auxiliary stream image according to preset transparency information to obtain image features of the processed auxiliary stream image, wherein the image features of the processed auxiliary stream image match the annotated information;
  • the processed annotated image is used as the background image, and the image features of the processed auxiliary stream image are superimposed on the processed annotated image to obtain a current frame composite image.
  • the processed auxiliary stream image is processed according to the preset transparency information, so that the image features of the processed auxiliary stream image can be obtained, and the image features of the processed auxiliary stream image match the annotation information, so that the characteristics of the annotation information can be represented, so that the processed auxiliary stream image can be further processed.
  • the image features of the processed auxiliary stream image are superimposed on the processed annotated image, so that the current frame synthetic image has both the annotated features in the processed annotated image and the image features of the processed auxiliary stream image.
  • the detection of the current frame composite image and the previous frame composite image to determine the difference information in step S102 can be implemented in the following manner: based on preset sizes, the current frame composite image and the pre-stored previous frame composite image are partitioned to obtain a first region image set corresponding to the current frame composite image and a second region image set corresponding to the previous frame composite image; based on the number of regions, the first region image and the second region image are compared to obtain the difference information.
  • the first region image set includes a plurality of first region images
  • the second region image set includes a plurality of second region images
  • the preset size can be a predefined minimum size for partitioning or blocking an image. For example, if the preset size is 16*16, the current frame composite image can be divided into multiple 16*16 first area images. At the same time, the previous frame composite image can also be divided into multiple 16*16 second area images. This allows the image to be divided in detail and the differences between different images to be more prominent.
  • the number of regions in the first region image set is the same as the number of regions in the second region image set, which can facilitate block-by-block comparison of the region images in the two region image sets, thereby making the obtained difference information more accurate.
  • the difference information includes: at least one difference region.
  • Encoding the current frame synthesized image according to the difference information to generate encoded data includes: determining difference contour information according to the at least one difference region; cropping the current frame synthesized image according to the difference contour information to obtain a changed region image; encoding the changed region image to generate encoded data.
  • the difference region is used to characterize an image region where the image features of the first region image are different from the image features of the second region image, and can accurately measure the difference between the two frames of images, making it convenient to process the current frame synthesized image.
  • At least one difference area is merged to the maximum extent to obtain difference contour information, and the image boundary with difference is clarified, so as to crop the current frame synthetic image based on the difference contour information to obtain the image change information that only includes the difference information.
  • the image of the changing area is merged to the maximum extent to obtain difference contour information, and the image boundary with difference is clarified, so as to crop the current frame synthetic image based on the difference contour information to obtain the image change information that only includes the difference information.
  • the encoded data can reflect the difference between the previous and next two frames of image, thereby improving the encoding speed of the current frame synthesized image.
  • Fig. 3 shows a schematic diagram of a process for detecting auxiliary stream images provided by the present application.
  • the input image of the region detection device 300 is an auxiliary stream annotated image F1, which is a processed image obtained after being processed by the image synthesis device 200, and can simultaneously reflect the features of the auxiliary stream image and the annotated image.
  • the region detection device 300 After acquiring the auxiliary stream annotated image F1, the region detection device 300 performs block (or partition) processing on the auxiliary stream annotated image F1 to obtain a first region image set, where the first region image set includes a plurality of first region images.
  • auxiliary stream annotated image F1 By dividing the auxiliary stream annotated image F1 into blocks (or partitions), local information of the auxiliary stream annotated image F1 can be reflected, so as to facilitate subsequent comparison of features of different local images and realize detection of changed areas.
  • the region detection device 300 also pre-stores a second region image set, which includes multiple second region images.
  • the second region image set is an image set obtained by performing block (or partition) processing on the previous frame synthetic image, which can reflect the image features in different regions of the previous frame synthetic image.
  • difference information for example, different feature information in a certain area, etc.
  • the image block in the area is cached and the area where the image block is located is recorded.
  • the process of caching image blocks can be performed synchronously by multiple threads or by scanning the image blocks with differences line by line.
  • the contour of the image with differences can be extracted (for example, the circumscribed rectangular contour of the image block is extracted, etc.), and then based on the contour, the image within the contour is cropped to generate a difference image corresponding to the difference information.
  • the difference image and the auxiliary stream annotated image F1 are both input to the encoding module 310 for encoding, so that the encoded data can be obtained quickly and accurately.
  • the image of the changed area within the contour is cropped to obtain difference information, thereby improving the accuracy of judging the difference changes of the auxiliary stream image.
  • the method further includes: skipping the current frame synthetic image when it is determined that the difference information indicates that there is no difference between the current frame synthetic image and the previous frame synthetic image.
  • sending the encoded data to the peer device includes: sending the encoded data to the peer device through a first channel; after sending the encoded data to the peer device, further includes: sending labeled data corresponding to the labeled information to the peer device through a second channel.
  • the annotation data corresponding to the annotation information may be data obtained by packaging the annotation information and complying with the transmission rules of the second channel.
  • the annotation information is represented by binary data, and a data packet header (e.g., a data packet header representing information such as the network address of the peer device) is added in front of the binary data, thereby obtaining the annotation data corresponding to the annotation information.
  • Sending encoded data and labeled data corresponding to the labeling information to the peer device through different transmission channels can facilitate the peer device's processing of different data, so that the peer device can analyze and process the obtained encoded data more quickly, thereby improving data processing efficiency.
  • FIG4 is a flow chart of the image processing method provided by the present application. The method can be applied to a decoding device. As shown in FIG4 , the image processing method in the embodiment of the present application includes but is not limited to the following steps S401 to S404.
  • Step S401 obtaining encoded data.
  • the coded data is sent by the peer device (such as a coding device) through the Data encoded by any image processing method in the application.
  • the encoded data is data obtained by the encoding device by encoding the current frame synthetic image based on the difference information
  • the difference information is information obtained by the encoding device by detecting the current frame synthetic image and the previous frame synthetic image
  • the current frame synthetic image is an image synthesized by the encoding device on the auxiliary stream image and its corresponding annotation information.
  • Step S402 decode the encoded data to obtain a decoded image.
  • the decoded image is an image that carries the auxiliary stream image and its corresponding annotation information.
  • the decoding device since the encoded data sent by the encoding device is data encoded by any image processing method in the present application, that is, the encoded data already carries the annotation information and the auxiliary stream image, the decoding device only needs to perform corresponding decoding on the encoded data, thereby ensuring that the decoded image includes the characteristics of the auxiliary stream image and the characteristics of the annotation information corresponding to the auxiliary stream image.
  • the decoding method used by the decoding device to decode the encoded data matches the encoding method of the encoded data to ensure that an accurate decoded image is obtained.
  • the encoding device can use a specific compression technology to encode the current frame synthetic image based on the difference information to obtain encoded data, and the decoding device needs to use the same compression technology to decode the encoded data so that the obtained decoded image can simultaneously include the characteristics of the auxiliary stream image and the characteristics of the annotation information corresponding to the auxiliary stream image.
  • Step S403 superimpose the decoded image and the previous frame of synthesized image to generate an image to be displayed.
  • the decoded image includes an auxiliary stream image and annotation information, and the decoded image can reflect the features of the annotation information and the auxiliary stream image.
  • the decoded image is superimposed with the previous frame composite image to generate an image to be displayed, so that the image to be displayed can reflect the annotation information.
  • Step S404 display the image to be displayed.
  • the method before performing step S404 of displaying the image to be displayed, the method further includes: rendering the image to be displayed to obtain a rendered image to be displayed.
  • the surface shading effect of the image to be displayed can be reflected intuitively and in real time, thereby showing the texture characteristics of the image to be displayed and the influence of the light source on the image to be displayed, so that the user can also view the rendered image to be displayed, thereby improving the user's viewing experience.
  • the invention relates to a method for processing a video image, wherein the coded data is data sent by the coding device and encoded by any one of the image processing methods in the present application, so as to facilitate subsequent processing; the coded data is decoded to obtain and display a decoded image, which is an image carrying an auxiliary stream image and its corresponding annotation information, so that the decoded image can reflect the annotation information and the characteristics of the auxiliary stream image; the decoded image is superimposed with the previous frame composite image to generate an image to be displayed, and the image to be displayed is displayed, so that the image to be displayed can reflect the annotation information.
  • the coded data is data sent by the coding device and encoded by any one of the image processing methods in the present application, so as to facilitate subsequent processing
  • the coded data is decoded to obtain and display a decoded image, which is an image carrying an auxiliary stream image and its corresponding annotation information, so that the decoded image can reflect the annotation information and the characteristics of the auxiliary stream image
  • obtaining the encoded data in step S401 includes: obtaining the encoded data, including: receiving the encoded data through a first channel, wherein the encoded data is data corresponding to a synthesized image of a current frame, and the synthesized image of the current frame is an image synthesized of an auxiliary stream image and its corresponding annotation information; decoding the encoded data, and before obtaining the decoded image, the method also includes: receiving annotation data corresponding to the annotation information through a second channel.
  • the annotation data is data corresponding to the annotation information.
  • the annotation data can be represented by binary data and is used to represent the information.
  • the specific meaning of the annotation information corresponding to the auxiliary stream image can be clarified, so as to facilitate the subsequent processing of the data to be analyzed and improve the data processing efficiency; and by separately processing the data transmitted in different channels, different types of data can be processed to improve the accuracy of data processing.
  • Fig. 5 is a block diagram of the image processing system provided by the present application. As shown in Fig. 5, a first terminal 510 is connected to a second terminal 520 for communication (eg, communicating via the Internet or a communication network, etc.).
  • a first terminal 510 is connected to a second terminal 520 for communication (eg, communicating via the Internet or a communication network, etc.).
  • the first terminal 510 includes: an image synthesis device 511, a region detection device 512, an encoding module 513 and an auxiliary stream data sending module 514.
  • the second terminal 520 includes: a receiving module 521, a decoding module 522 and an image rendering module 523. The functions of each module can refer to the description in the above embodiment.
  • the image synthesis device 511 can simultaneously obtain the annotation information and the auxiliary stream image, process the annotation information, generate the annotation image, and then synthesize the annotation image with the auxiliary stream image to generate the current frame synthesized image, so that the current frame synthesized image can simultaneously reflect the image features of the auxiliary stream image and the features corresponding to the annotation information.
  • the superimposed auxiliary stream image processed by the decoding module 522 can also reflect the characteristics of the annotation information, but the finally obtained image cannot accurately and clearly represent the characteristics of the annotation information.
  • Fig. 6 shows a block diagram of the image processing system provided by the present application. As shown in Fig. 6, a first terminal 610 is connected in communication with a second terminal 620 (eg, communicating via the Internet or a communication network, etc.).
  • a first terminal 610 is connected in communication with a second terminal 620 (eg, communicating via the Internet or a communication network, etc.).
  • the first terminal 610 includes: an image synthesis device 611, a region detection device 612, an encoding module 613, an auxiliary stream data sending module 614 and an annotation information sending module 615.
  • the second terminal 620 includes: a receiving module 621, a decoding module 622, an image rendering module 623 and an annotation information receiving module 624. The functions of each module can refer to the description in the above embodiment.
  • the image synthesis device 611 can simultaneously obtain the annotation information and the auxiliary stream image, process the annotation information, generate the annotation image, and then synthesize the annotation image with the auxiliary stream image to generate the current frame synthesized image, so that the current frame synthesized image can simultaneously reflect the image features of the auxiliary stream image and the features corresponding to the annotation information.
  • annotation information sending module 615 can also obtain annotation information and send the annotation information to the second terminal 620 to facilitate the second terminal 620 to analyze the superimposed auxiliary stream image output by the decoding module 622, so that the image input to the image rendering module 623 can clearly and accurately reflect the characteristics of the annotation information.
  • the terminals can all support decoding of the annotation information, so that the user can obtain the characteristics of the annotation information.
  • Fig. 7 shows a block diagram of the image processing system provided by the present application.
  • the first terminal 710 is connected to the second terminal 720 and the third terminal 730 for communication (eg, communicating via the Internet or a communication network, etc.).
  • the first terminal 710 includes: an image synthesis device 711, a region detection device 712, an encoding module 713, an auxiliary stream data sending module 714, and a label information sending module 715.
  • the second terminal 720 includes: a receiving module 721, a decoding module 722, an image rendering module 723, and a label information receiving module 724.
  • the third terminal 730 includes: a receiving module 731, a decoding module 732, and an image rendering module 733. The functions of each module can refer to the description in the above embodiment.
  • the encoding module 713 is responsible for converting the image output by the area detection device 712 into a suitable
  • the auxiliary stream data sending module 714 and the annotation information sending module 715 transmit the image data to the second terminal 720 (or the third terminal 730) through a wired communication network or a wireless communication network (such as an optical network composed of optical fibers).
  • the image data processing may be implemented in the following manner.
  • the image synthesis device 711 obtains the annotation information and the auxiliary stream image respectively, and the image synthesis device 711 can obtain the annotation information and the auxiliary stream image at the same time, and process the annotation information to generate the annotation image, and then synthesize the annotation image with the auxiliary stream image to generate the current frame synthesis image. So that the current frame synthesis image can simultaneously reflect the image features of the auxiliary stream image and the features corresponding to the annotation information.
  • the annotation information is a series of annotated point set data generated by the annotation source.
  • the area detection device 712 performs difference detection on different areas of the input current frame composite image to obtain difference information, and inputs both the difference information and the auxiliary stream image into the encoding module 713 for encoding, generates encoded data, and outputs the encoded data to the auxiliary stream data sending module 714, so that the auxiliary stream data sending module 714 sends the obtained encoded data to the second terminal 720 (and/or the third terminal 730) through the communication network, so that the second terminal 720 and/or the third terminal 730 can obtain the encoded data synchronized with the auxiliary stream image and annotation information.
  • the region detection device 712 may obtain a first region image set including a plurality of first region images by dividing the input current frame synthetic image into blocks, and then compare the plurality of first region images with a plurality of second region images cached therein to obtain the changed region information.
  • the plurality of second region images are images obtained by dividing the previous frame synthetic image into blocks by the region detection device 712.
  • the second terminal 720 and/or the third terminal 730 will decode the encoded data, but the difference is that the second terminal 720 can also obtain the original annotation information at the same time to facilitate its analysis of the encoded data and obtain accurate auxiliary stream images and annotation information.
  • the annotated image and the auxiliary stream image are superimposed and synthesized. It can ensure the content consistency of the synthesized image of the current frame, and by comparing the differences between adjacent image frames, it can limit the range of image superposition corresponding to the annotation information, thereby improving the speed of image synthesis. It can meet the needs of users in different application scenarios and improve product competitiveness. It solves the problem of inconsistent interaction of auxiliary stream content between the first terminal 710 that can be annotated and the third terminal 730 that cannot be annotated.
  • FIG8 shows a schematic diagram of a display interface for auxiliary stream images provided by the present application.
  • (A) in FIG8 represents a display interface of an auxiliary stream image with annotated information sent by the first terminal 710, or a display interface of an auxiliary stream image with annotated information displayed by the second terminal 720.
  • FIG. 8 shows a display interface in the prior art in which only an auxiliary stream image (ie, an auxiliary stream image without annotation information) is displayed by a terminal.
  • FIG. 8 shows a display interface of the auxiliary stream image displayed by the third terminal 730 .
  • the display status of the information can be clearly marked, which facilitates user viewing and improves the user experience.
  • Fig. 9 shows a block diagram of the coding device provided by the present application. As shown in Fig. 9, in one embodiment, the coding device 900 includes but is not limited to the following modules.
  • the synthesis module 901 is configured to synthesize the auxiliary stream image and its corresponding annotation information to generate a current frame synthesized image;
  • the detection module 902 is configured to detect the current frame synthesized image and the previous frame synthesized image to determine the difference information;
  • the encoding module 903 is configured to encode the current frame synthesized image according to the difference information to generate encoded data;
  • the sending module 904 is configured to send the encoded data to the opposite device so that the opposite device processes the encoded data to obtain and display a decoded image including the annotation information corresponding to the auxiliary stream image.
  • the encoding device 900 in this embodiment can implement any image processing method applied to the encoding device in the embodiments of the present application.
  • the synthesis module synthesizes the auxiliary stream image and its corresponding annotation information to generate the current frame synthetic image, which can clarify the information of the terminal annotating the auxiliary stream image;
  • the detection module detects the current frame synthetic image and the previous frame synthetic image to determine the difference information, so that the user can synchronously obtain the difference information between two consecutive frames, thereby improving the interactivity between the terminal and the user;
  • the encoding module detects the current frame synthetic image and the previous frame synthetic image according to the difference information, and determines the difference information between the two consecutive frames.
  • the auxiliary stream image is encoded to generate encoded data, which can speed up the encoding speed of the image to reduce the energy consumption of encoding; and the encoded data is sent to the opposite device through the sending module, so that the opposite device processes the encoded data, obtains and displays the decoded image including the annotation information corresponding to the auxiliary stream image, so that the opposite device can view the decoded image with the annotation information, and the opposite device can display the annotation information more clearly.
  • Fig. 10 shows a block diagram of a decoding device provided by the present application. As shown in Fig. 10, in one embodiment, the decoding device 1000 includes but is not limited to the following modules.
  • the acquisition module 1001 is configured to acquire encoded data, which is data sent by any image processing method adopted by the encoding device in the present application; the decoding module 1002 is configured to decode the encoded data to obtain a decoded image, which is an image carrying an auxiliary stream image and its corresponding annotation information; the generation module 1003 is configured to superimpose the decoded image and the previous frame composite image to generate an image to be displayed; the display module 1004 is configured to display the image to be displayed.
  • the decoding device 1000 in this embodiment can implement any image processing method applied to a decoding device in the embodiments of the present application.
  • the decoding device of the implementation mode of the present application by using the acquisition module to obtain the encoded data, it is possible to clarify the processing requirements for the encoded data, wherein the encoded data is the data sent by the encoding device and encoded by it using any one of the image processing methods in the present application, which is convenient for subsequent processing; the encoded data is decoded to obtain and display a decoded image, which is an image carrying an auxiliary stream image and its corresponding annotation information, so that the decoded image can reflect the characteristics of the annotation information and the auxiliary stream image, which is convenient for users to use.
  • the encoded data is the data sent by the encoding device and encoded by it using any one of the image processing methods in the present application, which is convenient for subsequent processing
  • the encoded data is decoded to obtain and display a decoded image, which is an image carrying an auxiliary stream image and its corresponding annotation information, so that the decoded image can reflect the characteristics of the annotation information and the auxiliary stream image, which is convenient for users to use.
  • FIG11 is a block diagram showing a terminal provided by the present application.
  • the terminal 1100 includes but is not limited to the following modules: an encoding device 1101 and/or a decoding device 1102 .
  • (A) in FIG. 11 indicates that the terminal 1100 includes only the encoding device 1101 ; (B) in FIG. 11 indicates that the terminal 1100 includes only the decoding device 1102 ; and (C) in FIG. 11 indicates that the terminal 1100 includes the encoding device 1101 and the decoding device 1102 .
  • the encoding device 1101 is configured to execute any image processing method applied to an encoding device in the embodiments of the present application.
  • the decoding device 1102 is configured to execute any image processing method applied to a decoding device in the embodiments of the present application.
  • the terminal 1100 may be a terminal supporting audio/video conferencing functions (such as a smart phone, etc.), or a tablet computer supporting online teaching (or a personal computer, etc.).
  • the above terminal categories are only examples, and specific settings can be made according to actual needs. Other unspecified terminal categories are also within the scope of protection of this application and will not be repeated here.
  • the auxiliary stream image and its corresponding annotation information are synthesized by the encoding device to generate the current frame synthetic image, which can clarify the information of the terminal on the auxiliary stream image; the current frame synthetic image and the previous frame synthetic image are detected to determine the difference information, so that the user can synchronously obtain the difference information between two consecutive frames, thereby improving the interactivity between the terminal and the user; the auxiliary stream image is encoded according to the difference information to generate encoding data, which can speed up the encoding speed of the image to reduce the energy consumption of encoding.
  • the encoding data and its corresponding annotation information are obtained by the encoding device, which can clarify the processing requirements of the encoding data
  • the encoding data is the image obtained by the encoding device encoding the auxiliary stream image according to the difference information
  • the difference information is the information obtained by the encoding device detecting the current frame synthetic image and the previous frame synthetic image, which can enable the user to synchronously obtain the difference information between two consecutive frames, thereby improving the interactivity between the terminal and the user
  • the encoding data is decoded to obtain the image to be analyzed, thereby speeding up the processing speed of the image to be analyzed
  • the image to be analyzed is processed according to the annotation information corresponding to the encoding data to obtain a decoded image, so that the decoded image can reflect the characteristics of the annotation information and the auxiliary stream image, which is convenient for users to use.
  • Fig. 12 shows a block diagram of the image processing system provided by the present application.
  • the image processing system includes a plurality of terminals connected in communication; wherein the terminals can implement any one of the image processing methods in the embodiments of the present application.
  • the image processing system includes but is not limited to the following devices: at least one transmitting terminal 1201 in communication connection, and at least one first receiving terminal 1202 and/or second receiving terminal 1203 .
  • (A) in Figure 12 indicates that the image processing system includes: a sending terminal 1201 and a first receiving terminal 1202 that are communicatively connected; (B) in Figure 12 indicates that the image processing system includes: a sending terminal 1201 and a second receiving terminal 1203 that are communicatively connected; (C) in Figure 12 indicates that the image processing system includes: a sending terminal 1201, and a first receiving terminal 1202 and a second receiving terminal 1203 that are respectively communicatively connected to the sending terminal 1201.
  • the first sending terminal 1201 is configured to execute any of the embodiments of the present application.
  • the first receiving terminal 1202 is configured to execute any one of the image processing methods applied to a decoding device in the embodiments of the present application.
  • the second receiving terminal 1203 is configured to obtain the encoded data sent by the first terminal, decode the encoded data, obtain and display a decoded image including the annotation information corresponding to the auxiliary stream image, wherein the encoded data is data obtained by the encoding device by encoding the current frame synthetic image according to the difference information, the difference information is information obtained by the encoding device by detecting the current frame synthetic image and the previous frame synthetic image, and the current frame synthetic image is an image synthesized by the encoding device on the auxiliary stream image and its corresponding annotation information.
  • the auxiliary stream image and its corresponding annotation information are synthesized by the sending terminal to generate the current frame synthesized image, which can clearly identify the information annotated by the sending terminal on the auxiliary stream image; the current frame synthesized image and the previous frame synthesized image are detected to determine the difference information, so that the user can synchronously obtain the difference information between two consecutive frames, thereby improving the interactivity between the terminal and the user; the auxiliary stream image is encoded according to the difference information to generate encoded data, which can speed up the encoding speed of the image to reduce the energy consumption of encoding.
  • different receiving terminals can receive the encoded data and process the encoded data to obtain and display the decoded image including the annotation information corresponding to the auxiliary stream image, so that the first receiving terminal and/or the second receiving terminal can view the decoded image with the annotation information, so as to display the annotation information more clearly.
  • FIG. 13 is a block diagram showing an exemplary hardware architecture of a computing device capable of implementing the image processing method and apparatus according to the present application.
  • the computing device 1300 includes an input device 1301, an input interface 1302, a central processing unit 1303, a memory 1304, an output interface 1305, and an output device 1306.
  • the input interface 1302, the central processing unit 1303, the memory 1304, and the output interface 1305 are interconnected via a bus 1307.
  • the device 1306 is connected to the bus 1307 through the input interface 1302 and the output interface 1305 respectively, and further connected to other components of the computing device 1300.
  • the input device 1301 receives input information from the outside and transmits the input information to the central processing unit 1303 through the input interface 1302; the central processing unit 1303 processes the input information based on the computer executable instructions stored in the memory 1304 to generate output information, temporarily or permanently stores the output information in the memory 1304, and then transmits the output information to the output device 1306 through the output interface 1305; the output device 1306 outputs the output information to the outside of the computing device 1300 for user use.
  • the computing device shown in Figure 13 can be implemented as an electronic device, which may include: a memory configured to store a program; a processor configured to run the program stored in the memory to execute the image processing method described in the above embodiment.
  • the computing device shown in Figure 13 can be implemented as an image processing system, which may include: a memory configured to store a program; a processor configured to run the program stored in the memory to execute the image processing method described in the above embodiment.
  • Embodiments of the present application may be implemented by executing computer program instructions by a data processor of a mobile device, for example in a processor entity, or by hardware, or by a combination of software and hardware.
  • the computer program instructions may be assembly instructions, instruction set architecture (ISA) instructions, machine instructions, machine-dependent instructions, microcode, firmware instructions, state setting data, or source code or object code written in any combination of one or more programming languages.
  • ISA instruction set architecture
  • the block diagrams of any logic flow in the drawings of this application may represent program steps, or may represent interconnected logic circuits, modules and functions, or may represent a combination of program steps and logic circuits, modules and functions.
  • the computer program may be stored in a memory.
  • the memory may be of any type suitable for the local technical environment and may use any Suitable data storage technology implementations include, but are not limited to, read-only memory (ROM), random access memory (RAM), optical storage devices and systems (digital versatile discs DVD or CD discs), etc.
  • Computer-readable media may include non-transitory storage media.
  • the data processor may be any type suitable for the local technical environment, such as, but not limited to, a general-purpose computer, a special-purpose computer, a microprocessor, a digital signal processor (DSP), an application-specific integrated circuit (ASIC), a programmable logic device (FGPA), and a processor based on a multi-core processor architecture.
  • a general-purpose computer such as, but not limited to, a general-purpose computer, a special-purpose computer, a microprocessor, a digital signal processor (DSP), an application-specific integrated circuit (ASIC), a programmable logic device (FGPA), and a processor based on a multi-core processor architecture.
  • DSP digital signal processor
  • ASIC application-specific integrated circuit
  • FGPA programmable logic device

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Image Processing (AREA)
  • Compression Or Coding Systems Of Tv Signals (AREA)

Abstract

Provided in the present application are an image processing method and apparatus, and a terminal. The method comprises: performing synthesis on an auxiliary stream image and labeling information corresponding to same, so as to generate the current frame of synthesized image; performing detection on the current frame of synthesized image and the previous frame of synthesized image, so as to determine difference information; encoding the current frame of synthesized image according to the difference information, so as to generate encoded data; and sending the encoded data to a peer device, such that the peer device processes the encoded data, so as to obtain and display a decoded image comprising the labeling information corresponding to the auxiliary stream image.

Description

图像处理方法、装置和终端Image processing method, device and terminal
相关申请的交叉引用CROSS-REFERENCE TO RELATED APPLICATIONS
本申请要求2022年10月11日提交给中国专利局的第202211239520.9号专利申请的优先权,其全部内容通过引用合并于此。This application claims priority to patent application No. 202211239520.9 filed with the China Patent Office on October 11, 2022, the entire contents of which are incorporated herein by reference.
技术领域Technical Field
本申请涉及但不限于图像处理技术领域。The present application relates to but is not limited to the field of image processing technology.
背景技术Background technique
目前,在进行视频会议的过程中,终端会通过展示多帧图像的方式,与用户进行人机交互。但是,随着视频会议的发展,用户对实时互动的需求越来越突出。而传统的视频会议仅展示获取到的某帧图像,无法对连续帧之间的差异进行同步更新,降低了终端与用户之间的互动性,无法满足用户的使用需求。At present, during video conferencing, the terminal will interact with the user by displaying multiple frames of images. However, with the development of video conferencing, users' demand for real-time interaction is becoming more and more prominent. Traditional video conferencing only displays a certain frame of image acquired, and cannot synchronously update the difference between consecutive frames, which reduces the interactivity between the terminal and the user and cannot meet the user's usage needs.
发明内容Summary of the invention
本申请提供一种图像处理方法、装置、终端、电子设备和存储介质。The present application provides an image processing method, device, terminal, electronic device and storage medium.
第一方面,本申请提供一种图像处理方法,方法包括:对辅流图像及其对应的标注信息进行合成,生成当前帧合成图像;对当前帧合成图像和前一帧合成图像进行检测,确定差异信息;依据差异信息对当前帧合成图像进行编码,生成编码数据;向对端设备发送编码数据,以使对端设备对编码数据进行处理,获得并显示包括辅流图像对应的标注信息的解码图像。In a first aspect, the present application provides an image processing method, the method comprising: synthesizing an auxiliary stream image and its corresponding annotation information to generate a current frame synthesized image; detecting the current frame synthesized image and the previous frame synthesized image to determine difference information; encoding the current frame synthesized image based on the difference information to generate encoded data; sending the encoded data to a peer device so that the peer device processes the encoded data to obtain and display a decoded image including the annotation information corresponding to the auxiliary stream image.
第二方面,本申请提供一种图像处理方法,方法包括:获取编码数据,该编码数据为第一方面中的图像处理方法所发送的数据;对编码数据进行解码,获得解码图像,解码图像为携带有辅流图像及其对应的标注信息的图像;显示解码图像。 In a second aspect, the present application provides an image processing method, the method comprising: obtaining encoded data, which is the data sent by the image processing method in the first aspect; decoding the encoded data to obtain a decoded image, which is an image carrying an auxiliary stream image and its corresponding annotation information; and displaying the decoded image.
第三方面,本申请提供一种编码装置,其包括:合成模块,被配置为对辅流图像及其对应的标注信息进行合成,生成当前帧合成图像;检测模块,被配置为对当前帧合成图像和前一帧合成图像进行检测,确定差异信息编码模块,被配置为依据差异信息对当前帧合成图像进行编码,生成编码数据;发送模块,被配置为向对端设备发送所述编码数据,以使所述对端设备对所述编码数据进行处理,获得并显示包括所述辅流图像对应的标注信息的解码图像。In a third aspect, the present application provides an encoding device, which includes: a synthesis module, configured to synthesize an auxiliary stream image and its corresponding annotation information to generate a current frame synthesized image; a detection module, configured to detect the current frame synthesized image and the previous frame synthesized image, and determine the difference information encoding module, configured to encode the current frame synthesized image according to the difference information to generate encoded data; a sending module, configured to send the encoded data to a peer device, so that the peer device processes the encoded data, obtains and displays a decoded image including the annotation information corresponding to the auxiliary stream image.
第四方面,本申请提供一种解码装置,其包括:获取模块,被配置为获取编码数据,编码数据为第一方面中的图像处理方法所发送的数据;解码模块,被配置为对编码数据进行解码,获得解码图像,解码图像为携带有辅流图像及其对应的标注信息的图像;显示模块,被配置为显示解码图像。In a fourth aspect, the present application provides a decoding device, comprising: an acquisition module, configured to acquire encoded data, the encoded data being the data sent by the image processing method in the first aspect; a decoding module, configured to decode the encoded data to obtain a decoded image, the decoded image being an image carrying an auxiliary stream image and its corresponding annotation information; and a display module, configured to display the decoded image.
第五方面,本申请提供一种终端,其包括:编码装置和/或解码装置;编码装置,被配置为执行本申请第一方面中的图像处理方法;解码装置,被配置为执行本申请第二方面中的图像处理方法。In a fifth aspect, the present application provides a terminal, comprising: an encoding device and/or a decoding device; the encoding device is configured to execute the image processing method in the first aspect of the present application; and the decoding device is configured to execute the image processing method in the second aspect of the present application.
第六方面,本申请提供一种图像处理系统,图像处理系统,系统包括:通信连接的多个终端,终端被配置为实现本申请中的任意一种图像处理方法。In a sixth aspect, the present application provides an image processing system, the image processing system comprising: a plurality of terminals connected in communication, the terminals being configured to implement any one of the image processing methods in the present application.
第七方面,本申请提供一种电子设备,包括:一个或多个处理器;存储器,其上存储有一个或多个程序,当一个或多个程序被一个或多个处理器执行,使得一个或多个处理器实现本申请中的任意一种图像处理方法。In a seventh aspect, the present application provides an electronic device, comprising: one or more processors; a memory on which one or more programs are stored, and when the one or more programs are executed by one or more processors, the one or more processors implement any image processing method in the present application.
第八方面,本申请提供了一种可读存储介质,该可读存储介质存储有计算机程序,计算机程序被处理器执行时实现本申请中的任意一种图像处理方法。In an eighth aspect, the present application provides a readable storage medium, which stores a computer program, and when the computer program is executed by a processor, any one of the image processing methods in the present application is implemented.
附图说明BRIEF DESCRIPTION OF THE DRAWINGS
图1示出本申请提供的图像处理方法的流程示意图。FIG1 is a schematic flow chart of an image processing method provided in the present application.
图2示出本申请提供的图像合成装置对图像进行处理的流程示意图。 FIG. 2 is a schematic diagram showing a flow chart of image processing by the image synthesis device provided in the present application.
图3示出本申请提供的对辅流图像进行检测的流程示意图。FIG. 3 is a schematic diagram showing a flow chart of detecting auxiliary stream images provided by the present application.
图4示出本申请提供的图像处理方法的流程示意图。FIG4 is a schematic flow chart showing the image processing method provided by the present application.
图5示出本申请提供的图像处理系统的组成方框图。FIG5 is a block diagram showing the composition of the image processing system provided by the present application.
图6示出本申请提供的图像处理系统的组成方框图。FIG6 shows a block diagram of the image processing system provided by the present application.
图7示出本申请提供的图像处理系统的组成方框图。FIG. 7 shows a block diagram of the image processing system provided by the present application.
图8示出本申请提供的对辅流图像的展示界面示意图。FIG8 shows a schematic diagram of a display interface for auxiliary stream images provided in the present application.
图9示出本申请提供的编码装置的组成方框图。FIG9 shows a block diagram of the encoding device provided in the present application.
图10示出本申请提供的解码装置的组成方框图。FIG10 is a block diagram showing the composition of the decoding device provided in the present application.
图11示出本申请提供的终端的组成方框图。FIG. 11 shows a block diagram of the components of the terminal provided in the present application.
图12示出本申请提供的图像处理系统的组成方框图。FIG. 12 is a block diagram showing the composition of the image processing system provided by the present application.
图13示出能够实现根据本申请的图像处理方法和装置的计算设备的示例性硬件架构的结构图。FIG. 13 is a block diagram showing an exemplary hardware architecture of a computing device capable of implementing the image processing method and apparatus according to the present application.
具体实施方式Detailed ways
为使本申请的目的、技术方案和优点更加清楚明白,下文中将结合附图对本申请的实施方式进行详细说明。需要说明的是,在不冲突的情况下,本申请中的实施方式及实施方式中的特征可以相互任意组合。In order to make the purpose, technical solution and advantages of this application more clear, the following will describe the implementation of this application in detail with reference to the accompanying drawings. It should be noted that the implementation of this application and the features in the implementation can be combined with each other arbitrarily without conflict.
图1示出本申请提供的图像处理方法的流程示意图。该方法可应用于编码装置。如图1所示,本申请中的图像处理方法包括但不限于以下步骤S101至S104。FIG1 is a flow chart of an image processing method provided by the present application. The method can be applied to an encoding device. As shown in FIG1 , the image processing method in the present application includes but is not limited to the following steps S101 to S104.
步骤S101,对辅流图像及其对应的标注信息进行合成,生成当前帧合成图像。Step S101 : synthesize the auxiliary stream image and its corresponding annotation information to generate a current frame synthesized image.
步骤S102,对当前帧合成图像和前一帧合成图像进行检测,确定差异信息。Step S102: Detect the current frame synthesized image and the previous frame synthesized image to determine difference information.
其中,前一帧合成图像是对辅流图像的前一帧图像,以及该辅流图像的前一帧图像对应的标注信息进行合成,生成的图像。The previous frame composite image is an image generated by synthesizing the previous frame image of the auxiliary stream image and the annotation information corresponding to the previous frame image of the auxiliary stream image.
步骤S103,依据差异信息对当前帧合成图像进行编码,生成编码数据。Step S103, encoding the current frame synthesized image according to the difference information to generate encoded data.
步骤S104,向对端设备发送编码数据,以使对端设备对编码数 据进行处理,获得并显示包括辅流图像对应的标注信息的解码图像。Step S104: Send the coded data to the peer device so that the peer device can The decoded image including the annotation information corresponding to the auxiliary stream image is obtained and displayed.
其中,对端设备是能够对编码数据进行处理,获得并显示包括辅流图像对应的标注信息的解码图像的设备,例如,该对端设备可以是解码装置、接收终端等设备,可以基于实际应用场景,对对端设备进行设置,其他未说明的对端设备也在本申请的保护范围之内,在此不再赘述。Among them, the counterpart device is a device that can process the encoded data, obtain and display the decoded image including the annotation information corresponding to the auxiliary stream image. For example, the counterpart device can be a decoding device, a receiving terminal and other devices. The counterpart device can be set based on the actual application scenario. Other unspecified counterpart devices are also within the scope of protection of this application and will not be repeated here.
在本申请中,通过对辅流图像及其对应的标注信息进行合成,生成当前帧合成图像,能够明确终端对辅流图像进行标注的信息;对当前帧合成图像和前一帧合成图像进行检测,确定差异信息,使用户可以同步获取连续两帧之间的差异信息,提升终端与用户之间的互动性;依据差异信息对当前帧合成图像进行编码,生成编码数据,可加快对图像的编码速度,以降低编码的能耗;并且,通过向对端设备发送编码数据,以使对端设备对编码数据进行处理,获得并显示包括辅流图像对应的标注信息的解码图像,方便对端设备能够查看到带有标注信息的解码图像,使对端设备可以更清晰明确的显示标注信息。In the present application, by synthesizing the auxiliary stream image and its corresponding annotation information to generate the current frame synthesized image, the information on the terminal's annotation of the auxiliary stream image can be clearly identified; the current frame synthesized image and the previous frame synthesized image are detected to determine the difference information, so that the user can synchronously obtain the difference information between two consecutive frames, thereby improving the interactivity between the terminal and the user; the current frame synthesized image is encoded based on the difference information to generate encoded data, which can speed up the encoding speed of the image to reduce the energy consumption of encoding; and, by sending the encoded data to the opposite device, so that the opposite device processes the encoded data, obtains and displays the decoded image including the annotation information corresponding to the auxiliary stream image, so that the opposite device can view the decoded image with the annotation information, so that the opposite device can display the annotation information more clearly.
在一些实施方式中,步骤S101中的对辅流图像及其对应的标注信息进行合成,生成当前帧合成图像,可以采用如下方式实现:基于多种帧频,获取辅流图像对应的标注信息;依据预设容器和预设图像格式对辅流图像对应的标注信息进行处理,生成标注图像;将辅流图像和标注图像进行图像整合,生成当前帧合成图像。In some embodiments, the synthesis of the auxiliary stream image and its corresponding annotation information to generate a current frame synthesized image in step S101 can be implemented in the following manner: based on multiple frame rates, obtaining the annotation information corresponding to the auxiliary stream image; processing the annotation information corresponding to the auxiliary stream image according to a preset container and a preset image format to generate an annotated image; and integrating the auxiliary stream image and the annotated image to generate a current frame synthesized image.
其中,辅流图像对应的标注信息,可以是以点集数据的形式呈现的基于多种帧频的信息。帧频(frame rate)是指每秒钟放映或显示的帧或图像的数量。帧频主要用于指在电影、电视或视频的同步音频和/或图像中,每秒播放的图像的帧数。例如,帧频可以是每秒120帧,也可以是每秒24帧(或,每秒25帧、每秒30帧)等。The annotation information corresponding to the auxiliary stream image may be information based on multiple frame rates in the form of point set data. Frame rate refers to the number of frames or images shown or displayed per second. Frame rate is mainly used to refer to the number of frames of an image played per second in the synchronized audio and/or image of a movie, television or video. For example, the frame rate may be 120 frames per second, or 24 frames per second (or 25 frames per second, 30 frames per second), etc.
通过多种不同的帧频,获取辅流图像对应的标注信息,能够明确辅流图像的实时变化情况,进而在依据预设容器(如,比特图容器等)和预设图像格式(例如,红绿蓝-透明度(Red Green Blue Alpha,RGBA)色彩空间的图像格式;YUV图像格式等)对辅流图像对应的标注信息进行处理,使获得的标注图像更能体现实时变化特征,满足用户的 实时性使用需求。By obtaining the annotation information corresponding to the auxiliary stream image through a variety of different frame rates, the real-time change of the auxiliary stream image can be clarified, and then the annotation information corresponding to the auxiliary stream image is processed according to a preset container (such as a bitmap container, etc.) and a preset image format (such as an image format of a red green blue alpha (RGBA) color space; a YUV image format, etc.), so that the obtained annotated image can better reflect the real-time change characteristics and meet the user's needs. Real-time usage requirements.
其中,YUV图像格式中的“Y”表示明亮度(Luminance或Luma),也就是灰阶值;而“U”和“V”表示的则是色度(Chrominance或Chroma),描述影像色彩及饱和度,用于指定像素的颜色。Among them, the "Y" in the YUV image format represents brightness (Luminance or Luma), which is the grayscale value; while "U" and "V" represent chrominance (Chrominance or Chroma), which describes the color and saturation of the image and is used to specify the color of the pixel.
进一步地,将辅流图像和标注图像进行图像整合(例如,可以是叠加合成,也可以是差异性合成等),生成当前帧合成图像,方便后续处理,提升对图像的处理效率。Furthermore, the auxiliary stream image and the annotated image are integrated (for example, it can be superimposed synthesis or differential synthesis, etc.) to generate a current frame composite image, which is convenient for subsequent processing and improves the image processing efficiency.
在一些实施方式中,将辅流图像和标注图像进行图像整合,生成当前帧合成图像,包括:分别对辅流图像和标注图像进行图像格式转换,获得转换图像集合;依据预设图像分辨率对转换图像集合中的各个图像进行缩放处理,获得缩放图像集合;依据预设帧频对缩放图像集合中的各个图像进行同步,获得处理后的辅流图像和处理后的标注图像;将处理后的辅流图像和处理后的标注图像进行叠加合成,生成当前帧合成图像。In some embodiments, the auxiliary stream image and the annotated image are integrated to generate a current frame composite image, including: converting the image formats of the auxiliary stream image and the annotated image respectively to obtain a converted image set; scaling each image in the converted image set according to a preset image resolution to obtain a scaled image set; synchronizing each image in the scaled image set according to a preset frame rate to obtain a processed auxiliary stream image and a processed annotated image; and superimposing and synthesizing the processed auxiliary stream image and the processed annotated image to generate a current frame composite image.
通过对辅流图像和标注图像进行多层次、不同维度的处理,能够使处理后的辅流图像和处理后的标注图像可以更方便进行图像的叠加合成,保证叠加后的图像的准确性,并提升图像的处理效率。By performing multi-level and different-dimensional processing on auxiliary stream images and annotated images, the processed auxiliary stream images and the processed annotated images can be more conveniently superimposed and synthesized, thereby ensuring the accuracy of the superimposed images and improving the image processing efficiency.
例如,图2示出本申请提供的图像合成装置对图像进行处理的流程示意图。如图2所示,图像合成装置200包括但不限于如下模块:标注采集器201、数据转换模块202、辅流图像采集模块203、图像格式转换模块204、图像缩放模块205、帧频同步模块206和图像叠加模块207。For example, FIG2 shows a schematic diagram of a process flow of an image synthesis device provided by the present application for processing an image. As shown in FIG2, the image synthesis device 200 includes but is not limited to the following modules: a label collector 201, a data conversion module 202, an auxiliary stream image acquisition module 203, an image format conversion module 204, an image scaling module 205, a frame rate synchronization module 206, and an image overlay module 207.
其中,标注采集器201,配置为对标注信息进行采集,并支持多种帧频的标注信息的采集,获取基于点集数据形式呈现的标注信息,或被动接收标注源推送的点集数据。The annotation collector 201 is configured to collect annotation information and support the collection of annotation information at multiple frame rates, obtain annotation information presented in the form of point set data, or passively receive point set data pushed by an annotation source.
辅流图像采集模块203,配置为对辅流图像进行采集,支持多种帧频的辅流图像的采集,且支持多种图像格式,可以主动获取辅流图像,也可以被动接收辅流图像的数据推送。The auxiliary stream image acquisition module 203 is configured to acquire auxiliary stream images, support acquisition of auxiliary stream images of various frame rates, and support various image formats, and can actively acquire auxiliary stream images or passively receive auxiliary stream image data push.
数据转换模块202,配置为对点集数据进行处理,例如,基于比特图(BitMap)等预设容器对点集数据进行数据转换,将其转换为适 合合成的标注图像,该标注图像支持RGBA色彩空间的图像格式、以及YUV图像格式等预设图像格式的输出。The data conversion module 202 is configured to process the point set data, for example, to convert the point set data based on a preset container such as a bitmap, and convert it into an appropriate The synthesized annotated image supports output in preset image formats such as the image format of the RGBA color space and the YUV image format.
图像格式转换模块204,配置为将辅流图像的格式和标注图像的格式转换为同一种类型的图像格式,以避免因图像格式的不同而导致的图像合成失败。The image format conversion module 204 is configured to convert the format of the auxiliary stream image and the format of the annotated image into the same type of image format to avoid image synthesis failure caused by different image formats.
图像缩放模块205,配置为依据预设图像分辨率,将辅流图像和标注图像的拉伸至相同的图像分辨率,应用于辅流图像的分辨率、标注图像分辨率和目标图像的分辨率不一致的场景中。The image scaling module 205 is configured to stretch the auxiliary stream image and the annotated image to the same image resolution according to a preset image resolution, and is applied to the scenario where the resolution of the auxiliary stream image, the resolution of the annotated image and the resolution of the target image are inconsistent.
帧频同步模块206,配置为依据预设帧频,将辅流图像和标注图像的采集频率进行同步,通过丢帧和/或插帧的方式,控制合成后的当前帧合成图像的频率,从而降低图像合成装置200的数据处理压力,提高图像合成的效率和稳定性。The frame rate synchronization module 206 is configured to synchronize the acquisition frequencies of the auxiliary stream image and the annotated image according to a preset frame rate, and control the frequency of the synthesized current frame synthesized image by dropping frames and/or inserting frames, thereby reducing the data processing pressure of the image synthesis device 200 and improving the efficiency and stability of image synthesis.
例如,图像合成装置200可采用如下方式对输入的辅流图像和标注信息进行处理。For example, the image synthesis device 200 may process the input auxiliary stream image and annotation information in the following manner.
首先,采用辅流图像采集模块203对辅流图像进行采集,并采用标注采集器201对标注信息进行采集。然后,通过数据转换模块202依据预设容器和预设图像格式对辅流图像对应的标注信息进行处理,生成标注图像。First, the auxiliary stream image is collected by the auxiliary stream image collection module 203, and the annotation information is collected by the annotation collector 201. Then, the annotation information corresponding to the auxiliary stream image is processed by the data conversion module 202 according to the preset container and the preset image format to generate an annotated image.
通过图像格式转换模块204分别对辅流图像和标注图像进行图像格式转换,获得转换图像集合,其中,该转换图像集合包括:图像格式转换后的辅流图像和图像格式转换后的标注图像。The image format conversion module 204 performs image format conversion on the auxiliary stream image and the annotated image respectively to obtain a converted image set, wherein the converted image set includes: the auxiliary stream image after image format conversion and the annotated image after image format conversion.
再通过图像缩放模块205分别对图像格式转换后的辅流图像和图像格式转换后的标注图像进行缩放处理,例如,依据预设图像分辨率,对图像格式转换后的辅流图像的分辨率进行调整,获得缩放后的辅流图像;依据预设图像分辨率,对图像格式转换后的标注图像进行调整,获得缩放后的标注图像。以保证缩放后的标注图像和缩放后的辅流图像的图像分辨率都是预设图像分辨率,方便后续对图像的处理。Then, the image scaling module 205 performs scaling processing on the auxiliary stream image after the image format conversion and the annotated image after the image format conversion, for example, according to the preset image resolution, the resolution of the auxiliary stream image after the image format conversion is adjusted to obtain the scaled auxiliary stream image; according to the preset image resolution, the annotated image after the image format conversion is adjusted to obtain the scaled annotated image. This ensures that the image resolutions of the scaled annotated image and the scaled auxiliary stream image are both the preset image resolutions, which facilitates the subsequent image processing.
进一步地,采用帧频同步模块206对缩放后的标注图像和缩放后的辅流图像进行同步,以获得帧频相同(例如,都是预设帧频)的处理后的辅流图像和处理后的标注图像。 Furthermore, the scaled annotated image and the scaled auxiliary stream image are synchronized by the frame rate synchronization module 206 to obtain a processed auxiliary stream image and a processed annotated image with the same frame rate (for example, both are preset frame rates).
例如,依据预设帧频对缩放图像集合中的各个图像进行同步,获得处理后的辅流图像和处理后的标注图像,包括:在确定缩放图像集合中的图像的实际帧频大于预设帧频的情况下,基于采样的方式对缩放图像集合中的各个图像进行丢帧处理,获得处理后的辅流图像和处理后的标注图像;在确定缩放图像集合中的图像的实际帧频小于预设帧频的情况下,采用内部插帧的方式对缩放图像集合中的各个图像进行处理,获得处理后的辅流图像和处理后的标注图像。For example, each image in the scaled image set is synchronized according to a preset frame rate to obtain a processed auxiliary stream image and a processed annotated image, including: when it is determined that the actual frame rate of the images in the scaled image set is greater than the preset frame rate, frame dropping processing is performed on each image in the scaled image set based on a sampling method to obtain a processed auxiliary stream image and a processed annotated image; when it is determined that the actual frame rate of the images in the scaled image set is less than the preset frame rate, internal interpolation is used to process each image in the scaled image set to obtain a processed auxiliary stream image and a processed annotated image.
通过不同的帧频同步方式,对辅流图像和标注图像进行处理,能够提升图像在进行叠加合成过程的成功比例,提升图像的处理效率。By processing auxiliary stream images and annotated images in different frame rate synchronization modes, the success rate of images in the superposition synthesis process can be increased, and the image processing efficiency can be improved.
最后,采用图像叠加模块207对处理后的辅流图像和处理后的标注图像进行叠加,生成当前帧合成图像。Finally, the image superposition module 207 is used to superimpose the processed auxiliary stream image and the processed annotated image to generate a current frame composite image.
在一些实施方式中,将处理后的辅流图像和处理后的标注图像进行叠加合成,生成当前帧合成图像,包括:以处理后的辅流图像作为背景图像,将处理后的标注图像中的标注特征叠加至处理后的辅流图像中,获得当前帧合成图像。In some embodiments, the processed auxiliary stream image and the processed annotated image are superimposed and synthesized to generate a current frame synthesized image, including: using the processed auxiliary stream image as the background image, superimposing the annotation features in the processed annotated image onto the processed auxiliary stream image, and obtaining the current frame synthesized image.
例如,通过对处理后的标注图像中的Alpha分量的分析,将标注图像的透明度设置为全透明,从而获得处理后的标注图像中的标注特征。然后,再将该处理后的标注图像中的标注特征叠加至处理后的辅流图像中,获得当前帧合成图像。能够使当前帧合成图像同时具备处理后的标注图像中的标注特征,以及处理后的辅流图像的图像特征,丰富当前帧合成图像的内容。For example, by analyzing the Alpha component in the processed annotated image, the transparency of the annotated image is set to be fully transparent, thereby obtaining the annotation features in the processed annotated image. Then, the annotation features in the processed annotated image are superimposed on the processed auxiliary stream image to obtain the current frame composite image. The current frame composite image can have both the annotation features in the processed annotated image and the image features of the processed auxiliary stream image, thereby enriching the content of the current frame composite image.
在一些实施方式中,将处理后的辅流图像和处理后的标注图像进行叠加合成,生成当前帧合成图像,包括:依据预设透明度信息对处理后的辅流图像进行处理,获得处理后的辅流图像的图像特征,处理后的辅流图像的图像特征与标注信息相匹配;In some embodiments, the processed auxiliary stream image and the processed annotated image are superimposed and synthesized to generate a current frame synthesized image, including: processing the processed auxiliary stream image according to preset transparency information to obtain image features of the processed auxiliary stream image, wherein the image features of the processed auxiliary stream image match the annotated information;
以处理后的标注图像作为背景图像,将处理后的辅流图像的图像特征叠加至处理后的标注图像中,获得当前帧合成图像。The processed annotated image is used as the background image, and the image features of the processed auxiliary stream image are superimposed on the processed annotated image to obtain a current frame composite image.
其中,依据预设透明度信息对处理后的辅流图像进行处理,能够获取处理后的辅流图像的图像特征,并且,该处理后的辅流图像的图像特征与标注信息相匹配,能够表征标注信息的特性,从而再将处理 后的辅流图像的图像特征叠加至处理后的标注图像中,使当前帧合成图像同时具备处理后的标注图像中的标注特征,以及处理后的辅流图像的图像特征。The processed auxiliary stream image is processed according to the preset transparency information, so that the image features of the processed auxiliary stream image can be obtained, and the image features of the processed auxiliary stream image match the annotation information, so that the characteristics of the annotation information can be represented, so that the processed auxiliary stream image can be further processed. The image features of the processed auxiliary stream image are superimposed on the processed annotated image, so that the current frame synthetic image has both the annotated features in the processed annotated image and the image features of the processed auxiliary stream image.
在一些实施方式中,步骤S102中的对当前帧合成图像和前一帧合成图像进行检测,确定差异信息,可以采用如下方式实现:依据预设尺寸,分别对当前帧合成图像和预先存储的前一帧合成图像进行分区处理,获得与当前帧合成图像对应的第一区域图像集合,以及与前一帧合成图像对应的第二区域图像集合;依据区域数量,分别对比第一区域图像和第二区域图像,获得差异信息。In some embodiments, the detection of the current frame composite image and the previous frame composite image to determine the difference information in step S102 can be implemented in the following manner: based on preset sizes, the current frame composite image and the pre-stored previous frame composite image are partitioned to obtain a first region image set corresponding to the current frame composite image and a second region image set corresponding to the previous frame composite image; based on the number of regions, the first region image and the second region image are compared to obtain the difference information.
其中,第一区域图像集合包括多个第一区域图像,第二区域图像集合包括多个第二区域图像。The first region image set includes a plurality of first region images, and the second region image set includes a plurality of second region images.
需要说明的是,预设尺寸可以是预先定义的对图像进行分区或分块的最小尺寸,例如,预设尺寸为16*16,则可以将当前帧合成图像划分为多个16*16的第一区域图像,同时,也可以将前一帧合成图像划分为多个16*16的第二区域图像,能够对图像进行细节上的划分,更突出不同图像之间的差异性。It should be noted that the preset size can be a predefined minimum size for partitioning or blocking an image. For example, if the preset size is 16*16, the current frame composite image can be divided into multiple 16*16 first area images. At the same time, the previous frame composite image can also be divided into multiple 16*16 second area images. This allows the image to be divided in detail and the differences between different images to be more prominent.
并且,第一区域图像集合中的区域数量与第二区域图像集合中的区域数量相同,可方便对两个区域图像集合中的区域图像进行分块对比,使获得的差异信息更准确。Furthermore, the number of regions in the first region image set is the same as the number of regions in the second region image set, which can facilitate block-by-block comparison of the region images in the two region image sets, thereby making the obtained difference information more accurate.
在一些实施方式中,差异信息包括:至少一个差异区域。依据差异信息对当前帧合成图像进行编码,生成编码数据,包括:依据至少一个差异区域,确定差异轮廓信息;依据差异轮廓信息对当前帧合成图像进行剪裁,获得变化区域图像;对变化区域图像进行编码,生成编码数据。In some embodiments, the difference information includes: at least one difference region. Encoding the current frame synthesized image according to the difference information to generate encoded data includes: determining difference contour information according to the at least one difference region; cropping the current frame synthesized image according to the difference contour information to obtain a changed region image; encoding the changed region image to generate encoded data.
其中,差异区域用于表征第一区域图像的图像特征和第二区域图像的图像特征不同的图像区域,能够准确衡量两帧图像的差异性,方便对当前帧合成图像进行处理。The difference region is used to characterize an image region where the image features of the first region image are different from the image features of the second region image, and can accurately measure the difference between the two frames of images, making it convenient to process the current frame synthesized image.
例如,将至少一个差异区域进行最大范围上的合并,获得差异轮廓信息,明确存在差异的图像边界,从而基于差异轮廓信息对当前帧合成图像进行裁剪,获得仅包括差异信息的、能够体现图像变化信息 的变化区域图像。For example, at least one difference area is merged to the maximum extent to obtain difference contour information, and the image boundary with difference is clarified, so as to crop the current frame synthetic image based on the difference contour information to obtain the image change information that only includes the difference information. The image of the changing area.
通过对变化区域图像进行编码,能够使编码数据体现前后两帧图像的差异性,提升对当前帧合成图像的编码速度。By encoding the changed area image, the encoded data can reflect the difference between the previous and next two frames of image, thereby improving the encoding speed of the current frame synthesized image.
例如,图3示出本申请提供的对辅流图像进行检测的流程示意图。如图3所示,区域检测装置300的输入图像为辅流标注图像F1,该辅流标注图像F1是经过图像合成装置200处理后获得的处理后的图像,能够同时体现辅流图像和标注图像的特征。For example, Fig. 3 shows a schematic diagram of a process for detecting auxiliary stream images provided by the present application. As shown in Fig. 3, the input image of the region detection device 300 is an auxiliary stream annotated image F1, which is a processed image obtained after being processed by the image synthesis device 200, and can simultaneously reflect the features of the auxiliary stream image and the annotated image.
当区域检测装置300获取到辅流标注图像F1后,会对该辅流标注图像F1进行分块(或分区)处理,以获得第一区域图像集合,该第一区域图像集合包括多个第一区域图像。After acquiring the auxiliary stream annotated image F1, the region detection device 300 performs block (or partition) processing on the auxiliary stream annotated image F1 to obtain a first region image set, where the first region image set includes a plurality of first region images.
通过对辅流标注图像F1进行分块(或分区)处理,能够体现辅流标注图像F1的局部信息,以方便后续对不同的局部图像的特征进行对比,以实现对变化区域的检测。By dividing the auxiliary stream annotated image F1 into blocks (or partitions), local information of the auxiliary stream annotated image F1 can be reflected, so as to facilitate subsequent comparison of features of different local images and realize detection of changed areas.
并且,区域检测装置300还会预先存储第二区域图像集合,该第二区域图像集合包括多个第二区域图像,并且,第二区域图像集合是通过对前一帧合成图像进行分块(或分区)处理获得的图像集合,能够体现前一帧合成图像的不同区域内的图像特征。In addition, the region detection device 300 also pre-stores a second region image set, which includes multiple second region images. Moreover, the second region image set is an image set obtained by performing block (or partition) processing on the previous frame synthetic image, which can reflect the image features in different regions of the previous frame synthetic image.
进一步地,通过将第二区域图像集合中的多个第二区域图像,分别与第一区域图像集合中的多个第一区域图像进行分块(或分区)对比,获得差异信息(例如,某个区域中不同的特征信息等)。Furthermore, by comparing the multiple second area images in the second area image set with the multiple first area images in the first area image set in blocks (or partitions), difference information (for example, different feature information in a certain area, etc.) is obtained.
需要说明的是,若发现某个区域内的图像存在差异,则将该区域内的图像块进行缓存,并且记录该图像块所在的区域。缓存图像块的过程可以是多线程同步进行的,也可以是通过对存在差异的图像块进行逐行扫描完成的。It should be noted that if a difference is found in an image in a certain area, the image block in the area is cached and the area where the image block is located is recorded. The process of caching image blocks can be performed synchronously by multiple threads or by scanning the image blocks with differences line by line.
通过将多个存在差异的图像块进行存储,并将多个存在差异的图像块进行整合,从而可以提取存在差异的图像的轮廓(例如,提取图像块的外接矩形轮廓等),然后基于该轮廓,将轮廓内的图像进行裁剪,生成差异信息对应的差异图像。将该差异图像和辅流标注图像F1都输入到编码模块310进行编码,从而可以快速准确的获得编码数据。 By storing a plurality of image blocks with differences and integrating the plurality of image blocks with differences, the contour of the image with differences can be extracted (for example, the circumscribed rectangular contour of the image block is extracted, etc.), and then based on the contour, the image within the contour is cropped to generate a difference image corresponding to the difference information. The difference image and the auxiliary stream annotated image F1 are both input to the encoding module 310 for encoding, so that the encoded data can be obtained quickly and accurately.
若确定第一区域图像集合和第二区域图像集合之间不存在差异,则无需再进行轮廓的提取,直接跳过对该帧辅流图像的处理。If it is determined that there is no difference between the first region image set and the second region image set, there is no need to extract the contour, and the processing of the auxiliary stream image of the frame is directly skipped.
通过将缓存的前一帧合成图像和辅流图像进行帧间对比,并提取发生变化的图像区域对应的轮廓,从而对该轮廓内的变化区域图像进行裁剪,获得差异信息,提升对辅流图像的差异变化的判断准确性。By comparing the previous cached synthetic image with the auxiliary stream image, and extracting the contour corresponding to the changed image area, the image of the changed area within the contour is cropped to obtain difference information, thereby improving the accuracy of judging the difference changes of the auxiliary stream image.
在一些实施方式中,对当前帧合成图像和前一帧合成图像进行检测,确定差异信息之后,所述方法还包括:在确定差异信息表征当前帧合成图像和前一帧合成图像之间无差异的情况下,跳过当前帧合成图像。In some embodiments, after detecting the current frame synthetic image and the previous frame synthetic image and determining the difference information, the method further includes: skipping the current frame synthetic image when it is determined that the difference information indicates that there is no difference between the current frame synthetic image and the previous frame synthetic image.
需要说明的是,由于当前帧合成图像和前一帧合成图像之间不存在差异,表征当前帧合成图像和前一帧合成图像是相同的两帧图像,则无需对当前帧图像进行处理,只需跳过该当前帧合成图像即可,以加快对图像的处理速度。It should be noted that since there is no difference between the current frame composite image and the previous frame composite image, which indicates that the current frame composite image and the previous frame composite image are the same two frame images, there is no need to process the current frame image, and only the current frame composite image needs to be skipped to speed up the image processing.
在一些实施方式中,向对端设备发送编码数据,包括:通过第一通道向对端设备发送编码数据;向对端设备发送编码数据之后,还包括:通过第二通道向对端设备发送与标注信息对应的标注数据。In some implementations, sending the encoded data to the peer device includes: sending the encoded data to the peer device through a first channel; after sending the encoded data to the peer device, further includes: sending labeled data corresponding to the labeled information to the peer device through a second channel.
其中,与标注信息对应的标注数据,可以是对标注信息进行组包,并符合第二通道的传输规则的数据。例如,将标注信息以二进制数据表示,并在该二进制数据前面添加数据包头(如,表征对端设备的网络地址等信息的数据包头),从而获得与标注信息对应的标注数据。The annotation data corresponding to the annotation information may be data obtained by packaging the annotation information and complying with the transmission rules of the second channel. For example, the annotation information is represented by binary data, and a data packet header (e.g., a data packet header representing information such as the network address of the peer device) is added in front of the binary data, thereby obtaining the annotation data corresponding to the annotation information.
通过不同的传输通道(如,第一通道和第二通道等),向对端设备发送编码数据,以及与标注信息对应的标注数据,能够方便对端设备的对不同数据的处理,以使对端设备更快捷的对获得的编码数据进行分析和处理,提升数据的处理效率。Sending encoded data and labeled data corresponding to the labeling information to the peer device through different transmission channels (such as the first channel and the second channel, etc.) can facilitate the peer device's processing of different data, so that the peer device can analyze and process the obtained encoded data more quickly, thereby improving data processing efficiency.
图4示出本申请提供的图像处理方法的流程示意图。该方法可应用于解码装置。如图4所示,本申请实施方式中的图像处理方法包括但不限于以下步骤S401至S404。FIG4 is a flow chart of the image processing method provided by the present application. The method can be applied to a decoding device. As shown in FIG4 , the image processing method in the embodiment of the present application includes but is not limited to the following steps S401 to S404.
步骤S401,获取编码数据。Step S401, obtaining encoded data.
其中,编码数据是对端设备(如编码装置)发送的经过其采用本 申请中任意一种图像处理方法进行编码的数据。The coded data is sent by the peer device (such as a coding device) through the Data encoded by any image processing method in the application.
例如,编码数据是编码装置依据差异信息对当前帧合成图像进行编码获得的数据,差异信息是编码装置对当前帧合成图像和前一帧合成图像进行检测获得的信息,当前帧合成图像是编码装置对辅流图像及其对应的标注信息进行合成的图像。For example, the encoded data is data obtained by the encoding device by encoding the current frame synthetic image based on the difference information, the difference information is information obtained by the encoding device by detecting the current frame synthetic image and the previous frame synthetic image, and the current frame synthetic image is an image synthesized by the encoding device on the auxiliary stream image and its corresponding annotation information.
步骤S402,对编码数据进行解码,获得解码图像。Step S402: decode the encoded data to obtain a decoded image.
其中,解码图像为携带有辅流图像及其对应的标注信息的图像。The decoded image is an image that carries the auxiliary stream image and its corresponding annotation information.
需要说明的是,由于编码装置发送的编码数据是经过其采用本申请中任意一种图像处理方法进行编码的数据,即编码数据已经携带了标注信息和辅流图像,因此,解码装置只需针对该编码数据做对应的解码即可,从而能够保证解码图像包括辅流图像的特征,以及该辅流图像对应的标注信息的特征。解码装置对编码数据进行解码的解码方式与编码数据的编码方式相匹配,以保证获取到准确的解码图像。It should be noted that, since the encoded data sent by the encoding device is data encoded by any image processing method in the present application, that is, the encoded data already carries the annotation information and the auxiliary stream image, the decoding device only needs to perform corresponding decoding on the encoded data, thereby ensuring that the decoded image includes the characteristics of the auxiliary stream image and the characteristics of the annotation information corresponding to the auxiliary stream image. The decoding method used by the decoding device to decode the encoded data matches the encoding method of the encoded data to ensure that an accurate decoded image is obtained.
例如,编码装置可以采用某个特定的压缩技术,依据差异信息对当前帧合成图像进行编码,获得编码数据,则解码装置需要采用相同的压缩技术对编码数据进行解码,以使获取到的解码图像能够同时包括辅流图像的特征,以及该辅流图像对应的标注信息的特征。For example, the encoding device can use a specific compression technology to encode the current frame synthetic image based on the difference information to obtain encoded data, and the decoding device needs to use the same compression technology to decode the encoded data so that the obtained decoded image can simultaneously include the characteristics of the auxiliary stream image and the characteristics of the annotation information corresponding to the auxiliary stream image.
步骤S403,将解码图像和前一帧合成图像进行叠加,生成待显示图像。Step S403: superimpose the decoded image and the previous frame of synthesized image to generate an image to be displayed.
其中,解码图像包括辅流图像和标注信息,该解码图像能够体现标注信息和辅流图像的特征,通过将解码图像和前一帧合成图像进行叠加,从而生成待显示图像,以使待显示图像能够体现标注信息。The decoded image includes an auxiliary stream image and annotation information, and the decoded image can reflect the features of the annotation information and the auxiliary stream image. The decoded image is superimposed with the previous frame composite image to generate an image to be displayed, so that the image to be displayed can reflect the annotation information.
步骤S404,显示待显示图像。Step S404: display the image to be displayed.
在一些实施方式中,在执行步骤S404中的显示待显示图像之前,所述方法还包括:对待显示图像进行渲染,获得渲染后的待显示图像。In some implementations, before performing step S404 of displaying the image to be displayed, the method further includes: rendering the image to be displayed to obtain a rendered image to be displayed.
通过对待显示图像进行渲染,能够直观、实时的体现对待显示图像的表面着色效果,从而显示出待显示图像的纹理特征,以及光源对待显示图像的影响效果,从而使用户也可观看到渲染后的待显示图像,提升用户的观看感受。By rendering the image to be displayed, the surface shading effect of the image to be displayed can be reflected intuitively and in real time, thereby showing the texture characteristics of the image to be displayed and the influence of the light source on the image to be displayed, so that the user can also view the rendered image to be displayed, thereby improving the user's viewing experience.
在本实施方式中,通过获取编码数据,能够明确对编码数据的处 理需求,其中的编码数据是如编码装置发送的经过其采用本申请中任意一种图像处理方法进行编码的数据,便于后续处理;对编码数据进行解码,获得并显示解码图像,该解码图像为携带有辅流图像及其对应的标注信息的图像,以使该解码图像能够体现标注信息和辅流图像的特征;解码图像和前一帧合成图像进行叠加,生成待显示图像,并显示该待显示图像,能够使待显示图像能够体现标注信息。In this embodiment, by obtaining the coded data, it is possible to clearly determine the processing of the coded data. The invention relates to a method for processing a video image, wherein the coded data is data sent by the coding device and encoded by any one of the image processing methods in the present application, so as to facilitate subsequent processing; the coded data is decoded to obtain and display a decoded image, which is an image carrying an auxiliary stream image and its corresponding annotation information, so that the decoded image can reflect the annotation information and the characteristics of the auxiliary stream image; the decoded image is superimposed with the previous frame composite image to generate an image to be displayed, and the image to be displayed is displayed, so that the image to be displayed can reflect the annotation information.
在一些实施方式中,步骤S401中的获取编码数据,包括:获取编码数据,包括:通过第一通道接收编码数据,其中,编码数据为与当前帧合成图像对应的数据,当前帧合成图像为对辅流图像及其对应的标注信息进行合成的图像;对编码数据进行解码,获得解码图像之前,所述方法还包括:通过第二通道接收与标注信息对应的标注数据。In some embodiments, obtaining the encoded data in step S401 includes: obtaining the encoded data, including: receiving the encoded data through a first channel, wherein the encoded data is data corresponding to a synthesized image of a current frame, and the synthesized image of the current frame is an image synthesized of an auxiliary stream image and its corresponding annotation information; decoding the encoded data, and before obtaining the decoded image, the method also includes: receiving annotation data corresponding to the annotation information through a second channel.
其中,标注数据是与标注信息对应的数据,例如,该标注数据可以采用二进制数据表示,用于表征信息的数据。The annotation data is data corresponding to the annotation information. For example, the annotation data can be represented by binary data and is used to represent the information.
通过对第二通道接收到的与标注信息对应的标注数据进行解析,能够明确与辅流图像对应的标注信息的具体含义,以便于后续对待分析数据进行处理,提升数据处理效率;并且,通过对不同的通道中传输的数据进行分别处理,能够针对不同类型的数据进行处理,提升对数据的处理准确性。By parsing the annotation data corresponding to the annotation information received by the second channel, the specific meaning of the annotation information corresponding to the auxiliary stream image can be clarified, so as to facilitate the subsequent processing of the data to be analyzed and improve the data processing efficiency; and by separately processing the data transmitted in different channels, different types of data can be processed to improve the accuracy of data processing.
图5示出本申请提供的图像处理系统的组成方框图。如图5所示,第一终端510与第二终端520通信连接(如,通过互联网或通信网络进行通信等)。Fig. 5 is a block diagram of the image processing system provided by the present application. As shown in Fig. 5, a first terminal 510 is connected to a second terminal 520 for communication (eg, communicating via the Internet or a communication network, etc.).
其中,第一终端510包括:图像合成装置511、区域检测装置512、编码模块513和辅流数据发送模块514。第二终端520包括:接收模块521、解码模块522和图像渲染模块523。各模块的作用可以参照上述实施方式中的描述。The first terminal 510 includes: an image synthesis device 511, a region detection device 512, an encoding module 513 and an auxiliary stream data sending module 514. The second terminal 520 includes: a receiving module 521, a decoding module 522 and an image rendering module 523. The functions of each module can refer to the description in the above embodiment.
其中,图像合成装置511能够同时获取标注信息和辅流图像,并对标注信息进行处理,生成标注图像,然后将该标注图像与辅流图像进行合成,生成当前帧合成图像。以使该当前帧合成图像能够同时体现辅流图像的图像特征,以及标注信息对应的特征。The image synthesis device 511 can simultaneously obtain the annotation information and the auxiliary stream image, process the annotation information, generate the annotation image, and then synthesize the annotation image with the auxiliary stream image to generate the current frame synthesized image, so that the current frame synthesized image can simultaneously reflect the image features of the auxiliary stream image and the features corresponding to the annotation information.
由于第二终端520仅能进行常规的图像的解码,因此,经过其 解码模块522处理的叠加后的辅流图像,也能体现标注信息的特征,但会使最终获得的图像不能准确清晰的表征标注信息的特征。Since the second terminal 520 can only perform conventional image decoding, The superimposed auxiliary stream image processed by the decoding module 522 can also reflect the characteristics of the annotation information, but the finally obtained image cannot accurately and clearly represent the characteristics of the annotation information.
图6示出本申请提供的图像处理系统的组成方框图。如图6所示,第一终端610与第二终端620通信连接(如,通过互联网或通信网络进行通信等)。Fig. 6 shows a block diagram of the image processing system provided by the present application. As shown in Fig. 6, a first terminal 610 is connected in communication with a second terminal 620 (eg, communicating via the Internet or a communication network, etc.).
其中,第一终端610包括:图像合成装置611、区域检测装置612、编码模块613、辅流数据发送模块614和标注信息发送模块615。第二终端620包括:接收模块621、解码模块622、图像渲染模块623和标注信息接收模块624。各模块的作用可以参照上述实施方式中的描述。The first terminal 610 includes: an image synthesis device 611, a region detection device 612, an encoding module 613, an auxiliary stream data sending module 614 and an annotation information sending module 615. The second terminal 620 includes: a receiving module 621, a decoding module 622, an image rendering module 623 and an annotation information receiving module 624. The functions of each module can refer to the description in the above embodiment.
其中,图像合成装置611能够同时获取标注信息和辅流图像,并对标注信息进行处理,生成标注图像,然后将该标注图像与辅流图像进行合成,生成当前帧合成图像。以使该当前帧合成图像能够同时体现辅流图像的图像特征,以及标注信息对应的特征。The image synthesis device 611 can simultaneously obtain the annotation information and the auxiliary stream image, process the annotation information, generate the annotation image, and then synthesize the annotation image with the auxiliary stream image to generate the current frame synthesized image, so that the current frame synthesized image can simultaneously reflect the image features of the auxiliary stream image and the features corresponding to the annotation information.
需要说明的是,标注信息发送模块615也可以获取标注信息,并将该标注信息发送至第二终端620,以方便第二终端620对解码模块622输出的叠加后的辅流图像的分析,从而使输入至图像渲染模块623的图像能够清晰准确的体现标注信息的特征。It should be noted that the annotation information sending module 615 can also obtain annotation information and send the annotation information to the second terminal 620 to facilitate the second terminal 620 to analyze the superimposed auxiliary stream image output by the decoding module 622, so that the image input to the image rendering module 623 can clearly and accurately reflect the characteristics of the annotation information.
在不同类型的终端之间进行通信时,终端都能够支持对标注信息的解码,使用户获得标注信息的特征。When communicating between different types of terminals, the terminals can all support decoding of the annotation information, so that the user can obtain the characteristics of the annotation information.
例如,图7示出本申请提供的图像处理系统的组成方框图。如图7所示,第一终端710分别与第二终端720和第三终端730通信连接(如,通过互联网或通信网络进行通信等)。For example, Fig. 7 shows a block diagram of the image processing system provided by the present application. As shown in Fig. 7, the first terminal 710 is connected to the second terminal 720 and the third terminal 730 for communication (eg, communicating via the Internet or a communication network, etc.).
其中,第一终端710包括:图像合成装置711、区域检测装置712、编码模块713、辅流数据发送模块714和标注信息发送模块715。第二终端720包括:接收模块721、解码模块722、图像渲染模块723和标注信息接收模块724。第三终端730包括:接收模块731、解码模块732和图像渲染模块733。各模块的作用可以参照上述实施方式中的描述。The first terminal 710 includes: an image synthesis device 711, a region detection device 712, an encoding module 713, an auxiliary stream data sending module 714, and a label information sending module 715. The second terminal 720 includes: a receiving module 721, a decoding module 722, an image rendering module 723, and a label information receiving module 724. The third terminal 730 includes: a receiving module 731, a decoding module 732, and an image rendering module 733. The functions of each module can refer to the description in the above embodiment.
编码模块713负责将区域检测装置712输出的图像转换为适合 网络传输的压缩格式(例如,H.264格式等)。辅流数据发送模块714和标注信息发送模块715分别通过有线通信网络,或无线通信网络(如,光纤构成的光网络等)将图像数据传输至第二终端720(或,第三终端730)。The encoding module 713 is responsible for converting the image output by the area detection device 712 into a suitable The auxiliary stream data sending module 714 and the annotation information sending module 715 transmit the image data to the second terminal 720 (or the third terminal 730) through a wired communication network or a wireless communication network (such as an optical network composed of optical fibers).
例如,可采用如下方式实现图像数据的处理。For example, the image data processing may be implemented in the following manner.
图像合成装置711分别获取标注信息和辅流图像,并且图像合成装置711能够同时获取标注信息和辅流图像,并对标注信息进行处理,生成标注图像,然后将该标注图像与辅流图像进行合成,生成当前帧合成图像。以使该当前帧合成图像能够同时体现辅流图像的图像特征,以及标注信息对应的特征。其中,标注信息是标注源产生的一系列标注的点集数据。The image synthesis device 711 obtains the annotation information and the auxiliary stream image respectively, and the image synthesis device 711 can obtain the annotation information and the auxiliary stream image at the same time, and process the annotation information to generate the annotation image, and then synthesize the annotation image with the auxiliary stream image to generate the current frame synthesis image. So that the current frame synthesis image can simultaneously reflect the image features of the auxiliary stream image and the features corresponding to the annotation information. Among them, the annotation information is a series of annotated point set data generated by the annotation source.
区域检测装置712对输入的当前帧合成图像进行不同区域的差异性检测,以获得差异信息,并将该差异信息和辅流图像都输入至编码模块713中进行编码,生成编码数据,并将该编码数据输出至辅流数据发送模块714,以使辅流数据发送模块714将获得的编码数据通过通信网络发送至第二终端720(和/或第三终端730),以使第二终端720和/或第三终端730能够获得辅流图像和标注信息同步的编码数据。The area detection device 712 performs difference detection on different areas of the input current frame composite image to obtain difference information, and inputs both the difference information and the auxiliary stream image into the encoding module 713 for encoding, generates encoded data, and outputs the encoded data to the auxiliary stream data sending module 714, so that the auxiliary stream data sending module 714 sends the obtained encoded data to the second terminal 720 (and/or the third terminal 730) through the communication network, so that the second terminal 720 and/or the third terminal 730 can obtain the encoded data synchronized with the auxiliary stream image and annotation information.
其中,区域检测装置712可以通过对输入的当前帧合成图像进行分块,获得包括多个第一区域图像的第一区域图像集合,然后将多个第一区域图像与其内部缓存的多个第二区域图像进行对比,获得变化区域信息。其中,多个第二区域图像是区域检测装置712对前一帧合成图像进行分块获得的图像。The region detection device 712 may obtain a first region image set including a plurality of first region images by dividing the input current frame synthetic image into blocks, and then compare the plurality of first region images with a plurality of second region images cached therein to obtain the changed region information. The plurality of second region images are images obtained by dividing the previous frame synthetic image into blocks by the region detection device 712.
需要说明的是,第二终端720和/或第三终端730在接收到编码数据后,都会对该编码数据进行解码,但不同的是,第二终端720还可以同时获得原始的标注信息,以方便其对编码数据进行分析,获得准确的辅流图像和标注信息。It should be noted that after receiving the encoded data, the second terminal 720 and/or the third terminal 730 will decode the encoded data, but the difference is that the second terminal 720 can also obtain the original annotation information at the same time to facilitate its analysis of the encoded data and obtain accurate auxiliary stream images and annotation information.
进一步地,还需要采用图像渲染模块733或图像渲染模块723对叠加后的辅流图像进行渲染,以保证用户获得图像更清晰。Furthermore, it is also necessary to use the image rendering module 733 or the image rendering module 723 to render the superimposed auxiliary stream image to ensure that the user obtains a clearer image.
在本实施方式中,通过将标注图像和辅流图像进行叠加合成,能 够保证当前帧合成图像的内容一致性,并且通过对比相邻图像帧之间的差异,以限定标注信息对应的图像叠加的范围,从而提高图像合成的速度。能够满足用户在不同应用场景下的需求,提升产品竞争力。解决了可标注的第一终端710和无法标注的第三终端730之间的辅流内容交互不一致的问题。In this embodiment, the annotated image and the auxiliary stream image are superimposed and synthesized. It can ensure the content consistency of the synthesized image of the current frame, and by comparing the differences between adjacent image frames, it can limit the range of image superposition corresponding to the annotation information, thereby improving the speed of image synthesis. It can meet the needs of users in different application scenarios and improve product competitiveness. It solves the problem of inconsistent interaction of auxiliary stream content between the first terminal 710 that can be annotated and the third terminal 730 that cannot be annotated.
图8示出本申请提供的对辅流图像的展示界面示意图。如图8所示,图8中的(A)表示第一终端710发送的具有标注信息的辅流图像的展示界面,或,第二终端720显示的具有标注信息的辅流图像的展示界面。FIG8 shows a schematic diagram of a display interface for auxiliary stream images provided by the present application. As shown in FIG8 , (A) in FIG8 represents a display interface of an auxiliary stream image with annotated information sent by the first terminal 710, or a display interface of an auxiliary stream image with annotated information displayed by the second terminal 720.
图8中的(B)表示现有技术中的终端展示的仅显示辅流图像(即没有标注信息的辅流图像)的展示界面。(B) in FIG. 8 shows a display interface in the prior art in which only an auxiliary stream image (ie, an auxiliary stream image without annotation information) is displayed by a terminal.
图8中的(C)表示第三终端730展示的辅流图像的展示界面。(C) in FIG. 8 shows a display interface of the auxiliary stream image displayed by the third terminal 730 .
通过对比图8中的三个展示界面,能够明确标注信息的显示状态,方便用户的查看,提升用户的使用体验。By comparing the three display interfaces in FIG8 , the display status of the information can be clearly marked, which facilitates user viewing and improves the user experience.
图9示出本申请提供的编码装置的组成方框图。如图9所示,在一个实施方式中,编码装置900包括但不限于如下模块。Fig. 9 shows a block diagram of the coding device provided by the present application. As shown in Fig. 9, in one embodiment, the coding device 900 includes but is not limited to the following modules.
合成模块901,被配置为对辅流图像及其对应的标注信息进行合成,生成当前帧合成图像;检测模块902,被配置为对当前帧合成图像和前一帧合成图像进行检测,确定差异信息;编码模块903,被配置为依据差异信息对当前帧合成图像进行编码,生成编码数据;发送模块904,被配置为向对端设备发送所述编码数据,以使所述对端设备对所述编码数据进行处理,获得并显示包括所述辅流图像对应的标注信息的解码图像。The synthesis module 901 is configured to synthesize the auxiliary stream image and its corresponding annotation information to generate a current frame synthesized image; the detection module 902 is configured to detect the current frame synthesized image and the previous frame synthesized image to determine the difference information; the encoding module 903 is configured to encode the current frame synthesized image according to the difference information to generate encoded data; the sending module 904 is configured to send the encoded data to the opposite device so that the opposite device processes the encoded data to obtain and display a decoded image including the annotation information corresponding to the auxiliary stream image.
需要说明的是,本实施方式中的编码装置900能够实现本申请实施方式中任一种应用于编码装置的图像处理方法。It should be noted that the encoding device 900 in this embodiment can implement any image processing method applied to the encoding device in the embodiments of the present application.
根据本申请实施方式的编码装置,通过合成模块对辅流图像及其对应的标注信息进行合成,生成当前帧合成图像,能够明确终端对辅流图像进行标注的信息;检测模块对当前帧合成图像和前一帧合成图像进行检测,确定差异信息,使用户可以同步获取连续两帧之间的差异信息,提升终端与用户之间的互动性;编码模块依据差异信息对 辅流图像进行编码,生成编码数据,可加快图像的编码速度,以降低编码的能耗;并且,通过发送模块向对端设备发送编码数据,以使对端设备对编码数据进行处理,获得并显示包括辅流图像对应的标注信息的解码图像,方便对端设备能够查看到带有标注信息的解码图像,使对端设备可以更清晰明确的显示标注信息。According to the encoding device of the embodiment of the present application, the synthesis module synthesizes the auxiliary stream image and its corresponding annotation information to generate the current frame synthetic image, which can clarify the information of the terminal annotating the auxiliary stream image; the detection module detects the current frame synthetic image and the previous frame synthetic image to determine the difference information, so that the user can synchronously obtain the difference information between two consecutive frames, thereby improving the interactivity between the terminal and the user; the encoding module detects the current frame synthetic image and the previous frame synthetic image according to the difference information, and determines the difference information between the two consecutive frames. The auxiliary stream image is encoded to generate encoded data, which can speed up the encoding speed of the image to reduce the energy consumption of encoding; and the encoded data is sent to the opposite device through the sending module, so that the opposite device processes the encoded data, obtains and displays the decoded image including the annotation information corresponding to the auxiliary stream image, so that the opposite device can view the decoded image with the annotation information, and the opposite device can display the annotation information more clearly.
图10示出本申请提供的解码装置的组成方框图。如图10所示,在一个实施方式中,解码装置1000包括但不限于如下模块。Fig. 10 shows a block diagram of a decoding device provided by the present application. As shown in Fig. 10, in one embodiment, the decoding device 1000 includes but is not limited to the following modules.
获取模块1001,被配置为获取编码数据,编码数据为本申请中的编码装置采用的任意一项图像处理方法所发送的数据;解码模块1002,被配置为对编码数据进行解码,获得解码图像,解码图像为携带有辅流图像及其对应的标注信息的图像;生成模块1003,被配置为将所述解码图像和前一帧合成图像进行叠加,生成待显示图像;显示模块1004,被配置为显示待显示图像。The acquisition module 1001 is configured to acquire encoded data, which is data sent by any image processing method adopted by the encoding device in the present application; the decoding module 1002 is configured to decode the encoded data to obtain a decoded image, which is an image carrying an auxiliary stream image and its corresponding annotation information; the generation module 1003 is configured to superimpose the decoded image and the previous frame composite image to generate an image to be displayed; the display module 1004 is configured to display the image to be displayed.
需要说明的是,本实施方式中的解码装置1000能够实现本申请实施方式中任一种应用于解码装置的图像处理方法。It should be noted that the decoding device 1000 in this embodiment can implement any image processing method applied to a decoding device in the embodiments of the present application.
根据本申请实施方式的解码装置,通过使用获取模块获取编码数据,能够明确对编码数据的处理需求,其中的编码数据是编码装置发送的经过其采用本申请中任意一种图像处理方法进行编码的数据,便于后续处理;对编码数据进行解码,获得并显示解码图像,该解码图像为携带有辅流图像及其对应的标注信息的图像,以使该解码图像能够体现标注信息和辅流图像的特征,方便用户的使用。According to the decoding device of the implementation mode of the present application, by using the acquisition module to obtain the encoded data, it is possible to clarify the processing requirements for the encoded data, wherein the encoded data is the data sent by the encoding device and encoded by it using any one of the image processing methods in the present application, which is convenient for subsequent processing; the encoded data is decoded to obtain and display a decoded image, which is an image carrying an auxiliary stream image and its corresponding annotation information, so that the decoded image can reflect the characteristics of the annotation information and the auxiliary stream image, which is convenient for users to use.
图11示出本申请提供的终端的组成方框图。如图11所示,在一个实施方式中,终端1100包括但不限于如下模块:编码装置1101,和/或,解码装置1102。FIG11 is a block diagram showing a terminal provided by the present application. As shown in FIG11 , in one embodiment, the terminal 1100 includes but is not limited to the following modules: an encoding device 1101 and/or a decoding device 1102 .
例如,图11中的(A)表示终端1100仅包括编码装置1101;图11中的(B)表示终端1100仅包括解码装置1102;图11中的(C)表示终端1100包括编码装置1101和解码装置1102。For example, (A) in FIG. 11 indicates that the terminal 1100 includes only the encoding device 1101 ; (B) in FIG. 11 indicates that the terminal 1100 includes only the decoding device 1102 ; and (C) in FIG. 11 indicates that the terminal 1100 includes the encoding device 1101 and the decoding device 1102 .
其中,编码装置1101,被配置为执行本申请实施方式中任一项应用于编码装置的图像处理方法。解码装置1102,被配置为执行本申请实施方式中任一项应用于解码装置的图像处理方法。 The encoding device 1101 is configured to execute any image processing method applied to an encoding device in the embodiments of the present application. The decoding device 1102 is configured to execute any image processing method applied to a decoding device in the embodiments of the present application.
例如,终端1100可以为支持音频/视频会议功能的终端(如,智能手机等),也可以为支持网络授课的平板电脑(或,个人电脑等)。以上对于终端的类别仅是举例说明,可根据实际需要进行具体设定,其他未说明的终端的类别也在本申请的保护范围之内,在此不再赘述。For example, the terminal 1100 may be a terminal supporting audio/video conferencing functions (such as a smart phone, etc.), or a tablet computer supporting online teaching (or a personal computer, etc.). The above terminal categories are only examples, and specific settings can be made according to actual needs. Other unspecified terminal categories are also within the scope of protection of this application and will not be repeated here.
根据本申请实施方式的终端,通过编码装置对辅流图像及其对应的标注信息进行合成,生成当前帧合成图像,能够明确终端对辅流图像进行标注的信息;对当前帧合成图像和前一帧合成图像进行检测,确定差异信息,使用户可以同步获取连续两帧之间的差异信息,提升终端与用户之间的互动性;依据差异信息对辅流图像进行编码,生成编码数据,可加快图像的编码速度,以降低编码的能耗。并通过编码装置获取编码数据及其对应的标注信息,能够明确对编码数据的处理需求,并且,其中的编码数据是编码装置依据差异信息对辅流图像进行编码获得的图像,差异信息是编码装置对当前帧合成图像和前一帧合成图像进行检测获得的信息,能够使用户可以同步获取连续两帧之间的差异信息,提升终端与用户之间的互动性;对编码数据进行解码,获得待分析图像,以加快对待分析图像的处理速度;依据编码数据对应的标注信息对待分析图像进行处理,获得解码图像,以使该解码图像能够体现标注信息和辅流图像的特征,方便用户的使用。According to the terminal of the embodiment of the present application, the auxiliary stream image and its corresponding annotation information are synthesized by the encoding device to generate the current frame synthetic image, which can clarify the information of the terminal on the auxiliary stream image; the current frame synthetic image and the previous frame synthetic image are detected to determine the difference information, so that the user can synchronously obtain the difference information between two consecutive frames, thereby improving the interactivity between the terminal and the user; the auxiliary stream image is encoded according to the difference information to generate encoding data, which can speed up the encoding speed of the image to reduce the energy consumption of encoding. And the encoding data and its corresponding annotation information are obtained by the encoding device, which can clarify the processing requirements of the encoding data, and the encoding data is the image obtained by the encoding device encoding the auxiliary stream image according to the difference information, and the difference information is the information obtained by the encoding device detecting the current frame synthetic image and the previous frame synthetic image, which can enable the user to synchronously obtain the difference information between two consecutive frames, thereby improving the interactivity between the terminal and the user; the encoding data is decoded to obtain the image to be analyzed, thereby speeding up the processing speed of the image to be analyzed; the image to be analyzed is processed according to the annotation information corresponding to the encoding data to obtain a decoded image, so that the decoded image can reflect the characteristics of the annotation information and the auxiliary stream image, which is convenient for users to use.
图12示出本申请提供的图像处理系统的组成方框图。图像处理系统包括通信连接的多个终端;其中,终端能够实现本申请实施方式中的任意一种图像处理方法。Fig. 12 shows a block diagram of the image processing system provided by the present application. The image processing system includes a plurality of terminals connected in communication; wherein the terminals can implement any one of the image processing methods in the embodiments of the present application.
例如,如图12所示,在一个实施方式中,该图像处理系统包括但不限于如下设备:通信连接的至少一个发送终端1201,以及至少一个第一接收终端1202和/或第二接收终端1203。For example, as shown in FIG. 12 , in one embodiment, the image processing system includes but is not limited to the following devices: at least one transmitting terminal 1201 in communication connection, and at least one first receiving terminal 1202 and/or second receiving terminal 1203 .
例如,图12中的(A)表示图像处理系统包括:通信连接的发送终端1201和第一接收终端1202;图12中的(B)表示图像处理系统包括:通信连接的发送终端1201和第二接收终端1203;图12中的(C)表示图像处理系统包括:发送终端1201、以及分别与发送终端1201通信连接的第一接收终端1202和第二接收终端1203。For example, (A) in Figure 12 indicates that the image processing system includes: a sending terminal 1201 and a first receiving terminal 1202 that are communicatively connected; (B) in Figure 12 indicates that the image processing system includes: a sending terminal 1201 and a second receiving terminal 1203 that are communicatively connected; (C) in Figure 12 indicates that the image processing system includes: a sending terminal 1201, and a first receiving terminal 1202 and a second receiving terminal 1203 that are respectively communicatively connected to the sending terminal 1201.
其中,第一发送终端1201,被配置为执行本申请实施方式中任 一种应用于编码装置的图像处理方法。The first sending terminal 1201 is configured to execute any of the embodiments of the present application. An image processing method applied to an encoding device.
第一接收终端1202,被配置为执行本申请实施方式中任一种应用于解码装置的图像处理方法。The first receiving terminal 1202 is configured to execute any one of the image processing methods applied to a decoding device in the embodiments of the present application.
第二接收终端1203,被配置为获取第一终端发送的编码数据,对编码数据进行解码,获得并显示包括辅流图像对应的标注信息的解码图像,其中,编码数据是编码装置依据差异信息对当前帧合成图像进行编码获得的数据,差异信息是编码装置对当前帧合成图像和前一帧合成图像进行检测获得的信息,当前帧合成图像是编码装置对辅流图像及其对应的标注信息进行合成的图像。The second receiving terminal 1203 is configured to obtain the encoded data sent by the first terminal, decode the encoded data, obtain and display a decoded image including the annotation information corresponding to the auxiliary stream image, wherein the encoded data is data obtained by the encoding device by encoding the current frame synthetic image according to the difference information, the difference information is information obtained by the encoding device by detecting the current frame synthetic image and the previous frame synthetic image, and the current frame synthetic image is an image synthesized by the encoding device on the auxiliary stream image and its corresponding annotation information.
根据本申请实施方式的图像处理系统,通过发送终端对辅流图像及其对应的标注信息进行合成,生成当前帧合成图像,能够明确发送终端对辅流图像进行标注的信息;对当前帧合成图像和前一帧合成图像进行检测,确定差异信息,使用户可以同步获取连续两帧之间的差异信息,提升终端与用户之间的互动性;依据差异信息对辅流图像进行编码,生成编码数据,可加快图像的编码速度,以降低编码的能耗。进一步地,通过至少一个发送终端向第一接收终端和/或第二接收终端发送编码数据,能够使不同的接收终端都接收到该编码数据,并对编码数据进行处理,从而获得并显示包括辅流图像对应的标注信息的解码图像,使第一接收终端和/或第二接收终端都能够查看到带有标注信息的解码图像,以便更清晰明确的显示标注信息。According to the image processing system of the implementation mode of the present application, the auxiliary stream image and its corresponding annotation information are synthesized by the sending terminal to generate the current frame synthesized image, which can clearly identify the information annotated by the sending terminal on the auxiliary stream image; the current frame synthesized image and the previous frame synthesized image are detected to determine the difference information, so that the user can synchronously obtain the difference information between two consecutive frames, thereby improving the interactivity between the terminal and the user; the auxiliary stream image is encoded according to the difference information to generate encoded data, which can speed up the encoding speed of the image to reduce the energy consumption of encoding. Furthermore, by sending the encoded data to the first receiving terminal and/or the second receiving terminal through at least one sending terminal, different receiving terminals can receive the encoded data and process the encoded data to obtain and display the decoded image including the annotation information corresponding to the auxiliary stream image, so that the first receiving terminal and/or the second receiving terminal can view the decoded image with the annotation information, so as to display the annotation information more clearly.
需要明确的是,本发明并不局限于上文实施方式中所描述并在图中示出的特定配置和处理。为了描述的方便和简洁,这里省略了对已知方法的详细描述,并且上述描述的系统、模块和单元的具体工作过程,可以参考前述方法实施方式中的对应过程,在此不再赘述。It should be clear that the present invention is not limited to the specific configurations and processes described in the above embodiments and shown in the figures. For the convenience and brevity of description, a detailed description of the known methods is omitted here, and the specific working processes of the systems, modules and units described above can refer to the corresponding processes in the above method embodiments, which will not be repeated here.
图13示出能够实现根据本申请的图像处理方法和装置的计算设备的示例性硬件架构的结构图。FIG. 13 is a block diagram showing an exemplary hardware architecture of a computing device capable of implementing the image processing method and apparatus according to the present application.
如图13所示,计算设备1300包括输入设备1301、输入接口1302、中央处理器1303、存储器1304、输出接口1305、以及输出设备1306。其中,输入接口1302、中央处理器1303、存储器1304、以及输出接口1305通过总线1307相互连接,输入设备1301和输出设 备1306分别通过输入接口1302和输出接口1305与总线1307连接,进而与计算设备1300的其他组件连接。As shown in FIG13 , the computing device 1300 includes an input device 1301, an input interface 1302, a central processing unit 1303, a memory 1304, an output interface 1305, and an output device 1306. The input interface 1302, the central processing unit 1303, the memory 1304, and the output interface 1305 are interconnected via a bus 1307. The device 1306 is connected to the bus 1307 through the input interface 1302 and the output interface 1305 respectively, and further connected to other components of the computing device 1300.
示例性地,输入设备1301接收来自外部的输入信息,并通过输入接口1302将输入信息传送到中央处理器1303;中央处理器1303基于存储器1304中存储的计算机可执行指令对输入信息进行处理以生成输出信息,将输出信息临时或者永久地存储在存储器1304中,然后通过输出接口1305将输出信息传送到输出设备1306;输出设备1306将输出信息输出到计算设备1300的外部供用户使用。Exemplarily, the input device 1301 receives input information from the outside and transmits the input information to the central processing unit 1303 through the input interface 1302; the central processing unit 1303 processes the input information based on the computer executable instructions stored in the memory 1304 to generate output information, temporarily or permanently stores the output information in the memory 1304, and then transmits the output information to the output device 1306 through the output interface 1305; the output device 1306 outputs the output information to the outside of the computing device 1300 for user use.
在一个实施方式中,图13所示的计算设备可以被实现为一种电子设备,该电子设备可以包括:存储器,被配置为存储程序;处理器,被配置为运行存储器中存储的程序,以执行上述实施方式描述的图像处理方法。In one embodiment, the computing device shown in Figure 13 can be implemented as an electronic device, which may include: a memory configured to store a program; a processor configured to run the program stored in the memory to execute the image processing method described in the above embodiment.
在一个实施方式中,图13所示的计算设备可以被实现为一种图像处理系统,该系统可以包括:存储器,被配置为存储程序;处理器,被配置为运行存储器中存储的程序,以执行上述实施方式描述的图像处理方法。In one embodiment, the computing device shown in Figure 13 can be implemented as an image processing system, which may include: a memory configured to store a program; a processor configured to run the program stored in the memory to execute the image processing method described in the above embodiment.
以上所述,仅为本申请的示例性实施方式而已,并非用于限定本申请的保护范围。一般来说,本申请的多种实施方式可以在硬件或专用电路、软件、逻辑或其任何组合中实现。例如,一些方面可以被实现在硬件中,而其它方面可以被实现在可以被控制器、微处理器或其它计算装置执行的固件或软件中,尽管本申请不限于此。The above is only an exemplary embodiment of the present application and is not intended to limit the scope of protection of the present application. In general, the various embodiments of the present application can be implemented in hardware or special circuits, software, logic or any combination thereof. For example, some aspects can be implemented in hardware, while other aspects can be implemented in firmware or software that can be executed by a controller, microprocessor or other computing device, although the present application is not limited thereto.
本申请的实施方式可以通过移动装置的数据处理器执行计算机程序指令来实现,例如在处理器实体中,或者通过硬件,或者通过软件和硬件的组合。计算机程序指令可以是汇编指令、指令集架构(ISA)指令、机器指令、机器相关指令、微代码、固件指令、状态设置数据、或者以一种或多种编程语言的任意组合编写的源代码或目标代码。Embodiments of the present application may be implemented by executing computer program instructions by a data processor of a mobile device, for example in a processor entity, or by hardware, or by a combination of software and hardware. The computer program instructions may be assembly instructions, instruction set architecture (ISA) instructions, machine instructions, machine-dependent instructions, microcode, firmware instructions, state setting data, or source code or object code written in any combination of one or more programming languages.
本申请附图中的任何逻辑流程的框图可以表示程序步骤,或者可以表示相互连接的逻辑电路、模块和功能,或者可以表示程序步骤与逻辑电路、模块和功能的组合。计算机程序可以存储在存储器上。存储器可以具有任何适合于本地技术环境的类型并且可以使用任何 适合的数据存储技术实现,例如但不限于只读存储器(ROM)、随机访问存储器(RAM)、光存储器装置和系统(数码多功能光碟DVD或CD光盘)等。计算机可读介质可以包括非瞬时性存储介质。数据处理器可以是任何适合于本地技术环境的类型,例如但不限于通用计算机、专用计算机、微处理器、数字信号处理器(DSP)、专用集成电路(ASIC)、可编程逻辑器件(FGPA)以及基于多核处理器架构的处理器。The block diagrams of any logic flow in the drawings of this application may represent program steps, or may represent interconnected logic circuits, modules and functions, or may represent a combination of program steps and logic circuits, modules and functions. The computer program may be stored in a memory. The memory may be of any type suitable for the local technical environment and may use any Suitable data storage technology implementations include, but are not limited to, read-only memory (ROM), random access memory (RAM), optical storage devices and systems (digital versatile discs DVD or CD discs), etc. Computer-readable media may include non-transitory storage media. The data processor may be any type suitable for the local technical environment, such as, but not limited to, a general-purpose computer, a special-purpose computer, a microprocessor, a digital signal processor (DSP), an application-specific integrated circuit (ASIC), a programmable logic device (FGPA), and a processor based on a multi-core processor architecture.
通过示范性和非限制性的示例,上文已提供了对本申请的示范实施方式的详细描述。但结合附图和权利要求来考虑,对以上实施方式的多种修改和调整对本领域技术人员来说是显而易见的,但不偏离本发明的范围。因此,本发明的恰当范围将根据权利要求确定。 By way of exemplary and non-limiting examples, a detailed description of exemplary embodiments of the present application has been provided above. However, in conjunction with the drawings and claims, various modifications and adjustments to the above embodiments will be apparent to those skilled in the art without departing from the scope of the present invention. Therefore, the proper scope of the present invention will be determined according to the claims.

Claims (17)

  1. 一种图像处理方法,包括:An image processing method, comprising:
    对辅流图像及其对应的标注信息进行合成,生成当前帧合成图像;Synthesize the auxiliary stream image and its corresponding annotation information to generate a current frame synthesized image;
    对所述当前帧合成图像和前一帧合成图像进行检测,确定差异信息;Detecting the current frame synthesized image and the previous frame synthesized image to determine difference information;
    依据所述差异信息对所述当前帧合成图像进行编码,生成编码数据;Encoding the current frame synthesized image according to the difference information to generate encoded data;
    向对端设备发送所述编码数据,以使所述对端设备对所述编码数据进行处理,获得并显示包括所述辅流图像对应的标注信息的解码图像。The encoded data is sent to a peer device, so that the peer device processes the encoded data, obtains and displays a decoded image including the annotation information corresponding to the auxiliary stream image.
  2. 根据权利要求1所述的方法,其中,所述对所述当前帧合成图像和前一帧合成图像进行检测,确定差异信息,包括:The method according to claim 1, wherein the detecting the current frame synthetic image and the previous frame synthetic image to determine the difference information comprises:
    依据预设尺寸,分别对所述当前帧合成图像和预先存储的所述前一帧合成图像进行分区处理,获得与所述当前帧合成图像对应的第一区域图像集合,以及与所述前一帧合成图像对应的第二区域图像集合;其中,所述第一区域图像集合包括多个第一区域图像,所述第二区域图像集合包括多个第二区域图像,所述第一区域图像集合中的区域数量与所述第二区域图像集合中的区域数量相同;According to a preset size, the current frame synthetic image and the pre-stored previous frame synthetic image are respectively partitioned to obtain a first region image set corresponding to the current frame synthetic image and a second region image set corresponding to the previous frame synthetic image; wherein the first region image set includes a plurality of first region images, the second region image set includes a plurality of second region images, and the number of regions in the first region image set is the same as the number of regions in the second region image set;
    依据所述区域数量,分别对比所述第一区域图像和所述第二区域图像,获得所述差异信息。According to the number of regions, the first region image and the second region image are compared respectively to obtain the difference information.
  3. 根据权利要求2所述的方法,其中,所述差异信息包括:至少一个差异区域,所述差异区域为所述第一区域图像的图像特征和所述第二区域图像的图像特征不同的图像区域;The method according to claim 2, wherein the difference information comprises: at least one difference region, the difference region being an image region where image features of the first region image and image features of the second region image are different;
    所述依据所述差异信息对所述当前帧合成图像进行编码,生成编码数据,包括:The step of encoding the current frame synthesized image according to the difference information to generate encoded data includes:
    依据至少一个所述差异区域,确定差异轮廓信息; Determining difference profile information according to at least one of the difference regions;
    依据所述差异轮廓信息对所述当前帧合成图像进行剪裁,获得变化区域图像;Cutting the current frame synthesized image according to the difference contour information to obtain a changed area image;
    对所述变化区域图像进行编码,生成所述编码数据。The changed region image is encoded to generate the encoded data.
  4. 根据权利要求1所述的方法,其中,所述对所述当前帧合成图像和前一帧合成图像进行检测,确定差异信息之后,还包括:The method according to claim 1, wherein after detecting the current frame synthetic image and the previous frame synthetic image and determining the difference information, the method further comprises:
    在确定所述差异信息表征所述当前帧合成图像和所述前一帧合成图像之间无差异的情况下,跳过所述当前帧合成图像。In a case where it is determined that the difference information represents that there is no difference between the current frame synthesized image and the previous frame synthesized image, the current frame synthesized image is skipped.
  5. 根据权利要求1所述的方法,其中,所述对辅流图像及其对应的标注信息进行合成,生成当前帧合成图像,包括:The method according to claim 1, wherein synthesizing the auxiliary stream image and its corresponding annotation information to generate a current frame synthesized image comprises:
    基于多种帧频,获取所述辅流图像对应的标注信息;Based on multiple frame rates, obtaining annotation information corresponding to the auxiliary stream image;
    依据预设容器和预设图像格式对所述辅流图像对应的标注信息进行处理,生成标注图像;Processing the annotation information corresponding to the auxiliary stream image according to a preset container and a preset image format to generate an annotated image;
    将所述辅流图像和所述标注图像进行图像整合,生成所述当前帧合成图像。The auxiliary stream image and the annotated image are integrated to generate the current frame composite image.
  6. 根据权利要求5所述的方法,其中,所述将所述辅流图像和所述标注图像进行图像整合,生成所述当前帧合成图像,包括:The method according to claim 5, wherein the step of integrating the auxiliary stream image and the annotated image to generate the current frame composite image comprises:
    分别对所述辅流图像和所述标注图像进行图像格式转换,获得转换图像集合;Performing image format conversion on the auxiliary stream image and the annotated image respectively to obtain a converted image set;
    依据预设图像分辨率对所述转换图像集合中的各个图像进行缩放处理,获得缩放图像集合;Scaling each image in the converted image set according to a preset image resolution to obtain a scaled image set;
    依据预设帧频对所述缩放图像集合中的各个图像进行同步,获得处理后的辅流图像和处理后的标注图像;Synchronizing each image in the zoomed image set according to a preset frame rate to obtain a processed auxiliary stream image and a processed annotated image;
    将所述处理后的辅流图像和所述处理后的标注图像进行叠加合成,生成所述当前帧合成图像。The processed auxiliary stream image and the processed annotated image are superimposed and synthesized to generate the current frame synthesized image.
  7. 根据权利要求6所述的方法,其中,所述依据预设帧频对所述缩放图像集合中的各个图像进行同步,获得处理后的辅流图像和处 理后的标注图像,包括:The method according to claim 6, wherein the images in the zoomed image set are synchronized according to a preset frame rate to obtain the processed auxiliary stream image and the processed auxiliary stream image. The processed annotated images include:
    在确定所述缩放图像集合中的图像的实际帧频大于所述预设帧频的情况下,基于采样的方式对所述缩放图像集合中的各个图像进行丢帧处理,获得所述处理后的辅流图像和所述处理后的标注图像;When it is determined that the actual frame rate of the image in the zoomed image set is greater than the preset frame rate, performing frame drop processing on each image in the zoomed image set based on a sampling manner to obtain the processed auxiliary stream image and the processed annotated image;
    在确定所述缩放图像集合中的图像的实际帧频小于所述预设帧频的情况下,采用内部插帧的方式对所述缩放图像集合中的各个图像进行处理,获得所述处理后的辅流图像和所述处理后的标注图像。When it is determined that the actual frame rate of the images in the scaled image set is less than the preset frame rate, each image in the scaled image set is processed by internal interpolation to obtain the processed auxiliary stream image and the processed annotated image.
  8. 根据权利要求6所述的方法,其中,所述将所述处理后的辅流图像和所述处理后的标注图像进行叠加合成,生成所述当前帧合成图像,包括:The method according to claim 6, wherein the step of superimposing and synthesizing the processed auxiliary stream image and the processed annotated image to generate the current frame synthesized image comprises:
    以所述处理后的辅流图像为背景图像,将所述处理后的标注图像中的标注特征叠加至所述处理后的辅流图像中,获得所述当前帧合成图像。The processed auxiliary stream image is used as a background image, and the annotation features in the processed annotation image are superimposed on the processed auxiliary stream image to obtain the current frame composite image.
  9. 根据权利要求6所述的方法,其中,所述将所述处理后的辅流图像和所述处理后的标注图像进行叠加合成,生成所述当前帧合成图像,包括:The method according to claim 6, wherein the step of superimposing and synthesizing the processed auxiliary stream image and the processed annotated image to generate the current frame synthesized image comprises:
    依据预设透明度信息对所述处理后的辅流图像进行处理,获得所述处理后的辅流图像的图像特征,所述处理后的辅流图像的图像特征与所述标注信息相匹配;Processing the processed auxiliary stream image according to preset transparency information to obtain image features of the processed auxiliary stream image, wherein the image features of the processed auxiliary stream image match the annotation information;
    以所述处理后的标注图像为背景图像,将所述处理后的辅流图像的图像特征叠加至所述处理后的标注图像中,获得所述当前帧合成图像。The processed annotated image is used as a background image, and image features of the processed auxiliary stream image are superimposed on the processed annotated image to obtain the current frame composite image.
  10. 一种图像处理方法,包括:An image processing method, comprising:
    获取编码数据,所述编码数据为权利要求1至9中任意一项所述的图像处理方法所发送的数据;Acquire coded data, wherein the coded data is data sent by the image processing method according to any one of claims 1 to 9;
    对所述编码数据进行解码,获得解码图像,所述解码图像为携带有辅流图像及其对应的标注信息的图像; Decoding the encoded data to obtain a decoded image, where the decoded image is an image carrying the auxiliary stream image and its corresponding annotation information;
    将所述解码图像和前一帧合成图像进行叠加,生成待显示图像;Superimposing the decoded image and the previous frame of synthesized image to generate an image to be displayed;
    显示所述待显示图像。The image to be displayed is displayed.
  11. 根据权利要求10所述的方法,其中,所述显示所述待显示图像之前,还包括:The method according to claim 10, wherein before displaying the image to be displayed, the method further comprises:
    对所述待显示图像进行渲染,获得渲染后的待显示图像。The image to be displayed is rendered to obtain a rendered image to be displayed.
  12. 一种编码装置,包括:A coding device, comprising:
    合成模块,被配置为对辅流图像及其对应的标注信息进行合成,生成当前帧合成图像;A synthesis module is configured to synthesize the auxiliary stream image and its corresponding annotation information to generate a current frame synthesized image;
    检测模块,被配置为对所述当前帧合成图像和前一帧合成图像进行检测,确定差异信息;A detection module, configured to detect the current frame synthetic image and the previous frame synthetic image to determine difference information;
    编码模块,被配置为依据所述差异信息对所述当前帧合成图像进行编码,生成编码数据;An encoding module, configured to encode the current frame synthesized image according to the difference information to generate encoded data;
    发送模块,被配置为向对端设备发送所述编码数据,以使所述对端设备对所述编码数据进行处理,获得并显示包括所述辅流图像对应的标注信息的解码图像。The sending module is configured to send the encoded data to a peer device, so that the peer device processes the encoded data, obtains and displays a decoded image including the annotation information corresponding to the auxiliary stream image.
  13. 一种解码装置,包括:A decoding device, comprising:
    获取模块,被配置为获取编码数据,所述编码数据为权利要求1至9中任意一项所述的图像处理方法所发送的数据;An acquisition module, configured to acquire coded data, wherein the coded data is data sent by the image processing method according to any one of claims 1 to 9;
    解码模块,被配置为对所述编码数据进行解码,获得解码图像,所述解码图像为携带有辅流图像及其对应的标注信息的图像;A decoding module, configured to decode the encoded data to obtain a decoded image, wherein the decoded image is an image carrying a secondary stream image and its corresponding annotation information;
    生成模块,被配置为将所述解码图像和前一帧合成图像进行叠加,生成待显示图像;A generating module, configured to superimpose the decoded image and a previous frame of synthesized image to generate an image to be displayed;
    显示模块,被配置为显示所述待显示图像。The display module is configured to display the image to be displayed.
  14. 一种终端,包括:编码装置和/或解码装置;A terminal, comprising: an encoding device and/or a decoding device;
    所述编码装置,被配置为执行如权利要求1至9中任一项所述的图像处理方法; The encoding device is configured to perform the image processing method according to any one of claims 1 to 9;
    所述解码装置,被配置为执行如权利要求10至11中任一项所述的图像处理方法。The decoding device is configured to execute the image processing method according to any one of claims 10 to 11.
  15. 一种图像处理系统,包括:通信连接的多个终端;An image processing system, comprising: a plurality of terminals in communication connection;
    所述终端,被配置为执行如权利要求1至9中任一项,或,如权利要求10至11中任一项所述的图像处理方法。The terminal is configured to execute the image processing method according to any one of claims 1 to 9, or according to any one of claims 10 to 11.
  16. 一种电子设备,包括:An electronic device, comprising:
    一个或多个处理器;one or more processors;
    存储器,其上存储有一个或多个程序,当所述一个或多个程序被所述一个或多个处理器执行,使得所述一个或多个处理器实现如权利要求1至9中任一项,或,如权利要求10至11中任一项所述的图像处理方法。A memory having one or more programs stored thereon, wherein when the one or more programs are executed by the one or more processors, the one or more processors implement the image processing method as claimed in any one of claims 1 to 9, or any one of claims 10 to 11.
  17. 一种可读存储介质,其中,所述可读存储介质存储有计算机程序,所述计算机程序被处理器执行时实现如权利要求1至9中任一项,或,如权利要求10至11中任一项所述的图像处理方法。 A readable storage medium, wherein the readable storage medium stores a computer program, and when the computer program is executed by a processor, the image processing method according to any one of claims 1 to 9 or any one of claims 10 to 11 is implemented.
PCT/CN2023/105927 2022-10-11 2023-07-05 Image processing method and apparatus, and terminal WO2024078064A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN202211239520.9A CN117915022A (en) 2022-10-11 2022-10-11 Image processing method, device and terminal
CN202211239520.9 2022-10-11

Publications (1)

Publication Number Publication Date
WO2024078064A1 true WO2024078064A1 (en) 2024-04-18

Family

ID=90668671

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2023/105927 WO2024078064A1 (en) 2022-10-11 2023-07-05 Image processing method and apparatus, and terminal

Country Status (2)

Country Link
CN (1) CN117915022A (en)
WO (1) WO2024078064A1 (en)

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20070183679A1 (en) * 2004-02-05 2007-08-09 Vodafone K.K. Image processing method, image processing device and mobile communication terminal
JP2012156820A (en) * 2011-01-27 2012-08-16 Nippon Telegr & Teleph Corp <Ntt> Video communication system, and operation method of the same
CN103281539A (en) * 2013-06-07 2013-09-04 华为技术有限公司 Method, device and terminal for image encoding and decoding processing
CN105744281A (en) * 2016-03-28 2016-07-06 飞依诺科技(苏州)有限公司 Continuous image processing method and device
CN106791937A (en) * 2016-12-15 2017-05-31 广东威创视讯科技股份有限公司 The mask method and system of video image
CN114419502A (en) * 2022-01-12 2022-04-29 深圳力维智联技术有限公司 Data analysis method and device and storage medium

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20070183679A1 (en) * 2004-02-05 2007-08-09 Vodafone K.K. Image processing method, image processing device and mobile communication terminal
JP2012156820A (en) * 2011-01-27 2012-08-16 Nippon Telegr & Teleph Corp <Ntt> Video communication system, and operation method of the same
CN103281539A (en) * 2013-06-07 2013-09-04 华为技术有限公司 Method, device and terminal for image encoding and decoding processing
CN105744281A (en) * 2016-03-28 2016-07-06 飞依诺科技(苏州)有限公司 Continuous image processing method and device
CN106791937A (en) * 2016-12-15 2017-05-31 广东威创视讯科技股份有限公司 The mask method and system of video image
CN114419502A (en) * 2022-01-12 2022-04-29 深圳力维智联技术有限公司 Data analysis method and device and storage medium

Also Published As

Publication number Publication date
CN117915022A (en) 2024-04-19

Similar Documents

Publication Publication Date Title
US11200426B2 (en) Video frame extraction method and apparatus, computer-readable medium
US11729465B2 (en) System and method providing object-oriented zoom in multimedia messaging
CN112565627B (en) Multi-channel video centralized display design method based on bitmap superposition
EP3751862A1 (en) Display method and device, television set, and storage medium
US10720091B2 (en) Content mastering with an energy-preserving bloom operator during playback of high dynamic range video
CN108235055B (en) Method and device for realizing transparent video in AR scene
US20120013717A1 (en) Moving picture decoding method, moving picture decoding program, moving picture decoding apparatus, moving picture encoding method, moving picture encoding program, and moving picture encoding apparatus
CN110868625A (en) Video playing method and device, electronic equipment and storage medium
CN102474639A (en) Transforming video data in accordance with human visual system feedback metrics
TWI539790B (en) Apparatus, method and software product for generating and rebuilding a video stream
CN111479154B (en) Equipment and method for realizing sound and picture synchronization and computer readable storage medium
WO2023138226A1 (en) Sending card and receiving card for display system, display control method, and storage medium
US8854435B2 (en) Viewpoint navigation
JP2012522285A (en) System and format for encoding data and 3D rendering
WO2024078064A1 (en) Image processing method and apparatus, and terminal
US20230325987A1 (en) Tone mapping method and apparatus
WO2011134373A1 (en) Method, device and system for synchronous transmmision of multi-channel videos
CN111406404B (en) Compression method, decompression method, system and storage medium for obtaining video file
CN111970564B (en) Optimization method and device for HDR video display processing, storage medium and terminal
CN116962805A (en) Video synthesis method, device, electronic equipment and readable storage medium
CN116962742A (en) Live video image data transmission method, device and live video system
WO2015132957A1 (en) Video device and video processing method
US11706394B1 (en) Method of real-time variable refresh rate video processing for ambient light rendering and system using the same
CN115118922B (en) Method and device for inserting motion picture in real-time video screen combination in cloud conference
US20220247891A1 (en) Processing method and processing device