WO2014000238A1 - 终端和视频图像压缩方法 - Google Patents

终端和视频图像压缩方法 Download PDF

Info

Publication number
WO2014000238A1
WO2014000238A1 PCT/CN2012/077807 CN2012077807W WO2014000238A1 WO 2014000238 A1 WO2014000238 A1 WO 2014000238A1 CN 2012077807 W CN2012077807 W CN 2012077807W WO 2014000238 A1 WO2014000238 A1 WO 2014000238A1
Authority
WO
WIPO (PCT)
Prior art keywords
target area
code stream
terminal
video image
allocated
Prior art date
Application number
PCT/CN2012/077807
Other languages
English (en)
French (fr)
Inventor
张焰焰
陈维山
马晓燕
Original Assignee
宇龙计算机通信科技(深圳)有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 宇龙计算机通信科技(深圳)有限公司 filed Critical 宇龙计算机通信科技(深圳)有限公司
Priority to PCT/CN2012/077807 priority Critical patent/WO2014000238A1/zh
Priority to CN201280073344.4A priority patent/CN104322065B/zh
Publication of WO2014000238A1 publication Critical patent/WO2014000238A1/zh

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/23Processing of content or additional data; Elementary server operations; Server middleware
    • H04N21/234Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs
    • H04N21/2343Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs involving reformatting operations of video signals for distribution or compliance with end-user requests or end-user device requirements
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/102Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
    • H04N19/115Selection of the code volume for a coding unit prior to coding
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/134Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
    • H04N19/167Position within a video image, e.g. region of interest [ROI]
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/169Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
    • H04N19/17Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object
    • H04N19/172Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a picture, frame or field
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/23Processing of content or additional data; Elementary server operations; Server middleware
    • H04N21/234Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs
    • H04N21/2343Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs involving reformatting operations of video signals for distribution or compliance with end-user requests or end-user device requirements
    • H04N21/234345Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs involving reformatting operations of video signals for distribution or compliance with end-user requests or end-user device requirements the reformatting operation being performed only on part of the stream, e.g. a region of the image or a time segment
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/102Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
    • H04N19/124Quantisation
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/134Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
    • H04N19/162User input
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/60Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding
    • H04N19/61Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding in combination with predictive coding

Definitions

  • the present invention relates to the field of video image processing technologies, and in particular, to a terminal and a video image compression method. Background technique
  • video transmission cannot be smoothly viewed due to network bandwidth limitations or other unstable factors in the transmission of video data over the network.
  • the prior art can ensure normal transmission and playback of video data through two schemes.
  • Solution 1 Increase the compression ratio to reduce the code rate of video compression without changing the image resolution, and ensure that the video bit rate can be transmitted within the bandwidth. For example, a video with a high resolution picture quality of 800 X 600 is adjusted to a low-definition video of 800 X 600 after being disturbed by the network.
  • This method improves the compression ratio while guaranteeing the image resolution, that is, sacrificing image quality to reduce the code rate, which makes more image details partially ignored, and the decoder at the video receiving end cannot truly restore the details of the image, resulting in reception.
  • the quality of the video quality viewed at the end is lower overall. In severe cases, a large number of mosaics can occur.
  • Option 2 Change the resolution of the video without changing the image quality. As the resolution decreases, the amount of video data will decrease. At the same compression ratio, the code rate will also decrease to ensure the video.
  • the code rate can be transmitted within the bandwidth allowable range. For example, a video with a high resolution image with a resolution of 800 X 600 is adjusted to a video of the same definition quality of 320 X 240.
  • scenario 1 There is a common disadvantage between scenario 1 and scenario 2:
  • the video picture is made for bandwidth changes.
  • the adjustment is to transform the entire picture to achieve the purpose of reducing the video bit rate.
  • This method treats all elements in the video picture equally, without distinguishing between important and secondary elements in the video.
  • the present invention is based on the above problems, and proposes a terminal that can adopt different compression strategies for different areas of a video image, thereby ensuring transmission bandwidth requirements and distinguishing between important elements and secondary elements in the video picture.
  • a terminal comprising: an identification unit that identifies a target area selected in a currently displayed video image according to the received selection signal; and a compression unit connected to the identification unit And increasing a first code stream to which the target area is allocated, compressing the target area according to the first code stream, and/or reducing a number assigned to a non-target area in the video image
  • the second code stream compresses the non-target area according to the second code stream.
  • the video image compression mode of the terminal includes the following three compression modes: First: After the user selects the target area, the code stream to which the target area is allocated may be increased, and the code stream allocated by the non-target area does not change. That is, the image of the target area becomes clear, and the image resolution of the non-target area is unchanged; Second: When the user selects the target area, not only the code stream allocated to the target area is increased, but also the non-target is reduced.
  • the code stream allocated by the area that is, the image of the target area becomes clear, and the image clarity of the non-target area is reduced;
  • the third type after the user selects the target area, the code stream allocated by the target area is not Change, but reduce the code stream allocated by the non-target area, that is, the definition of the image of the target area is unchanged, and the image clarity of the non-target area is reduced.
  • These methods can change the sharpness of some images to distinguish between important and secondary elements.
  • the method further includes: a calculating unit, connected to the identifying unit, calculating, according to the allocation algorithm, the first code stream to which the target area is allocated and the non-target area to which the target area is allocated a second code stream; an obtaining unit, connected to the calculating unit, acquiring a first quantized value corresponding to the target area according to the first code stream, and acquiring the non-target area according to the second code stream Corresponding second quantized value; the compressing unit is further configured to compress the target area according to the first quantized value and compress the non-target area i according to the second quantized value.
  • the code stream of different regions in the video image is obtained by the distribution algorithm, and the quantized values of different regions in the video image are calculated according to the code stream, and finally different regions in the video image are compressed according to different quantization values.
  • This method can distinguish between important elements (ie, target areas) and secondary elements (excluding other areas of the target area, ie, non-target areas) in the video picture, so that without changing the resolution and reducing the image quality of important elements, Ensure that the total bit stream meets the bandwidth requirements, which does not affect the user's viewing effect, and achieves the purpose of compressing video.
  • the allocation algorithm used by the calculating unit includes: the number of operations when the target area is selected according to the identified area of the target area and the area of the non-target area And a code stream limit value of the current network bandwidth, and calculating a first code stream to which the target area is allocated and a second code stream to which the non-target area is allocated.
  • BR L S L / ( n SH+S l ) BR LIMIT , where BR H is the assigned code stream of the target area, and BRL is the assigned code stream of the non-target area, where n is The number of operations when the target area is selected, S H is the area of the target area identified by the identification unit, and S L is the area of the non-target area identified by the identification unit, and BR LIMIT is the current network.
  • the code stream limit value of the bandwidth is the code stream limit value of the bandwidth.
  • the method further includes: a restoring unit connected to the identifying unit, when receiving the preset operation signal, the code stream distribution ratio of the target area and the non-target area is 1: 1.
  • the resolutions of the target area and the non-target area can be restored to the same state according to the user's operation, so that the user reselects the target area or re-operates the video image.
  • the identifying unit includes a detecting subunit, detecting a displacement change state of the target area when the target area moves, and adjusting the target area and the position according to the displacement change state. Demarcation of non-target areas.
  • the target area in the video image moves, the target area is adjusted to the position of the target area, and other areas different from the target area are automatically divided into non-target areas, and the target area can be dynamically divided.
  • the acquiring unit includes a recording subunit, and the first quantized value is recorded in the first code stream, and the second quantized value is recorded in the second code stream.
  • the terminal further includes a decompression unit that performs inverse quantization calculation on the target region according to the first quantization value when decompressing the video image, and pairs the non-target according to the second quantization value The area is inverse quantized. Decompression is the inverse of compression. By recording the quantized value in the code stream, the inverse quantization operation can be performed.
  • the method further includes: a sending unit, connected to the identifying unit, and sending the identified target area information to the other terminal, where the other terminal simultaneously processes with the terminal The video image; the terminal further includes: a receiving unit, receiving target area information recognized by the other terminal, acquiring the selected target area in the video image according to the target area information, and the compressing unit increases the target a first code stream to which the area is allocated, compressing the target area according to the first code stream, and/or reducing a second code stream to which the non-target area in the video image is allocated, according to the The second code stream compresses the non-target area i.
  • the information of the target area selected by the local terminal can be sent to other terminals, so that other terminals can increase the definition of the selected target area, so that the target area highlighted by the local terminal can also be Highlight on other terminals.
  • the sending unit is further configured to send operation information when the target area is selected to the other terminal; the receiving unit further receives the selection sent by the other terminal Operation information at the target area, and obtaining local information according to the operation information The number of operations when the terminal selects the target area; the calculating unit calculates the location according to the area of the target area, the number of operations when the target area is selected, and the code stream limit value of the current network bandwidth of the other terminal. Decoding a first code stream to which the target area is allocated and a second code stream to which the non-target area is allocated.
  • the local terminal can also send the operation information of the target area to other terminals, so that the operation information of the video image of the local terminal can be reflected to other terminals at the same time, for example, the operation times of the target area, the operation information of the switching target area, and the like. Can be reflected on other terminals.
  • the other terminal operates the video image based on the operation information, highlights the target area to which the sharpness is to be increased, or reselects the target area.
  • a video image compression method including: identifying a target area selected in a video image; increasing a first code stream to which the target area is allocated, according to the first code The stream compresses the target area, and/or reduces a second code stream to which the non-target area in the video image is allocated, and compresses the non-target area i according to the second code stream.
  • the first code stream to which the target area is allocated and the second code stream to which the non-target area is allocated are calculated according to an allocation algorithm; and acquiring according to the first code stream a first quantized value corresponding to the target area, and acquiring a second quantized value corresponding to the non-target area according to the second code stream; compressing the target area according to the first quantized value and according to the The second quantized value compresses the non-target area.
  • the code stream of different regions in the video image is obtained by the distribution algorithm, and the quantized values of different regions in the video image are calculated according to the code stream, and finally different regions in the video image are compressed according to different quantization values.
  • This method can distinguish between important elements and secondary elements in the video picture, so that the total bit stream can satisfy the bandwidth requirement without changing the resolution and without reducing the image quality of the important elements, and the user's viewing effect is not affected. It also achieved the purpose of compressing video.
  • the allocation algorithm includes: a code stream according to the identified area of the target area and the area of the non-target area, the number of operations when the target area is selected, and the current network bandwidth. The limit value is calculated, and the first code stream to which the target area is allocated and the second code stream to which the non-target area is allocated are calculated.
  • the allocation algorithm includes:
  • BR H n XS H / ( n XS H + S L ) x BR limit ,
  • BR L S L / ( n S H + S L ) BR limit , where BR H is the assigned code stream of the target area, and BRL is the allocated code stream of the non-target area, n
  • the S H is the identified area of the target area
  • the identified area of the non-target area the BR limit is a code stream limit value of the current network bandwidth.
  • the method further includes: when receiving the preset operation signal, the code stream distribution ratio of the target area and the non-target area is 1:1.
  • the resolutions of the target area and the non-target area can be restored to the same state according to a user operation, so that the user reselects the target area or re-operates the video image.
  • a displacement change state of the target area is detected, and a division of the target area and the non-target area is adjusted according to the displacement change state.
  • the target area in the video image moves, the target area is adjusted to the position of the target area, and other areas different from the target area are automatically divided into non-target areas.
  • the method further includes: recording the first quantized value in the first code stream, where the second quantized value is recorded in the second code stream;
  • the target region is inverse quantized according to the first quantized value
  • the non-target region is inverse quantized according to the second quantized value.
  • the method further includes the following steps: when the other terminal processes the video image simultaneously with the local terminal, the local terminal sends the identified target area information to the other terminal;
  • the other terminal acquires the selected target area in the video image according to the target area information, increases the first code stream to which the target area is allocated, and compresses the target area according to the first code stream, And/or reducing a second code stream to which the non-target area in the video image is allocated, and compressing the non-target area according to the second code stream.
  • the target selected at the local terminal can be selected.
  • the information of the area is sent to other terminals for other terminals to increase the definition of the selected target area, so that the target area highlighted at the local terminal can also be highlighted on other terminals.
  • the method further includes the following steps: the local terminal sends operation information when the target area is selected to the other terminal, where the other terminal acquires the local terminal according to the operation information. The number of operations when the target area is selected; the other terminal calculates the code according to the area of the target area, the number of operations when the target area is selected, and the code stream limit value of the current network bandwidth of the other terminal. The first code stream to which the target area is allocated and the second code stream to which the non-target area is allocated.
  • the local terminal can also send the operation information of the target area to other terminals, so that the operation information of the video image of the local terminal can be reflected to other terminals at the same time, for example, the operation times of the target area, the operation information of the switching target area, and the like. Can be reflected on other terminals.
  • the other terminal operates the video image based on the operation information, highlights the target area to which the sharpness is to be increased, or reselects the target area.
  • FIG. 1 shows a block diagram of a terminal in accordance with an embodiment of the present invention
  • FIG. 2 is a flow chart showing a video image compression method according to an embodiment of the present invention
  • FIG. 3 is a flowchart showing a video image compression method according to still another embodiment of the present invention
  • FIGS. 4A to 4E are shown A schematic diagram of operating a video image in accordance with an embodiment of the present invention. detailed description
  • FIG. 1 shows a block diagram of a terminal in accordance with an embodiment of the present invention.
  • a terminal 100 may include: an identification unit
  • the compression unit 108 is connected to the identification unit 102, and adds the first code stream to which the target area is allocated, according to the first code stream.
  • the area is compressed, and/or used to reduce the second code stream to which the non-target area in the video image is allocated, and the non-target area is compressed according to the second code stream.
  • the video image compression mode of the terminal 100 includes the following three compression modes: First: After the user selects the target area, the code stream to which the target area is allocated may be increased, and the code stream allocated by the non-target area is unchanged. , that is, the image of the target area becomes clear, and the image resolution of the non-target area is unchanged; Second: When the user selects the target area, not only the code stream allocated to the target area is increased, but also the non-decrease is reduced.
  • the code stream to which the target area is assigned that is, the image of the target area becomes clear, and the image clarity of the non-target area is reduced;
  • These methods can change the sharpness of some images to distinguish between important and secondary elements.
  • the terminal 100 may further include: a calculating unit 104, calculating, according to the allocation algorithm, the first code stream to which the target area is allocated and the second code stream to which the non-target area is allocated; the obtaining unit 106, acquiring and acquiring according to the first code stream a first quantized value corresponding to the target area, and acquiring a second quantized value corresponding to the non-target area according to the second code stream; the compressing unit 108 is further configured to compress the target area according to the first quantized value and according to the second quantized value pair Non-target areas are compressed.
  • the code stream of different regions in the video image is obtained by the distribution algorithm, and the quantized values of different regions in the video image are calculated according to the code stream, and finally different regions in the video image are compressed according to different quantization values.
  • This method is able to distinguish between important and secondary elements in the video, so that the resolution is not changed and the quality of important elements is not reduced.
  • the total bit stream is guaranteed to meet the bandwidth requirement, and the user's viewing effect is not affected, and the purpose of compressing the video is achieved.
  • the allocation algorithm adopted by the calculating unit 104 includes: calculating a target area according to the identified area of the target area and the area of the non-target area, the number of operations when the target area is selected, and the code stream limit value of the current network bandwidth. The first code stream allocated and the second code stream to which the non-target area is allocated.
  • the allocation algorithm adopted by the computing unit 104 includes:
  • BR H n XS H / ( n XS H + S L ) x BR limit ,
  • BR L S L / ( n SH+S l ) BR limit , where BR H is the assigned code stream of the target area, B is the assigned code stream of the non-target area, and n is the selected target area
  • S H is the area of the target area identified by the identification unit 102
  • S L is the area of the non-target area identified by the identification unit 102
  • BR limit is the code stream limit value of the current network bandwidth.
  • BR H is proportional to the number of operations when selecting the target area
  • BR L is inversely proportional to the number of operations when the target area is selected, that is, the more operations are performed on the target area
  • the assigned code stream of the target area is The larger the non-target area, the smaller the code stream is assigned, the clearer the image quality of the target area, the more blurred the image quality of the non-target area, and the more distinct the difference between the target area and the non-target area.
  • the terminal 100 may further include: a detecting unit 1 18, configured to detect a change of a code stream limit value of the current network bandwidth; and the calculating unit 104 is further configured to: when the code stream limit value changes, And re-allocating the first code stream of the target area and the second code stream of the non-target area according to the allocation algorithm and the re-acquired code stream limit value.
  • a detecting unit 1 18, configured to detect a change of a code stream limit value of the current network bandwidth
  • the calculating unit 104 is further configured to: when the code stream limit value changes, And re-allocating the first code stream of the target area and the second code stream of the non-target area according to the allocation algorithm and the re-acquired code stream limit value.
  • the code stream allocated by the target area and the code stream allocated by the non-target area can be recalculated, and the highlighted image is not affected by the bandwidth limitation.
  • the method further includes: a restoration unit 1 10 connected to the identification unit 102, and when the preset operation signal is received, the code stream distribution ratio of the target area and the non-target area is 1:1.
  • the definition of the target area and the non-target area can be restored to the same state according to a user operation, so that the user reselects the target area or re-images the video image. Operation.
  • the identification unit 102 may include a detection sub-unit 1022 that detects a displacement change state of the target area when the target area moves, and adjusts the division of the target area and the non-target area according to the displacement change state.
  • the target area in the video image moves, the target area is adjusted to the position of the target area, and other areas different from the target area are automatically divided into non-target areas, and the target area can be dynamically adjusted.
  • the obtaining unit 106 includes a recording subunit 1062, the first quantized value is recorded in the first code stream, and the second quantized value is recorded in the second code stream; the terminal 100 further includes a decompressing unit 1 12 When the video image is decompressed, the target region is inverse quantized based on the first quantized value, and the non-target region is inverse quantized based on the second quantized value. Decompression is the inverse of compression. By recording the quantized value in the code stream, the inverse quantization operation can be performed.
  • terminal A and terminal B make a video call, and both terminal A and terminal B display real-time images of the two terminals.
  • terminal A user S performs a selection operation on his avatar, and terminal A compresses the avatar video data according to the allocation algorithm, and records the quantized value in the compressed output stream, and then the code stream
  • the signal is transmitted to the terminal B, and the terminal B decodes the code stream to obtain a quantized value used for compression.
  • the video image can be inverse quantized to obtain an image with higher definition.
  • the display effect in terminal B is that the user's avatar is clearer.
  • the terminal 100 may further include: a sending unit 114, connected to the identifying unit 102, and transmitting the identified target area information to other terminals, where the other terminal simultaneously processes the video image with the terminal;
  • the method further includes: receiving unit 116, receiving target area information recognized by other terminals, acquiring the selected target area in the video image according to the target area information, and compressing unit 108 increases the first code stream allocated by the target area in the video image, according to The first code stream compresses the target area, and/or reduces the second code stream to which the non-target area in the video image is allocated, and compresses the non-target area according to the second code stream.
  • the sending unit 114 is further configured to: send the operation information when the target area is selected to the other terminal; the receiving unit 116 further receives the operation information when the selected target area is sent by the other terminal, and acquires the local terminal when the target area is selected according to the operation information. Number of operations; calculation unit
  • the second code assigned to the first code stream and the non-target area allocated by the target area is calculated according to the area of the target area, the number of operations when the target area is selected, and the code stream limit value of the current network bandwidth of other terminals. flow.
  • the local terminal can also send the operation information of the target area to other terminals, so that the operation information of the video image of the local terminal can be reflected to other terminals at the same time, for example, the operation times of the target area, the operation information of the switching target area, and the like. Can be reflected on other terminals.
  • the other terminal operates the video image based on the operation information, highlights the target area to which the sharpness is to be increased, or reselects the target area.
  • terminal A and terminal B make a video call, and both terminal A and terminal B display real-time images of the two terminals.
  • the terminal A user feels that the avatar of the terminal B user is unclear in the image transmitted by the terminal B, and then the user avatar of the terminal B is selected on the terminal A, and the information of the selected target area and the selected number of times are transmitted to the terminal B.
  • the terminal B compresses the video of the collected local terminal according to the allocation algorithm, highlights the avatar of the terminal B user, and transmits the compressed video data to the terminal A, the terminal.
  • A decompresses and displays the video data, and the display effect is that the avatar of the terminal B user is clearer.
  • the video image compression method includes: Step 202: Identify a target area selected in a video image; Step 204, add a first code stream to which the target area is allocated, and compress the target area according to the first code stream. And/or reducing the second code stream to which the non-target area in the video image is allocated, and compressing the non-target area according to the second code stream.
  • the video image compression method includes the following three compression methods: First: After the user selects the target area, the code stream to which the target area is allocated may be increased, and the code stream allocated by the non-target area is unchanged, that is, the target The image of the area becomes clear, and the image clarity of the non-target area is unchanged. Second: When the user selects the target area, it not only increases the code stream allocated by the target area, but also reduces the non-target area. The assigned code stream, that is, the image of the target area Clear, the image clarity of the non-target area is reduced; Third: When the user selects the target area, the code stream assigned to the target area does not change, but the code stream allocated by the non-target area is reduced.
  • the sharpness of the image of the target area is unchanged, and the sharpness of the image of the non-target area is lowered.
  • a compression method opposite to the above compression method. For example, after the user selects the target area, the code stream to which the target area is allocated may be reduced, and the code stream allocated by the non-target area increases or remains unchanged. That is, the image of the target area becomes blurred, and the image of the non-target area becomes clear or remains unchanged.
  • the area when the user wants to increase the sharpness of an area in the video image, the area only needs to be selected in a certain way, thereby increasing the sharpness of the image of the area and highlighting important elements in the video image.
  • the code stream of the non-target area can be reduced, and the clarity of the non-target area can be reduced, thereby distinguishing between important elements and secondary elements in the video image.
  • the method further includes: calculating, according to the allocation algorithm, the first code stream to which the target area is allocated and the second code stream to which the non-target area is allocated; acquiring the first quantized value corresponding to the target area according to the first code stream Obtaining, according to the second code stream, a second quantized value corresponding to the non-target area; compressing the target area according to the first quantized value; and compressing the non-target area i or according to the second quantized value.
  • the code stream of different regions in the video image is obtained by the distribution algorithm, and the quantized values of different regions in the video image are calculated according to the code stream, and finally different regions in the video image are compressed according to different quantization values.
  • This method can distinguish between important elements and secondary elements in the video picture, so that the total bit stream can satisfy the bandwidth requirement without changing the resolution and without reducing the image quality of the important elements, and does not affect the user's viewing effect. It also achieved the purpose of compressing video.
  • the allocation algorithm comprises: calculating the first code assigned to the target area according to the identified area of the target area and the area of the non-target area, the number of operations when the target area is selected, and the code stream limit value of the current network bandwidth.
  • BR H is proportional to the number of operations when selecting the target area
  • BR L is inversely proportional to the number of operations when the target area is selected, that is, the more operations are performed on the target area
  • the assigned code stream of the target area is The larger the non-target area, the smaller the code stream is allocated, the clearer the image quality of the target area, the more blurred the image quality of the non-target area, and the clearer the difference between the target area and the non-target area.
  • the video image compression method may further include: detecting a change of a code stream limit value of the current network bandwidth; and when the code stream limit value changes, according to the allocation algorithm and reacquiring The code stream defines a value, reallocating the first code stream of the target area and the second code stream of the non-target area.
  • the code stream allocated by the target area and the code stream allocated by the non-target area can be recalculated, and the highlighted image is not affected by the bandwidth limitation.
  • the method further includes: when receiving the preset operation signal, the code stream distribution ratio of the target area and the non-target area is 1:1.
  • the sharpness of the target area and the non-target area can be restored to the same state according to user operations, so that the user can reselect the target area or re-operate the video image.
  • the displacement change state of the target area is detected, and the division of the target area and the non-target area is adjusted according to the displacement change state.
  • the target area in the video image is released, the target area is adjusted to the position of the target area, and other areas different from the target area are automatically divided into non-target areas.
  • the method further includes: recording the first quantized value in the first code stream, the second quantized value being recorded in the second code stream; and when decoding the video image, according to the first quantized value pair
  • the target region is subjected to inverse quantization calculation, and the non-target region is inverse quantized based on the second quantized value.
  • the method further includes the following steps: when the other terminal processes the video image simultaneously with the local terminal, the local terminal sends the identified target area information to other terminals; the other terminal acquires the selected target area in the video image according to the target area information. And increasing a first code stream to which the target area is allocated, compressing the target area according to the first code stream, and/or reducing a second code stream to which the non-target area in the video image is allocated, according to the second code stream pair Non-target areas are compressed.
  • the information of the target area selected by the local terminal can be sent to other terminals, so that other terminals can increase the definition of the selected target area, so that the target area highlighted by the local terminal can also be Highlight on other terminals.
  • the method further includes the following steps: the local terminal sends the operation information when the target area is selected to the other terminal, and the other terminal acquires the operation times when the local terminal selects the target area according to the operation information; the other terminal selects the target according to the area of the target area.
  • the number of operations in the area and the code stream limit value of the current network bandwidth of the other terminal, and the first code stream to which the target area is allocated and the second code stream to which the non-target area is allocated are calculated.
  • the local terminal can also send the operation information of the target area to other terminals, so that the operation information of the video image of the local terminal can be reflected to other terminals at the same time, for example, the operation times of the target area, the operation information of the switching target area, and the like. Can be reflected on other terminals.
  • the other terminal operates the video image based on the operation information, highlights the target area to which the sharpness is to be increased, or reselects the target area. Then, the other terminal transmits the video compressed by the compression policy to the local terminal, so that the target area selected by the local terminal user is highlighted.
  • the compression scheme may be changed, the video image is compressed at the local terminal, and then the compressed video image is transmitted to other terminals, or local operation instructions are sent to other terminals, and other terminals compress the video according to the operation instruction.
  • Image the compressed video image is transmitted to the local terminal.
  • Fig. 3 is a flow chart showing a method of adjusting a video image compression according to still another embodiment of the present invention.
  • step 302 the user clicks on the touch screen to highlight the target. Area.
  • a target area e.g., a human face
  • the area of the highlighted area and the weakened area is calculated, and the highlighted area is the selected target area, and the weakened area, that is, the remaining area of the video image from which the target area is removed (each area includes a plurality of macroblocks).
  • the code stream corresponding to the divided area is allocated according to the allocation policy, and the corresponding quantized value is obtained according to the allocated code stream.
  • the corresponding macroblock is encoded according to the corresponding quantized value.
  • step 312 when compressing the video image, it is necessary to adjust the compression policy according to real-time conditions, such as target motion, target switching, bandwidth change, number of operations, long buttons, and the like.
  • the user When it is detected that the target in the video image is moving, then returning to step 302, the user is required to click the screen again to detect and identify the target; when the user switches the target in the video image, the process returns to step 304 and needs to be restarted.
  • the target is detected and identified to re-divide the target area and the non-target area; when the transmission bandwidth changes or the number of operations of the target changes, then return to step 308, and the code stream needs to be re-allocated according to the allocation algorithm and the quantized value is calculated;
  • the long button operation can be set as the restore operation. When the user performs the long button operation, the code stream distribution ratio of the target area and the non-target area is 1:1, and the video image is restored to the initial state, and the definition of all areas is consistent.
  • the user can also set the conditions for adjusting the compression policy according to specific needs.
  • the image area where the object to be highlighted is located is first determined by the touch screen, and then the higher code rate is assigned to the macroblock in the area, and the weakened area is displayed for the weakened area.
  • a macroblock of a region allocates a lower code rate when performing video compression.
  • the bit rate allocation of the highlighted area and the weakened display area can be further adjusted according to the number of clicks of the touch screen, and the average code stream (the number of bits allocated per pixel) is gradually increased as the number of clicks increases, and the sharpness and contrast are also increased. More and more obvious.
  • the highlighted area and the weakened display area will have a more distinct difference in sharpness; when clicked twice, the sharpness of the highlighted area is further improved, and the definition of the weakened display area is lower and lower; if clicked three times, The image of the weakened display area is hardly seen.
  • the highlighted area and the weakened display area can be switched depending on the click object.
  • the scheme can ensure that the video image is compressed, the code rate is within the allowable range of the bandwidth, and the image size, ie, the resolution, can be ensured, and the important elements in the image can be highlighted to weaken the secondary elements, so the image quality of the highlighted image portion is not Will be affected by bandwidth limitations.
  • the target area is identified by clicking on the desired target.
  • the area of the video image that is different from the target area 402 is the non-target area 404.
  • the code stream corresponding to the target area 402 according to the allocation algorithm will increase, and the code stream allocated to the non-target area 404 will decrease.
  • the display effect is that the sharpness of the target area 402 rises, and the sharpness of the non-target area 404 decreases.
  • the code stream of the target area 402 will further increase, and the code stream of the non-target area 404 will be further reduced, and the display effect is clear of the target area 402.
  • the degree is further increased, and the definition of the non-target area 404 is further lowered.
  • the mobile terminal can simultaneously view the other party and the user's own avatar screen.
  • These avatar images generally include a background image, and the background image may be a residence, an office or a classroom, or may include other people around the user. .
  • the bandwidth is limited or the other party cannot see the user's avatar, or the user wants to intentionally block the background image, the user can highlight the avatar with a higher definition by clicking on his avatar, and the person or object in the background image It will be weakened and displayed. If the other party feels that the avatar picture is not clear enough, you can continue to click on the avatar and send it to the other party with higher definition quality.
  • the reselected area becomes the target area 402, and the other areas are automatically divided into the non-target areas 404.
  • this compression method in the video conferencing application, the presenter can be highlighted, so that the other party can see the expression and action of the presenter more clearly.
  • this image compression method can also be used to highlight important elements in a video image and weaken secondary elements. Users no longer have to worry about the picture being watched, which will affect the viewing effect as the network bandwidth changes.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Compression Or Coding Systems Of Tv Signals (AREA)

Abstract

本发明提供了一种终端,包括:识别单元,识别在当前显示的视频图像中选择的目标区域;压缩单元,连接至所述识别单元,增加所述目标区域所分配到的第一码流,根据所述第一码流对所述目标区域进行压缩,和/或所述压缩单元还用于减少所述视频图像中的非目标区域所分配到的第二码流,根据所述第二码流对所述非目标区域进行压缩。本发明还提出了一种视频图像压缩方法。本发明的技术方案针对区域的视频图像的不同区域采取不同的压缩策略,既保证了传输带宽要求,又可以区分视频图像中的重要元素和次要元素。

Description

说 明 书
终端和视频图像压缩方法
技术领域
本发明涉及视频图像处理技术领域, 具体而言, 涉及一种终端和一种 视频图像压缩方法。 背景技术
在流媒体、 可视电话、 视频会议等视频应用中, 在通过网络传输视频 数据的过程中, 由于受到网络带宽的限制或其他不稳定因素的影响, 视频 将无法流畅观看。 现有技术可通过两种方案保证视频数据的正常传输及播 放。
方案 1 : 在不改变图像分辨率的情况下, 提高压缩比以降低视频压缩 的码率, 保证视频码率能够在带宽允许范围内进行传输。 例如, 原有分辨 率为 800 X 600 的高清晰度画质的视频, 受到网络干扰后调整为 800 X 600 的低清晰度画质的视频。
这种方法在保证图像分辨率的同时提高了压缩比, 即牺牲图像质量以 降低码率, 这使得更多的图像细节部分被忽略, 视频接收端的解码器将无 法真实还原图像的细节, 造成接收端观看的视频画质清晰度整体变低。 在 严重的情况下会出现大量的马赛克现象。
方案 2: 在不改变图像画质的情况下, 改变视频的分辨率, 这样随着 分辨率的降低, 视频数据量将减少, 在相同压缩比下, 码率也会随之降 低, 以保证视频码率能够在带宽允许范围内进行传输。 例如原有分辨率为 800 X 600 的高清晰度画质的视频, 受到网络干 4尤后调整为 320 X 240 的相 同清晰度画质的视频。
这种方法虽然可以保证画质的清晰度, 但是可见视频画面整体缩小 了, 势必会影响接收端的整体观看效果。
方案 1和方案 2存在一个共同的缺点: 视频画面针对带宽变化所做的 调整都是对整个画面进行变换, 以达到降低视频码率的目的。 该方法对视 频画面中的所有元素都同等对待, 没有区分视频画面中的重要元素和次要 元素。
因此, 需要一种新的视频压缩方法, 能够针对视频图像的不同区域采 取不同的压缩策略, 既保证了传输带宽要求, 又可以区分视频画面中的重 要元素和次要元素。 发明内容
本发明正是基于上述问题, 提出了一种终端, 能够针对视频图像的不 同区域采取不同的压缩策略, 既保证了传输带宽要求, 又可以区分视频画 面中的重要元素和次要元素。
有鉴于此, 根据本发明的一个方面, 提出了一种终端, 包括: 识别单 元, 根据接收的选择信号, 识别在当前显示的视频图像中选择的目标区 域; 压缩单元, 连接至所述识别单元, 增加所述目标区域所分配到的第一 码流, 根据所述第一码流对所述目标区域进行压缩, 和 /或用于减少所述 视频图像中的非目标区域所分配到的第二码流, 根据所述第二码流对所述 非目标区域进行压缩。
该终端的视频图像压缩方式包括以下三种压缩方式: 第一种: 当用户 选择了目标区域之后, 可以增加该目标区域所分配到的码流, 非目标区域 所分配到的码流不变, 即目标区域的图像变清晰了, 而非目标区域的图像 清晰度不变; 第二种: 当用户选择了目标区域之后, 不仅增加了该目标区 域所分配到的码流, 也减少了非目标区域所分配到的码流, 即目标区域的 图像变清晰了, 而非目标区域的图像清晰度降低了; 第三种: 当用户选择 了目标区域之后, 该目标区域所分配到的码流不变, 但减少了非目标区域 所分配到的码流, 即目标区域的图像的清晰度不变, 而非目标区域的图像 清晰度降低了。 这些方式均可以改变部分图像的清晰度, 来区分重要元素 和次要元素。 当然还可以采用与上述压缩方式相反的压缩方式, 例如: 当 用户选择了目标区域之后, 可以减少该目标区域所分配到的码流, 非目标 区域所分配到的码流增加或保持不变, 即目标区域的图像变模糊了, 而非 目标区域的图像变清晰了或者保持不变等等。
在上述技术方案中, 优选的, 还可以包括: 计算单元, 连接至所述识 别单元, 根据分配算法计算出所述目标区域所分配到的第一码流和所述非 目标区域所分配到的第二码流; 获取单元, 连接至所述计算单元, 根据所 述第一码流获取与所述目标区域对应的第一量化值, 以及根据所述第二码 流获取与所述非目标区域对应的第二量化值; 压缩单元还用于根据所述第 一量化值对所述目标区域进行压缩以及根据所述第二量化值对所述非目标 区 i或进行压缩。
在该技术方案中, 通过分配算法得到视频图像中不同区域的码流, 再 根据码流计算得到视频图像中不同区域的量化值, 最后根据不同的量化值 对视频图像中不同区域进行压缩处理, 这种方法能够区分视频画面中的重 要元素 (即目标区域) 和次要元素 (除去目标区域的其他区域即非目标区 域) , 使得在不改变分辨率和不降低重要元素画质的前提下, 保证总的比 特流满足带宽的要求, 既没有影响用户的观看效果, 又达到了压缩视频的 目的。
在上述技术方案中, 优选的, 所述计算单元所采用的所述分配算法包 括: 根据识别出的所述目标区域的面积和所述非目标区域的面积、 选择所 述目标区域时的操作次数以及当前网络带宽的码流限定值, 计算出所述目 标区域所分配到的第一码流和所述非目标区域所分配到的第二码流。
在上述技术方案中, 优选的, 所述计算单元所采用的所述分配算法包 括: BRH=n SH/ ( n SH +Sl ) x BRLIMIT ,
BRL=SL/ ( n SH+Sl ) BRLIMIT , 其中, BRH为所述目标区域的所分配 到的码流, BRL为所述非目标区域的所分配到的码流, n 为选择所述目标 区域时的操作次数, SH为所述识别单元识别出的所述目标区域的面积, SL 为所述识别单元识别出的所述非目标区域的面积, BRLIMIT为当前网络带宽 的码流限定值。
在上述技术方案中, 优选的, 还可以包括: 还原单元, 连接至所述识 别单元, 在接收到预设操作信号时, 所述目标区域与所述非目标区域的码 流分配比为 1 : 1。 在该技术方案中, 可以根据用户的操作, 将目标区域与非目标区域的 清晰度还原为相同的状态, 以便用户重新选择目标区域或重新对视频图像 进行操作。
在上述技术方案中, 优选的, 所述识别单元包括检测子单元, 在所述 目标区域发生移动时, 检测所述目标区域的位移变化状态, 根据所述位移 变化状态调整所述目标区域和所述非目标区域的划分。
在该技术方案中, 当视频图像中的目标区域发生移动时, 那么目标区 域将调整到目标区域的位置, 异于目标区域的其他区域自动划分为非目标 区域, 即可动态划分出目标区域。
在上述任一技术方案中, 优选的, 所述获取单元包括记录子单元, 将 所述第一量化值记录在所述第一码流, 所述第二量化值记录在所述第二码 流中; 所述终端还包括解压单元, 在对所述视频图像进行解压时, 根据所 述第一量化值对所述目标区域进行反量化计算, 以及根据所述第二量化值 对所述非目标区域进行反量化计算。 解压是压缩的逆过程, 将量化值记录 在码流中, 就可以进行反量化操作。
在上述任一技术方案中, 优选的, 还包括: 发送单元, 连接至所述识 别单元, 将识别出的目标区域信息发送至所述其他终端, 其中, 所述其他 终端与所述终端同时处理所述视频图像; 所述终端还包括: 接收单元, 接 收其他终端识别出的目标区域信息, 根据所述目标区域信息获取所述视频 图像中被选择的目标区域, 所述压缩单元增加所述目标区域所分配到的第 一码流, 根据所述第一码流对所述目标区域进行压缩, 和 /或减少所述视 频图像中的非目标区域所分配到的第二码流, 根据所述第二码流对所述非 目标区 i或进行压缩。
当本地终端与其他终端进行视频通信时, 可将在本地终端选择的目标 区域的信息发送至其他终端, 供其他终端增加选择的目标区域的清晰度, 这样, 在本地终端凸显的目标区域也能够在其他终端上进行凸显。
在上述任一技术方案中, 优选的, 所述发送单元还用于将选择所述目 标区域时的操作信息发送至所述其他终端; 所述接收单元还接收所述其他 终端发送的选择所述目标区域时的操作信息, 根据所述操作信息获取本地 终端选择所述目标区域时的操作次数; 所述计算单元根据所述目标区域的 面积、 选择所述目标区域时的操作次数以及所述其他终端的当前网络带宽 的码流限定值, 计算出所述目标区域所分配到的第一码流和所述非目标区 域所分配到的第二码流。
因此, 本地终端也可以将对目标区域的操作信息发送至其他终端, 使 本地终端对视频图像的操作信息同时能够反映到其他终端上, 例如, 对目 标区域的操作次数、 切换目标区域等操作信息均可以反映到其他终端上。 其他终端根据这些操作信息对视频图像进行操作, 对需增加清晰度的目标 区域进行凸显, 或者重新选择目标区域。
根据本发明的又一方面, 还提出了一种视频图像压缩方法, 包括: 识 别在视频图像中选择的目标区域; 增加所述目标区域所分配到的第一码 流, 根据所述第一码流对所述目标区域进行压缩, 和 /或减少所述视频图 像中的非目标区域所分配到的第二码流, 根据所述第二码流对所述非目标 区 i或进行压缩。
在上述技术方案中, 优选的, 根据分配算法计算出所述目标区域所分 配到的第一码流和所述非目标区域所分配到的第二码流; 根据所述第一码 流获取与所述目标区域对应的第一量化值, 以及根据所述第二码流获取与 所述非目标区域对应的第二量化值; 根据所述第一量化值对所述目标区域 进行压缩以及根据所述第二量化值对所述非目标区域进行压缩。
在该技术方案中, 通过分配算法得到视频图像中不同区域的码流, 再 根据码流计算得到视频图像中不同区域的量化值, 最后根据不同的量化值 对视频图像中不同区域进行压缩处理, 这种方法能够区分视频画面中的重 要元素和次要元素, 使得在不改变分辨率和不降低重要元素画质的前提下 保证总的比特流满足带宽的要求, 既没有影响用户的观看效果, 又达到了 压缩视频的目的。
在上述技术方案中, 优选的, 所述分配算法包括: 根据识别出的所述 目标区域的面积和所述非目标区域的面积、 选择所述目标区域时的操作次 数以及当前网络带宽的码流限定值, 计算出所述目标区域所分配到的第一 码流和所述非目标区域所分配到的第二码流。 在上述技术方案中, 优选的, 所述分配算法包括:
BRH=n X SH/ ( n X SH + SL ) x BRlimit,
BRL=SL/ ( n SH + SL ) BRlimit, 其中, BRH为所述目标区域的所分 配到的码流, BRL为所述非目标区域的所分配到的码流, n 为选择所述目 标区域时的操作次数, SH为识别出的所述目标区域的面积, 为识别出 的所述非目标区域的面积, BRlimit为当前网络带宽的码流限定值。
在上述技术方案中, 优选的, 还包括: 在接收到预设操作信号时, 所 述目标区域与所述非目标区域的码流分配比为 1 : 1。
在该技术方案中, 在该技术方案中, 可以根据用户操作将目标区域与 非目标区域的清晰度还原为相同的状态, 以便用户重新选择目标区域或重 新对视频图像进行操作。
在上述技术方案中, 优选的, 在所述目标区域发生移动时, 检测所述 目标区域的位移变化状态, 根据所述位移变化状态调整所述目标区域和所 述非目标区域的划分。
在该技术方案中, 当视频图像中的目标区域发生移动时, 那么目标区 域将调整到目标区域的位置, 异于目标区域的其他区域自动划分为非目标 区域。
在上述任一技术方案中, 优选的, 还包括: 将所述第一量化值记录在 所述第一码流, 所述第二量化值记录在所述第二码流中; 在对所述视频图 像进行解码时, 根据所述第一量化值对所述目标区域进行反量化计算, 以 及根据所述第二量化值对所述非目标区域进行反量化计算。
在上述任一技术方案中, 优选的, 还可以包括以下步骤: 在其他终端 与本地终端同时处理所述视频图像时, 所述本地终端将识别出的目标区域 信息发送至所述其他终端; 所述其他终端根据所述目标区域信息获取所述 视频图像中被选择的目标区域, 增加所述目标区域所分配到的第一码流, 根据所述第一码流对所述目标区域进行压缩, 和 /或减少所述视频图像中 的非目标区域所分配到的第二码流, 根据所述第二码流对所述非目标区域 进行压缩。
当本地终端与其他终端进行视频通信时, 可将在本地终端选择的目标 区域的信息发送至其他终端, 供其他终端增加选择的目标区域的清晰度, 这样, 在本地终端凸显的目标区域也能够在其他终端上进行凸显。
在上述任一技术方案中, 优选的, 还可以包括以下步骤: 所述本地终 端将选择所述目标区域时的操作信息发送至所述其他终端, 所述其他终端 根据所述操作信息获取本地终端选择所述目标区域时的操作次数; 所述其 他终端根据所述目标区域的面积、 选择所述目标区域时的操作次数以及所 述其他终端的当前网络带宽的码流限定值, 计算出所述目标区域所分配到 的第一码流和所述非目标区域所分配到的第二码流。
因此, 本地终端也可以将对目标区域的操作信息发送至其他终端, 使 本地终端对视频图像的操作信息同时能够反映到其他终端上, 例如, 对目 标区域的操作次数、 切换目标区域等操作信息均可以反映到其他终端上。 其他终端根据这些操作信息对视频图像进行操作, 对需增加清晰度的目标 区域进行凸显, 或者重新选择目标区域。
通过以上技术方案, 针对视频图像的不同区域采取不同的压缩策略, 既保证了传输带宽要求, 又可以区分视频画面中的重要元素和次要元素。 附图说明
图 1示出了根据本发明的实施例的终端的框图;
图 2示出了根据本发明的一个实施例的视频图像压缩方法的流程图; 图 3示出了根据本发明的又一实施例的视频图像压缩方法的流程图; 图 4A 至图 4E 示出了根据本发明的实施例的操作视频图像的示意 图。 具体实施方式
为了能够更清楚地理解本发明的上述目的、 特征和优点, 下面结合附 图和具体实施方式对本发明进行进一步的详细描述。 需要说明的是, 在不 沖突的情况下, 本申请的实施例及实施例中的特征可以相互组合。
在下面的描述中阐述了很多具体细节以便于充分理解本发明, 但是, 本发明还可以采用其他不同于在此描述的其他方式来实施, 因此, 本发明 的保护范围并不受下面公开的具体实施例的限制。
图 1示出了根据本发明的实施例的终端的框图。
如图 1 所示, 根据本发明的实施例的终端 100 可以包括: 识别单元
102 , 根据接收的选择信号, 识别在当前显示的视频图像中选择的目标区 域; 压缩单元 108 , 连接至识别单元 102 , 增加目标区域所分配到的第一 码流, 根据第一码流对目标区域进行压缩, 和 /或用于减少视频图像中的 非目标区域所分配到的第二码流, 根据第二码流对非目标区域进行压缩。
该终端 100的视频图像压缩方式包括以下三种压缩方式: 第一种: 当 用户选择了目标区域之后, 可以增加该目标区域所分配到的码流, 非目标 区域所分配到的码流不变, 即目标区域的图像变清晰了, 而非目标区域的 图像清晰度不变; 第二种: 当用户选择了目标区域之后, 不仅增加了该目 标区域所分配到的码流, 也减少了非目标区域所分配到的码流, 即目标区 域的图像变清晰了, 而非目标区域的图像清晰度降低了; 第三种: 当用户 选择了目标区域之后, 该目标区域所分配到的码流不变, 但减少了非目标 区域所分配到的码流, 即目标区域的图像的清晰度不变, 而非目标区域的 图像清晰度降低了。 这些方式均可以改变部分图像的清晰度, 来区分重要 元素和次要元素。 当然还可以采用与上述压缩方式相反的压缩方式, 例 如: 当用户选择了目标区域之后, 可以减少该目标区域所分配到的码流, 非目标区域所分配到的码流增加或保持不变, 即目标区域的图像变模糊 了, 而非目标区域的图像变清晰了或者保持不变等等。
该终端 100还可以包括: 计算单元 104 , 根据分配算法计算出目标区 域所分配到的第一码流和非目标区域所分配到的第二码流; 获取单元 106 , 根据第一码流获取与目标区域对应的第一量化值, 以及根据第二码 流获取与非目标区域对应的第二量化值; 压缩单元 108还用于根据第一量 化值对目标区域进行压缩以及根据第二量化值对非目标区域进行压缩。
在该技术方案中, 通过分配算法得到视频图像中不同区域的码流, 再 根据码流计算得到视频图像中不同区域的量化值, 最后根据不同的量化值 对视频图像中不同区域进行压缩处理, 这种方法能够区分视频画面中的重 要元素和次要元素, 使得在不改变分辨率和不降低重要元素画质的前提 下, 保证总的比特流满足带宽的要求, 既不影响用户的观看效果, 又达到 了压缩视频的目的。
优选的, 该计算单元 104所采用的分配算法包括: 根据识别出的目标 区域的面积和非目标区域的面积、 选择目标区域时的操作次数以及当前网 络带宽的码流限定值, 计算出目标区域所分配到的第一码流和非目标区域 所分配到的第二码流。
在上述技术方案中, 计算单元 104所采用的分配算法包括:
BRH=n X SH/ ( n X SH+SL ) x BRlimit,
BRL=SL/ ( n SH+Sl ) BRlimit, 其中, BRH为目标区域的所分配到的 码流, B 为非目标区域的所分配到的码流, n 为选择目标区域时的操作 次数, SH为识别单元 102 识别出的目标区域的面积, SL为识别单元 102 识别出的非目标区域的面积, BRlimit为当前网络带宽的码流限定值。
可见, BRH与选择目标区域时的操作次数成正比, BRL与选择目标区 域时的操作次数成反比, 也就是说, 对目标区域操作次数越多, 目标区域 的所分配到的码流就越大, 非目标区域的所分配到的码流就越小, 目标区 域的画质就越清晰, 非目标区域的画质就越模糊, 目标区域与非目标区域 的清晰度差异也就越明显。
优选的, 该终端 100还可以包括: 检测单元 1 18 , 用于检测所述当前 网络带宽的码流限定值的变化情况; 所述计算单元 104还用于在所述码流 限定值变化时, 根据所述分配算法和重新获取的码流限定值, 重新分配所 述目标区域的第一码流和所述非目标区域的第二码流。
这样, 当带宽码流限定值受网络环境影响发生变化时, 可重新计算目 标区域所分配的码流和非目标区域所分配的码流, 凸显的图像不会因为带 宽的限制而受影响。
在上述技术方案中, 优选的, 还可以包括: 还原单元 1 10 , 连接至识 别单元 102 , 在接收到预设操作信号时, 目标区域与非目标区域的码流分 配比为 1 : 1。
在该技术方案中, 可以根据用户操作将目标区域与非目标区域的清晰 度还原为相同的状态, 以便用户重新选择目标区域或重新对视频图像进行 操作。
在上述任一技术方案中, 识别单元 102 可包括检测子单元 1022 , 在 目标区域发生移动时, 检测目标区域的位移变化状态, 根据位移变化状态 调整目标区域和非目标区域的划分。
在该技术方案中, 当视频图像中的目标区域发生移动时, 那么目标区 域将调整到目标区域的位置, 异于目标区域的其他区域自动划分为非目标 区域, 即可动态调整目标区域。
在上述任一技术方案中, 获取单元 106 包括记录子单元 1062 , 将第 一量化值记录在第一码流, 第二量化值记录在第二码流中; 终端 100还包 括解压单元 1 12 , 在对视频图像进行解压时, 根据第一量化值对目标区域 进行反量化计算, 以及根据第二量化值对非目标区域进行反量化计算。 解 压是压缩的逆过程, 将量化值记录在码流中, 就可以进行反量化操作。
例如, 终端 A和终端 B进行视频通话, 在终端 A和终端 B中都显示 了两个终端的实时图像。 在终端 A 中, 用户 S 对自己的头像进行了选择 操作, 那么终端 A根据分配算法对该头像视频数据进行了压缩处理, 并将 量化值记录在压缩输出的码流中, 然后将该码流传输至终端 B , 终端 B对 该码流进行解码, 获取压缩时采用的量化值, 根据该量化值便可以对视频 图像进行反量化计算, 得到清晰度更高的图像。 在终端 B中的显示效果就 是用户 S的头像更清晰了。
在一种实施方式中, 该终端 100还可以包括: 发送单元 114, 连接至 识别单元 102 , 将识别出的目标区域信息发送至其他终端, 其中, 其他终 端与终端同时处理上述视频图像; 终端 100还包括: 接收单元 116 , 接收 其他终端识别出的目标区域信息, 根据目标区域信息获取视频图像中被选 择的目标区域, 压缩单元 108增加视频图像中目标区域所分配到的第一码 流, 根据第一码流对目标区域进行压缩, 和 /或减少视频图像中的非目标 区域所分配到的第二码流, 根据第二码流对非目标区域进行压缩。
当本地终端与其他终端进行视频通信时, 可将在本地终端选择的目标 区域的信息发送至其他终端, 供其他终端增加选择的目标区域的清晰度, 这样, 在本地终端凸显的目标区域也能够在其他终端上进行凸显。 优选的, 发送单元 114还用于将选择目标区域时的操作信息发送至其 他终端; 接收单元 116 还接收其他终端发送的选择目标区域时的操作信 息, 根据操作信息获取本地终端选择目标区域时的操作次数; 计算单元
104 根据目标区域的面积、 选择目标区域时的操作次数以及其他终端的当 前网络带宽的码流限定值, 计算出目标区域所分配到的第一码流和非目标 区域所分配到的第二码流。
因此, 本地终端也可以将对目标区域的操作信息发送至其他终端, 使 本地终端对视频图像的操作信息同时能够反映到其他终端上, 例如, 对目 标区域的操作次数、 切换目标区域等操作信息均可以反映到其他终端上。 其他终端根据这些操作信息对视频图像进行操作, 对需增加清晰度的目标 区域进行凸显, 或者重新选择目标区域。
继续以上一示例进行说明, 终端 A和终端 B 进行视频通话, 在终端 A和终端 B 中都显示了两个终端的实时图像。 终端 A用户觉得终端 B传 过来的图像中终端 B用户的头像不清晰, 于是在终端 A上对终端 B的用 户头像进行了选择操作, 将选择的目标区域的信息以及选择次数传输至终 端 B。 终端 B接收到选择的目标区域的信息以及选择次数之后, 根据分配 算法对采集的本地终端的视频进行压缩处理, 凸显终端 B用户的头像, 并 将压缩处理后的视频数据传输至终端 A, 终端 A对该视频数据进行解压并 显示, 显示效果就是终端 B用户的头像更清晰了。
图 2示出了根据本发明的一个实施例的视频图像压缩方法的流程图。 如图 2所示, 视频图像压缩方法包括: 步骤 202, 识别在视频图像中 选择的目标区域; 步骤 204, 增加目标区域所分配到的第一码流, 根据第 一码流对目标区域进行压缩, 和 /或减少视频图像中的非目标区域所分配 到的第二码流, 根据第二码流对非目标区域进行压缩。
该视频图像压缩方法包括以下三种压缩方式: 第一种: 当用户选择了 目标区域之后, 可以增加该目标区域所分配到的码流, 非目标区域所分配 到的码流不变, 即目标区域的图像变清晰了, 而非目标区域的图像清晰度 不变; 第二种: 当用户选择了目标区域之后, 不仅增加了该目标区域所分 配到的码流, 也减少了非目标区域所分配到的码流, 即目标区域的图像变 清晰了, 而非目标区域的图像清晰度降低了; 第三种: 当用户选择了目标 区域之后, 该目标区域所分配到的码流不变, 但减少了非目标区域所分配 到的码流, 即目标区域的图像的清晰度不变, 而非目标区域的图像清晰度 降低了。 当然还可以采用与上述压缩方式相反的压缩方式, 例如: 当用户 选择了目标区域之后, 可以减少该目标区域所分配到的码流, 非目标区域 所分配到的码流增加或保持不变, 即目标区域的图像变模糊了, 而非目标 区域的图像变清晰了或者保持不变等等。 这些方式均可以改变部分图像的 清晰度, 来区分重要元素和次要元素。
在采用第二种方式时, 当用户希望增加视频图像中某个区域的清晰度 时, 只需要按照某种方式选择该区域, 便可以增加该区域图像的清晰度, 凸显视频图像中的重要元素。 在凸显用户选择的目标区域时, 考虑到网络 带宽等因素的限制, 可以降低非目标区域的码流, 降低非目标区域的清晰 度, 从而区别视频图像中的重要元素和次要元素。
优选的, 还可以包括: 根据分配算法计算出目标区域所分配到的第一 码流和非目标区域所分配到的第二码流; 根据第一码流获取与目标区域对 应的第一量化值, 以及根据第二码流获取与非目标区域对应的第二量化 值; 根据第一量化值对目标区域进行压缩以及根据第二量化值对非目标区 i或进行压缩。
在该技术方案中, 通过分配算法得到视频图像中不同区域的码流, 再 根据码流计算得到视频图像中不同区域的量化值, 最后根据不同的量化值 对视频图像中不同区域进行压缩处理, 这种方法能够区分视频画面中的重 要元素和次要元素, 使得在不改变分辨率和不降低重要元素画质的前提下 保证总的比特流满足带宽的要求, 既不影响用户的观看效果, 又达到了压 缩视频的目的。
优选的, 分配算法包括: 根据识别出的目标区域的面积和非目标区域 的面积、 选择目标区域时的操作次数以及当前网络带宽的码流限定值, 计 算出目标区域所分配到的第一码流和非目标区域所分配到的第二码流。
其中, 分配算法可以包括: BRH=n SH/ ( n SH + SL ) BRlimit, BRL=SL/ ( n SH + SL ) BRlimit, 其中, BRH为目标区域的所分配到 的码流, BRL为非目标区域的所分配到的码流, n 为选择目标区域时的操 作次数, SH为识别出的目标区域的面积, 为识别出的非目标区域的面 积, BRLIMIT为当前网络带宽的码流限定值。
可见, BRH与选择目标区域时的操作次数成正比, BRL与选择目标区 域时的操作次数成反比, 也就是说, 对目标区域操作次数越多, 目标区域 的所分配到的码流就越大, 非目标区域的所分配到的码流就越小, 目标区 域的画质就越清晰, 非目标区域的画质就越模糊, 目标区域与非目标区域 的清晰图差异也就越明显。
在一种具体实施方式中, 该视频图像压缩方法还可以包括: 检测所述 当前网络带宽的码流限定值的变化情况; 在所述码流限定值变化时, 根据 所述分配算法和重新获取的码流限定值, 重新分配所述目标区域的第一码 流和所述非目标区域的第二码流。
这样, 当带宽码流限定值受网络环境影响发生变化时, 可重新计算目 标区域所分配的码流和非目标区域所分配的码流, 凸显的图像不会因为带 宽的限制而受影响。
在上述技术方案中, 优选的, 还可以包括: 在接收到预设操作信号 时, 目标区域与非目标区域的码流分配比为 1 : 1。
在该技术方案中, 可以根据用户操作将目标区域与非目标区域的清晰 度还原为相同的状态, 以便用户重新选择目标区域或重新对视频图像进行 操作。
在上述技术方案中, 优选的, 在目标区域发生移动时, 检测目标区域 的位移变化状态, 根据位移变化状态调整目标区域和非目标区域的划分。
在该技术方案中, 当视频图像中的目标区域放生发生移动时, 那么目 标区域将调整到目标区域的位置, 异于目标区域的其他区域自动划分为非 目标区域。
在上述任一技术方案中, 还可以包括: 将第一量化值记录在第一码 流, 第二量化值记录在第二码流中; 在对视频图像进行解码时, 根据第一 量化值对目标区域进行反量化计算, 以及根据第二量化值对非目标区域进 行反量化计算。 这样, 当其他终端接收到该视频并进行解码时, 就可以根据这些量化 值进行解压, 真实还原视频图像, 区分凸显图像和弱化图像。
优选的, 还可以包括以下步骤: 在其他终端与本地终端同时处理视频 图像时, 本地终端将识别出的目标区域信息发送至其他终端; 其他终端根 据目标区域信息获取视频图像中被选择的目标区域, 增加目标区域所分配 到的第一码流, 根据第一码流对目标区域进行压缩, 和 /或减少视频图像 中的非目标区域所分配到的第二码流, 根据第二码流对非目标区域进行压 缩。
当本地终端与其他终端进行视频通信时, 可将在本地终端选择的目标 区域的信息发送至其他终端, 供其他终端增加选择的目标区域的清晰度, 这样, 在本地终端凸显的目标区域也能够在其他终端上进行凸显。
优选的, 还可以包括以下步骤: 本地终端将选择目标区域时的操作信 息发送至其他终端, 其他终端根据操作信息获取本地终端选择目标区域时 的操作次数; 其他终端根据目标区域的面积、 选择目标区域时的操作次数 以及其他终端的当前网络带宽的码流限定值, 计算出目标区域所分配到的 第一码流和非目标区域所分配到的第二码流。
因此, 本地终端也可以将对目标区域的操作信息发送至其他终端, 使 本地终端对视频图像的操作信息同时能够反映到其他终端上, 例如, 对目 标区域的操作次数、 切换目标区域等操作信息均可以反映到其他终端上。 其他终端根据这些操作信息对视频图像进行操作, 对需增加清晰度的目标 区域进行凸显, 或者重新选择目标区域。 然后其他终端将按照压缩策略压 缩处理后的视频传输至本地终端, 使本地终端用户选择的目标区域被凸 显。 因此, 根据不同的应用场景, 可以改变压缩方案, 在本地终端压缩视 频图像, 然后经压缩后的视频图像传送至其他终端, 或将本地的操作指令 给其他终端, 其他终端根据该操作指令压缩视频图像, 将压缩后的视频图 像传送至本地终端。
图 3示出了根据本发明的又一实施例的调整视频图像压缩方法的流程 图。
如图 3 所示, 在步骤 302 , 用户在触摸屏上点击需要进行凸显的目标 区域。 在步骤 304 , 检测并识别用户选择的目标区域 (例如人脸) 。 在步 骤 306 , 计算凸显区域和弱化区域的面积, 凸显区域即选择的目标区域, 弱化区域即视频图像中除去该目标区域的剩余区域 (每个区域包括多个宏 块) 。
在步骤 308按照分配策略分配划分的区域所对应的码流, 并根据分配 的码流获取相应的量化值。 在步骤 310 , 根据相应的量化值对相应的宏块 进行编码。
在步骤 312 , 在对视频图像进行压缩时, 需要根据实时条件调整压缩 策略, 实时条件例如目标运动、 目标切换、 带宽变化、 操作次数、 长按键 等。
当检测到在视频图像中的目标发生运动时, 则回到步骤 302 , 需要用 户重新点击屏幕选择检测和识别目标; 当用户对视频图像中的目标进行切 换时, 则回到步骤 304 , 需要重新对目标进行检测和识别以重新划分目标 区域与非目标区域; 当传输带宽发生变化或对目标的操作次数发生变化 时, 则回到步骤 308 , 需要根据分配算法重新分配码流并计算量化值; 可 以设置长按键操作为还原操作, 当用户进行长按键操作时, 目标区域与非 目标区域的码流分配比为 1 : 1 , 及视频图像还原为初始状态, 所有区域清 晰度一致。
当然, 用户也可以根据具体需要设置调整压缩策略的条件。
因此, 根据上述流程, 首先由触屏点击确定需要凸显的对象所在的图 像区域, 然后对该区域内的宏块进行视频压缩时分配更高的码率, 其它区 域则为弱化区域, 针对弱化显示区域的宏块进行视频压缩时分配较低的码 率。 凸显区域和弱化显示区域的码率分配, 根据触屏的点击次数可以进一 步调整, 平均码流(每个像素分配的比特数) 分配比, 随点击次数的增加 而逐渐增加, 清晰度和对比度也越来越明显。 例如, 点击一次, 凸显区域 和弱化显示区域将有一个较明显的清晰度的差别; 点击两次, 凸显区域的 清晰度进一步提高, 弱化显示区域的清晰度则越来越低; 若点击三次, 则 几乎看不到弱化显示区域的图像。
另外, 凸显区域和弱化显示区域可根据点击对象的不同进行切换。 该 方案能保证视频图像压缩后, 码率在带宽允许范围内, 并且能保证图像大 小即分辨率不变, 而且可以凸显图像中的重要元素而弱化次要元素, 因此 凸显的图像部分的画质不会因为受带宽的限制而受到影响。
为便于理解本发明, 下面继续结合图 4A 至图 4D 进一步说明根据本 发明的技术方案。
如图 4A 所示, 在屏幕的视频图像中, 点击所需目标, 则识别目标区 域 402。
如图 4B所示, 在识别出目标区域 402后, 视频图像中异于目标区域 402的区域为非目标区域 404。
如图 4C所示, 在对目标区域 402进行操作后 (例如点击一次) , 根 据分配算法分配给目标区域 402对应的码流将会增加, 且分配给非目标区 域 404的码流将会减小, 显示效果上就是目标区域 402的清晰度上升, 且 非目标区域 404的清晰度降低。
如图 4D所示, 当对目标区域 402再次进行操作后, 目标区域 402的 码流将会进一步增加, 且非目标区域 404的码流将会进一步减小, 显示效 果上就是目标区域 402的清晰度进一步上升, 且非目标区域 404的清晰度 进一步降低。
例如手机视频电话应用中, 手机终端可以同时看到对方和用户自己的 头像画面, 这些头像画面中一般都包含背景画面, 背景画面可能是住所、 办公室或教室等, 也可能包含用户身边的其他人。 当带宽受限制或对方看 不清用户的头像, 或用户想有意屏蔽掉背景画面, 则用户可以通过点击自 己的头像, 使头像以较高的清晰度凸显出来, 而背景画面中的人或物就会 被弱化显示。 如果对方觉得头像画面还不够清晰, 则可以继续点击头像, 以更高清晰度的画质发送给对方。
如图 4E 所示, 当用户进行目标切换时, 被重新选择的区域将变为目 标区域 402 , 其他区域自动划分为非目标区域 404。
例如若本地的多个人与异地的某个人进行视频电话时, 只凸显本地的 正在讲话的人的头像, 而使其他人的头像被弱化显示, 当讲话的人需要更 换时, 则可以点击头像进行切换。 当本地有两人在和异地的某个人同时通 话时, 该异地的某个人也可以同时选中本地的这两个人, 将其划分到目标 区域。 这样正在讲话的人就会被凸显出来, 而其他没有讲话的人被弱化显 示。
通过这种压缩方法, 在视频会议应用中, 可以凸显主讲人, 使对方可 以更加清晰的看到主讲人的表情和动作。 在视频聊天、 流媒体播放等视频 类通信应用中也可以采用此图像压缩方法凸显视频图像中的重要元素, 弱 化次要元素。 用户不用再担心所关注的画面会随网络带宽的变化而影响观 看效果。
以上结合附图详细说明了本发明的技术方案, 针对视频图像的不同区 域采取不同的压缩策略, 既保证了传输带宽要求, 又可以区分视频画面中 重要元素和次要元素。
以上所述仅为本发明的优选实施例而已, 并不用于限制本发明, 对于 本领域的技术人员来说, 本发明可以有各种更改和变化。 凡在本发明的精 神和原则之内, 所作的任何修改、 等同替换、 改进等, 均应包含在本发明 的保护范围之内。

Claims

权 利 要 求 书
1. 一种终端, 其特征在于, 包括:
识别单元, 根据接收的选择信号, 识别在当前显示的视频图像中选择 的目标区 i或;
压缩单元, 连接至所述识别单元, 增加所述目标区域所分配到的第一 码流, 根据所述第一码流对所述目标区域进行压缩, 和 /或用于减少所述 视频图像中的非目标区域所分配到的第二码流, 根据所述第二码流对所述 非目标区域进行压缩。
2. 根据权利要求 1所述的终端, 其特征在于, 还包括:
计算单元, 根据分配算法计算出所述目标区域所分配到的第一码流和 所述非目标区域所分配到的第二码流;
获取单元, 根据所述第一码流获取与所述目标区域对应的第一量化 值, 以及根据所述第二码流获取与所述非目标区域对应的第二量化值; 所述压缩单元还用于根据所述第一量化值对所述目标区域进行压缩以 及根据所述第二量化值对所述非目标区域进行压缩。
3. 根据权利要求 2 所述的终端, 其特征在于, 所述计算单元所采用 的所述分配算法包括: 根据识别出的所述目标区域的面积和所述非目标区 域的面积、 选择所述目标区域时的操作次数以及当前网络带宽的码流限定 值, 计算出所述目标区域所分配到的第一码流和所述非目标区域所分配到 的第二码流。
4. 根据权利要求 3所述的终端, 其特征在于, 还包括:
检测单元, 用于检测所述当前网络带宽的码流限定值的变化情况; 所述计算单元还用于在所述码流限定值变化时, 根据所述分配算法和 重新获取的码流限定值, 重新分配所述目标区域的第一码流和所述非目标 区域的第二码流。
5. 根据权利要求 1所述的终端, 其特征在于, 还包括:
还原单元, 连接至所述识别单元, 在接收到预设操作信号时, 所述目 标区域与所述非目标区域的码流分配比为 1 : 1。
6. 根据权利要求 1 所述的终端, 其特征在于, 所述识别单元包括检 测子单元, 在所述目标区域发生移动时, 检测所述目标区域的位移变化状 态, 根据所述位移变化状态调整所述目标区域和所述非目标区域的划分。
7. 根据权利要求 2 至 6 中任一项所述的终端, 其特征在于, 所述获 取单元包括记录子单元, 将所述第一量化值记录在所述第一码流中, 将所 述第二量化值记录在所述第二码流中;
所述终端还包括解压单元, 在对所述视频图像进行解压时, 根据所述 第一量化值对所述目标区域进行反量化计算, 以及根据所述第二量化值对 所述非目标区域进行反量化计算。
8. 根据权利要求 2 至 6 中任一项所述的终端, 其特征在于, 还包 括:
发送单元, 连接至所述识别单元, 将识别出的目标区域信息发送至所 述其他终端, 其中, 所述其他终端与所述终端同时处理所述视频图像; 接收单元, 连接至所述压缩单元, 接收其他终端识别出的目标区域信 息, 根据所述目标区域信息获取所述视频图像中被选择的目标区域, 所述 压缩单元对所述视频图像中的所述目标区域和所述非目标区域进行压缩。
9. 根据权利要求 8 所述的终端, 其特征在于, 所述发送单元还用于 将选择所述目标区域时的操作信息发送至所述其他终端;
所述接收单元还接收所述其他终端发送的选择所述目标区域时的操作 信息, 根据所述操作信息获取本地终端选择所述目标区域时的操作次数; 所述计算单元根据所述目标区域的面积、 选择所述目标区域时的操作 次数以及所述其他终端的当前网络带宽的码流限定值, 计算出所述目标区 域所分配到的第一码流和所述非目标区域所分配到的第二码流。
10. 一种视频图像压缩方法, 其特征在于, 包括以下步骤:
识别在视频图像中选择的目标区域,
增加所述目标区域所分配到的第一码流, 根据所述第一码流对所述目 标区域进行压缩, 和 /或减少所述视频图像中的非目标区域所分配到的第 二码流, 根据所述第二码流对所述非目标区域进行压缩。
11. 根据权利要求 10 所述的视频图像压缩方法, 其特征在于, 还包 括:
根据分配算法计算出所述目标区域所分配到的第一码流和所述非目标 区域所分配到的第二码流;
根据所述第一码流获取与所述目标区域对应的第一量化值, 以及根据 所述第二码流获取与所述非目标区域对应的第二量化值;
根据所述第一量化值对所述目标区域进行压缩以及根据所述第二量化 值对所述非目标区域进行压缩。
12. 根据权利要求 11 所述的视频图像压缩方法, 其特征在于, 所述 分配算法包括: 根据识别出的所述目标区域的面积和所述非目标区域的面 积、 选择所述目标区域时的操作次数以及当前网络带宽的码流限定值, 计 算出所述目标区域所分配到的第一码流和所述非目标区域所分配到的第二 码流。
13. 根据权利要求 12 所述的视频图像压缩方法, 其特征在于, 还包 括: 检测所述当前网络带宽的码流限定值的变化情况;
在所述码流限定值变化时, 根据所述分配算法和重新获取的码流限定 值, 重新分配所述目标区域的第一码流和所述非目标区域的第二码流。
14. 根据权利要求 11 所述的视频图像压缩方法, 其特征在于, 还包 括: 在接收到预设操作信号时, 所述目标区域与所述非目标区域的码流分 配比为 1 : 1。
15. 根据权利要求 11 所述的视频图像压缩方法, 其特征在于, 在所 述目标区域发生移动时, 检测所述目标区域的位移变化状态, 根据所述位 移变化状态调整所述目标区域和所述非目标区域的划分。
16. 根据权利要求 11至 16中任一项所述的视频图像压缩方法, 其特 征在于, 还包括: 将所述第一量化值记录在所述第一码流中, 将所述第二 量化值记录在所述第二码流中;
在对所述视频图像进行解码时, 根据所述第一量化值对所述目标区域 进行反量化计算, 以及根据所述第二量化值对所述非目标区域进行反量化 计算。
17. 根据权利要求 11至 16中任一项所述的视频图像压缩方法, 其特 征在于, 还包括: 在其他终端与本地终端同时处理所述视频图像时, 所述 本地终端将识别出的目标区域信息发送至所述其他终端;
所述其他终端根据所述目标区域信息获取所述视频图像中被选择的目 标区域, 增加所述目标区域所分配到的第一码流, 根据所述第一码流对所 述目标区域进行压缩, 和 /或减少所述视频图像中的非目标区域所分配到 的第二码流, 根据所述第二码流对所述非目标区域进行压缩。
18. 根据权利要求 17 所述的视频图像压缩方法, 其特征在于, 还包 括: 所述本地终端将选择所述目标区域时的操作信息发送至所述其他终 端, 所述其他终端根据所述操作信息获取本地终端选择所述目标区域时的 操作次数;
所述其他终端根据所述目标区域的面积、 选择所述目标区域时的操作 次数以及所述其他终端的当前网络带宽的码流限定值, 计算出所述目标区 域所分配到的第一码流和所述非目标区域所分配到的第二码流。
PCT/CN2012/077807 2012-06-28 2012-06-28 终端和视频图像压缩方法 WO2014000238A1 (zh)

Priority Applications (2)

Application Number Priority Date Filing Date Title
PCT/CN2012/077807 WO2014000238A1 (zh) 2012-06-28 2012-06-28 终端和视频图像压缩方法
CN201280073344.4A CN104322065B (zh) 2012-06-28 2012-06-28 终端和视频图像压缩方法

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
PCT/CN2012/077807 WO2014000238A1 (zh) 2012-06-28 2012-06-28 终端和视频图像压缩方法

Publications (1)

Publication Number Publication Date
WO2014000238A1 true WO2014000238A1 (zh) 2014-01-03

Family

ID=49782089

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2012/077807 WO2014000238A1 (zh) 2012-06-28 2012-06-28 终端和视频图像压缩方法

Country Status (2)

Country Link
CN (1) CN104322065B (zh)
WO (1) WO2014000238A1 (zh)

Cited By (16)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104967867A (zh) * 2015-06-19 2015-10-07 美国掌赢信息科技有限公司 一种即时视频的传输方法和电子设备
CN104967914A (zh) * 2015-06-19 2015-10-07 美国掌赢信息科技有限公司 一种即时视频的编码方法和电子设备
CN104994405A (zh) * 2015-06-19 2015-10-21 美国掌赢信息科技有限公司 一种即时视频的传输方法和电子设备
CN105049947A (zh) * 2015-06-19 2015-11-11 美国掌赢信息科技有限公司 一种即时视频的传输方法和电子设备
CN105163197A (zh) * 2015-06-19 2015-12-16 美国掌赢信息科技有限公司 一种即时视频的编码方法和电子设备
CN105163125A (zh) * 2015-06-19 2015-12-16 美国掌赢信息科技有限公司 一种即时视频的编码方法和电子设备
CN105163199A (zh) * 2015-06-19 2015-12-16 美国掌赢信息科技有限公司 一种即时视频的编码方法和电子设备
CN105163195A (zh) * 2015-06-19 2015-12-16 美国掌赢信息科技有限公司 一种即时视频的编码方法和电子设备
CN105163196A (zh) * 2015-06-19 2015-12-16 美国掌赢信息科技有限公司 一种即时视频的编码方法和电子设备
CN105163198A (zh) * 2015-06-19 2015-12-16 美国掌赢信息科技有限公司 一种即时视频的编码方法和电子设备
CN105721872A (zh) * 2016-01-29 2016-06-29 广东欧珀移动通信有限公司 一种图像压缩的方法及终端
WO2016202287A1 (zh) * 2015-06-19 2016-12-22 美国掌赢信息科技有限公司 一种即时视频的传输方法和电子设备
WO2016202285A1 (zh) * 2015-06-19 2016-12-22 美国掌赢信息科技有限公司 一种即时视频的传输方法和电子设备
CN111031325A (zh) * 2019-10-28 2020-04-17 西安万像电子科技有限公司 数据处理方法及系统
CN113824967A (zh) * 2021-11-23 2021-12-21 环球数科集团有限公司 一种基于深度学习的视频压缩方法
CN114520873A (zh) * 2021-12-31 2022-05-20 广州文远知行科技有限公司 一种清扫车视频加速方法及系统

Families Citing this family (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108366244A (zh) * 2018-03-16 2018-08-03 北京虚拟映画科技有限公司 影视图像传输方法及装置
CN108366245B (zh) * 2018-03-16 2020-09-11 北京虚拟映画科技有限公司 影像图像传输方法及装置
CN109769143A (zh) * 2019-02-03 2019-05-17 广州视源电子科技股份有限公司 视频图像处理方法、装置、视频系统、设备和存储介质
CN111193911B (zh) * 2020-01-15 2021-12-14 未来新视界文化科技(嘉善)有限公司 大数据视频的速传处理方法和装置

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101547365A (zh) * 2009-05-08 2009-09-30 北京北纬通信科技股份有限公司 保证特定区域显示质量的视频编码的方法和装置
CN101742296A (zh) * 2008-11-14 2010-06-16 北京中星微电子有限公司 降低码流数据量波动的视频编解码方法及装置
CN101860757A (zh) * 2010-06-03 2010-10-13 无锡中星微电子有限公司 智能监控系统及其图像编解码方法
US20110235706A1 (en) * 2010-03-25 2011-09-29 Texas Instruments Incorporated Region of interest (roi) video encoding

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101742296A (zh) * 2008-11-14 2010-06-16 北京中星微电子有限公司 降低码流数据量波动的视频编解码方法及装置
CN101547365A (zh) * 2009-05-08 2009-09-30 北京北纬通信科技股份有限公司 保证特定区域显示质量的视频编码的方法和装置
US20110235706A1 (en) * 2010-03-25 2011-09-29 Texas Instruments Incorporated Region of interest (roi) video encoding
CN101860757A (zh) * 2010-06-03 2010-10-13 无锡中星微电子有限公司 智能监控系统及其图像编解码方法

Cited By (17)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105163196A (zh) * 2015-06-19 2015-12-16 美国掌赢信息科技有限公司 一种即时视频的编码方法和电子设备
CN105163197A (zh) * 2015-06-19 2015-12-16 美国掌赢信息科技有限公司 一种即时视频的编码方法和电子设备
CN104967867A (zh) * 2015-06-19 2015-10-07 美国掌赢信息科技有限公司 一种即时视频的传输方法和电子设备
CN105049947A (zh) * 2015-06-19 2015-11-11 美国掌赢信息科技有限公司 一种即时视频的传输方法和电子设备
CN105163198A (zh) * 2015-06-19 2015-12-16 美国掌赢信息科技有限公司 一种即时视频的编码方法和电子设备
CN105163125A (zh) * 2015-06-19 2015-12-16 美国掌赢信息科技有限公司 一种即时视频的编码方法和电子设备
CN105163199A (zh) * 2015-06-19 2015-12-16 美国掌赢信息科技有限公司 一种即时视频的编码方法和电子设备
CN105163198B (zh) * 2015-06-19 2019-03-19 美国掌赢信息科技有限公司 一种即时视频的编码方法和电子设备
CN104994405A (zh) * 2015-06-19 2015-10-21 美国掌赢信息科技有限公司 一种即时视频的传输方法和电子设备
CN104967914A (zh) * 2015-06-19 2015-10-07 美国掌赢信息科技有限公司 一种即时视频的编码方法和电子设备
CN105163195A (zh) * 2015-06-19 2015-12-16 美国掌赢信息科技有限公司 一种即时视频的编码方法和电子设备
WO2016202287A1 (zh) * 2015-06-19 2016-12-22 美国掌赢信息科技有限公司 一种即时视频的传输方法和电子设备
WO2016202285A1 (zh) * 2015-06-19 2016-12-22 美国掌赢信息科技有限公司 一种即时视频的传输方法和电子设备
CN105721872A (zh) * 2016-01-29 2016-06-29 广东欧珀移动通信有限公司 一种图像压缩的方法及终端
CN111031325A (zh) * 2019-10-28 2020-04-17 西安万像电子科技有限公司 数据处理方法及系统
CN113824967A (zh) * 2021-11-23 2021-12-21 环球数科集团有限公司 一种基于深度学习的视频压缩方法
CN114520873A (zh) * 2021-12-31 2022-05-20 广州文远知行科技有限公司 一种清扫车视频加速方法及系统

Also Published As

Publication number Publication date
CN104322065A (zh) 2015-01-28
CN104322065B (zh) 2018-10-26

Similar Documents

Publication Publication Date Title
WO2014000238A1 (zh) 终端和视频图像压缩方法
EP1856913B1 (en) Region-of-interest extraction for video telephony
US8977063B2 (en) Region-of-interest extraction for video telephony
US8514265B2 (en) Systems and methods for selecting videoconferencing endpoints for display in a composite video image
US8274577B2 (en) Methods of processing digital image and/or video data including luminance filtering based on chrominance data and related systems and computer program products
US7953283B2 (en) Portable terminal
US20110026591A1 (en) System and method of compressing video content
WO2010037310A1 (zh) 多路视频通信方法与系统
WO2021057481A1 (zh) 视频编解码方法和相关装置
WO2012057733A1 (en) Systems, methods, and devices for adjusting video conference parameters to maintain system performance
US9306987B2 (en) Content message for video conferencing
WO2021057705A1 (zh) 视频编解码方法和相关装置
JP2011029868A (ja) 端末装置、遠隔会議システム、端末装置の制御方法、端末装置の制御プログラム、及び端末装置の制御プログラムを記録したコンピュータ読み取り可能な記録媒体
US11849127B2 (en) Video encoding method, video decoding method, and related apparatuses
US20220038708A1 (en) Video encoding method, video decoding method, and related apparatuses
JP2009038604A (ja) 映像信号処理装置
US20220078454A1 (en) Video encoding method, video decoding method, and related apparatuses
KR20090125556A (ko) 영상 코덱 방법, 영상 코덱 장치 및 그 영상 서비스 시스템
CN116962613A (zh) 数据传输方法及装置、计算机设备、存储介质
JPH11177518A (ja) データ通信装置
JP2005341347A (ja) 多地点接続装置および方法
JPH11177436A (ja) データ通信装置
JP4735558B2 (ja) 情報処理装置および情報処理方法
CN118044191A (zh) 用于对视频帧进行编码的动态量化参数
JP2008227937A (ja) 情報処理装置および情報処理方法

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 12879950

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 12879950

Country of ref document: EP

Kind code of ref document: A1