WO2020125631A1 - Procédé et appareil de compression vidéo, et support de stockage lisible par ordinateur - Google Patents
Procédé et appareil de compression vidéo, et support de stockage lisible par ordinateur Download PDFInfo
- Publication number
- WO2020125631A1 WO2020125631A1 PCT/CN2019/126016 CN2019126016W WO2020125631A1 WO 2020125631 A1 WO2020125631 A1 WO 2020125631A1 CN 2019126016 W CN2019126016 W CN 2019126016W WO 2020125631 A1 WO2020125631 A1 WO 2020125631A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- image
- frame
- blur
- video compression
- pixel
- Prior art date
Links
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/85—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using pre-processing or post-processing specially adapted for video compression
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/169—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
- H04N19/186—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being a colour or a chrominance component
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/42—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by implementation details or hardware specially adapted for video compression or decompression, e.g. dedicated software implementation
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/47—End-user applications
- H04N21/478—Supplemental services, e.g. displaying phone caller identification, shopping application
- H04N21/4788—Supplemental services, e.g. displaying phone caller identification, shopping application communicating with other users, e.g. chatting
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N7/00—Television systems
- H04N7/14—Systems for two-way working
- H04N7/141—Systems for two-way working between two video terminals, e.g. videophone
Definitions
- This application relates to the field of video processing, and in particular, to a video compression method, device, and computer-readable storage medium.
- Video chat has gradually replaced ordinary language and text, and has become an important way for people to communicate and communicate in work or daily life.
- the video information in the video chat process is composed of one frame of image, but the current video image contains a lot of redundant information, which takes up a lot of bandwidth during the digital transmission of video, which affects the cost of operation and the video’s Fluency.
- many compression methods have emerged, such as JPEG and MPEG methods.
- the current compression method will compress all the content in each frame of the image, which leads to the reduction of the clarity of the face in the video, which cannot meet the requirements of people in the process of video chatting.
- the main purpose of the present application is to provide a video compression method, device and computer-readable storage medium, aiming to realize the recognition of the area where the face is located in the video chat and blur the compression of the part outside the area where the face is located, Not only can reduce the bandwidth cost of the entire process, but also ensure the clarity and fluency of face information in the video.
- the present application provides a video compression method, which includes the following steps:
- Gaussian blur processing is performed on the background image to obtain a blurred background image
- the step of dividing each frame of the video image into a face image and a background image includes:
- the step of performing skin color analysis on each frame of the video image to obtain the face image and background image in each frame includes:
- Morphological processing is performed on the skin color region of each frame image to obtain a face image and a background image in each frame image.
- the first preset algorithm is used to perform brightness compensation on each frame image according to the brightness and chromaticity of the pixels in each frame image to obtain the compensated pixels in each frame image
- the chroma steps include:
- the chromaticity of the compensated pixels in each frame image is calculated according to the illumination compensation coefficient and the brightness of the pixels in each frame image.
- the step of performing Gaussian blur processing on each frame image to obtain a blurred background image includes:
- Gaussian blur processing is performed on each image to obtain a background image that is gradually blurred from a direction away from the center point of the blur processing.
- the step of performing Gaussian blur processing on each image to obtain a progressively blurred background image includes:
- Gaussian blur processing is performed on each pixel of the blur area according to the variable radius of the Gaussian blur processing to obtain a gradually blurred blur area.
- the transition area is located between the face image and the blurred area.
- the step of determining the blur processing center point of the blur area according to the second preset algorithm includes:
- the blur processing center point of the blur area is calculated according to the coordinates of each pixel in the blur area.
- the second preset algorithm is:
- x is the horizontal coordinate of the pixel
- y is the vertical coordinate of the pixel
- target is the target function
- background is the background function
- x0 is the horizontal coordinate of the blur processing center point
- y0 is the vertical coordinate of the blur processing center point.
- the present application also provides a video compression device
- the video compression device includes: a memory, a processor, and a video compression program stored on the memory and executable on the processor, so When the video compression program is executed by the processor, the following steps are realized:
- Gaussian blur processing is performed on the background image to obtain a blurred background image
- the present application also provides a computer-readable storage medium on which a video compression program is stored.
- a video compression program is executed by a processor, the video compression method described above is implemented. step.
- This application obtains a video image and performs face recognition on each frame image in the video image to obtain a face image and a background image in each frame image; Gaussian blur processing is performed on the background image to obtain a blurred background Image; Compress the blurred background image.
- the application can determine the face area in the image according to skin color, and perform blur compression processing on the part outside the face area, so that during the video transmission process, the user can be guaranteed during the video process
- the unprocessed face image can meet the requirements of people's face clarity in the video chat process; at the same time, relative The clear background image requires less bandwidth, which can reduce the bandwidth cost of the entire system.
- FIG. 1 is a schematic structural diagram of a video compression terminal involved in an embodiment of the present application
- FIG. 2 is a schematic flowchart of a first embodiment of a video compression method of this application
- FIG. 3 is a schematic diagram of a detailed process of performing face recognition on each frame image in the video image in the embodiment of the present application to obtain the face image and the background image in each frame image;
- FIG. 4 is a schematic diagram of a detailed process of performing Gaussian blur processing on each frame image to obtain a blurred background image in an embodiment of the present application;
- FIG. 5 is a schematic diagram of a progressive blurring method in an embodiment of a video compression method of this application.
- the commonly used video compression algorithm directly compresses the entire video picture, which not only reduces the video transmission cost, but also affects the clarity of the face in the video.
- the present application provides a video compression method, by obtaining a video image, and performing face recognition on each frame image in the video image to obtain a face image and a background image in each frame image; Gaussian blur processing is performed on the background image to obtain a blurred background image; and compression processing is performed on the blurred background image.
- This application can realize the recognition of the area of the face in the image in the process of video chat, and the area other than the face is blurred and compressed, thereby ensuring the clarity of the face information in the process of video chat and reducing the cost of video bandwidth .
- FIG. 1 is a schematic structural diagram of a video compression terminal according to an embodiment of the present application.
- the terminal may be a PC, or may be a terminal device that can have a video compression function, such as a smart phone, tablet computer, video recorder, or portable computer.
- a video compression function such as a smart phone, tablet computer, video recorder, or portable computer.
- the terminal may include: a processor 1001, such as a CPU, a communication bus 1002, a user interface 1003, a network interface 1004, a memory 1005, a camera 1006, stored on the memory, and may be on the processor Video compression program running.
- the communication bus 1002 is used to implement connection communication between these components.
- the user interface 1003 may include a display (Display), an input unit such as a keyboard (Keyboard), and the optional user interface 1003 may also include a standard wired interface and a wireless interface.
- the network interface 1004 may optionally include a standard wired interface and a wireless interface (such as a WI-FI interface).
- the memory 1005 may be a high-speed RAM memory, or may be a non-volatile memory (non-volatile memory), such as a disk memory.
- the memory 1005 may optionally be a storage device independent of the foregoing processor 1001.
- the terminal may further include a microphone, an RF (Radio Frequency) circuit, a sensor, an audio circuit, a WiFi module, and so on.
- sensors such as light sensors, motion sensors and other sensors.
- the terminal can also be configured with other sensors such as gyroscopes, barometers, hygrometers, thermometers, and infrared sensors, which will not be repeated here.
- FIG. 1 does not constitute a limitation on the terminal, and may include more or less components than those illustrated, or combine certain components, or have different component arrangements.
- the memory 1005 as a computer storage medium may include an operating system, a network communication module, a user interface module, and a video compression program.
- the network interface 1004 is mainly used to connect to the background server and perform data communication with the background server;
- the user interface 1003 is mainly used to connect to the client (user) and perform data communication with the client;
- the processor 1001 is mainly used to execute the video compression program stored in the memory 1005 to achieve the following steps:
- Gaussian blur processing is performed on the background image to obtain a blurred background image
- processor 1002 is mainly used to execute the video compression program stored in the memory 1001 to implement the following steps:
- processor 1002 is mainly used to execute the video compression program stored in the memory 1001 to implement the following steps:
- Morphological processing is performed on the skin color region of each frame image to obtain a face image and a background image in each frame image.
- processor 1002 is mainly used to execute the video compression program stored in the memory 1001 to implement the following steps:
- the chromaticity of the compensated pixels in each frame image is calculated according to the illumination compensation coefficient and the brightness of the pixels in each frame image.
- processor 1002 is mainly used to execute the video compression program stored in the memory 1001 to implement the following steps:
- Gaussian blur processing is performed on each frame image to get gradually away from the direction of the center point of the blur processing
- processor 1002 is mainly used to execute the video compression program stored in the memory 1001 to implement the following steps:
- Gaussian blur processing is performed on each pixel of the blur area according to the variable radius of the Gaussian blur processing to obtain a gradually blurred blur area.
- processor 1002 is mainly used to execute the video compression program stored in the memory 1001 to implement the following steps:
- the transition area is located between the face image and the blur area.
- processor 1002 is mainly used to execute the video compression program stored in the memory 1001 to implement the following steps:
- processor 1002 is mainly used to execute the video compression program stored in the memory 1001 to implement the following steps:
- the blur processing center point of the blur area is calculated according to the coordinates of each pixel in the blur area.
- processor 1002 is mainly used to execute the video compression program stored in the memory 1001 to implement the following steps:
- the second preset algorithm is:
- x is the horizontal coordinate of the pixel
- y is the vertical coordinate of the pixel
- target is the target function
- background is the background function
- x0 is the horizontal coordinate of the blur processing center point
- y0 is the vertical coordinate of the blur processing center point.
- FIG. 2 is a schematic flowchart of a first embodiment of a video compression method according to this application.
- the video compression method includes:
- Step S100 obtaining a video image
- the video compression method mainly recognizes the area of the face in the image during the video chat process, and blurs the area other than the face to ensure the face information during the video chat process Sharpness also reduces the cost of video bandwidth.
- the video compression method can be completed by a device with an image processing function such as a terminal or a server.
- the device with image processing function may be a PC, a server connected to the terminal, or a mobile terminal device such as a smart phone, tablet computer, video recorder, portable computer, or the like.
- This embodiment uses a mobile terminal as an example for description.
- This application can be applied to video recording scenarios, as well as video chat sessions through chat software, such as video calls in WeChat or QQ.
- the embodiment of the present application collects video images through a camera, where the camera may be a camera of the mobile terminal, such as a mobile phone camera or a computer camera.
- Step S200 Perform face recognition on each frame image in the video image to obtain a face image and a background image in each frame image;
- This application divides each frame of images into face images and background images through face recognition, mainly by extracting the face features in each frame image to achieve face recognition.
- face recognition mainly by extracting the face features in each frame image to achieve face recognition.
- the step of dividing each frame of the video image into a face image and a background image includes:
- face recognition is mainly performed by skin color recognition.
- Skin color is important information in face information. Skin color recognition can speed up the recognition process of face recognition, and is a relatively quick and concise method for face recognition. Specific embodiments may also recognize the human face through other techniques, so that each frame of image is divided into a human face image and a background image.
- skin color analysis is performed on each frame of the video image, which mainly utilizes the clustering characteristics of skin color in a specific color space, that is, skin color in a certain color space, people of different ages, races and genders
- the chromaticity of the skin tone is stably distributed within a certain range, which can be described by a correlation distribution model.
- Commonly used distribution models are regional models, Gaussian models, histogram models, etc.
- the skin color analysis for each frame of image is to first project each frame of image obtained in the instant messaging software video chat to the color space to obtain the brightness and chromaticity of the pixels in each frame of image, and then for each frame Image pixels are compensated for brightness to eliminate differences in brightness of people's skin tones of different ages, races, and genders, and then the chromaticity of each frame of image pixels after compensation is compared with the skin color chromaticity model constructed by a large number of skin color samples. Determine the skin color area of each frame image, and then segment the skin color area to obtain the face image and background image of each frame image. Through this method of skin color recognition, the face image and the background image of each image can be separated.
- the method for segmenting the skin color region may be a threshold method, a boundary detection method, a matching method, and the like.
- Step S300 performing Gaussian blur processing on the background image to obtain a blurred background image
- Gaussian blur is a widely practical technology in image processing, and is a data smoothing technology that is suitable for multiple occasions.
- the Gaussian fuzzy calculation process is the same as that in the prior art, and no redundant description is given here. To put it simply, the average value of the pixels of a certain pixel of the background image in step S200 is taken, so that the pixel loses details, and the effect of smoothing the pixel is achieved, so that the image becomes blurry. During the blurring process, the larger the range of surrounding pixels, the stronger the blurring effect.
- the progressive blurring effect may be from the blurred area close to the face image to away from the face image
- the blur area gradually blurs, that is, the area close to the face image is clearer, and the area farther away from the face image is more blurry; it can also start from the blur area away from the face image to the blur area close to the face image, that is, The area near the face image is blurry, and the area away from the face image is clearer; it can also start from the area in the middle of the blur area and gradually blur toward the two sides, that is, near the face area and away from the face area, that is, in the middle of the background image The area is clearer, and the area near and far away from the face is blurry.
- Step S400 Compress the blurred background image.
- each frame image is divided into two images: a face image and a background image, and the background image is Gaussian blurred, and then compressed, so that each frame image is divided into two images: human face
- the image and the background image are transmitted, and the background image occupies less bandwidth after blurring and compression.
- the receiving client decompresses the compressed background image to obtain a blurred background image, and merges it with the face image to form an image. In this way, this process realizes the functions to be implemented in this application, effectively reduces the cost of bandwidth, and at the same time ensures the clarity of face information in video chat.
- the application by acquiring a video image, according to skin color analysis of each frame of the video image, the face image and background image in each frame of image are obtained; Gaussian blur processing is performed on the background image to obtain a blurred background image ; Compress the blurred background image.
- the application can determine the face area in the image according to skin color, and perform blur compression processing on the part outside the face area, so that during the video transmission process, the user can be guaranteed during the video process
- the most concerned part the clearness of the face area, but it will not be very important part: the background image is blurred.
- the unprocessed face image can meet the requirements of people's face clarity in the video chat process; at the same time, relative The clear background image requires less bandwidth, which can reduce the bandwidth cost of the entire system.
- FIG. 3 is a schematic diagram of a detailed process of performing a skin color analysis on each frame of the video image in the embodiment of the present application to obtain a face image and a background image in each frame of the image.
- step S200 includes:
- Step S201 Project the image of each frame into a preset color space to obtain the brightness and chromaticity of pixels in the image of each frame;
- the preset color spaces are: YCrCb color space, RGB color space, HSI color space, YIQ color space, etc.
- the projection onto the YCrCb color space is used as an example for description.
- the YCbCr color space represents color as three sub-views, namely brightness Y, blue chroma Cb and red chroma Cr. In this way, by projecting each frame of image into the YCbCr color space, you can get the two kinds of information of the brightness Y and chroma CrCb of each frame of image pixels.
- step S202 the first preset algorithm is used to perform brightness compensation on each frame image according to the brightness and chroma of the pixels in each frame image to obtain the chroma of the compensated pixels in each frame image;
- the difference in skin color between people of different ages, races, and genders is mainly in brightness, but the difference in chroma is small. That is, different skin tones differ greatly in brightness, but are relatively close in chroma.
- the reference white algorithm is used as an example for description. Specifically, the brightness compensation for each frame image is performed by referring to the white algorithm, that is, in step S201 The luminance Y and chromaticity CrCb of each pixel of the image are calculated and compensated by the reference white algorithm, and the influence of the luminance Y of the pixel in the YCbCr color space is eliminated to obtain the CbCr chromaticity of the pixel of the video image.
- Step Reference White algorithm is a commonly used brightness compensation algorithm.
- the brightness compensation process includes:
- the chromaticity of the compensated pixel in each frame of image is calculated according to the illumination compensation coefficient.
- Step S203 comparing the chromaticity of the compensated pixels in each frame image with a preset skin color chromaticity Gaussian model to obtain a skin color probability map of pixels in each frame image;
- the distribution range of skin color on chromaticity can be described by establishing a skin color model.
- Common skin color models include regional models, Gaussian models, and histogram models.
- a Gaussian model is used to describe the distribution range of different skin tones in chroma. Specifically, in the CbCr space, a large number of skin color samples are selected for statistics to obtain the distribution range of each skin color in chroma, and then a Gaussian model is used for calculation and description.
- the Gaussian model is preset in the embodiments of the present application Skin color chroma Gaussian model. Specifically, the chromaticity of the compensated pixels in each frame of image is compared with the preset skin color chromaticity Gaussian model to obtain a skin color probability map of the pixels in each frame of image.
- the CbCr chromaticity value is compared with the preset skin tone chromaticity Gaussian model, that is, the chromaticity of each pixel in the YCrCb chromaticity space and the distance between the distribution center of the preset skin tone chromaticity Gaussian model and The similarity of skin color, and then the probability that the point is skin color.
- the chromaticity of the pixel in the YCrCb chromaticity space is The farther the distribution center of the preset skin tone chroma Gaussian model, the smaller the probability that the pixel is skin color, and the chroma of the pixel in the YCrCb chromaticity space is not in the preset skin tone chroma Gaussian model Within the distribution range, the probability that the pixel is skin color is 0. In this way, by comparing the chromaticity of all compensated pixels with a preset skin color chromaticity Gaussian model, a skin color probability map of the entire image can be obtained.
- Step S204 determine the skin color region of each frame image using an adaptive threshold according to the skin color probability map of pixels in each frame image
- each frame image After obtaining the skin color probability map of each frame image, the entire image needs to be segmented, and then it is determined that each frame image is divided into a skin color area and a background area.
- segmentation methods include threshold method, boundary detection method, matching method and so on.
- an adaptive threshold method is used for segmentation. Specifically, the skin color probability map of each pixel in each frame of image is compared with the preset adaptive threshold. If the skin color probability of the pixel is greater than the set adaptive threshold, the pixel is the skin color area, and the skin color of the pixel If the probability is less than the set adaptive threshold, the pixel is a non-skin area. In this way, the adaptive threshold can be used to determine the skin area and non-skin area of each image.
- the adaptive threshold segmentation method is a more common segmentation method for image segmentation.
- the adaptive threshold segmentation method is simpler to calculate and less affected by other factors.
- step S205 morphological processing is performed on the skin color region of each frame image to obtain a face image and a background image in each frame image.
- the skin color area of each frame of the image segmented is not a coherent whole, more scattered, or even composed of multiple unconnected image areas.
- morphological processing needs to be used to improve the segmentation effect .
- the segmented skin color region is processed with a morphological processing-related algorithm to make the skin color region boundary smoother, and the dispersed skin color regions are completely connected as a whole.
- the overall skin color area is the face image, and the other area is the background image. In this way, the whole frame image is divided into two parts: a face image and a background image.
- the morphological processing method is a commonly used image processing method, and the image processing process is the same as in the prior art, and no redundant description is provided here.
- FIG. 4 is a schematic diagram of a detailed process of performing Gaussian blur processing on each frame image to obtain a blurred background image in an embodiment of the present application.
- step S300 further includes:
- Step S301 Perform Gaussian filtering on the background image to obtain the transition area and the blur area;
- the blur processing adopts a progressive blur processing method, which divides each frame of the image into three areas, namely a clear area, a transition area, and a blur area.
- the transition area is located between the clear area and the blur area.
- the clear area The resolution of is greater than the resolution of the transition area, and the resolution of the transition area is greater than the resolution of the blurred area.
- FIG. 5 is a schematic diagram of the progressive blurring method in the embodiment of the video compression method of the present application.
- the face image and the background image of each frame image in the chat video are cut and separated, and the background image is subjected to Gaussian filter processing.
- the separated face image is our clear area a, and the background image after Gaussian filtering is divided into transition area b and blur area c.
- Gaussian filter processing is a common method for image processing, which is the same as in the prior art, and will not be redundantly described here.
- Step S302 Determine a blur processing center point of the blur area according to a second preset algorithm
- the second preset algorithm is:
- x is the horizontal coordinate of the pixel
- y is the vertical coordinate of the pixel
- target is the target function
- background is the background function
- x0 is the horizontal coordinate of the blur processing center point
- y0 is the vertical coordinate of the blur processing center point.
- the blur processing center point of the blur area is determined according to the second preset algorithm, and the blur processing center point is located in the clear area.
- the position coordinates of the blur processing center point of the blur area c are calculated by the above formula, and then the position of the blur processing center point is obtained, and the blur processing center point is located in the clear area a.
- Step S303 Calculate the distance between each pixel in the blur area and the blur processing center point according to the blur processing center point;
- the distance between each pixel in the blur area c and the blur processing center point can be obtained.
- Step S304 Calculate the variable radius of each pixel in the blur area for Gaussian blur processing according to the distance between each pixel in the blur area and the center point of the blur processing;
- the size of the distance between each pixel point in the blur area c and the center point of the blur processing is related calculated to obtain the variable Gaussian blur processing radius.
- Step S305 Perform Gaussian blur processing on each pixel of the blur area according to the variable radius of the Gaussian blur processing to obtain a gradually blurred blur area.
- the larger the variable radius of the Gaussian blur processing the more blurred the image blur processing effect after the Gaussian blur processing; on the contrary, the smaller the variable radius of the Gaussian blur processing, the Gaussian blur processing The later the image blurring effect is clearer, that is, the pixels near the blurring center point are clearer, and the pixels farther from the blurring center point are blurry.
- each pixel in the blur area c is Gaussian blurred according to the size of the variable radius of the Gaussian blur processing.
- the larger the variable radius of the Gaussian blur processing of each pixel in the blur area c the more blurred the Gaussian blur.
- the smaller the variable radius of the Gaussian blur processing of each pixel in the blur area c the clearer the Gaussian blur. That is, in the blur area c, the closer to the clear area a, the clearer the image blur effect, and the farther away from the clear area a, the more blurred the image blur effect, thus forming a gradual blur from the clear area to the blur area Effect.
- compressing the blur area c with the progressive blur effect together with the transition area b can reduce the bandwidth of the video chat during transmission, and at the same time ensure the clarity of face information that has not been processed by blur compression.
- the blurring process from the clear area to the blurry area can have a more natural transition, so that the processed image in the video chat process does not appear too abrupt, and the background blur is improved. effect.
- the face image and the background image in each frame image are separated to obtain the face image and the background image, and the background image is subjected to Gaussian filter processing to obtain the transition area and the blur area;
- Two preset algorithms determine the blur processing center point of the blur area; according to the blur processing center point, calculate the distance between each pixel point in the blur area and the blur processing center point; according to the blur area
- the distance between the pixel point and the center point of the blur processing calculates the variable radius of each pixel point of the blur area for Gaussian blur processing; the Gaussian blur processing of each pixel point of the blur area according to the variable radius of the Gaussian blur processing, Obtain progressively blurred areas of blur.
- the image is divided into two face image parts and a background image part.
- the background image is blurred and compressed, which can reduce the bandwidth of the entire frame image during transmission, and can also ensure that the image is not blurred and compressed.
- the clarity of the face image The method of progressive blur processing from the clear area to the blur area can make the blur effect more natural and the transition more coherent.
- embodiments of the present application also provide a computer-readable storage medium that stores a video compression program on the computer-readable storage medium, and when the video compression program is executed by a processor, implements the steps of the video compression method described above .
- the methods in the above embodiments can be implemented by means of software plus a necessary general hardware platform, and of course, can also be implemented by hardware, but in many cases the former is better Implementation.
- the technical solution of the present application can be embodied in the form of a software product in essence or part that contributes to the existing technology, and the computer software product is stored in a storage medium (such as ROM/RAM) as described above , Disk, CD), including several instructions to make a terminal device (which can be a mobile phone, computer, server, air conditioner, or network equipment, etc.) to perform the method described in each embodiment of the present application.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- General Engineering & Computer Science (AREA)
- Image Processing (AREA)
- Image Analysis (AREA)
Abstract
L'invention concerne un procédé de compression vidéo. Le procédé consiste à : acquérir une image vidéo ; effectuer une reconnaissance de visage sur chaque trame d'une image dans l'image vidéo pour obtenir une image de visage et une image d'arrière-plan dans chaque trame de l'image ; effectuer un traitement de flou Gaussien sur l'image d'arrière-plan pour obtenir une image d'arrière-plan floue ; et effectuer un traitement de compression sur l'image d'arrière-plan floue. L'invention concerne en outre un appareil de compression vidéo, et un support de stockage lisible par ordinateur. La présente invention peut atteindre les effets suivants : dans un processus de chat vidéo, la clarté des informations de visage dans un processus vidéo peut être garantie, et la bande passante d'une vidéo peut également être réduite.
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201811546877.5A CN109618173B (zh) | 2018-12-17 | 2018-12-17 | 视频压缩方法、装置和计算机可读存储介质 |
CN201811546877.5 | 2018-12-17 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2020125631A1 true WO2020125631A1 (fr) | 2020-06-25 |
Family
ID=66010246
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/CN2019/126016 WO2020125631A1 (fr) | 2018-12-17 | 2019-12-17 | Procédé et appareil de compression vidéo, et support de stockage lisible par ordinateur |
Country Status (2)
Country | Link |
---|---|
CN (1) | CN109618173B (fr) |
WO (1) | WO2020125631A1 (fr) |
Cited By (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN112307968A (zh) * | 2020-10-30 | 2021-02-02 | 天地伟业技术有限公司 | 一种人脸识别特征压缩方法 |
CN112651883A (zh) * | 2021-01-18 | 2021-04-13 | 广东工业大学 | 一种图像高速场景恢复方法、设备及介质 |
CN112907457A (zh) * | 2021-01-19 | 2021-06-04 | Tcl华星光电技术有限公司 | 图像处理方法、图像处理装置及计算机设备 |
CN113824967A (zh) * | 2021-11-23 | 2021-12-21 | 环球数科集团有限公司 | 一种基于深度学习的视频压缩方法 |
CN114565966A (zh) * | 2022-04-26 | 2022-05-31 | 全时云商务服务股份有限公司 | 一种人脸视频图像处理方法及装置 |
CN116228606A (zh) * | 2023-05-09 | 2023-06-06 | 南京茂聚智能科技有限公司 | 一种基于大数据的图像优化处理系统 |
CN117857808A (zh) * | 2024-03-06 | 2024-04-09 | 深圳市旭景数字技术有限公司 | 一种基于数据分类压缩的高效视频传输方法及系统 |
CN118015108A (zh) * | 2024-04-10 | 2024-05-10 | 深圳市丽台电子有限公司 | 基于物联网的智慧物业智能管理方法及系统 |
CN118524258A (zh) * | 2024-07-25 | 2024-08-20 | 浙江嗨皮网络科技有限公司 | 离线视频背景处理方法、系统及可读存储介质 |
Families Citing this family (10)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN109618173B (zh) * | 2018-12-17 | 2021-09-28 | 深圳Tcl新技术有限公司 | 视频压缩方法、装置和计算机可读存储介质 |
CN110059642B (zh) * | 2019-04-23 | 2020-07-31 | 北京海益同展信息科技有限公司 | 人脸图像筛选方法与装置 |
WO2021102702A1 (fr) * | 2019-11-26 | 2021-06-03 | 深圳市大疆创新科技有限公司 | Procédé et appareil de traitement d'image |
CN112334942A (zh) * | 2019-11-26 | 2021-02-05 | 深圳市大疆创新科技有限公司 | 图像处理方法和装置 |
CN111246093B (zh) * | 2020-01-16 | 2021-07-20 | Oppo广东移动通信有限公司 | 图像处理方法、装置、存储介质及电子设备 |
CN112639867A (zh) * | 2020-05-07 | 2021-04-09 | 深圳市大疆创新科技有限公司 | 图像处理方法和装置 |
CN111626310B (zh) * | 2020-05-27 | 2023-08-29 | 百度在线网络技术(北京)有限公司 | 图像比对方法、装置、设备以及存储介质 |
CN111626959A (zh) * | 2020-05-28 | 2020-09-04 | 深圳康佳电子科技有限公司 | 一种图像处理方法、终端及存储介质 |
CN112785512B (zh) * | 2020-06-30 | 2023-05-12 | 青岛经济技术开发区海尔热水器有限公司 | 一种高斯模糊图像处理的优化方法 |
CN112488964B (zh) * | 2020-12-18 | 2024-04-16 | 深圳市镜玩科技有限公司 | 针对滑动列表的图像处理方法、相关装置、设备及介质 |
Citations (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20100266207A1 (en) * | 2009-04-21 | 2010-10-21 | ArcSoft ( Hangzhou) Multimedia Technology Co., Ltd | Focus enhancing method for portrait in digital image |
CN102348116A (zh) * | 2010-08-03 | 2012-02-08 | 株式会社理光 | 视频处理方法、视频处理装置以及视频处理系统 |
CN103152521A (zh) * | 2013-01-30 | 2013-06-12 | 广东欧珀移动通信有限公司 | 一种移动终端中实现景深效果的方法及移动终端 |
CN103632132A (zh) * | 2012-12-11 | 2014-03-12 | 广西工学院 | 一种基于肤色分割和模板匹配的人脸检测与识别方法 |
CN104361558A (zh) * | 2014-10-24 | 2015-02-18 | 小米科技有限责任公司 | 图像处理方法、装置及设备 |
CN104966266A (zh) * | 2015-06-04 | 2015-10-07 | 福建天晴数码有限公司 | 自动模糊身体部位的方法及系统 |
CN107637072A (zh) * | 2015-03-18 | 2018-01-26 | 阿凡达合并第二附属有限责任公司 | 视频会议中的背景修改 |
CN108122194A (zh) * | 2017-11-28 | 2018-06-05 | 沈阳美行科技有限公司 | 一种图像发光的方法及装置 |
CN109618173A (zh) * | 2018-12-17 | 2019-04-12 | 深圳Tcl新技术有限公司 | 视频压缩方法、装置和计算机可读存储介质 |
Family Cites Families (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20060140445A1 (en) * | 2004-03-22 | 2006-06-29 | Cusack Francis J Jr | Method and apparatus for capturing digital facial images optimally suited for manual and automated recognition |
CN100568262C (zh) * | 2007-12-29 | 2009-12-09 | 浙江工业大学 | 基于多摄像机信息融合的人脸识别检测装置 |
WO2015009188A1 (fr) * | 2013-07-19 | 2015-01-22 | Limited Liability Company Mail.Ru | Systèmes et procédés de production automatisée d'interfaces graphiques d'utilisateur |
CN103945118B (zh) * | 2014-03-14 | 2017-06-20 | 华为技术有限公司 | 图像虚化方法、装置及电子设备 |
CN106993112B (zh) * | 2017-03-09 | 2020-01-10 | Oppo广东移动通信有限公司 | 基于景深的背景虚化方法及装置和电子装置 |
CN106971165B (zh) * | 2017-03-29 | 2018-08-10 | 武汉斗鱼网络科技有限公司 | 一种滤镜的实现方法及装置 |
US10147216B1 (en) * | 2017-11-01 | 2018-12-04 | Essential Products, Inc. | Intelligent camera |
CN108230243B (zh) * | 2018-02-09 | 2021-04-27 | 福州大学 | 一种基于显著性区域检测模型背景虚化方法 |
-
2018
- 2018-12-17 CN CN201811546877.5A patent/CN109618173B/zh active Active
-
2019
- 2019-12-17 WO PCT/CN2019/126016 patent/WO2020125631A1/fr active Application Filing
Patent Citations (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20100266207A1 (en) * | 2009-04-21 | 2010-10-21 | ArcSoft ( Hangzhou) Multimedia Technology Co., Ltd | Focus enhancing method for portrait in digital image |
CN102348116A (zh) * | 2010-08-03 | 2012-02-08 | 株式会社理光 | 视频处理方法、视频处理装置以及视频处理系统 |
CN103632132A (zh) * | 2012-12-11 | 2014-03-12 | 广西工学院 | 一种基于肤色分割和模板匹配的人脸检测与识别方法 |
CN103152521A (zh) * | 2013-01-30 | 2013-06-12 | 广东欧珀移动通信有限公司 | 一种移动终端中实现景深效果的方法及移动终端 |
CN104361558A (zh) * | 2014-10-24 | 2015-02-18 | 小米科技有限责任公司 | 图像处理方法、装置及设备 |
CN107637072A (zh) * | 2015-03-18 | 2018-01-26 | 阿凡达合并第二附属有限责任公司 | 视频会议中的背景修改 |
CN104966266A (zh) * | 2015-06-04 | 2015-10-07 | 福建天晴数码有限公司 | 自动模糊身体部位的方法及系统 |
CN108122194A (zh) * | 2017-11-28 | 2018-06-05 | 沈阳美行科技有限公司 | 一种图像发光的方法及装置 |
CN109618173A (zh) * | 2018-12-17 | 2019-04-12 | 深圳Tcl新技术有限公司 | 视频压缩方法、装置和计算机可读存储介质 |
Cited By (13)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN112307968A (zh) * | 2020-10-30 | 2021-02-02 | 天地伟业技术有限公司 | 一种人脸识别特征压缩方法 |
CN112651883B (zh) * | 2021-01-18 | 2023-04-07 | 广东工业大学 | 一种图像高速场景恢复方法、设备及介质 |
CN112651883A (zh) * | 2021-01-18 | 2021-04-13 | 广东工业大学 | 一种图像高速场景恢复方法、设备及介质 |
CN112907457A (zh) * | 2021-01-19 | 2021-06-04 | Tcl华星光电技术有限公司 | 图像处理方法、图像处理装置及计算机设备 |
US12063335B2 (en) | 2021-01-19 | 2024-08-13 | Tcl China Star Optoelectronics Technology Co., Ltd. | Image processing method, image processing device, and computer device |
CN113824967A (zh) * | 2021-11-23 | 2021-12-21 | 环球数科集团有限公司 | 一种基于深度学习的视频压缩方法 |
CN113824967B (zh) * | 2021-11-23 | 2022-02-15 | 环球数科集团有限公司 | 一种基于深度学习的视频压缩方法 |
CN114565966A (zh) * | 2022-04-26 | 2022-05-31 | 全时云商务服务股份有限公司 | 一种人脸视频图像处理方法及装置 |
CN116228606A (zh) * | 2023-05-09 | 2023-06-06 | 南京茂聚智能科技有限公司 | 一种基于大数据的图像优化处理系统 |
CN117857808A (zh) * | 2024-03-06 | 2024-04-09 | 深圳市旭景数字技术有限公司 | 一种基于数据分类压缩的高效视频传输方法及系统 |
CN117857808B (zh) * | 2024-03-06 | 2024-06-04 | 深圳市旭景数字技术有限公司 | 一种基于数据分类压缩的高效视频传输方法及系统 |
CN118015108A (zh) * | 2024-04-10 | 2024-05-10 | 深圳市丽台电子有限公司 | 基于物联网的智慧物业智能管理方法及系统 |
CN118524258A (zh) * | 2024-07-25 | 2024-08-20 | 浙江嗨皮网络科技有限公司 | 离线视频背景处理方法、系统及可读存储介质 |
Also Published As
Publication number | Publication date |
---|---|
CN109618173B (zh) | 2021-09-28 |
CN109618173A (zh) | 2019-04-12 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2020125631A1 (fr) | Procédé et appareil de compression vidéo, et support de stockage lisible par ordinateur | |
CN109272459B (zh) | 图像处理方法、装置、存储介质及电子设备 | |
KR101802146B1 (ko) | 화상처리장치 및 화상처리방법 | |
US8244003B2 (en) | Image preprocessing | |
US8559708B2 (en) | Image preprocessing | |
US10565742B1 (en) | Image processing method and apparatus | |
EP3327668A1 (fr) | Dispositif de traitement d'images et procédé de correction de masque d'avant-plan pour segmentation d'objet | |
CN103973977B (zh) | 一种预览界面的虚化处理方法、装置及电子设备 | |
US8358812B2 (en) | Image Preprocessing | |
US20140079319A1 (en) | Methods for enhancing images and apparatuses using the same | |
US20200126193A1 (en) | Method and device for video processing, electronic device, and storage medium | |
US11917158B2 (en) | Static video recognition | |
US9330340B1 (en) | Noise estimation for images using polynomial relationship for pixel values of image features | |
CN112839167B (zh) | 图像处理方法、装置、电子设备及计算机可读介质 | |
US8244004B2 (en) | Image preprocessing | |
CN115471413A (zh) | 图像处理方法及装置、计算机可读存储介质和电子设备 | |
CN115375592A (zh) | 图像处理方法及装置、计算机可读存储介质和电子设备 | |
WO2022111269A1 (fr) | Procédé et dispositif pour améliorer des détails vidéo, terminal mobile et support de stockage | |
CN114298935A (zh) | 图像增强方法、装置及计算机可读存储介质 | |
CN110796629B (zh) | 一种图像融合方法及系统 | |
CN111797694A (zh) | 一种车牌检测方法及装置 | |
CN113781330B (zh) | 图像处理方法、装置及电子系统 | |
CN112215237B (zh) | 图像处理方法及装置、电子设备和计算机可读存储介质 | |
KR20030091471A (ko) | YCrCb 칼라 기반 얼굴 영역 추출 방법 | |
CN113453017A (zh) | 视频处理方法、装置、设备及计算机程序产品 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 19900350 Country of ref document: EP Kind code of ref document: A1 |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 19900350 Country of ref document: EP Kind code of ref document: A1 |