US12094089B2 - Intelligent video enhancement system - Google Patents
Intelligent video enhancement system Download PDFInfo
- Publication number
- US12094089B2 US12094089B2 US17/474,736 US202117474736A US12094089B2 US 12094089 B2 US12094089 B2 US 12094089B2 US 202117474736 A US202117474736 A US 202117474736A US 12094089 B2 US12094089 B2 US 12094089B2
- Authority
- US
- United States
- Prior art keywords
- video
- features
- user
- operators
- analysis module
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active, expires
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T5/00—Image enhancement or restoration
- G06T5/70—Denoising; Smoothing
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T3/00—Geometric image transformations in the plane of the image
- G06T3/40—Scaling of whole images or parts thereof, e.g. expanding or contracting
- G06T3/4053—Scaling of whole images or parts thereof, e.g. expanding or contracting based on super-resolution, i.e. the output image resolution being higher than the sensor resolution
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T5/00—Image enhancement or restoration
- G06T5/60—Image enhancement or restoration using machine learning, e.g. neural networks
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T5/00—Image enhancement or restoration
- G06T5/90—Dynamic range modification of images or parts thereof
- G06T5/92—Dynamic range modification of images or parts thereof based on global image properties
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10016—Video; Image sequence
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20084—Artificial neural networks [ANN]
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20172—Image enhancement details
- G06T2207/20208—High dynamic range [HDR] image processing
Definitions
- the present invention relates to a system and method for enhancing a video. More specifically, an intelligent video enhancement system that aims to improve visual quality of the video for seamless experience by analysing and processing the video collectively based on preferences of a user.
- Video enhancement is a systematic task, which aims to improve visual quality for better user experience.
- the aim of video enhancement is to improve the visual appearance of the video, or to provide a better transformation representation for future automated video processing, such as analysis, detection, segmentation, recognition, surveillance, traffic, etc.
- the video enhancement techniques also helps analyses background information that is essential to understand object behaviour without requiring expensive human visual inspection.
- digital video is acquired, processed and used, such as surveillance, general identity verification, traffic, criminal justice systems, civilian or military video processing.
- a U.S. Pat. No. 7,746,382 assigned to Microsoft Corp describes an automated video enhancement system and method.
- the automated video enhancement method uses frame-to-frame motion estimation as the basis of the video enhancement.
- the automated video enhancement method includes generating global alignment transforms, generating optic flow vectors, and using these global alignment transforms and optic flow vectors to enhance the video.
- estimation includes the computation of global motion and the computation of local motion.
- the approach is traditional for improving the video for locating the local motion within the video.
- U.S. Pat. No. 7,548,659 assigned to Microsoft Corp discloses a video enhancement technique for improving visual experience of multiple image frames of a video sequence using a video completion technique.
- the missing image areas in the video sequence filled in using a video completion technique that involves motion in painting.
- Local motion data of a motion map from known image areas propagated into the missing image areas to produce an extended motion map.
- the patent uses traditional approach to obtain optical flow, which helps to locate the local motion within the video.
- the system disclosed improves the quality of a video to some extent, but the system explained has limited flexibility.
- a video format converter may be used to process a low-resolution video signal from a media-providing device before the video is displayed.
- the video format converter may detect the true resolution of the video and DE interlaces the video signal accordingly.
- the video format converter may reduce compression artifacts and apply techniques to enhance the appearance of the video.
- Video real resolution is detected to determine how to reduce compression artifacts and apply techniques to enhance the appearance of the video.
- U.S. Pat. No. 10,523,955 assigned to Magic Pony Technology Ltd described a method for enhancing at least a section of lower-quality visual data.
- the visual data enhances a hierarchical algorithm is then selected from a plurality of hierarchical algorithms, wherein the step of selection is based on a predetermined metric and wherein the hierarchical algorithms were developed using a learned approach and at least one of the hierarchical algorithms is operable to increase the quality of the lower-quality visual data.
- the selected hierarchical algorithm is then used to increase the quality of the lower-quality visual data to create a higher-quality visual data.
- the patent involves a learned approach and at least one of hierarchical algorithms to only increase the video resolution.
- WO2012068008 assigned to Cisco Technology Inc. discloses a method for providing enhanced graphics in a video environment. The method described evaluates a plurality of pixels within the video input, and determines if a particular pixel among the pixels is associated with a colour that matches a designated value associated with a chroma-key to adjust the corresponding colour by a certain percentage.
- Another PCT Publication WO2008011202 assigned to Qualcomm Inc discloses a devices and methods for processing multimedia data to generate enhanced multimedia data at a receiver based on encoder assisted pre-processing. Although the patent described processes multimedia data, which includes identifying at least a one-pixel intensity range. This method is able to modify dynamic range in pixel level.
- the video enhancing system includes a dynamic analysis module and an adaptive processing module for analysing and processing the video respectively for enhancing the quality of the video.
- the dynamic analysis module includes a scanner for scanning fundamental information related to the video.
- the scanner scans the video to generate a scanned video.
- the dynamic analysis module includes a pair of extractors for extracting multiple features associated with the video.
- the pair of extractors includes a low-level feature extractor and a high-level feature extractor.
- the low-level feature extractor extracts multiple inherent features.
- the high-level feature extractor extracts multiple unique features.
- the dynamic analysis module includes a feature analysis module.
- the feature analysis unit analyzes the scanned video and the multiple features to develop a recommended workflow.
- the video enhancement system includes the adaptive processing core module for triggering multiple operators in reference to the recommended workflow.
- the adaptive processing controller includes an operator pool for storing multiple operators.
- the adaptive processing module includes a processing controller.
- the processing controller commands and controls an operator pool.
- the processing controller commands the operator pool for selecting the multiple operators and further triggers the multiple operators from the operator pool based on the recommended workflow.
- the dynamic analysis module and the adaptive processing unit module collectively enhance the video quality.
- a primary objective of the present invention is to provide a video enhancement system for enhancing a video with better visual quality for improved user experience.
- Another objective of the present invention is to provide a video enhancing system that is intelligent and self-learning.
- Yet another objective of the present invention is to provide a flexible video enhancing system based on video corresponding features and personal preference.
- Another objective of the present invention is to provide the video enhancing system that is flexible and intelligent system with Dynamic Analysis module (DAM) or Dynamic Analysis Model to recommend an enhancement workflow based on analysis of the original video and users' preference.
- DAM Dynamic Analysis module
- Dynamic Analysis Model to recommend an enhancement workflow based on analysis of the original video and users' preference.
- Yet another objective of the present invention is to provide the video enhancing system that proposes Adaptive Processing Core (APC) module that triggers appropriate operators for processing the video to enhance the quality of the video.
- APC Adaptive Processing Core
- the dynamic analysis module based on deep learning model that recommend an enhancement workflow based on analysis of the original video and users' preference.
- the scanner scans video resolution, video compressing rate, video colour space and video bit depth.
- the multiple inherent features extracted by the low-level feature extractor pixel intensity and global gradient.
- the low-level feature extractor further measures pixel intensity and global gradient of the video.
- the multiple unique features are video content, video semantic, video scenario and video style extracted by the high-level feature extractor.
- the dynamic analysis module further includes a user preference unit.
- the user preference unit analyzes multiple preferences of a user.
- the user preference unit includes a feature analyzer and a feedback analyzer.
- the feature analyzer analyzes multiple features associated with the multiple preferences of the user.
- the user preference unit includes a feedback analyzer unit.
- the feedback analyzer unit analyzes a feedback provided by the user.
- the dynamic analysis module includes a feature analysis module. The feature analysis unit analyzes the scanned video, the inherent features, the unique features and the multiple preferences of the user to develop a recommended workflow with respect to the original video.
- the multiple user preferences are mood of the user, liking of the user, disliking of the user.
- the dynamic analysis module unit receives a user feedback periodically.
- the user feedback tends the intelligent video enhancing system to improve multiple functionalities of the video enhancing system.
- the adaptive processing core module works on basis of deep learning model, which triggers appropriate operators, provides each with corresponding parameters, communicates between all processing units, and enables them work cooperatively to outperform a simple collection of these processing units.
- the multiple operators are super resolution, video de-noise, high dynamic range, de-flicker, slow motion, colour enhancement.
- the adaptive processing module further includes an operator pool that includes multiple operators based on the multiple preferences of the user. In one another embodiment, the adaptive processing module further includes a plurality of controllers to control the operator pool.
- the adaptive processing module triggers multiple operators from the operator pool based on the recommended work-flow.
- the dynamic analysis module and the adaptive processing module enhance the video intelligently and flexibly based on multiple user preferences.
- the adaptive processing module works on basis of deep learning model, which triggers appropriate operators, provides each with corresponding parameters, communicates between all processing units, and enables them work cooperatively to outperform a simple collection of these processing units.
- the multiple operators are super resolution, video de-noise, high dynamic range, de-flicker, slow motion, colour enhancement.
- the adaptive processing module further includes an operator pool that includes multiple operators based on the multiple preferences of the user.
- the adaptive processing controller includes an operator pool for storing multiple operators.
- the adaptive processing module includes a processing controller.
- the processing controller controls an operator pool.
- the processing controller commands the operator pool for selecting the multiple operators and further triggers the multiple operators from the operator pool based on the recommended workflow.
- the dynamic analysis module and the adaptive processing unit module collectively enhance the video quality.
- the dynamic analysis module and the adaptive processing module enhance the video intelligently and flexibly based on multiple user preferences.
- FIG. 1 (A) illustrates an intelligent video enhancement system (IVES) based on a user preferences in accordance with the present invention
- FIG. 1 (B) illustrates an intelligent video enhancement system (IVES) in accordance with the present invention
- FIG. 2 illustrates a dynamic analysis module (DAM) in accordance with the present invention
- FIG. 3 illustrates an adaptive processing core (APC) module in accordance with the present invention
- FIG. 4 illustrates a workflow diagram of the adaptive processing core (APC) module in accordance with the present invention.
- FIG. 5 illustrates a method of enhancing quality of a video intelligently in accordance with the present invention.
- the present invention discloses a flexible, self-learning and an intelligent video enhancing system based on video corresponding features and personal preference with better visual quality for improved user experience.
- the video enhancing system includes a dynamic analysis module and an adaptive processing module for analysing and processing the video respectively.
- the dynamic analysis module includes a scanner for scanning the video to generate a scanned video. Further, the dynamic analysis module includes a pair of extractors for extracting multiple features associated with the video. Further, the dynamic analysis module includes a feature analysis module. The feature analysis unit analyzes the scanned video and the multiple features to develop a recommended workflow.
- the video enhancement system includes an adaptive processing module for triggering multiple operators in reference to the recommended workflow.
- the adaptive processing core module includes an operator pool for storing multiple operators.
- the adaptive processing module includes a processing controller. The processing controller controls an operator pool. The processing controller commands the operator pool to select multiple operators and then the processing controller triggers the selected operator in accordance with the recommended workflow generated by the dynamic analysis module to enhance the video.
- FIG. 1 (A) illustrates a framework of a video enhancing system ( 100 ) for enhancing the video ( 102 ) for improved user experience.
- the video enhancing system ( 100 ) is a part of whole video system, indicating that it does not involve any video capture, recording, encoding-decoding, display, transportation, etc., and only concerns with video enhancement.
- a video signal or command first provided or inputted to the video enhancing system ( 100 ).
- the video enhancing system ( 100 ) includes a dynamic analysis module ( 104 ) to recommend an enhancement workflow based on analysis of the original video and users' preference.
- the video ( 102 ) firstly goes through the dynamic analysis module ( 104 ).
- the dynamic analysis module ( 104 ) analysis runs analysis of the video dynamically to recommend a workflow upon which further processing of the video ( 102 ).
- the dynamic analysis module ( 104 ) involves scanning of video fundamental information, low-level feature extraction, high-level feature extraction, etc.
- the video enhancing system ( 100 ) further includes a user preference ( 106 ) providing unit to analyze multiple preferences of a user with respect to the video.
- the user preference ( 106 ) is taken in account to improve user experience.
- the dynamic analysis module ( 104 ) analyses the video ( 102 ) with respect to the user preferences ( 106 ) to provide a recommended workflow. To determine the target video quality and amount of enhancement, the dynamic analysis module ( 104 ) needs to send information to an Adaptive Processing Core (APC) after completing the corresponding analysis.
- API Adaptive Processing Core
- the video enhancing system ( 100 ) further includes an adaptive processing core module ( 108 ) for triggering multiple operators in reference to the recommended work-flow.
- the adaptive processing module ( 108 ) includes an operator pool for storing multiple operators in correspondence to the video enhancement.
- the adaptive processing module ( 108 ) includes a processing controller.
- the processing controller controls the operator pool.
- the processing controller commands the operator pool for selecting the multiple operators and further triggers the multiple operators from the operator pool based on the recommended workflow.
- the dynamic analysis module ( 104 ) and the adaptive processing module ( 108 ) collectively enhance the video quality.
- the dynamic analysis module ( 104 ) and the adaptive processing module ( 108 ) collectively enhance the video ( 102 ) based on users preferences ( 106 ).
- the enhanced video ( 110 ) outputted to next step, such as video display, storage, processing system, etc.
- FIG. 1 (B) illustrates a framework of a video enhancing system ( 100 ) for enhancing the video ( 102 ) for improved user experience.
- the video enhancing system ( 100 ) includes a dynamic analysis module ( 104 ) to recommend an enhancement workflow based on analysis of the original video and users' preference.
- the video ( 102 ) firstly goes through the dynamic analysis module ( 104 ).
- the dynamic analysis module ( 104 ) analysis runs analysis of the video dynamically to recommend a workflow upon which further processing of the video ( 102 ).
- the dynamic analysis module ( 104 ) involves scanning of video fundamental information, low-level feature extraction, high-level feature extraction, etc.
- the video enhancing system ( 100 ) further includes an adaptive processing core module ( 108 ) for triggering multiple operators in reference to the recommended work-flow.
- the adaptive processing module ( 108 ) includes an operator pool for storing multiple operators in correspondence to the video enhancement.
- the adaptive processing module ( 108 ) includes a processing controller. The processing controller controls the operator pool.
- the processing controller commands the operator pool for selecting the multiple operators and further triggers the multiple operators from the operator pool based on the recommended workflow.
- the dynamic analysis module ( 104 ) and the adaptive processing module ( 108 ) collectively enhance the video quality.
- the dynamic analysis module ( 104 ) and the adaptive processing module ( 108 ) collectively enhance the video ( 102 ). Finally, the enhanced video ( 110 ) outputted to next step, such as video display, storage, processing system, etc.
- FIG. 2 illustrates recommended dynamic analysis module ( 104 ) framework including flow of the video ( 102 ) through different sub-blocks or components along with user's preference to generate a recommended workflow.
- the main purpose of this dynamic analysis module ( 104 ) is video analysis in which huge video information extracted before being processed.
- the dynamic analysis module ( 104 ) includes several sub-blocks to analyze not only concrete but also abstract features.
- the dynamic analysis module ( 104 ) includes a scanner ( 112 ) for scanning a video ( 102 ).
- the scanner ( 112 ) scans the video ( 102 ) for scanning fundamental information with respect to the video ( 102 ).
- the fundamental information includes video resolution, video compression rate, video color space, video bit depth and other features. In one example, for videos compressed by H.264, H.265, MPEG, etc., much data could usually be extracted from their compression information. This includes video resolution, video compression rate, video color space, video bit depth, etc., which could be organized into a table for the joint inference later.
- the dynamic analysis module ( 104 ) further includes a pair of extractor ( 114 , 116 ) for extracting features from the video ( 102 ).
- the dynamic analysis module ( 104 ) includes a low-level feature extractor ( 114 ) for extracting multiple inherent features of the video ( 102 ).
- the low-level feature extractor ( 114 ) measures the multiple inherent features firstly and performs operation to extract features from the video ( 102 ).
- the low-level feature extractor ( 114 ) selects some low frequency patches from the video frames to calculate variance of pixel intensity, which is a reference noise level useful for the de-noising operation. Additionally, a resolution level of the video ( 102 ) evaluated by calculating the global gradient as resolution does not only refer to its size but also actual manifested resolution.
- the low-level feature extractor ( 114 ) figures out the actual enhancement operation required by each video in more comprehensive way.
- the dynamic analysis module ( 104 ) includes a high-level feature extractor ( 116 ) for extracting unique features of the video ( 102 ).
- the unique features include content, semantic, scenario, style, and alike of the video ( 102 ). Unlike the inherent features, these unique features are indescribable and unquantifiable.
- the high-level feature extractor ( 116 ) efficient to be extracted by adopting deep learning model as it is known that deep-learning network pays attention to object abstract features. Network could recognize or localize corresponding content in sequence. Furthermore, if a temporal model equipped to expand analysis on time dimension, the moving object would be recognized and predicted. With this abstract information, the specific content in video can be reasonably enhanced.
- the dynamic analysis module ( 104 ) includes a user preferences analysis unit ( 118 ) to analyze multiple preferences of a user.
- the user preference unit ( 118 ) includes a feature analyzer ( 120 ) to analyze multiple features associated with the multiple preferences of the user.
- the user preference unit ( 118 ) further includes a feedback analyzer unit ( 122 ) to analyze a feedback provided by the user.
- a recommended approach conducted to a quick video performance survey. For instance, at the end of each video, a few randomly selected clips could be displayed and graded by the users in several aspects such as aesthetic, resolution, comfort ability, view definition, etc. These subjective quality evaluations can be structured for DAM analysis, similar to other extracted features.
- the dynamic analysis module includes a feature analysis module ( 124 ).
- the feature analysis unit ( 124 ) analyzes the scanned video, the multiple features, and the multiple preferences of the user to develop a recommended workflow ( 126 ) with respect to the original video ( 102 ).
- the multiple user preferences are mood of the user, liking of the user, disliking of the user.
- the dynamic analysis module ( 104 ) unit receives a user feedback periodically.
- the user feedback tends the intelligent video enhancing system ( 100 ) to improve multiple functionalities of the intelligent video enhancing system ( 100 ).
- the purpose of gathering features using the respective sub-units described above is to help understand the video and recommend the best recommended workflow to APC ( 108 ).
- the feature analysis module ( 124 ) analyses the unstructured data, such as high-level features, and structured data, such as fundamental information of the video, low-level features and user preference encoded by the convolution neural networks (CNN) respective sub-units would be further processed. All encoded information further embedded and passed through a classifier, which further followed by identification of the best workflow to use in improving the video quality.
- CNN convolution neural networks
- a network with two VGG-style convolution neural networks (CNN) branches are to be set up, one branch for extracting video sequence features while the other for extracting synchronous audio features.
- Features resulting from the two branches are concatenated and passed through the fully connected layers to analyze video scenes representation, such as video captioning also adopts a deep-learning network to extract video and audio features for video scene recognition task.
- 3D-CNN concatenates short-term features and the long-term feature bank to attention blocks in order to analyze video scene features comprehensively to reveal short-term relationship between features, considers long-term relationship through a long-term feature bank.
- the DAM incorporates these features and reveals their implied relationship using a deep learning model, efficiently trained with a well-designed target.
- FIG. 3 illustrates an adaptive processing module ( 108 ) for processing the video ( 102 ) based on recommended workflow ( 126 ) generated through the dynamic analysis module ( 104 ) for producing the enhanced video ( 112 ).
- the adaptive processing module ( 108 ) works on basis of deep learning model, which triggers appropriate operators as DAM ( 104 ) required, provides each with corresponding parameters, communicates between all processing units, and enables them work cooperatively to outperform a simple collection of these processing units.
- the adaptive processing module ( 108 ) run the selected operator with specific parameters, cache intermediate data on which next activated operator makes progress, etc.
- the adaptive processing module ( 108 ) includes a processing controller ( 128 ) for activating and controlling specific operators from an operator pool ( 130 ) for video enhancement.
- the operator pool ( 130 ) includes multiple operators based on the user's preferences and the features related to the video ( 102 ).
- the adaptive processing module ( 108 ) enhances the video intelligently and flexibly based on multiple user preferences.
- the operator pools ( 130 ) includes the multiple operators based on deep learning model such as but not limited to super resolution, video de-noise, high dynamic range, de-flicker, slow motion, colour enhancement explained below in detail.
- the operator named as super resolution from the operator pool for enhancing resolution of the video is discussed.
- the recommended network explained may contain many popular blocks to consider varied situations of video. If inherent pattern needs to be reconstructed by stacking the residual block or by back-projection, the deep learning model within the APC ( 108 ) model would be a good choice to achieve large receptive field.
- deformable convolutional network instead of traditional convolution network within the APC ( 108 ) overcome the shortcoming of the latter, i.e. unable to process deformable object and features.
- the operator named as video de-noising from the operator pool for enhancing the video by removing unwanted noises discussed.
- the noise model estimated by DAM ( 104 ) used as a factor that adaptively controls de-noise effect with a deep-learning network.
- U-Net based network has been proved useful for de-noising at most times. The two approaches suitable for de-noising the video, firstly we can reduce noise directly or by using, predict pixel-wise noise kernel.
- a high dynamic range (HDR) operator of the video from the operator pool ( 130 ) for improving visual experience discussed.
- HDR high dynamic range
- a high dynamic range video gives a better visual experience.
- the video dynamic range enhanced at corresponding area using the abstract feature model.
- the diverse frequency information takes into account for not only overexposure but also underexposure.
- the network also learns non-linear transformation such as tone-mapping curve.
- De-flicker associated with the video ( 102 ) need to be considered for improving or enhancing the video ( 102 ).
- Video flickering harms user experience, especially in extreme motion scenes such as those in action movies.
- a temporal model used within the APC ( 108 ) to improve the video quality by reducing the de-flickering A 3D convolution network or recurrent network also used for this purpose.
- the operator known as slow motion associated with the video ( 102 ) used for improving or enhancing the video ( 102 ).
- the method used for improving motion within the video ( 102 ) is to develop a depth-aware flow projection layer to synthesize intermediate flows that preferably sample closer objects than further ones.
- the learned intermediate flow maps help deep learning network to interpolate frames properly.
- a one-stage space-time video super resolution framework approach can be used approach used a one-stage space-time video super resolution framework, which directly synthesizes an HR slow-motion video from a low frame rate.
- the operator know color enhancement used for enhancing color experience of the video.
- color enhancement used for enhancing color experience of the video.
- an approach used which uses real old images and synthetic color degraded images as inputs to train the first encoder such that it maps them to the latent space for corrupted images.
- the second encoder within the APC ( 108 ) works on synthetic old clean images and maps them to another latent space for clean images.
- the CNN network would then learn the mapping between the two latent spaces through adversarial learning.
- the APC ( 108 ) uses the mapping to recover the color of the real old image.
- FIG. 4 illustrates an example of working of APC ( 108 ) based on recommended workflow ( 126 ).
- dynamic analysis module ( 104 ) analyzes a specific video ( 102 ) source, it will send an instruction which involves de-noise, colour enhancement, super resolution, and HDR in form of recommended workflow ( 126 ).
- the Slow-motion operator ( 140 ) is an optional processing, as it based on personal preference, the user only wants to increase the frame rate at certain scenes.
- the processing controller ( 128 ) inside APC ( 108 ) triggered.
- the processing controller ( 128 ) will first activate the de-noise operator ( 132 ), which requires the noise model information identified earlier by DAM ( 104 ) to reduce noise properly.
- the noise-free sequence separated into low-frequency portion and high-frequency portion by a filter such as bilateral filter.
- the adaptive colour enhancement operator ( 134 ) will enhance the low frequency portion of the noise-free sequence accordingly.
- the processing controller ( 128 ) uses the filter from the previous step to extract high frequency portions from the original noisy video.
- All three sequences will be concatenated and sent to a super-resolution operator ( 136 ) because the super resolution model is trained to have the ability to refer the noise-free sequence and filter only useful information from the original high-frequency portion to reconstruction the resolution.
- the super resolution operator ( 136 ) needs both target resolution factor and the video style information to select the proper deep learning model.
- HDR operator ( 138 ) takes into account the video style and personal preference to give the sequence a high dynamic range so that the scene becomes more visually pleasing.
- the processing controller ( 128 ) at the end of recommended workflow ( 126 ) will activate the slow-motion operator ( 140 ). Based on the reasons mentioned above, this slow-motion operator ( 140 ) would also require the video style and personal preference information from dynamic analysis module (DAM) ( 104 ).
- DAM dynamic analysis module
- FIG. 5 illustrates a method ( 200 ) for enhancing a video ( 102 ) based on a user's preference and features to be enhanced or improved for providing better video experience of the video.
- step 202 firstly, the original video ( 102 ) inputted to a video enhancing system ( 100 ).
- the video ( 102 ) inputted to a dynamic analysis module ( 104 ) of the IVES ( 100 ).
- the video ( 102 ) scanned by a scanner ( 112 ) of the dynamic analysis module ( 104 ).
- the pair of extractor ( 114 , 116 ) extracts multiple inherent features and the features associated with the video ( 102 ).
- the low-level feature extractor ( 114 ) measures the multiple inherent features firstly and performs operation to extract features from the video ( 102 ).
- the low-level feature extractor ( 114 ) selects some low frequency patches from the video frames to calculate variance of pixel intensity, which is a reference noise level useful for the de-noising operation.
- the high-level feature extractor ( 116 ) extracts unique features of the video ( 102 ).
- the unique features include content, semantic, scenario, style, and alike of the video ( 102 ).
- a user preferences analysis unit ( 118 ) adds multiple preferences of a user associated with the video ( 102 ).
- the user preference unit ( 118 ) includes a feature analyzer ( 120 ) to analyze multiple features associated with the multiple preferences of the user.
- the user preference unit ( 118 ) further includes a feedback analyzer unit ( 122 ) to periodically analyze a feedback provided by the user.
- the feature analysis unit ( 124 ) analyzes the scanned video and the multiple features and the multiple preferences of the user to develop a recommended workflow ( 126 ) with respect to the original video ( 102 ).
- a recommended workflow generated based on analysis of the DAM ( 104 ).
- selecting one or more operators form an operator pool ( 130 ) based on the recommended work-flow by the processing controller ( 128 ) of the adaptive processing module ( 108 ), the adaptive processing module ( 108 ) on basis of deep learning model, which triggers appropriate operators as DAM ( 104 ) required, provides each with corresponding parameters, communicates between all processing units, and enables them work cooperatively to outperform a simple collection of these processing units.
- the adaptive processing module ( 108 ) run the selected operator with specific parameters, cache intermediate data on which next activated operator makes progress, etc.
- triggering the multiple operators associated with the recommended workflow ( 126 ) to intelligently enhance the video The enhanced video outputted to video playing devices such as video display, storage, processing system, etc. According to the user's satisfaction with the resulting enhanced video, their feedback sent back to DAM ( 104 ) for the use of the next personal preference analysis.
Landscapes
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Image Analysis (AREA)
Abstract
Description
Claims (14)
Priority Applications (2)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| US17/474,736 US12094089B2 (en) | 2021-09-14 | 2021-09-14 | Intelligent video enhancement system |
| CN202211116440.4A CN115665477A (en) | 2021-09-14 | 2022-09-14 | Video enhancement system, method and device |
Applications Claiming Priority (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| US17/474,736 US12094089B2 (en) | 2021-09-14 | 2021-09-14 | Intelligent video enhancement system |
Publications (2)
| Publication Number | Publication Date |
|---|---|
| US20230081916A1 US20230081916A1 (en) | 2023-03-16 |
| US12094089B2 true US12094089B2 (en) | 2024-09-17 |
Family
ID=84983269
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| US17/474,736 Active 2042-11-13 US12094089B2 (en) | 2021-09-14 | 2021-09-14 | Intelligent video enhancement system |
Country Status (2)
| Country | Link |
|---|---|
| US (1) | US12094089B2 (en) |
| CN (1) | CN115665477A (en) |
Families Citing this family (2)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US12250400B2 (en) * | 2022-02-14 | 2025-03-11 | Microsoft Technology Licensing, Llc | Unified space-time interpolation of video information |
| CN119583851A (en) * | 2024-12-23 | 2025-03-07 | 上海幻电信息科技有限公司 | Video conversion method and device |
Citations (15)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US7262818B2 (en) | 2004-01-02 | 2007-08-28 | Trumpion Microelectronic Inc. | Video system with de-motion-blur processing |
| WO2008011502A2 (en) | 2006-07-20 | 2008-01-24 | Qualcomm Incorporated | Dynamic range reduction before video encoding |
| US7548659B2 (en) | 2005-05-13 | 2009-06-16 | Microsoft Corporation | Video enhancement |
| US7746382B2 (en) | 2002-06-28 | 2010-06-29 | Microsoft Corporation | Video processing system and method for automatic enhancement of digital video |
| WO2012068008A1 (en) | 2010-11-15 | 2012-05-24 | Cisco Technology, Inc. | System and method for providing enhanced graphics in a video environment |
| US8269886B2 (en) | 2007-01-05 | 2012-09-18 | Marvell World Trade Ltd. | Methods and systems for improving low-resolution video |
| US10523955B2 (en) | 2015-02-19 | 2019-12-31 | Magic Pony Technology Limited | Enhancement of visual data |
| US20210097288A1 (en) * | 2019-09-26 | 2021-04-01 | Samsung Electronics Co., Ltd. | Method and system for generating video |
| US20210160571A1 (en) * | 2019-11-27 | 2021-05-27 | Rovi Guides, Inc. | Systems and methods for deep recommendations using signature analysis |
| WO2021108707A1 (en) * | 2019-11-27 | 2021-06-03 | Rovi Guides, Inc. | Systems and methods for deep recommendations using signature analysis |
| US20210281918A1 (en) * | 2019-04-23 | 2021-09-09 | Tencent Technology (Shenzhen) Company Limited | Video recommendation method and device, computer device and storage medium |
| US20210312318A1 (en) * | 2020-04-02 | 2021-10-07 | Rovi Guides, Inc. | Systems and methods for automated content curation using signature analysis |
| US20220159086A1 (en) * | 2020-11-17 | 2022-05-19 | Rovi Guides, Inc. | Systems and methods for adjusting storage based on determining content item popularity |
| US20220398402A1 (en) * | 2021-06-15 | 2022-12-15 | Lemon Inc. | Detecting objects in a video using attention models |
| US20230024037A1 (en) * | 2021-07-21 | 2023-01-26 | Avago Technologies International Sales Pte. Limited | Methods for non-reference video-quality prediction |
Family Cites Families (9)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN102158699B (en) * | 2011-03-28 | 2012-12-12 | 广州市聚晖电子科技有限公司 | Embedded video compression coding system with image enhancement function |
| CN103793888B (en) * | 2014-02-18 | 2017-01-11 | 厦门美图网科技有限公司 | Image enhancing method based on main colors of reference image |
| CN106603859A (en) * | 2016-12-30 | 2017-04-26 | 努比亚技术有限公司 | Photo filter processing method, device and terminal |
| CN109002857B (en) * | 2018-07-23 | 2020-12-29 | 厦门大学 | A method and system for video style transformation and automatic generation based on deep learning |
| CN109525901B (en) * | 2018-11-27 | 2020-08-25 | Oppo广东移动通信有限公司 | Video processing method, apparatus, electronic device and computer readable medium |
| CN110766639B (en) * | 2019-10-30 | 2022-09-27 | 北京迈格威科技有限公司 | Image enhancement method, apparatus, mobile device, and computer-readable storage medium |
| CN112019827B (en) * | 2020-09-02 | 2022-01-25 | 上海网达软件股份有限公司 | Method, device, equipment and storage medium for enhancing video image color |
| CN113038000A (en) * | 2021-02-26 | 2021-06-25 | 惠州Tcl移动通信有限公司 | Shooting prompt processing method, device, equipment and medium based on image analysis |
| CN112991203B (en) * | 2021-03-08 | 2024-05-07 | Oppo广东移动通信有限公司 | Image processing method, device, electronic equipment and storage medium |
-
2021
- 2021-09-14 US US17/474,736 patent/US12094089B2/en active Active
-
2022
- 2022-09-14 CN CN202211116440.4A patent/CN115665477A/en active Pending
Patent Citations (15)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US7746382B2 (en) | 2002-06-28 | 2010-06-29 | Microsoft Corporation | Video processing system and method for automatic enhancement of digital video |
| US7262818B2 (en) | 2004-01-02 | 2007-08-28 | Trumpion Microelectronic Inc. | Video system with de-motion-blur processing |
| US7548659B2 (en) | 2005-05-13 | 2009-06-16 | Microsoft Corporation | Video enhancement |
| WO2008011502A2 (en) | 2006-07-20 | 2008-01-24 | Qualcomm Incorporated | Dynamic range reduction before video encoding |
| US8269886B2 (en) | 2007-01-05 | 2012-09-18 | Marvell World Trade Ltd. | Methods and systems for improving low-resolution video |
| WO2012068008A1 (en) | 2010-11-15 | 2012-05-24 | Cisco Technology, Inc. | System and method for providing enhanced graphics in a video environment |
| US10523955B2 (en) | 2015-02-19 | 2019-12-31 | Magic Pony Technology Limited | Enhancement of visual data |
| US20210281918A1 (en) * | 2019-04-23 | 2021-09-09 | Tencent Technology (Shenzhen) Company Limited | Video recommendation method and device, computer device and storage medium |
| US20210097288A1 (en) * | 2019-09-26 | 2021-04-01 | Samsung Electronics Co., Ltd. | Method and system for generating video |
| US20210160571A1 (en) * | 2019-11-27 | 2021-05-27 | Rovi Guides, Inc. | Systems and methods for deep recommendations using signature analysis |
| WO2021108707A1 (en) * | 2019-11-27 | 2021-06-03 | Rovi Guides, Inc. | Systems and methods for deep recommendations using signature analysis |
| US20210312318A1 (en) * | 2020-04-02 | 2021-10-07 | Rovi Guides, Inc. | Systems and methods for automated content curation using signature analysis |
| US20220159086A1 (en) * | 2020-11-17 | 2022-05-19 | Rovi Guides, Inc. | Systems and methods for adjusting storage based on determining content item popularity |
| US20220398402A1 (en) * | 2021-06-15 | 2022-12-15 | Lemon Inc. | Detecting objects in a video using attention models |
| US20230024037A1 (en) * | 2021-07-21 | 2023-01-26 | Avago Technologies International Sales Pte. Limited | Methods for non-reference video-quality prediction |
Also Published As
| Publication number | Publication date |
|---|---|
| CN115665477A (en) | 2023-01-31 |
| US20230081916A1 (en) | 2023-03-16 |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| US10586312B2 (en) | Method for image processing and video compression with sparse zone salient features | |
| Yang et al. | Spatiotemporal generative adversarial network-based dynamic texture synthesis for surveillance video coding | |
| Li et al. | Weight-based R-λ rate control for perceptual HEVC coding on conversational videos | |
| CN108337465B (en) | Video processing method and device | |
| CN114339030B (en) | Network live video image stabilizing method based on self-adaptive separable convolution | |
| US20070147504A1 (en) | Selecting key frames from video frames | |
| CN110620924B (en) | Method and device for processing coded data, computer equipment and storage medium | |
| US12094089B2 (en) | Intelligent video enhancement system | |
| JP2011517228A (en) | System and method for improving visibility of objects in digital images | |
| CN106550244A (en) | The picture quality enhancement method and device of video image | |
| CN119450126B (en) | Image processing method and device for intelligent interaction control unit of television box | |
| CN109640151A (en) | Video processing method, video processing device, electronic equipment and storage medium | |
| CN120321436A (en) | Video stream transmission method, device and storage medium | |
| US11570465B2 (en) | Machine-learned in-loop predictor for video compression | |
| CN114500879A (en) | Video data processing method, device, equipment and storage medium | |
| CN120410929A (en) | Facial image enhancement method, device, equipment and storage medium | |
| CN118233675B (en) | Live image optimization method and system based on artificial intelligence | |
| US20180090172A1 (en) | Automated seamless video loop | |
| Wang et al. | Deep intensity guidance based compression artifacts reduction for depth map | |
| US12452485B2 (en) | Electronic device, system, and method for intelligent horizontal-vertical image transform | |
| US12159443B2 (en) | Previewing and generating stroke outlines in association with images | |
| Chen et al. | Model-based portrait video compression with spatial constraint and adaptive pose processing | |
| Gong et al. | Temporal transformer-based video super-resolution reconstruction with cross-modal attention | |
| Zhou et al. | Removing banding artifacts in hdr videos generated from inverse tone mapping | |
| Kumar et al. | Adaptive fusion of dual SRGANs and top-hat gradient domain filtering for enhancing low-light videos on mobile devices |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| FEPP | Fee payment procedure |
Free format text: ENTITY STATUS SET TO UNDISCOUNTED (ORIGINAL EVENT CODE: BIG.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY |
|
| FEPP | Fee payment procedure |
Free format text: ENTITY STATUS SET TO SMALL (ORIGINAL EVENT CODE: SMAL); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY |
|
| AS | Assignment |
Owner name: BLACK SESAME TECHNOLOGIES INC., CALIFORNIA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:CHEE, EVELYN;SHEN, SHANLAN;DUAN, YUBO;REEL/FRAME:057923/0492 Effective date: 20210901 |
|
| FEPP | Fee payment procedure |
Free format text: ENTITY STATUS SET TO UNDISCOUNTED (ORIGINAL EVENT CODE: BIG.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY |
|
| FEPP | Fee payment procedure |
Free format text: ENTITY STATUS SET TO SMALL (ORIGINAL EVENT CODE: SMAL); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY |
|
| FEPP | Fee payment procedure |
Free format text: ENTITY STATUS SET TO UNDISCOUNTED (ORIGINAL EVENT CODE: BIG.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY |
|
| FEPP | Fee payment procedure |
Free format text: ENTITY STATUS SET TO SMALL (ORIGINAL EVENT CODE: SMAL); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY Free format text: ENTITY STATUS SET TO UNDISCOUNTED (ORIGINAL EVENT CODE: BIG.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: NON FINAL ACTION MAILED |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: NOTICE OF ALLOWANCE MAILED -- APPLICATION RECEIVED IN OFFICE OF PUBLICATIONS |
|
| ZAAB | Notice of allowance mailed |
Free format text: ORIGINAL CODE: MN/=. |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: PUBLICATIONS -- ISSUE FEE PAYMENT RECEIVED |
|
| STCF | Information on status: patent grant |
Free format text: PATENTED CASE |