CN113973175A - Rapid HDR video reconstruction method - Google Patents
Rapid HDR video reconstruction method Download PDFInfo
- Publication number
- CN113973175A CN113973175A CN202110993299.5A CN202110993299A CN113973175A CN 113973175 A CN113973175 A CN 113973175A CN 202110993299 A CN202110993299 A CN 202110993299A CN 113973175 A CN113973175 A CN 113973175A
- Authority
- CN
- China
- Prior art keywords
- hdr
- foreground
- background
- video
- frame
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/95—Computational photography systems, e.g. light-field imaging systems
- H04N23/951—Computational photography systems, e.g. light-field imaging systems by using two or more images to influence resolution, frame rate or aspect ratio
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- Computing Systems (AREA)
- General Health & Medical Sciences (AREA)
- General Engineering & Computer Science (AREA)
- Biophysics (AREA)
- Computational Linguistics (AREA)
- Data Mining & Analysis (AREA)
- Evolutionary Computation (AREA)
- Artificial Intelligence (AREA)
- Molecular Biology (AREA)
- Life Sciences & Earth Sciences (AREA)
- Biomedical Technology (AREA)
- General Physics & Mathematics (AREA)
- Mathematical Physics (AREA)
- Software Systems (AREA)
- Health & Medical Sciences (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Image Analysis (AREA)
Abstract
High quality video is important in SLAM and video surveillance, which can clearly monitor objects or activities of interest and improve the positioning accuracy of SLAM. The conversion of low quality video to high quality video is becoming more and more a key to the robotic task, and High Dynamic Range (HDR) imaging technology represents high quality images and is finding wider application. The pixels contained in HDR images and videos represent a larger range of colors and luminances than the pixels provided by conventional Low Dynamic Range (LDR) images and videos. HDR imaging is a technique used to take pictures that captures a greater dynamic luminance range than standard digital cameras. The goal of HDR imaging is to produce a range of real-world luminances similar to those observed by the human visual system, resulting in a more realistic and engaging experience for the observer.
Description
One, the technical field
The invention provides a method for accelerating an LDR (low density direct) video to reconstruct an HDR (high dynamic range) video by adopting a divide-and-conquer strategy, belonging to the field of HDR video reconstruction.
Second, background Art
High quality video is important in SLAM and video surveillance, which can clearly monitor objects or activities of interest and improve the positioning accuracy of SLAM. The conversion of low quality video to high quality video is becoming more and more a key to the robotic task, and High Dynamic Range (HDR) imaging technology represents high quality images and is finding wider application. The pixels contained in HDR images and videos represent a larger range of colors and luminances than the pixels provided by conventional Low Dynamic Range (LDR) images and videos. The brightness of a real scene may be directed from bright sunlight to extreme shadows. Standard digital cameras typically use 8 bits to represent each color channel of an image, one of the main disadvantages being that this representation cannot capture all the luminance range in a real scene. HDR imaging is a technique used to take pictures that captures a greater dynamic luminance range than standard digital cameras. The goal of HDR imaging is to produce a range of real-world luminances similar to those observed by the human visual system, resulting in a more realistic and engaging experience for the observer.
To achieve HDR reconstruction from LDR images, inverse tone mapping algorithms have great potential, converting a large number of original LDR images into HDR images by recovering the missing signals in a given image, but these algorithms can only be performed in certain types of scenes. In recent years, with the rapid development of deep learning techniques, a high dynamic range image reconstruction method based on a neural network has been proposed. Endo et al, Lee et al, and Eiletsen et al successfully restored the lost dynamic range using a deep neural network.
The fast moving object monitoring/static camera HDR video frame reconstruction method based on the divide and conquer strategy is provided. The divide and conquer strategy is also applied in other fields and research subjects, but has never been discussed in the HDR video reconstruction problem. We explore this efficient strategy in HDR video reconstruction to explore temporal information in video, which has never been efficiently explored before. Specifically, we first use a target detection method to separate the foreground from the entire image, and then train a specified network of background and foreground, connected by context-aware constraints. Once both are trained, this framework combines the background and foreground into an HDR video frame. In order to overcome the problem of inconsistent foreground and background color tones in the synthesis process, a context perception loss constraint is designed to provide context information for the background training of the foreground HDR. Experimental results show that the method can reconstruct high-quality HDR video frames in a shorter time.
Third, the invention
Aiming at a monitoring camera with a moving target, the invention provides a method for accelerating HDR video frame reconstruction by adopting a divide-and-conquer strategy. The framework consists of two connected CNN branches, modeling the entire HDR video reconstruction process. In this process, we use an object detection algorithm to separate the foreground and background of a video frame, then train the CNN branch to reconstruct the background and foreground frames, and connect the two branches. And synthesizing the enhanced HDR foreground and background to reconstruct a final HDR video frame. And the context perception loss constraint under an end-to-end framework is provided so as to eliminate the problem of inconsistent foreground and background colors in the synthesis process. The method provided by the invention is verified on a reference data set, and the result shows that the method can accurately reconstruct HDR video frames, and greatly shortens the time for reconstructing HDR video while ensuring the HDR image effect of each video frame.
TABLE-quantitative comparison of different methods
Fifth, detailed description of the invention
1. And making a target video scene. The invention provides a rapid method for reconstructing HDR video frames from LDR video frames, for a video sequence, a rapid target detection algorithm is firstly used for separating a background and a foreground; and then training the background network and the foreground network to generate a background HDR frame and a foreground HDR frame, synthesizing the frames and reconstructing a final HDR video frame.
2. Background/foreground separation. And extracting the boundary of each frame in the video sequence by using the trained model, obtaining position coordinates in the boundary, and segmenting the foreground from the whole scene. If no object is detected, the background HDR image will be the current HDR frame output, which is the easiest and fastest processing scheme.
3. Background/foreground reconstruction. After the foreground is separated from the scene with yolov5 algorithm, the background and foreground are trained separately with the network. Since the scene to be processed is a surveillance/still camera with moving objects, the background is reconstructed only once for a scene, while the number of foreground reconstructions is determined based on the number of frames contained in the scene.
4. Background/foreground frame synthesis. For a video scene, firstly reconstructing a background once, then fusing a reconstructed foreground HDR frame and a reconstructed background HDR frame, and recording the coordinate position of a foreground in the scene when the foreground is extracted by using a yolov5 algorithm. After the foreground reconstruction is completed, the scene in the background HDR is replaced according to the position given by the scene boundary, so as to form an HDR video frame corresponding to the original LDR video frame.
Claims (4)
1. The invention provides a method for converting an LDR video into an HDR video and accelerating the reconstruction of the HDR video, which is characterized by comprising the following steps:
1) a fast method for reconstructing HDR video frames from LDR video frames is proposed;
2) the invention uses the target detection algorithm of foreground and background, can improve the reconstruction time of HDR video frame;
3) after the foreground and the background are separated, the background and the foreground are trained respectively by using a network;
4) for a video scene, firstly, reconstructing a background once, and then fusing a reconstructed foreground HDR frame and a reconstructed background HDR frame.
2. A fast method for reconstructing HDR video frames from LDR video frames as claimed in claim 1, wherein step 1) comprises:
(1) firstly, separating a foreground and a background by using a rapid target detection algorithm;
(2) training a background network and a foreground network to generate background HDR and foreground HDR frames;
(3) and synthesizing the generated background HDR and foreground HDR frames to reconstruct a final HDR video frame.
3. The method for separately training the background and the foreground according to claim 1, wherein step 3) comprises:
(1) in order to avoid the difference between the background and the foreground boundary after synthesis, the invention provides a loss algorithm based on context sensing, for a video scene, the background is only reconstructed once, and the frequency of foreground reconstruction is determined according to the frame number contained in the scene;
(2) the reconstruction network proposed by the present invention trains the foreground and background separately,is the LDR video of the t-th frame, the corresponding HDR video frame at this moment isThe invention is toPredicting as inputTo achieve the goal of reconstructing HDR video;
(3) the foreground network and the background network have the same network structure, and the only difference is the difference between the training data and the loss function;
(4) after the foreground reconstruction is completed, the scene in the background HDR is replaced according to the position given by the scene bounding box, so as to form an HDR video frame corresponding to the original LDR video frame.
4. Fusing the reconstructed foreground HDR frame and the reconstructed background HDR frame as claimed in claim 1, wherein the step 4) comprises:
(1) when the background and the foreground are fused, spliced light beams of the boundary are eliminated so as to ensure the consistency of the color of the boundary, and a specified loss function corresponding to the image segmentation and combination stage is designed so as to output a more natural and clearer image;
(2) after the foreground reconstruction is completed, the scene in the background HDR is replaced according to the position given by the scene boundary, so as to form an HDR video frame corresponding to the original LDR video frame.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202110993299.5A CN113973175A (en) | 2021-08-27 | 2021-08-27 | Rapid HDR video reconstruction method |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202110993299.5A CN113973175A (en) | 2021-08-27 | 2021-08-27 | Rapid HDR video reconstruction method |
Publications (1)
Publication Number | Publication Date |
---|---|
CN113973175A true CN113973175A (en) | 2022-01-25 |
Family
ID=79586382
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202110993299.5A Pending CN113973175A (en) | 2021-08-27 | 2021-08-27 | Rapid HDR video reconstruction method |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN113973175A (en) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN115100043A (en) * | 2022-08-25 | 2022-09-23 | 天津大学 | HDR image reconstruction method based on deep learning |
Citations (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20070237393A1 (en) * | 2006-03-30 | 2007-10-11 | Microsoft Corporation | Image segmentation using spatial-color gaussian mixture models |
US20110044537A1 (en) * | 2009-08-18 | 2011-02-24 | Wesley Kenneth Cobb | Background model for complex and dynamic scenes |
CN104052905A (en) * | 2013-03-12 | 2014-09-17 | 三星泰科威株式会社 | Method and apparatus for processing image |
US20190347776A1 (en) * | 2018-05-08 | 2019-11-14 | Altek Corporation | Image processing method and image processing device |
CN110677558A (en) * | 2018-07-02 | 2020-01-10 | 华晶科技股份有限公司 | Image processing method and electronic device |
CN111709896A (en) * | 2020-06-18 | 2020-09-25 | 三星电子(中国)研发中心 | Method and equipment for mapping LDR video into HDR video |
CN113096029A (en) * | 2021-03-05 | 2021-07-09 | 电子科技大学 | High dynamic range image generation method based on multi-branch codec neural network |
CN113112452A (en) * | 2021-03-09 | 2021-07-13 | 北京迈格威科技有限公司 | Image processing method and device, electronic equipment and computer readable storage medium |
-
2021
- 2021-08-27 CN CN202110993299.5A patent/CN113973175A/en active Pending
Patent Citations (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20070237393A1 (en) * | 2006-03-30 | 2007-10-11 | Microsoft Corporation | Image segmentation using spatial-color gaussian mixture models |
US20110044537A1 (en) * | 2009-08-18 | 2011-02-24 | Wesley Kenneth Cobb | Background model for complex and dynamic scenes |
CN104052905A (en) * | 2013-03-12 | 2014-09-17 | 三星泰科威株式会社 | Method and apparatus for processing image |
US20190347776A1 (en) * | 2018-05-08 | 2019-11-14 | Altek Corporation | Image processing method and image processing device |
CN110677558A (en) * | 2018-07-02 | 2020-01-10 | 华晶科技股份有限公司 | Image processing method and electronic device |
CN111709896A (en) * | 2020-06-18 | 2020-09-25 | 三星电子(中国)研发中心 | Method and equipment for mapping LDR video into HDR video |
CN113096029A (en) * | 2021-03-05 | 2021-07-09 | 电子科技大学 | High dynamic range image generation method based on multi-branch codec neural network |
CN113112452A (en) * | 2021-03-09 | 2021-07-13 | 北京迈格威科技有限公司 | Image processing method and device, electronic equipment and computer readable storage medium |
Non-Patent Citations (2)
Title |
---|
李琢等: "多场景下复杂监控视频的前景目标提取", 《数学的实践与认识》 * |
范劲松等: "高动态范围图像(HDRI)编码及色调映射技术研究", 《工程图学学报》 * |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN115100043A (en) * | 2022-08-25 | 2022-09-23 | 天津大学 | HDR image reconstruction method based on deep learning |
CN115100043B (en) * | 2022-08-25 | 2022-11-15 | 天津大学 | HDR image reconstruction method based on deep learning |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
Johnston et al. | A review of digital video tampering: From simple editing to full synthesis | |
CN111489372B (en) | Video foreground and background separation method based on cascade convolution neural network | |
US20190379883A1 (en) | Stereoscopic video generation method based on 3d convolution neural network | |
US11037308B2 (en) | Intelligent method for viewing surveillance videos with improved efficiency | |
WO2021177324A1 (en) | Image generating device, image generating method, recording medium generating method, learning model generating device, learning model generating method, learning model, data processing device, data processing method, inferring method, electronic instrument, generating method, program, and non-transitory computer-readable medium | |
CN111539884A (en) | Neural network video deblurring method based on multi-attention machine mechanism fusion | |
KR102142567B1 (en) | Image composition apparatus using virtual chroma-key background, method and computer program | |
KR20110084025A (en) | Apparatus and method for image fusion | |
CN115393227B (en) | Low-light full-color video image self-adaptive enhancement method and system based on deep learning | |
CN113034413A (en) | Low-illumination image enhancement method based on multi-scale fusion residual error codec | |
Zhou et al. | Evunroll: Neuromorphic events based rolling shutter image correction | |
Han et al. | Hybrid high dynamic range imaging fusing neuromorphic and conventional images | |
CN114494050A (en) | Self-supervision video deblurring and image frame inserting method based on event camera | |
CN113973175A (en) | Rapid HDR video reconstruction method | |
Shaw et al. | Hdr reconstruction from bracketed exposures and events | |
US11044399B2 (en) | Video surveillance system | |
CN110062132B (en) | Theater performance reconstruction method and device | |
CN111626944A (en) | Video deblurring method based on space-time pyramid network and natural prior resistance | |
CN111429375A (en) | Night monitoring video quality improving method assisted by daytime image reference | |
CN111161189A (en) | Single image re-enhancement method based on detail compensation network | |
WO2022257184A1 (en) | Method for acquiring image generation apparatus, and image generation apparatus | |
CN115100218A (en) | Video consistency fusion method based on deep learning | |
CN113077385A (en) | Video super-resolution method and system based on countermeasure generation network and edge enhancement | |
KR102496362B1 (en) | System and method for producing video content based on artificial intelligence | |
Lin et al. | RE2L: A real-world dataset for outdoor low-light image enhancement |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
WD01 | Invention patent application deemed withdrawn after publication |
Application publication date: 20220125 |
|
WD01 | Invention patent application deemed withdrawn after publication |