CN113973175A - Rapid HDR video reconstruction method - Google Patents

Rapid HDR video reconstruction method Download PDF

Info

Publication number
CN113973175A
CN113973175A CN202110993299.5A CN202110993299A CN113973175A CN 113973175 A CN113973175 A CN 113973175A CN 202110993299 A CN202110993299 A CN 202110993299A CN 113973175 A CN113973175 A CN 113973175A
Authority
CN
China
Prior art keywords
hdr
foreground
background
video
frame
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202110993299.5A
Other languages
Chinese (zh)
Inventor
张涛
梁杰
王昊
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Tianjin University
Original Assignee
Tianjin University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Tianjin University filed Critical Tianjin University
Priority to CN202110993299.5A priority Critical patent/CN113973175A/en
Publication of CN113973175A publication Critical patent/CN113973175A/en
Pending legal-status Critical Current

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/95Computational photography systems, e.g. light-field imaging systems
    • H04N23/951Computational photography systems, e.g. light-field imaging systems by using two or more images to influence resolution, frame rate or aspect ratio
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Computing Systems (AREA)
  • General Health & Medical Sciences (AREA)
  • General Engineering & Computer Science (AREA)
  • Biophysics (AREA)
  • Computational Linguistics (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Computation (AREA)
  • Artificial Intelligence (AREA)
  • Molecular Biology (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Biomedical Technology (AREA)
  • General Physics & Mathematics (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Health & Medical Sciences (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Image Analysis (AREA)

Abstract

High quality video is important in SLAM and video surveillance, which can clearly monitor objects or activities of interest and improve the positioning accuracy of SLAM. The conversion of low quality video to high quality video is becoming more and more a key to the robotic task, and High Dynamic Range (HDR) imaging technology represents high quality images and is finding wider application. The pixels contained in HDR images and videos represent a larger range of colors and luminances than the pixels provided by conventional Low Dynamic Range (LDR) images and videos. HDR imaging is a technique used to take pictures that captures a greater dynamic luminance range than standard digital cameras. The goal of HDR imaging is to produce a range of real-world luminances similar to those observed by the human visual system, resulting in a more realistic and engaging experience for the observer.

Description

Rapid HDR video reconstruction method
One, the technical field
The invention provides a method for accelerating an LDR (low density direct) video to reconstruct an HDR (high dynamic range) video by adopting a divide-and-conquer strategy, belonging to the field of HDR video reconstruction.
Second, background Art
High quality video is important in SLAM and video surveillance, which can clearly monitor objects or activities of interest and improve the positioning accuracy of SLAM. The conversion of low quality video to high quality video is becoming more and more a key to the robotic task, and High Dynamic Range (HDR) imaging technology represents high quality images and is finding wider application. The pixels contained in HDR images and videos represent a larger range of colors and luminances than the pixels provided by conventional Low Dynamic Range (LDR) images and videos. The brightness of a real scene may be directed from bright sunlight to extreme shadows. Standard digital cameras typically use 8 bits to represent each color channel of an image, one of the main disadvantages being that this representation cannot capture all the luminance range in a real scene. HDR imaging is a technique used to take pictures that captures a greater dynamic luminance range than standard digital cameras. The goal of HDR imaging is to produce a range of real-world luminances similar to those observed by the human visual system, resulting in a more realistic and engaging experience for the observer.
To achieve HDR reconstruction from LDR images, inverse tone mapping algorithms have great potential, converting a large number of original LDR images into HDR images by recovering the missing signals in a given image, but these algorithms can only be performed in certain types of scenes. In recent years, with the rapid development of deep learning techniques, a high dynamic range image reconstruction method based on a neural network has been proposed. Endo et al, Lee et al, and Eiletsen et al successfully restored the lost dynamic range using a deep neural network.
The fast moving object monitoring/static camera HDR video frame reconstruction method based on the divide and conquer strategy is provided. The divide and conquer strategy is also applied in other fields and research subjects, but has never been discussed in the HDR video reconstruction problem. We explore this efficient strategy in HDR video reconstruction to explore temporal information in video, which has never been efficiently explored before. Specifically, we first use a target detection method to separate the foreground from the entire image, and then train a specified network of background and foreground, connected by context-aware constraints. Once both are trained, this framework combines the background and foreground into an HDR video frame. In order to overcome the problem of inconsistent foreground and background color tones in the synthesis process, a context perception loss constraint is designed to provide context information for the background training of the foreground HDR. Experimental results show that the method can reconstruct high-quality HDR video frames in a shorter time.
Third, the invention
Aiming at a monitoring camera with a moving target, the invention provides a method for accelerating HDR video frame reconstruction by adopting a divide-and-conquer strategy. The framework consists of two connected CNN branches, modeling the entire HDR video reconstruction process. In this process, we use an object detection algorithm to separate the foreground and background of a video frame, then train the CNN branch to reconstruct the background and foreground frames, and connect the two branches. And synthesizing the enhanced HDR foreground and background to reconstruct a final HDR video frame. And the context perception loss constraint under an end-to-end framework is provided so as to eliminate the problem of inconsistent foreground and background colors in the synthesis process. The method provided by the invention is verified on a reference data set, and the result shows that the method can accurately reconstruct HDR video frames, and greatly shortens the time for reconstructing HDR video while ensuring the HDR image effect of each video frame.
TABLE-quantitative comparison of different methods
Figure BSA0000251157140000021
Fifth, detailed description of the invention
1. And making a target video scene. The invention provides a rapid method for reconstructing HDR video frames from LDR video frames, for a video sequence, a rapid target detection algorithm is firstly used for separating a background and a foreground; and then training the background network and the foreground network to generate a background HDR frame and a foreground HDR frame, synthesizing the frames and reconstructing a final HDR video frame.
2. Background/foreground separation. And extracting the boundary of each frame in the video sequence by using the trained model, obtaining position coordinates in the boundary, and segmenting the foreground from the whole scene. If no object is detected, the background HDR image will be the current HDR frame output, which is the easiest and fastest processing scheme.
3. Background/foreground reconstruction. After the foreground is separated from the scene with yolov5 algorithm, the background and foreground are trained separately with the network. Since the scene to be processed is a surveillance/still camera with moving objects, the background is reconstructed only once for a scene, while the number of foreground reconstructions is determined based on the number of frames contained in the scene.
4. Background/foreground frame synthesis. For a video scene, firstly reconstructing a background once, then fusing a reconstructed foreground HDR frame and a reconstructed background HDR frame, and recording the coordinate position of a foreground in the scene when the foreground is extracted by using a yolov5 algorithm. After the foreground reconstruction is completed, the scene in the background HDR is replaced according to the position given by the scene boundary, so as to form an HDR video frame corresponding to the original LDR video frame.

Claims (4)

1. The invention provides a method for converting an LDR video into an HDR video and accelerating the reconstruction of the HDR video, which is characterized by comprising the following steps:
1) a fast method for reconstructing HDR video frames from LDR video frames is proposed;
2) the invention uses the target detection algorithm of foreground and background, can improve the reconstruction time of HDR video frame;
3) after the foreground and the background are separated, the background and the foreground are trained respectively by using a network;
4) for a video scene, firstly, reconstructing a background once, and then fusing a reconstructed foreground HDR frame and a reconstructed background HDR frame.
2. A fast method for reconstructing HDR video frames from LDR video frames as claimed in claim 1, wherein step 1) comprises:
(1) firstly, separating a foreground and a background by using a rapid target detection algorithm;
(2) training a background network and a foreground network to generate background HDR and foreground HDR frames;
(3) and synthesizing the generated background HDR and foreground HDR frames to reconstruct a final HDR video frame.
3. The method for separately training the background and the foreground according to claim 1, wherein step 3) comprises:
(1) in order to avoid the difference between the background and the foreground boundary after synthesis, the invention provides a loss algorithm based on context sensing, for a video scene, the background is only reconstructed once, and the frequency of foreground reconstruction is determined according to the frame number contained in the scene;
(2) the reconstruction network proposed by the present invention trains the foreground and background separately,
Figure FSA0000251157130000011
is the LDR video of the t-th frame, the corresponding HDR video frame at this moment is
Figure FSA0000251157130000012
The invention is to
Figure FSA0000251157130000013
Predicting as input
Figure FSA0000251157130000014
To achieve the goal of reconstructing HDR video;
(3) the foreground network and the background network have the same network structure, and the only difference is the difference between the training data and the loss function;
(4) after the foreground reconstruction is completed, the scene in the background HDR is replaced according to the position given by the scene bounding box, so as to form an HDR video frame corresponding to the original LDR video frame.
4. Fusing the reconstructed foreground HDR frame and the reconstructed background HDR frame as claimed in claim 1, wherein the step 4) comprises:
(1) when the background and the foreground are fused, spliced light beams of the boundary are eliminated so as to ensure the consistency of the color of the boundary, and a specified loss function corresponding to the image segmentation and combination stage is designed so as to output a more natural and clearer image;
(2) after the foreground reconstruction is completed, the scene in the background HDR is replaced according to the position given by the scene boundary, so as to form an HDR video frame corresponding to the original LDR video frame.
CN202110993299.5A 2021-08-27 2021-08-27 Rapid HDR video reconstruction method Pending CN113973175A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202110993299.5A CN113973175A (en) 2021-08-27 2021-08-27 Rapid HDR video reconstruction method

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202110993299.5A CN113973175A (en) 2021-08-27 2021-08-27 Rapid HDR video reconstruction method

Publications (1)

Publication Number Publication Date
CN113973175A true CN113973175A (en) 2022-01-25

Family

ID=79586382

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202110993299.5A Pending CN113973175A (en) 2021-08-27 2021-08-27 Rapid HDR video reconstruction method

Country Status (1)

Country Link
CN (1) CN113973175A (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115100043A (en) * 2022-08-25 2022-09-23 天津大学 HDR image reconstruction method based on deep learning

Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20070237393A1 (en) * 2006-03-30 2007-10-11 Microsoft Corporation Image segmentation using spatial-color gaussian mixture models
US20110044537A1 (en) * 2009-08-18 2011-02-24 Wesley Kenneth Cobb Background model for complex and dynamic scenes
CN104052905A (en) * 2013-03-12 2014-09-17 三星泰科威株式会社 Method and apparatus for processing image
US20190347776A1 (en) * 2018-05-08 2019-11-14 Altek Corporation Image processing method and image processing device
CN110677558A (en) * 2018-07-02 2020-01-10 华晶科技股份有限公司 Image processing method and electronic device
CN111709896A (en) * 2020-06-18 2020-09-25 三星电子(中国)研发中心 Method and equipment for mapping LDR video into HDR video
CN113096029A (en) * 2021-03-05 2021-07-09 电子科技大学 High dynamic range image generation method based on multi-branch codec neural network
CN113112452A (en) * 2021-03-09 2021-07-13 北京迈格威科技有限公司 Image processing method and device, electronic equipment and computer readable storage medium

Patent Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20070237393A1 (en) * 2006-03-30 2007-10-11 Microsoft Corporation Image segmentation using spatial-color gaussian mixture models
US20110044537A1 (en) * 2009-08-18 2011-02-24 Wesley Kenneth Cobb Background model for complex and dynamic scenes
CN104052905A (en) * 2013-03-12 2014-09-17 三星泰科威株式会社 Method and apparatus for processing image
US20190347776A1 (en) * 2018-05-08 2019-11-14 Altek Corporation Image processing method and image processing device
CN110677558A (en) * 2018-07-02 2020-01-10 华晶科技股份有限公司 Image processing method and electronic device
CN111709896A (en) * 2020-06-18 2020-09-25 三星电子(中国)研发中心 Method and equipment for mapping LDR video into HDR video
CN113096029A (en) * 2021-03-05 2021-07-09 电子科技大学 High dynamic range image generation method based on multi-branch codec neural network
CN113112452A (en) * 2021-03-09 2021-07-13 北京迈格威科技有限公司 Image processing method and device, electronic equipment and computer readable storage medium

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
李琢等: "多场景下复杂监控视频的前景目标提取", 《数学的实践与认识》 *
范劲松等: "高动态范围图像(HDRI)编码及色调映射技术研究", 《工程图学学报》 *

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115100043A (en) * 2022-08-25 2022-09-23 天津大学 HDR image reconstruction method based on deep learning
CN115100043B (en) * 2022-08-25 2022-11-15 天津大学 HDR image reconstruction method based on deep learning

Similar Documents

Publication Publication Date Title
Johnston et al. A review of digital video tampering: From simple editing to full synthesis
CN111489372B (en) Video foreground and background separation method based on cascade convolution neural network
US20190379883A1 (en) Stereoscopic video generation method based on 3d convolution neural network
US11037308B2 (en) Intelligent method for viewing surveillance videos with improved efficiency
WO2021177324A1 (en) Image generating device, image generating method, recording medium generating method, learning model generating device, learning model generating method, learning model, data processing device, data processing method, inferring method, electronic instrument, generating method, program, and non-transitory computer-readable medium
CN111539884A (en) Neural network video deblurring method based on multi-attention machine mechanism fusion
KR102142567B1 (en) Image composition apparatus using virtual chroma-key background, method and computer program
KR20110084025A (en) Apparatus and method for image fusion
CN115393227B (en) Low-light full-color video image self-adaptive enhancement method and system based on deep learning
CN113034413A (en) Low-illumination image enhancement method based on multi-scale fusion residual error codec
Zhou et al. Evunroll: Neuromorphic events based rolling shutter image correction
Han et al. Hybrid high dynamic range imaging fusing neuromorphic and conventional images
CN114494050A (en) Self-supervision video deblurring and image frame inserting method based on event camera
CN113973175A (en) Rapid HDR video reconstruction method
Shaw et al. Hdr reconstruction from bracketed exposures and events
US11044399B2 (en) Video surveillance system
CN110062132B (en) Theater performance reconstruction method and device
CN111626944A (en) Video deblurring method based on space-time pyramid network and natural prior resistance
CN111429375A (en) Night monitoring video quality improving method assisted by daytime image reference
CN111161189A (en) Single image re-enhancement method based on detail compensation network
WO2022257184A1 (en) Method for acquiring image generation apparatus, and image generation apparatus
CN115100218A (en) Video consistency fusion method based on deep learning
CN113077385A (en) Video super-resolution method and system based on countermeasure generation network and edge enhancement
KR102496362B1 (en) System and method for producing video content based on artificial intelligence
Lin et al. RE2L: A real-world dataset for outdoor low-light image enhancement

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
WD01 Invention patent application deemed withdrawn after publication

Application publication date: 20220125

WD01 Invention patent application deemed withdrawn after publication