WO2020007165A1 - 一种视频信号处理的方法及装置 - Google Patents
一种视频信号处理的方法及装置 Download PDFInfo
- Publication number
- WO2020007165A1 WO2020007165A1 PCT/CN2019/090566 CN2019090566W WO2020007165A1 WO 2020007165 A1 WO2020007165 A1 WO 2020007165A1 CN 2019090566 W CN2019090566 W CN 2019090566W WO 2020007165 A1 WO2020007165 A1 WO 2020007165A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- signal
- brightness
- linear
- display
- rgb
- Prior art date
Links
- 238000000034 method Methods 0.000 title claims abstract description 107
- 238000012545 processing Methods 0.000 title claims abstract description 72
- 238000013507 mapping Methods 0.000 claims abstract description 281
- 238000006243 chemical reaction Methods 0.000 claims description 121
- 230000006870 function Effects 0.000 claims description 79
- 238000004364 calculation method Methods 0.000 claims description 40
- 238000012546 transfer Methods 0.000 claims description 29
- 238000003860 storage Methods 0.000 claims description 18
- 238000004590 computer program Methods 0.000 claims description 6
- 238000013139 quantization Methods 0.000 claims 1
- 230000008569 process Effects 0.000 abstract description 33
- 230000000694 effects Effects 0.000 abstract description 14
- 230000003068 static effect Effects 0.000 description 33
- 230000005540 biological transmission Effects 0.000 description 26
- 238000010586 diagram Methods 0.000 description 20
- 230000003287 optical effect Effects 0.000 description 17
- 230000006978 adaptation Effects 0.000 description 15
- 239000003086 colorant Substances 0.000 description 11
- 238000004519 manufacturing process Methods 0.000 description 10
- 238000003672 processing method Methods 0.000 description 9
- 238000005516 engineering process Methods 0.000 description 8
- 239000000243 solution Substances 0.000 description 8
- 238000012360 testing method Methods 0.000 description 7
- 238000005070 sampling Methods 0.000 description 6
- 239000004065 semiconductor Substances 0.000 description 5
- 230000000007 visual effect Effects 0.000 description 5
- 230000008859 change Effects 0.000 description 4
- 230000011218 segmentation Effects 0.000 description 4
- 230000006835 compression Effects 0.000 description 3
- 238000007906 compression Methods 0.000 description 3
- 238000012937 correction Methods 0.000 description 3
- 238000009826 distribution Methods 0.000 description 3
- 230000014509 gene expression Effects 0.000 description 3
- 239000011159 matrix material Substances 0.000 description 3
- 238000012805 post-processing Methods 0.000 description 3
- 238000007781 pre-processing Methods 0.000 description 3
- 241000023320 Luma <angiosperm> Species 0.000 description 2
- 238000004891 communication Methods 0.000 description 2
- 238000011156 evaluation Methods 0.000 description 2
- 239000004973 liquid crystal related substance Substances 0.000 description 2
- OSWPMRLSEDHDFF-UHFFFAOYSA-N methyl salicylate Chemical group COC(=O)C1=CC=CC=C1O OSWPMRLSEDHDFF-UHFFFAOYSA-N 0.000 description 2
- 238000010606 normalization Methods 0.000 description 2
- 230000000717 retained effect Effects 0.000 description 2
- 239000000758 substrate Substances 0.000 description 2
- 239000012085 test solution Substances 0.000 description 2
- 238000011144 upstream manufacturing Methods 0.000 description 2
- 101100117775 Arabidopsis thaliana DUT gene Proteins 0.000 description 1
- 101150091805 DUT1 gene Proteins 0.000 description 1
- 241000282412 Homo Species 0.000 description 1
- 230000009471 action Effects 0.000 description 1
- 230000009286 beneficial effect Effects 0.000 description 1
- 239000003990 capacitor Substances 0.000 description 1
- 230000000052 comparative effect Effects 0.000 description 1
- 230000023077 detection of light stimulus Effects 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 230000018109 developmental process Effects 0.000 description 1
- 238000005538 encapsulation Methods 0.000 description 1
- 238000002474 experimental method Methods 0.000 description 1
- 238000000605 extraction Methods 0.000 description 1
- 230000008570 general process Effects 0.000 description 1
- 230000006872 improvement Effects 0.000 description 1
- 239000003550 marker Substances 0.000 description 1
- 239000000463 material Substances 0.000 description 1
- 238000013178 mathematical model Methods 0.000 description 1
- 238000005259 measurement Methods 0.000 description 1
- 230000007246 mechanism Effects 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 238000003032 molecular docking Methods 0.000 description 1
- 230000006855 networking Effects 0.000 description 1
- 239000013307 optical fiber Substances 0.000 description 1
- 239000005022 packaging material Substances 0.000 description 1
- 230000008447 perception Effects 0.000 description 1
- 230000002093 peripheral effect Effects 0.000 description 1
- 238000007639 printing Methods 0.000 description 1
- 238000000926 separation method Methods 0.000 description 1
- 229910052710 silicon Inorganic materials 0.000 description 1
- 239000010703 silicon Substances 0.000 description 1
- 238000010129 solution processing Methods 0.000 description 1
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N9/00—Details of colour television systems
- H04N9/64—Circuits for processing colour signals
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N5/00—Details of television systems
- H04N5/14—Picture signal circuitry for video frequency region
- H04N5/20—Circuitry for controlling amplitude response
- H04N5/202—Gamma control
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N9/00—Details of colour television systems
- H04N9/64—Circuits for processing colour signals
- H04N9/67—Circuits for processing colour signals for matrixing
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N11/00—Colour television systems
- H04N11/06—Transmission systems characterised by the manner in which the individual colour picture signal components are combined
- H04N11/20—Conversion of the manner in which the individual colour picture signal components are combined, e.g. conversion of colour television standards
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T5/00—Image enhancement or restoration
- G06T5/90—Dynamic range modification of images or parts thereof
- G06T5/92—Dynamic range modification of images or parts thereof based on global image properties
-
- G—PHYSICS
- G09—EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
- G09G—ARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
- G09G5/00—Control arrangements or circuits for visual indicators common to cathode-ray tube indicators and other visual indicators
- G09G5/10—Intensity circuits
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N5/00—Details of television systems
- H04N5/44—Receiver circuitry for the reception of television signals according to analogue transmission standards
- H04N5/57—Control of contrast or brightness
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N7/00—Television systems
- H04N7/01—Conversion of standards, e.g. involving analogue television standards or digital television standards processed at pixel level
- H04N7/0135—Conversion of standards, e.g. involving analogue television standards or digital television standards processed at pixel level involving interpolation processes
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N9/00—Details of colour television systems
- H04N9/64—Circuits for processing colour signals
- H04N9/68—Circuits for processing colour signals for controlling the amplitude of colour signals, e.g. automatic chroma control circuits
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N9/00—Details of colour television systems
- H04N9/64—Circuits for processing colour signals
- H04N9/68—Circuits for processing colour signals for controlling the amplitude of colour signals, e.g. automatic chroma control circuits
- H04N9/69—Circuits for processing colour signals for controlling the amplitude of colour signals, e.g. automatic chroma control circuits for modifying the colour signals by gamma correction
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N9/00—Details of colour television systems
- H04N9/77—Circuits for processing the brightness signal and the chrominance signal relative to each other, e.g. adjusting the phase of the brightness signal relative to the colour signal, correcting differential gain or differential phase
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10024—Color image
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20172—Image enhancement details
- G06T2207/20208—High dynamic range [HDR] image processing
-
- G—PHYSICS
- G09—EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
- G09G—ARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
- G09G2320/00—Control of display operating conditions
- G09G2320/02—Improving the quality of display appearance
- G09G2320/0271—Adjustment of the gradation levels within the range of the gradation scale, e.g. by redistribution or clipping
- G09G2320/0276—Adjustment of the gradation levels within the range of the gradation scale, e.g. by redistribution or clipping for the purpose of adaptation to the characteristics of a display device, i.e. gamma correction
-
- G—PHYSICS
- G09—EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
- G09G—ARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
- G09G2340/00—Aspects of display data processing
- G09G2340/04—Changes in size, position or resolution of an image
- G09G2340/0407—Resolution change, inclusive of the use of different resolutions for different screen areas
- G09G2340/0428—Gradation resolution change
-
- G—PHYSICS
- G09—EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
- G09G—ARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
- G09G2340/00—Aspects of display data processing
- G09G2340/06—Colour space transformation
-
- G—PHYSICS
- G09—EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
- G09G—ARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
- G09G2360/00—Aspects of the architecture of display systems
- G09G2360/16—Calculation or use of calculated indices related to luminance levels in display data
-
- G—PHYSICS
- G09—EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
- G09G—ARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
- G09G2370/00—Aspects of data communication
- G09G2370/04—Exchange of auxiliary data, i.e. other than image data, between monitor and graphics controller
Definitions
- the present application relates to the field of multimedia communications, and in particular, to a method and a device for video signal processing.
- High Dynamic Range (HDR) video technology expands the brightness range of the image that can be displayed, so it can record larger brightness range information and show more details of bright and dark parts.
- HDR is a popular technology that has appeared in the video industry in recent years, and it is also the direction of the future development of the video industry.
- the real picture seen by the human eye in the real world has a large dynamic range.
- the traditional Standard Dynamic Range (SDR) display device has low brightness and small dynamic range.
- Traditional SDR video technology continuously compresses the dynamic range of the captured picture during camera capture, production editing, and encoding.
- HDR video brightness is usually much larger than the brightness range that SDR display equipment or a large number of other existing HDR display equipment can display; therefore, when displaying HDR video signals on existing display equipment, The device's ability to process the brightness of the HDR video signal to match the range of brightness that the display device can display, suitable for display on existing devices. When the video signal is processed for brightness, unreasonable brightness processing causes the display effect of the HDR video to be ineffective. good.
- the embodiments of the present application provide a method and a device for processing video signals, which improve the display effect of video signals on a display device.
- a first aspect of the present application provides a video signal processing method, which includes: obtaining a first linear luminance signal, the first linear luminance signal being obtained based on a first linear red-green-blue RGB signal corresponding to a video signal to be processed; The first linear luminance signal is converted into a first non-linear luminance signal; segmented luminance mapping is performed on the first non-linear luminance signal to obtain a second non-linear luminance signal; and the second non-linear luminance signal is converted into a second non-linear luminance signal.
- the embodiment of the present application converts the luminance signal to a non-linear space and performs segmented luminance mapping, which can reasonably map the display luminance range of the video signal to the luminance range that can be displayed by the display device, and improve the contrast, brightness and detail performance of the picture, especially In the case of low-brightness display, the display brightness distribution after mapping is reasonable, and the screen display will not be dark; because the brightness mapping is performed in a non-linear space, the errors introduced by the brightness mapping are evenly distributed, and the final display effect of the video signal is The impact is small. In addition, because the brightness range of the video signal to be processed is large, the contribution of video signals in different brightness regions to the video display effect is also different. Different brightness is used based on the characteristics of the brightness interval of the video signal to be processed. The mapping relationship performs segmentation mapping on the brightness value of the video signal to be processed, which improves the flexibility and rationality of the brightness mapping.
- the video signal to be processed is a perceptually quantized PQ signal
- obtaining the first linear luminance signal includes: performing color space conversion on the PQ signal to obtain a first non-linear RGB signal; according to the PQ
- the electro-optical transfer function converts the first non-linear RGB signal into the first linear RGB signal; and performs calculations based on the primary color signals of the first linear RGB signal to obtain the first linear luminance signal.
- the video signal to be processed is a mixed-log-gamma HLG signal
- obtaining the first linear luminance signal includes: performing color space conversion on the HLG signal to obtain a second nonlinear RGB signal. ; According to the HLG photoelectric transfer inverse function, convert the second non-linear RGB signal into a second linear RGB signal; perform calculations based on the primary color signals of the second linear RGB signal to obtain a third linear luminance signal; The linear luminance signal is converted into a luminance signal type to obtain the first linear luminance signal.
- the brightness signal obtained based on the HLG signal is a scene light brightness signal.
- the scene light brightness signal needs to be converted into a display light brightness signal, and after being converted into a display light brightness signal, it is not directly displayed, but
- the brightness signal is converted into a non-linear space for segmented brightness mapping, as much as possible to preserve brightness details, improve the rationality of the brightness mapping, and improve the display effect of the HLG signal.
- the first linear brightness signal is a linear display light brightness signal
- the third brightness signal is a linear scene light brightness signal
- the method further includes: performing color space conversion on the RGB display signal to obtain a target display signal, wherein the target display signal is The color format is the same as that of the display device.
- the method further includes: superimposing a black level level on each primary color value of the RGB display signal to increase BlackLevelLift to obtain processed RGB For a display signal, the BlackLevelLift is the minimum value of the display brightness of the display device.
- the color space conversion of the RGB display signal includes color space conversion of the processed RGB display signal.
- the embodiment of the present application considers the influence of the black level of the display device on the brightness mapping curve, and retains the brightness details of the low brightness portion.
- performing segmented brightness mapping on the first non-linear luminance signal to obtain a second non-linear luminance signal includes: determining a first threshold and a second threshold, the first threshold The threshold value is less than the second threshold value; when the brightness value of the first nonlinear brightness signal is less than or equal to the first threshold value, the brightness value of the second nonlinear brightness signal is equal to the brightness value of the first nonlinear brightness signal; When the brightness value of the first non-linear brightness signal is greater than the first threshold value and less than or equal to the second threshold value, the brightness value of the second non-linear brightness signal is based on the brightness value of the first non-linear brightness signal since A fitting curve of the variable is obtained; when the brightness value of the first non-linear brightness signal is greater than the second threshold, the brightness value of the second non-linear brightness signal is equal to the maximum non-linear display brightness value corresponding to the display device.
- the video signal to be processed is divided into three segments according to two luminance value thresholds, and a portion smaller than the first threshold is taken as the first segment of the video signal.
- the luminance after the luminance mapping is equal to the luminance before the luminance mapping, that is, for low luminance Part of the video signal is not compressed, which can retain the image details of the low-brightness part most completely.
- the high-brightness part is divided into two sections. Among them, the part larger than the first threshold value and smaller than the second threshold value is compressed based on the fitting curve and retained as much as possible. Details of the brightness of this part; parts larger than the second threshold are compressed to the second brightness threshold; segmented mapping of the brightness takes full account of the characteristics of each brightness, retains brightness details as much as possible, and improves the rationality of the brightness mapping.
- the fitting curve is obtained by performing Hermite interpolation on the first threshold value and the second threshold value.
- performing segmented brightness mapping on the first non-linear luminance signal to obtain a second non-linear luminance signal includes: using the following segmentation function on the first non-linear luminance signal The signal performs this brightness mapping:
- e is the first non-linear brightness signal
- f tm (e) is the second non-linear brightness signal
- KP1 is the first threshold value
- KP2 is the second threshold value
- maxDL is the maximum non-linear display brightness value of the display device
- MaxSL is the maximum non-linear source luminance value
- x 0 KP1
- x 1 maxSL
- y 0 KP1
- y 1 maxDL
- the determining the first threshold and the second threshold includes: determining the first threshold according to a relationship between a display brightness range of the first non-linear brightness signal and a display brightness range of the display device. A threshold value; the maximum brightness value of the first non-linear brightness signal is used as the second threshold value.
- the selection of the brightness threshold is related to the difference between the source brightness and the display brightness of the display device.
- the first threshold is equal to the second
- the threshold is equal to the non-linear maximum brightness value of the source signal.
- the step of performing segmented brightness mapping on the first nonlinear brightness signal to obtain a second nonlinear brightness signal includes: based on the preset first nonlinear brightness signal. And a mapping relationship with the brightness value of the second non-linear brightness signal to determine the brightness value of the second non-linear brightness signal corresponding to the brightness value of the first non-linear brightness signal.
- the converting the first linear luminance signal into a first non-linear luminance signal includes: converting the first linear luminance signal into the first non-linear luminance signal according to an inverse PQ electro-optical transfer function.
- a non-linear luminance signal; correspondingly, converting the second non-linear luminance signal into a second linear luminance signal includes: converting the second non-linear luminance signal into the second linear luminance signal according to a PQ electro-optical transfer function .
- a second aspect of the present application provides a video signal processing apparatus, which is characterized in that the apparatus includes: a brightness obtaining unit, configured to obtain a first linear brightness signal based on a first linear brightness signal corresponding to a video signal to be processed; A linear red-green-blue RGB signal is obtained; a first conversion unit is configured to convert the first linear brightness signal into a first non-linear brightness signal; a brightness mapping unit is used to perform segmented brightness on the first non-linear brightness signal Mapping to obtain a second non-linear luminance signal; a second conversion unit for converting the second non-linear luminance signal into a second linear luminance signal; a gain calculation unit for calculating the second linear luminance signal and the first linear luminance signal A brightness gain of a linear brightness signal; a display signal acquisition unit is configured to obtain an RGB display signal corresponding to the video signal to be processed based on a product of the brightness gain and the first linear RGB signal.
- the video signal to be processed is a perceptually quantized PQ signal
- the brightness obtaining unit is specifically configured to: perform color space conversion on the PQ signal to obtain a first non-linear RGB signal; according to the PQ electro-optic A transfer function converts the first non-linear RGB signal into the first linear RGB signal; and performs calculations based on the primary color signals of the first linear RGB signal to obtain the first linear luminance signal.
- the video signal to be processed is a mixed logarithmic gamma HLG signal
- the brightness obtaining unit is specifically configured to perform color space conversion on the HLG signal to obtain a second nonlinear RGB signal; Convert the second non-linear RGB signal into a second linear RGB signal according to the inverse HLG photoelectric transfer function; perform calculations based on the primary color signals of the second linear RGB signal to obtain a third linear luminance signal; The brightness signal is converted into a brightness signal type to obtain the first linear brightness signal.
- the device further includes: a color space conversion unit, configured to perform color space conversion on the RGB display signal to obtain a target display signal, wherein the color format and display device of the target display signal The corresponding color format is the same.
- the device further includes: a compensation unit, configured to: superimpose a black level level on each of the primary color values of the RGB display signal and raise BlackLevelLift to obtain a processed RGB display signal, where the BlackLevelLift is the The minimum value of the display brightness of the display device; correspondingly, the color space conversion unit is specifically configured to perform color space conversion on the processed RGB display signal.
- a compensation unit configured to: superimpose a black level level on each of the primary color values of the RGB display signal and raise BlackLevelLift to obtain a processed RGB display signal, where the BlackLevelLift is the The minimum value of the display brightness of the display device; correspondingly, the color space conversion unit is specifically configured to perform color space conversion on the processed RGB display signal.
- the brightness mapping unit is specifically configured to determine a first threshold value and a second threshold value, where the first threshold value is less than the second threshold value; and when the brightness value of the first non-linear brightness signal is less than or When equal to the first threshold, the luminance value of the second non-linear luminance signal is equal to the luminance value of the first non-linear luminance signal; when the luminance value of the first non-linear luminance signal is greater than the first threshold value, and is less than or equal to When the second threshold value, the brightness value of the second non-linear brightness signal is obtained based on a fitting curve using the brightness value of the first non-linear brightness signal as an independent variable; when the brightness value of the first non-linear brightness signal is greater than the At the second threshold, the brightness value of the second non-linear brightness signal is equal to the maximum non-linear display brightness value corresponding to the display device.
- the fitting curve is obtained by performing Hermite interpolation on the first threshold value and the second threshold value.
- the brightness mapping unit is specifically configured to perform the brightness mapping on the first nonlinear brightness signal by using the following piecewise function:
- e is the first non-linear brightness signal
- f tm (e) is the second non-linear brightness signal
- KP1 is the first threshold value
- KP2 is the second threshold value
- maxDL is the maximum non-linear display brightness value of the display device
- MaxSL is the maximum non-linear source luminance value
- x 0 KP1
- x 1 maxSL
- y 0 KP1
- y 1 maxDL
- determining the first threshold and the second threshold includes: determining the first threshold according to a relationship between a display brightness range of the first non-linear brightness signal and a display brightness range of the display device; and The maximum brightness value of the first non-linear brightness signal is used as the second threshold.
- the brightness mapping unit is specifically configured to determine a relationship with the first non-linear brightness signal based on a preset mapping relationship between the brightness value of the first non-linear brightness signal and the second non-linear brightness signal.
- the brightness value of the second non-linear brightness signal corresponding to the brightness value of the linear brightness signal.
- the first conversion unit is specifically configured to convert the first linear luminance signal into the first non-linear luminance signal according to an inverse PQ electro-optical transfer function; correspondingly, the second conversion The unit is specifically configured to convert the second non-linear luminance signal into the second linear luminance signal according to a PQ electro-optical transfer function.
- a third aspect of the present application provides a video signal processing device, which includes: a processor and a memory; the processor calls software instructions in the memory to perform the following steps: obtaining a first linear luminance signal, the first The linear luminance signal is obtained based on the first linear red-green-blue RGB signal corresponding to the video signal to be processed; converting the first linear luminance signal into a first non-linear luminance signal; and performing segmented luminance mapping on the first non-linear luminance signal, To obtain a second non-linear brightness signal; converting the second non-linear brightness signal into a second linear brightness signal; calculating a brightness gain of the second linear brightness signal and the first linear brightness signal; based on the brightness gain and the first A product of a linear RGB signal to obtain an RGB display signal corresponding to the video signal to be processed.
- the video signal to be processed is a perceptually quantized PQ signal
- the processor is specifically configured to: perform color space conversion on the PQ signal to obtain a first nonlinear RGB signal; and perform electro-optical transfer according to the PQ Function to convert the first non-linear RGB signal into the first linear RGB signal; perform calculations based on the primary color signals of the first linear RGB signal to obtain the first linear luminance signal.
- the video signal to be processed is a mixed log-gamma HLG signal
- the processor is specifically configured to: perform color space conversion on the HLG signal to obtain a second nonlinear RGB signal; according to the HLG
- the inverse photoelectric transfer function converts the second nonlinear RGB signal into a second linear RGB signal; calculates based on the primary color signals of the second linear RGB signal to obtain a third linear luminance signal; and the third linear luminance signal Perform a luminance signal type conversion to obtain the first linear luminance signal.
- the processor is further configured to perform color space conversion on the RGB display signal to obtain a target display signal, wherein the color format of the target display signal is the same as the color format corresponding to the display device.
- the processor is configured to superimpose a black level level on each primary color value of the RGB display signal and increase BlackLevelLift to obtain a processed RGB display signal, where the BlackLevelLift is a display brightness of the display device.
- Minimum value correspondingly, the processor is specifically configured to perform color space conversion on the processed RGB display signal.
- the processor is specifically configured to determine a first threshold value and a second threshold value, where the first threshold value is less than the second threshold value; and when the brightness value of the first non-linear brightness signal is less than or equal to the When the first threshold value, the brightness value of the second nonlinear brightness signal is equal to the brightness value of the first nonlinear brightness signal; when the brightness value of the first nonlinear brightness signal is greater than the first threshold value and less than or equal to the first When the threshold value is two, the brightness value of the second nonlinear brightness signal is obtained based on a fitting curve using the brightness value of the first nonlinear brightness signal as an independent variable; when the brightness value of the first nonlinear brightness signal is greater than the second At the threshold, the brightness value of the second nonlinear brightness signal is equal to the maximum nonlinear display brightness value corresponding to the display device.
- the fitting curve is obtained by performing Hermite interpolation on the first threshold value and the second threshold value.
- the fitted curve is stored in the memory.
- the processor is specifically configured to perform the brightness mapping on the first non-linear brightness signal by using the following piecewise function:
- e is the first non-linear brightness signal
- f tm (e) is the second non-linear brightness signal
- KP1 is the first threshold value
- KP2 is the second threshold value
- maxDL is the maximum non-linear display brightness value of the display device
- MaxSL is the maximum non-linear source luminance value
- x 0 KP1
- x 1 maxSL
- y 0 KP1
- y 1 maxDL
- the piecewise function is stored in the memory.
- the processor is specifically configured to: determine the first threshold value according to a relationship between a display brightness range of the first non-linear brightness signal and a display brightness range of the display device; The maximum brightness value of the brightness signal is used as the second threshold.
- the processor is specifically configured to determine a relationship with the first non-linear brightness based on a preset mapping relationship between the first non-linear brightness signal and the second non-linear brightness signal ’s brightness value.
- the brightness value of the signal corresponds to the brightness value of the second non-linear brightness signal.
- the mapping relationship is stored in the memory.
- the processor is specifically configured to convert the first linear luminance signal into the first non-linear luminance signal according to an inverse PQ electro-optical transfer function; correspondingly, the processor is specifically configured to : Converting the second non-linear luminance signal into the second linear luminance signal according to a PQ electro-optical transfer function.
- a fourth aspect of the present application provides a computer-readable storage medium, where the computer-readable storage medium stores instructions, and when it runs on a computer or a processor, the computer or processor executes the first aspect as described above. Or the method described in any of its possible embodiments.
- a fifth aspect of the present application provides a computer program product containing instructions, which when executed on a computer or processor, causes the computer or processor to execute the first aspect or any possible implementation manner described above. The method described.
- FIG. 1 is a schematic diagram of an exemplary application scenario provided by an embodiment of the present application.
- FIG. 2 is a schematic structural diagram of an exemplary application scenario according to an embodiment of the present application.
- FIG. 3 is a schematic structural diagram of another exemplary application scenario according to an embodiment of the present application.
- FIG. 4 is a schematic diagram of a hardware architecture of an exemplary playback device and display device according to an embodiment of the present application
- FIG. 5 is a flowchart of an exemplary video signal processing method according to an embodiment of the present application.
- FIG. 6 is a flowchart of an exemplary method for processing brightness of an HDR video signal according to an embodiment of the present application
- FIG. 7 is a schematic diagram of an exemplary brightness mapping curve provided by an embodiment of the present application.
- FIG. 8 (a) is a schematic diagram of an exemplary static metadata HDR tone mapping provided by an embodiment of the present application.
- FIG. 8 (b) is a schematic diagram of an exemplary dynamic metadata HDR tone mapping provided by an embodiment of the present application.
- FIG. 9 is a flowchart of an exemplary method for obtaining a brightness mapping curve according to an embodiment of the present application.
- FIG. 10 is a flowchart of another exemplary HDR signal brightness processing method according to an embodiment of the present application.
- FIG. 11 is an exemplary PQ EOTF curve (left) and an exemplary PQ EOTF -1 curve (right) provided by an embodiment of the present application;
- FIG. 12 is an exemplary HLG OETF curve (left) and an exemplary HLG OETF -1 curve (right) provided by an embodiment of the present application;
- FIG. 13 is an exemplary apparatus for processing a video signal according to an embodiment of the present application.
- FIG. 14 is an exemplary HDR terminal technical solution processing flowchart according to an embodiment of the present application.
- 15 is a schematic diagram of an exemplary color gamut conversion process according to an embodiment of the present application.
- 16 is a schematic diagram of an exemplary signal conversion process according to an embodiment of the present application.
- 17 is a schematic diagram of an exemplary test network provided by an embodiment of the present application.
- FIG. 18 is a schematic diagram of another exemplary test network provided by an embodiment of the present application.
- FIG. 20 is a schematic diagram of an exemplary 1000 cd / m2 HLG curve HDR video end-to-end system according to an embodiment of the present application.
- FIG. 21 is an exemplary non-1000cd / m2 HLG curve HDR video end-to-end system provided by an embodiment of the present application.
- At least one (a), a, b, or c can represent: a, b, c, "a and b", “a and c", “b and c", or "a and b and c" ", Where a, b, and c can be single or multiple.
- Color value a value corresponding to a specific image color component (such as R, G, B or Y).
- Digital code value A digital expression value of an image signal.
- a digital code value is used to represent a non-linear primary color value.
- Linear color value (linear color value): The linear color value is proportional to the light intensity. In an optional case, its value should be normalized to [0,1], referred to as E.
- Non-linear primary color value is the normalized digital expression value of the image information, which is proportional to the digital encoding value. In an optional case, its value should be normalized to [0,1], referred to as E ′.
- Electro-optical transfer function A conversion relationship from a non-linear primary color value to a linear primary color value.
- Optical-electric transfer function A conversion relationship from a linear primary color value to a non-linear primary color value.
- Metadata Data describing the video source information carried in the video signal.
- Dynamic metadata Metadata associated with each frame of image. This metadata changes from picture to picture.
- Static metadata Metadata associated with an image sequence, which remains unchanged within the image sequence.
- Luma signal (luma) represents a combination of non-linear primary color signals, the symbol is Y '.
- Brightness mapping The brightness of the source image is mapped to the brightness of the target system.
- Chroma volume The volume formed by the chromaticity and brightness of a display in chromaticity space.
- Display adaptation processing video signals to adapt to the display characteristics of the target display.
- Source image The image input during the HDR pre-processing stage.
- Mastering Monitor (Mastering Display): The reference display used in the editing and production of video signals to determine the effect of video editing and production;
- HDR video signal with content as scene light in HDR video technology refers to the scene light captured by the camera / camera sensor, which is generally a relative value; after the linear scene light signal is HLG encoded The HLG signal is obtained.
- the HLG signal is a scene light signal, and the HLG signal is nonlinear.
- the scene light signal generally needs to be converted to a display light signal through OOTF and displayed on a display device.
- Linear display light signal that uses content as display light in HDR video technology refers to the display light emitted by the display device, generally an absolute value, in units of nits; the linear display light signal is PQ coded After that, a PQ signal is obtained.
- the PQ signal is a display light signal and the PQ signal is a non-linear signal.
- the general standard of the display light signal is displayed on the display device according to its absolute brightness.
- OOTF Light-to-light conversion curve
- Dynamic Range The ratio of the maximum brightness to the minimum brightness in a video signal
- Luma-Chroma-Chroma Bright color separates the three components of the video signal
- PQ Perceptual Quantizer
- PQEOTF curve converts the PQ-coded electrical signal into a linear optical signal, in units of nits; the conversion formula is:
- E ′ is the input electric signal, and the value range is [0,1]; the fixed parameter values are as follows:
- the PQEOTF curve is shown in the left figure of Figure 11: the input is an electrical signal in the range [0,1], and the output is a linear optical signal at [0,10000] nits;
- PQ EOTF -1 curve the inverse curve of PQ EOTF; the physical meaning is to convert the linear optical signal of [0,10000] nits into a PQ-coded electrical signal; the conversion formula is:
- the PQ EOTF -1 curve is shown in the right figure of Figure 11.
- the input is a linear optical signal with [0,10000] nits, and the output is an electrical signal with a range of [0,1];
- Color Gamut A color space contains the range of colors.
- the relevant color gamut standards are BT.709, BT.2020.
- Hybrid Log Gamma An HDR standard. Video signals collected by cameras, cameras, image sensors, or other types of image acquisition equipment are video signals in the HLG encoding format.
- HLG OETF curve A curve that transforms a linear scene light signal into a non-linear electrical signal by HLG encoding.
- the conversion formula is as follows:
- E is the input linear scene light signal, the range [0,1];
- E ′ is the output non-linear electrical signal, the range [0,1];
- HLG OETF -1 curve The inverse curve of HLG OETF converts HLG-coded non-linear electrical signals into linear scene light signals.
- the conversion formula is as follows:
- E ′ is the input non-linear electrical signal with a range of [0,1]
- E is the output linear scene light signal with a range of [0 ,1].
- Linear space In this application, linear space refers to the space where the linear optical signal is located
- Non-linear space refers to the space where the linear optical signal is transformed by using a non-linear curve; the non-linear curves commonly used in HDR include PQ, EOTF-1, HLG, OETF, etc.
- the linear curve has a gamma curve; it is generally considered that the linear optical signal is visually linear with respect to the human eye after being encoded by the above-mentioned non-linear curve. It should be understood that the non-linear space can be considered as a visual linear space.
- Gamma correction is a method for non-linear tone editing of an image. It can detect the dark and light parts of the image signal and increase the ratio of the two, thereby improving the contrast of the image. .
- the non-linear conversion of the color values output by the device is because the human visual system is not linear, and humans perceive visual stimuli through comparison.
- the outside world strengthens the stimulus by a certain proportion. For people, this stimulus is evenly increased. Therefore, for human perception, the physical quantity added in a series of equal proportions is uniform.
- the value of gamma can be determined according to the photoelectric conversion curve of the color space.
- Color space can be the eyes' different perception of light at different frequencies, and it can also represent objectively existing light at different frequencies.
- Color space is a color range defined by the coordinate system that people establish to represent color.
- the color gamut together with the color model, defines a color space.
- the color model is an abstract mathematical model that uses a set of color components to represent colors.
- the color model may include, for example, a three-primary-color light mode (red, green, blue, RGB), and a printing four-color mode (cyan, magenta, yellow, or key plate) (CMYK).
- Color gamut is the sum of colors that a system can produce.
- AdobeRGB and sRGB are two different color spaces based on the RGB model.
- Each device such as a monitor or printer, has its own color space and can only generate colors within its color gamut.
- the color of the image on different devices may change because each device converts and displays RGB or CMYK according to its own color space.
- sRGB color space (standard Red Green Blue color space): is a standard RGB color space developed by HP and Microsoft in 1996 for monitors, printers and the Internet. It provides a standard way to define color, allowing a variety of computer peripherals and applications, such as display, print, and scan, to have a common language for color.
- the sRGB color space is based on independent color coordinates, which can make colors correspond to the same color coordinate system in the transmission of different devices without being affected by different color coordinates of these devices.
- the color gamut space of sRGB is relatively small. sRGB defines the colors of the three primary colors of red, green, and blue.
- the color value of one of the three primary colors takes the maximum value, and the color corresponding to the color value of the other two colors is zero represents the one color.
- the values of the color values R, G, and B are all 0-255, and when the values of R and G are all zero, and the value of B is 255, Color indicates blue.
- YCC color space The color space for bright color separation in this application.
- the three components of YCC represent Luma-Chroma-Chroma respectively.
- Common YCC space video signals include YUV, YCbCr, ICtCp, etc .;
- Reserved_bits "Reserved bits" in the bit stream indicate that some syntax units are reserved for future extensions to this section. These bits should be ignored during decoding. The "reserved bits” should not have more than 21 consecutive '0's starting from any byte-aligned position.
- Marker bit (marker_bit): It means that the value of this bit should be '1'.
- the HDR video can display a large brightness.
- the maximum brightness of the SDR video signal is 100nits, and the maximum brightness of the HDR video signal is more than 1000nits.
- a large number of existing display devices can display a brightness range that is less than that of HDR video. Brightness; therefore, when displaying HDR video signals, the brightness of the HDR video signal needs to be processed according to the display capability of the display device, so that the HDR video signal matches the range of brightness that the display device can display, suitable for display on current devices; ITU.BT.2100
- the defined PQ signal and HLG signal are two internationally recognized HDR signal sources, and have been included in the standard by many countries and regions.
- FIG. 1 it is a schematic diagram of an exemplary application scenario provided by an embodiment of the present application.
- the playback device 102 finishes receiving and decoding the video stream 101, and the playback device 102 sends the decoded video or audio data to the display device 103 for display through a high definition multimedia interface (High Definition Multimedia Interface). Or play so that users can enjoy video or audio content.
- a high definition multimedia interface High Definition Multimedia Interface
- the video stream 101 may come from a website streaming media, a remote network device, the Internet, an optical fiber network, and the like.
- the video stream may be dynamic metadata HDR video data or static metadata HDR video data.
- the video stream 101 may be a data stream in the Transport Stream (TS) format.
- the TS may include a video stream, an audio stream, a subtitle data packet, and the like.
- the video stream may also use other similar format data streams, such as streaming media. You can use the (Matroska Video File, MKV) format to encapsulate audio data, video data, and subtitle data at the same time.
- MKV Meatroska Video File
- the transmission format of the audio and video streams in this application is not Make a limitation; for example, the video stream may include: HDR video data and metadata for describing the HDR video. In this case, both metadata and HDR video data are compressed in the video stream.
- the TS may include a video stream, an audio stream, a subtitle data packet, and metadata used to describe the HDR video. In this case, the metadata used to describe the HDR video data is placed in the TS. Without compression in the video stream.
- the metadata contains a description of the video image data.
- the static metadata describes the production environment of the entire video, which may include information about the monitor used for color adjustment and correction of video production, peak brightness, black level, RGB three-color coordinates, and white point coordinates.
- the dynamic metadata generally includes a description of each frame of the video image, for example, it may include the highest brightness, the lowest brightness, and the average brightness of the image; optionally, the dynamic metadata may also include a certain frame of image
- the reference mapping curve of the display screen it should be understood that the reference mapping curve included in the dynamic metadata varies with the change of the video image.
- the playback device 102 may be a set-top box (STB), a multimedia player, etc.
- the STB mainly includes more security functions, such as card charging, video encryption and decryption, etc.
- Some high-quality videos have a digital rights protection mechanism, which needs to be decrypted by the STB side before they can be viewed on the TV side.
- the video data received by the STB side is usually encoded video data.
- the STB also has a decoding function. After the data is decoded, it is sent to the TV side for display.
- FIG. 3 a schematic diagram of an application scenario in which a playback device completes processing video data and then sends it to a display device according to an embodiment of the present application.
- the STB parses the received TS to obtain video data. , Audio data, metadata metadata, etc.
- the video data received by the STB may be an HDR HLG video signal, an HDR PQ video signal, or an SDR video signal.
- the STB decodes the video data, displays brightness processing, color saturation processing, color gamut processing, etc., so that the HDR video signal adapts to the brightness range of the display screen, and then passes the processed video data through wired or wireless HDMI, Display Port transmission to the display device 103.
- the video data processed by the playback device matches the brightness range that the display device can display.
- the type of processed video data is related to the type of display device and can be HDR.
- the video data may also be SDR video data. As shown in FIG.
- the calculation of the brightness mapping curve is implemented by the main processor and is generally completed by software; the brightness mapping curve calculated in advance is stored in the memory, and the brightness processing unit in the video processor calls the brightness mapping curve in the memory.
- the video processor may be dedicated integrated hardware or dedicated circuits or multiple software modules running on a dedicated chip; it should be understood that FIG. 3 is only an example for a playback device.
- the architecture is not limited. In actual use, the implementation of each functional module can be adjusted according to the actual application scenario. It can be implemented by either the video processor or the main processor software, or other dedicated chips such as DSP, FPGA and other hardware are combined with software.
- a video stream or the like containing video data may be directly transmitted to a display device 103, such as a TV, etc., and then the video data is subjected to video decoding, brightness mapping processing, color saturation processing, and color gamut within the display device.
- a display device 103 such as a TV, etc.
- the video data is subjected to video decoding, brightness mapping processing, color saturation processing, and color gamut within the display device.
- FIG. 2 is a schematic diagram of an application scenario architecture provided by a display device after processing video data by a display device according to an embodiment of the present application.
- the calculation of the brightness mapping curve is realized by the main processor, which is generally completed by software instructions; the brightness mapping curve calculated in advance is stored in the memory, and the brightness processing unit in the video processor calls the memory
- the brightness mapping curve completes the processing of the video picture; it should be understood that in actual use, the implementation of each functional module can be adjusted according to the actual application scenario, which can be implemented by either the video processor or the main processor software. It can also be implemented by using other special-purpose chips such as DSP, FPGA and other hardware combined with software.
- the display device 103 may be a TV, a computer monitor, or any fixed terminal or mobile terminal having a display function.
- TV displays the video data sent by STB and displays it on the screen.
- the display processing may be that the TV adjusts the brightness of the video according to a preset brightness mapping relationship, so that the video content is compatible with the display capability of the TV.
- the TV uses different sets of brightness mapping relationships contained in the dynamic metadata to different videos.
- the picture is processed for brightness mapping tone mapping so that various pictures can be displayed and presented in the best way.
- the TV may also be an SDR TV or an HDR PQ TV.
- the following describes a playback device and a display device in the embodiments of the present application from the perspective of processor hardware.
- FIG. 4 it is a schematic diagram of a hardware architecture of a playback device and a display device according to an embodiment of the present application.
- the playback device 102 includes at least one central processing unit CPU, at least one memory, GPU, decoder, dedicated video / graphics processor, input interface, HDMI transmitter, and the like.
- the playback device may further include a microprocessor, a microcontroller (Microcontroller Unit, MCU), and the like.
- MCU microcontroller Unit
- the foregoing parts of the playback device are coupled through a connector, and the connector may include various interfaces, transmission lines, or buses, which is not limited in this embodiment.
- the connector may include various interfaces, transmission lines, or buses, which are not limited in this embodiment.
- the above parts are integrated on the same chip and together form the core chip of the playback device; in another optional case, the CPU, GPU, decoder, input interface, and HDMI transmitter Integrated on a chip, the internal parts of the chip access the external memory through the bus.
- the dedicated video / graphics processor can be integrated on the same chip as the CPU, or it can exist as a separate processor chip.
- the chip involved in the embodiment of the present application is a system manufactured on the same semiconductor substrate by an integrated circuit process, also called a semiconductor chip, which can be manufactured on the substrate (usually, for example, silicon or the like) by using an integrated circuit process. Semiconductor material), the outer layer of which is usually encapsulated by a semiconductor packaging material.
- the integrated circuit may include various types of functional devices, and each type of functional devices includes logic gate circuits, Metal-Oxide-Semiconductor (MOS) transistors, bipolar transistors, or transistors such as diodes, and may also include capacitors and resistors. Or other components such as inductors.
- MOS Metal-Oxide-Semiconductor
- bipolar transistors bipolar transistors
- transistors such as diodes
- capacitors and resistors or other components such as inductors.
- Each functional device can work independently or under the action of necessary driver software, and can achieve various functions such as communication, calculation, or storage.
- the CPU may be used to implement part or all of the operations in the embodiments of the present application, for example, it may implement tone mapping of an image, demultiplexing and decoding of video data, calculation of a brightness-mapping curve, and a color saturation curve. Calculation, etc .; optionally, the CPU may be a single-CPU processor or a multi-CPU processor; optionally, the CPU may be a processor group composed of multiple processors, and multiple The processors are coupled to each other through one or more buses. In an optional case, the processing of images and videos is partly performed by the GPU, partly by a dedicated video / graphics processor, and possibly by software code running on a general-purpose CPU or GPU.
- Memory which can be used to store computer program instructions, including various operating system (Operation System, OS), various user application programs, and various computer program codes including program code for executing the scheme of the present application.
- OS Operating System
- various user application programs and various computer program codes including program code for executing the scheme of the present application.
- it can be used to store brightness Mapping curve or brightness mapping lookup table LUT;
- the memory can also be used to store video data, audio data, subtitle data, etc .;
- the CPU can be used to execute computer program code stored in the memory to implement the method in the embodiment of the present application and be executed
- the various types of program code can be considered as the driver of the CPU.
- the memory 302 may be a non-power-loss volatile memory, such as an Embedded Multimedia Card (EMMC), a Universal Flash Memory (UFS), or a Read-Only Memory (Read-Only Memory, ROM), or other types of static storage devices that can store static information and instructions, can also be volatile memory (volatile memory), such as Random Access Memory (Random Access Memory, RAM) or can store information and Other types of dynamic storage devices for instructions can also be Electrically Erasable Programmable Read-Only Memory (EEPROM), Compact Disc Read-Only Memory (CD-ROM), or other optical disks Storage, optical disc storage (including compact discs, laser discs, optical discs, digital versatile discs, Blu-ray discs, etc.), magnetic disk storage media or other magnetic storage devices, or can be used to carry or store program code in the form of instructions or data structures and can Any other computer-readable storage medium accessed by the computer, but is not limited thereto.
- EMMC Embedded Multimedia Card
- UFS Universal Flash Memory
- the input interface is used to receive the transport stream.
- the input interface of the playback device can be a network interface, such as a WIFI or Ethernet interface.
- the input interface can also be a terminal of a broadcast television such as a tuner.
- the input interface can also be a universal serial bus. (Universal Serial Bus, USB) interface.
- the CPU executes related code to demultiplex the transport stream to obtain video data and subtitle data, etc.
- the decoder decodes the video data stream to obtain video data and metadata
- the video / graphics processor completes For video data brightness mapping processing, color saturation processing, color space conversion, color preprocessing, scene light and display light signal conversion, linear space and non-linear space conversion, etc., optionally, it can also complete the brightness mapping curve Calculation of saturation mapping curve.
- the HDMI transmitter encapsulates the decoded video data, metadata and subtitle data respectively, and transmits the encapsulated data packets / information frames to the display device 103 through the HDMI data channel.
- the display device 103 includes an HDMI receiver, at least one central processing unit CPU, at least one memory, a GPU, a decoder, a dedicated video / graphics processor, and a ByOne interface.
- the display device also includes a display screen (not shown in the figure).
- the VByOne interface is coupled to the display screen.
- the VByOne interface is a digital interface standard developed for image transmission.
- the display device 103 shown in FIG. 4 may be an integrated display chip, and the received video data is processed on the display chip and sent to a display screen for display.
- the HDMI transmitter separates video data frames, metadata information frames, subtitle information frames, and other information frames or data packets and transmits them to the display device. It should be understood that there are multiple channels inside HDMI, some of which are used to transmit data information, and some are used to transmit control information such as clock, check signal, power signal, and ground signal.
- the data channel is time-multiplexed, and various data packets cannot be transmitted simultaneously.
- the amount of data transmitted in a unit time of a channel is limited by the operating frequency.
- the maximum amount of data that can be transmitted in a unit time of a channel is the bandwidth of the HDMI channel.
- the bandwidth of HDMI 2.1 is 18Gbps (bit second).
- the transmission interface transmits the HDR video data frame, metadata information frame, and subtitle information frame in a time-sharing manner.
- the transmission interface corresponds to multiple transmission frequency bands, and the transmission interface divides and transmits video data frames, metadata frames, and subtitle information frames; optionally, the transmission interface corresponds to multiple transmission channels, and the transmission interface transmits video data frames in different channels. Metadata frame and subtitle information frame.
- Tone mapping of video data can be done by the GPU or by a dedicated video / graphics processor.
- Luminance mapping can be done by a dedicated video / graphics processor, or by software code running on a CPU or GPU.
- the video / image processor transmits the video data after display brightness processing to the display screen for display through the VByOne interface.
- the display screen may be a liquid crystal display (Liquid Crystal Display, LCD), a light emitting diode (Light Emitting Diode, LED) display, an organic light emitting diode (Organic Light-Emitting Diode, OLED) display, or a cathode ray tube ( Cathode Ray Tube (CRT) display and so on.
- LCD Liquid Crystal Display
- LED Light Emitting Diode
- OLED Organic Light-Emitting Diode
- CTR Cathode Ray Tube
- FIG. 5 it is a flowchart of a video signal processing method according to an embodiment of the present application. It should be understood that for the convenience of description, FIG. 5 describes the method in the form of steps. Although the method sequence is shown in the method flowchart 5, in some cases, the description may be performed in a different order than here. A step of.
- the video signal processing method includes:
- the first linear luminance signal is obtained based on a first linear RGB signal corresponding to a video signal to be processed.
- the first linear RGB signal is a linear display light signal
- the first linear luminance signal is a linear display light luminance signal.
- the luminance signal is a luminance component of a video signal to be processed.
- the first linear luminance signal is calculated based on the three primary color signals R, G, and B of the first linear RGB signal.
- the video signal to be processed may be a PQ signal
- the PQ signal may be a PQ signal in a YUV space.
- the PQ signal needs to be converted from the YUV space to the RGB space to obtain a first nonlinear RGB signal.
- the first non-linear RGB signal is converted into the first linear RGB signal, and then calculation is performed based on each primary color signal of the first linear RGB signal to obtain a first linear luminance signal.
- the first linear luminance signal is a display luminance signal.
- the video signal to be processed may be an HLG signal
- the HLG signal may be an HLG signal in a YUV space.
- the HLG signal needs to be converted from the YUV space to the RGB space to obtain a second nonlinear RGB signal
- the second non-linear RGB signal is converted into a second linear RGB signal
- the second linear RGB signal is a linear scene light signal, and then based on each primary color signal of the second linear RGB signal Perform calculation to obtain the third linear luminance signal.
- the third linear luminance signal needs to be converted into a luminance signal type to obtain the first linear luminance signal.
- Linear brightness signal is a display light brightness signal.
- the embodiment of the present application performs brightness mapping in a non-linear space.
- the input of the brightness mapping is a linear brightness signal, so the linear brightness signal needs to be converted into a non-linear brightness signal.
- the first linear luminance signal may be converted into the first non-linear luminance signal according to an inverse PQ electro-optical transfer function.
- Other conversion functions or conversion curves may also be used to complete the conversion between the first linear luminance signal and the first non-linear luminance signal.
- the first non-linear brightness signal is a non-linear brightness signal before brightness mapping.
- segmented brightness mapping may be performed on the first non-linear brightness signal based on a preset brightness mapping lookup table.
- the brightness mapping lookup table may be stored in a memory, and the brightness mapping lookup table includes several discrete sets of coordinate points.
- the mapping data in the brightness mapping lookup table may be obtained through prior experiments;
- the first non-linear luminance signal may be segmented in luminance mapping based on a piecewise function.
- the first non-linear luminance signal may be divided into three segments according to the luminance value, and the luminance of each segment is The mapping functions are all different. For example, the first threshold value and the second threshold value are determined, and the first threshold value is smaller than the second threshold value. When the brightness value of the first nonlinear brightness signal is less than or equal to the first threshold value, the second nonlinear brightness signal is determined.
- the brightness value of is equal to the brightness value of the first non-linear brightness signal; when the brightness value of the first non-linear brightness signal is greater than the first threshold value and less than or equal to the second threshold value, the brightness value of the second non-linear brightness signal is based on the
- the brightness value of a non-linear brightness signal is obtained by a fitting curve of an independent variable; when the brightness value of the first non-linear brightness signal is greater than a second threshold, the brightness value of the non-linear brightness signal is equal to the maximum non-linearity corresponding to the display device Displays the brightness value.
- the fitted curve is obtained by performing Hermite interpolation on the first threshold and the second threshold.
- the piecewise brightness mapping curve can be expressed as the following piecewise function:
- the piecewise brightness mapping curve can be expressed as the following piecewise function:
- the piecewise brightness mapping curve can be expressed as the following piecewise function:
- the piecewise brightness mapping curve can be expressed as the following piecewise function:
- the piecewise brightness mapping curve can be expressed as the following piecewise function:
- the piecewise brightness mapping curve can be expressed as the following piecewise function:
- the piecewise brightness mapping curve can be expressed as the following piecewise function:
- segmented brightness mapping may be performed on the first non-linear brightness signal based on the brightness mapping curve, and the brightness mapping curve may be a segmented curve. It should be understood that the segmented brightness mapping curve may be considered as an illustration of a segmented function.
- the discrete data in the lookup table may be coordinate points on a brightness mapping curve.
- the second non-linear luminance signal may be converted into the second linear luminance signal according to the PQ electro-optical transfer function.
- Other conversion functions or conversion curves may also be used to complete the conversion of the second non-linear luminance signal and the second linear luminance signal.
- the second non-linear brightness signal is a non-linear brightness signal after brightness mapping.
- the brightness gain is a ratio of the second linear brightness signal to the first linear brightness signal.
- the brightness gain is multiplied with the three primary color components R, G, and B of the first linear RGB signal to obtain an RGB display signal.
- the RGB display signal can be used for display on a display device.
- the display device can display a color format different from RGB, and the method further includes:
- a color space conversion is performed on the RGB display signal to obtain a target display signal.
- the color format of the target display signal is the same as the color format corresponding to the display device.
- RGB display signal after obtaining the RGB display signal, superimpose the black level level on each of the primary color components R, G, and B of the RGB display signal to raise the BlackLevelLift, and BlackLevelLift is the minimum value of the display brightness of the display device. . Further, color space conversion is performed on the processed display RGB signal to obtain a target display signal with the same color format as that of the display device.
- the video signal processing method provided in the embodiment of the present application is described below by taking the input as an HDR PQ signal and the input as an HDR HLG signal as examples.
- FIG. 6 it is a schematic flowchart of a method for processing brightness of HDR video data according to an embodiment of the present application.
- the linear scene light signal and the linear display light signal are two types of video signals.
- the linear scene light signal is a video signal captured by a camera or other image acquisition device or video acquisition device, and the linear display light signal is displayed by the display device.
- the luminance signal is the component that represents the luminance in the video signal.
- a linear scene luminance signal is obtained based on the linear scene light signal, and a linear display light luminance signal is obtained based on the linear display light signal.
- the method includes:
- the brightness mapping curve may be any brightness mapping curve acting on the selected non-linear space.
- the non-linear space may be a PQ curve space.
- Figure 7 shows an example of a brightness mapping curve generated in a non-linear space (PQ EOTF -1 curve).
- the horizontal axis of the brightness mapping curve is a non-linearly encoded brightness signal before the brightness mapping, and the vertical axis is the brightness mapping.
- the range of the abscissa and ordinate values are both [0,1], which represents the electric power after the linear light signal with the luminance range of [0,10000] nits is encoded by the PQ curve.
- the range of the signal is [0,1] (refer to the right figure in Figure 11, the brightness signal in the brightness range of 0-10000nits is encoded by the PQ curve to obtain the electrical signal of 0-1), that is, the abscissa represents the value before the brightness mapping.
- the brightness range of the brightness signal is [0,10000] nits
- the ordinate represents the brightness range of the brightness signal after the brightness mapping
- the curve shown in FIG. 7 represents mapping the brightness signal with the brightness range of [0,10000] nits to the brightness range. [0,300] nits.
- the curve of the brightness mapping is a piecewise function.
- the formula of the brightness mapping curve can be as follows:
- a one-dimensional lookup table LUT is used to represent the brightness mapping curve.
- yn is the ordinate of the n sampling points, that is, the output of the brightness mapping curve, and represents the brightness signal after the brightness mapping. It should be understood that since the mapping relationship in the lookup table is a set of discrete points, if the input brightness value is not among these discrete points, the input brightness value can be obtained by interpolation based on known brightness values; in an optional In the case, the luminance values of the inputs within a range correspond to the luminance values of the same output, for example, x0-x3 all correspond to y0, x4-x6 all correspond to y1, and so on.
- These discrete coordinate points can be obtained through prior experimental measurements, such as inputting the brightness value, and measuring the display brightness value corresponding to the input brightness value on the display screen; in an optional case, the input brightness value and the The fitting function of the corresponding relationship of the brightness values displayed on the display screen, and the discrete coordinate points in the lookup table are calculated according to the fitting function.
- the essence of brightness mapping is to adjust the display brightness range of a video image, and map from one display brightness range to another display brightness range.
- the brightness range of a video image can be adjusted to a brightness that can be displayed on a display screen.
- the input signal of the TM_Curve curve is the brightness signal before brightness mapping
- the output is the brightness signal after brightness mapping.
- tone mapping can be an implementation of brightness mapping, or tone mapping In some cases, the process can be equivalent to the process of brightness mapping.
- the static metadata HDR video uses a fixed mapping method to process all video images.
- the static metadata HDR video has a fixed mapping curve.
- the input brightness range of the mapping curve is 0-4000 nits.
- Special nit the output brightness range is 0-500nit, TV for the maximum brightness of 200nit, the maximum brightness of 1500nit and the maximum brightness of 4000nit three images, all use this mapping curve for tone mapping, after processing and display on the display. Because the mapping relationship curve does not match the actual brightness range of the first two frames of the image, most of the image details are lost in the first two frames of the image, the overall image is dark, and the display effect is poor.
- Dynamic metadata HDR video has multiple sets of mapping curves.
- mapping curves used are adjusted according to the actual brightness of the image.
- Figure 8 (b) three different mapping curves are given.
- the input brightness range for curve 1 is 0-500 nits, and the output brightness range is 0-500nits.
- the brightness range for input curve 2 is 0-1500 nits, and the output brightness range is 0-500 nit.
- Mapping curve 3 The input brightness range is 0-4000 nits, and the output brightness range is 0-500nits.
- TV selects a suitable mapping curve for tone mapping based on the actual brightness range of the image frame, so that images of different brightness can be optimally used. It is displayed on the screen.
- the mapping curve 2 is selected to perform brightness mapping on an image with a maximum brightness of 1500nit.
- the processed image retains the image details well.
- the input brightness of the above mapping curve represents the brightness represented by the video signal source; and the output brightness is the brightness that a display device such as a TV can actually display.
- the brightness range represented by video signals produced in a professional film and television production environment is generally larger than the brightness range that consumer televisions can display.
- Tone mapping is a technology that maps and matches the brightness range of an input video signal to the brightness range displayed by a display device.
- the input video signal in the embodiment of the present application is an HDR PQ signal in the YUV space.
- PQ coding is performed on the linear display light signal to obtain a PQ signal
- the PQ signal is a display light signal
- the PQ signal is a kind of non-linear signal.
- the input video signal is converted from YUV space to RGB space to obtain the non-linear display light signal R'dG'dB'd in the RGB color space; further, the non-linear display light signal R'dG ' dB'd is converted into a linear display optical signal RdGdBd.
- the nonlinear display optical signal R'dG'dB'd is converted into a linear display optical signal RdGdBd based on the PQ EOTF curve.
- the conversion from YUV to RGB is considered to be a color space conversion, and the non-linear RGB to linear RGB signals are not converted in color space.
- the non-linear RGB signals and the linear RGB signals Both belong to the RGB color space.
- converting the input video signal from a YUV signal to a non-linear display light signal R'dG'dB'd is a color space conversion to convert the non-linear display light signal R
- the conversion of 'dG'dB'd to the linear display light signal RdGdBd can also be considered as a color space conversion.
- the nonlinear RGB display light signal and the linear RGB display light signal belong to different color spaces.
- Yd is a linear luminance signal
- the luminance signal is the amount that represents the brightness of the display light signal, that is, the luminance signal can also be considered as the component that represents the brightness of the video signal; among which, the selection of the parameters cr, cg, cb and the linear display optical signal RdGdBd
- the corresponding calculation parameters will also be different. Therefore, when calculating the brightness, it is necessary to select the linear brightness calculation parameter in the corresponding color gamut according to the color gamut in which the light signal is located.
- Yd is the display brightness in a linear space.
- the display brightness Yd is converted to a non-linear space using a PQ EOTF -1 curve to obtain a non-linear display brightness.
- NL_Yd PQ_EOTF -1 (Yd), which is equivalent to
- the linear display light luminance signal Yd with a luminance range of 0-10,000 nits is encoded by a PQ curve to obtain an electrical signal of 0-1.
- the electrical signal is the non-linear display light luminance NL_Yd in a non-linear space;
- the curve can also use other non-linear conversion curves.
- the brightness mapping curve TM_Curve can be TM_Curve in step 500; the non-linear display brightness
- the signal NL_Yd is the brightness signal before the brightness mapping, which can also be said to be a non-linear display light brightness signal of the source image signal, and NL_Yt is the non-linear display brightness signal after the brightness mapping, which can also be said to be compatible with the display capability of the display device.
- a non-linear display light brightness signal NL_Yd is input, and a corresponding brightness brightness non-linear brightness NL_Yt is obtained based on the calculation formula of the brightness map curve.
- a one-dimensional lookup table LUT is selected to implement the brightness mapping curve.
- a lookup table linear interpolation method may be adopted in actual mapping.
- Linear interpolation is an interpolation method for one-dimensional lookup table LUTs, which estimates values based on two data points adjacent to the left and right of the point in the one-dimensional data sequence that needs to be interpolated.
- the non-linear electrical signal is converted to a linear display light luminance signal in the range [0,10000] nits. It should be understood that NL_Yt can be converted to Yt by using a curve other than the PQ EOTF curve.
- the ratio of K to the linear display luminance signal Yt after the luminance mapping and the linear display luminance signal Yd before the luminance mapping can be used to measure the change of the display luminance before and after the luminance mapping;
- Rt, Gt, and Bt are the red, green, and blue components of the linear display light signal after the luminance mapping process, respectively.
- BLoffset is the black level of the display device. In theory, when the screen displays black, it corresponds to The brightness should be 0, but in actual application, when the screen displays black, the corresponding brightness value is not 0, but a relatively small brightness value, that is, the black level is the minimum brightness that the display device can display; in this step BLoffset is the display brightness in linear space. It should be understood that the BLoffset of each display device may be different, and the BLoffset can be obtained by measuring the display device.
- the processed linear display light signal RtGtBt After obtaining the processed linear display light signal RtGtBt, according to the color space of the actual display device, color space conversion is performed on the linear display light signal RtGtBt, and the RtGtBt signal is converted to the color space of the display device to obtain the display brightness range with the display device.
- the processed video signal that matches the color space; for example, if the color space of the display device is sRGB space, it can be displayed directly without color space conversion; if the color space of the display device is YUV, RtGtBt The signal is converted into a video signal YUV1 in YUV space.
- the brightness signal of the display light signal is converted to a non-linear space, and the brightness of the display brightness is mapped in the non-linear space, so that the display brightness range of the HDR video signal can be reasonably mapped to a display device capable of displaying In the brightness range, the contrast, brightness, and detail performance of the picture are improved. Since the brightness mapping is performed in a non-linear space, the errors introduced by the brightness mapping are evenly distributed, which has a small impact on the final display effect of the video signal, especially at low brightness. In the case of display, the display brightness distribution after mapping is reasonable, and the screen display will not be dark.
- the embodiment of the present application considers the influence of the black level of the display device on the brightness mapping curve, and retains the brightness details of the low brightness portion. Further, the HDR signal brightness processing method provided in the embodiment of the present application can convert an HDR signal into an SDR signal, thereby improving the compatibility of the SDR display device with the HDR signal.
- FIG. 9 a flowchart of a method for calculating a brightness mapping curve TM_Curve according to an embodiment of the present application. It should be understood that the method shown in FIG. 9 can be used to calculate the brightness mapping curve TM_Curve in step 500, and the brightness mapping curve TM_Curve in 500 can also be calculated by other methods, and is not limited to the method shown in FIG. 9.
- the method for calculating the brightness mapping curve TM_Curve may include:
- Source maximum brightness MaxSrcLuma, source minimum brightness MinSrcLuma, display device maximum brightness MaxDispLuma, display device minimum brightness MinDispLuma, units are nits; source maximum / minimum brightness can be determined according to actual conditions and experience, or metadata information carried from HDR signals To obtain the maximum / minimum brightness of the main control monitor as the maximum / minimum brightness of the source; the maximum / minimum brightness of the display device is measured according to the actual display device. Optionally, the brightness of the actual display device cannot be measured in special application scenarios. The value can be set based on experience.
- the same nonlinear space PQ EOTF -1 curve as in 503 may be selected, and the maximum nonlinear brightness maxSL, the minimum nonlinear brightness minSL, the maximum nonlinear brightness maxDL, Non-linear display minimum brightness minDL;
- step 901 may be considered as mapping a luminance signal in a linear space with a luminance range of [0,10000] nits to an electrical signal in a range of [0,1].
- KP is the inflection point of the brightness mapping curve, KneePoint. Brightness values lower than KneePoint are not compressed, and brightness values higher than KneePoint are compressed; for example, KP is selected from (minDL, maxDL), and KP can be adjusted according to the actual effect. In an optional case, the selection of KP is related to the difference between the source brightness and the display brightness of the display device. When the source brightness is less than or equal to the display brightness, there is no need to compress the source brightness.
- KP MaxSL; when the source brightness is much larger than the display brightness, the brightness interval to be compressed is larger, and KP chooses a smaller value; when the source brightness is greater than the display brightness and the difference between the two is not large, the value of KP can be taken as a larger value.
- Hermite interpolation uses cubic double Hermite interpolation. It should be understood that the interpolation method can also use other interpolation methods. The formula is as follows:
- x KP
- x1 maxSL
- y0 KP
- y1 maxDL
- y0 ′ 1
- y1 ′ 0.
- the display device Because the display device has a minimum brightness, even if a pure black signal is given to the display device, it will be displayed as the minimum brightness value on the display device. This minimum brightness is called the black level. Therefore, the display brightness needs to be raised to a certain level by BlackLevelLift, which is abbreviated as BLL to protect the detail brightness whose brightness is less than the threshold level.
- the threshold level boost BLL value is based on the input of the brightness mapping curve TM_Curve. Calculation:
- the black level rise BLL here is the brightness value in the non-linear space
- the BLoffset in the aforementioned 507 is the brightness value in the linear space to ensure that the brightness mapping curve TM_Curve can map the minimum brightness of the source to the minimum brightness of the screen
- the BLL increase TM_Curve has a small change, and the BLL needs to be normalized to obtain the normalized BLLnorm; the normalization result is as follows:
- BLLnorm MAX ((BLL-norm_y1) / (norm_y0-norm_y1) * (minDL-maxSL), 0) (35)
- norm_x0 minSL
- norm_y0 minDL * (1-minSL) ⁇ n;
- norm_x1 KP
- norm_y1 minDL * (1-KP) ⁇ n;
- the brightness mapping curve TM_Curve outputs e2 e1 + BLLnorm, where BLLnorm is the brightness value improvement caused by the black level after normalization;
- TM_Curve_x (x0, x1, ... xn) is the abscissa of n sampling points, that is, the input of the curve, represents the brightness signal before the brightness mapping;
- TM_Curve_y (y0, y1, ... yn) is the ordinate of n sampling points, that is, the output of the curve, which represents the brightness signal after brightness mapping;
- mapping curve TM_Curve can be expressed in other ways (such as formulating) according to requirements; for details, refer to the description of TM_Curve in 500.
- the brightness mapping curve is a relative value mapping curve.
- the input of the mapping curve is [0,1] for [minSL, maxSL]
- the output of the mapping curve is [0, 1] represents [minDL, maxDL]; in the embodiment of the present application, the brightness mapping curve directly calculates the output brightness based on the input brightness.
- the brightness mapping curve is an absolute value mapping curve.
- the mapping curve input [0,1] represents [0,10000]. ] nits
- the mapping curve output [0,1] means [0,10000] nits.
- the traditional brightness mapping curve is divided into three parts by selecting two thresholds.
- the low-luminance part that is, the brightness part below the first threshold
- the high-luminance part that is, the brightness part that is greater than the first threshold
- Compression, and the high-brightness part is further divided into two parts, in which the luminance part that is greater than the first threshold value and less than the second threshold value is flexibly compressed by a curve that is obtained by fitting the first threshold value and the second threshold value.
- the part greater than the second brightness inflection point is subjected to the second compression, that is, the brightness greater than the second threshold is mapped to the second threshold; segmentation mapping of the brightness, fully taking into account the individual brightness Characteristics, retain brightness details as much as possible, and improve the rationality of brightness mapping.
- the brightness mapping curve takes into account the influence of the black level of the display device on the brightness mapping curve, and retains the brightness details of the low brightness portion.
- FIG. 10 it is a flowchart of another method for processing brightness of an HDR video signal according to an embodiment of the present application.
- the source video signal input in the embodiment of the present application is a scene light signal HDR and an HLG signal.
- the method may include:
- the input video signal in the embodiment of the present application is HDR in YUV space, and the HLG signal YUV0.
- a nonlinear scene light signal R'sG'sB's in RGB color space is obtained; further, the nonlinearity
- the scene light signal R'sG'sB's is converted into a linear scene light signal RsGsBs.
- the nonlinear scene light signal R'sG'sB's is converted into a linear scene light signal RsGsBs based on the HLG OETF-1 curve.
- the conversion from YUV to RGB is considered to be a color space conversion, and the non-linear RGB to linear RGB signals are not converted in color space.
- the non-linear RGB signals and the linear RGB signals Both belong to the RGB color space.
- the input video signal is converted from a YUV signal to a non-linear scene light signal R'sG'sB's into a color space conversion and the non-linear scene light signal R'dG
- the conversion of 'sB's to the linear scene light signal RsGsBs can also be considered as a color space conversion.
- the nonlinear RGB scene light signal and the linear RGB scene light signal belong to different color spaces.
- Ys is the linear scene brightness signal
- the scene brightness signal is the brightness component of the scene light signal.
- the choice of the parameters cr, cg, and cb is related to the color gamut of the linear scene light signal RsGsBs.
- the color gamut of the light signal of the linear scene is different, the corresponding calculation parameters will also be different. Therefore, the brightness needs to be calculated according to where the light signal is located.
- the HLG OOTF defined in ITU BT.2100 is used to convert the linear scene light signal HLG to a linear display light signal for display. This method calculates the maximum brightness and minimum brightness of the actual display device. "System Gamma" related parameters convert linear scene light signals into linear display light signals within the dynamic range of the display device.
- HLG OOTF is defined as follows:
- L W and L B are the maximum brightness and the minimum brightness of the display device, respectively.
- the scene brightness Y S is obtained, it is converted into display brightness, and brightness mapping processing is performed on the display brightness, or tone mapping processing is performed.
- the brightness obtained based on the scene light signal is a scene light brightness signal, and the scene light brightness signal needs to be converted into a display light brightness signal.
- This step is an exemplary implementation of the brightness signal type conversion.
- Linear scene brightness Ys is converted to linear display brightness Yd;
- the maximum brightness of the display signal LW 1000 nits is set; LW may also set other values.
- Converting scene brightness to display brightness improves compatibility with the processing of scene signals and display signals.
- step 603 Please refer to the description of step 603, which will not be repeated here.
- step 604 Same as step 604, please refer to the description of step 604, which will not be repeated here.
- step 605 Same as step 605, please refer to the description in step 605, which will not be repeated here.
- K Yt / Ys, that is, K represents the ratio of the luminance signal Yt after luminance mapping in the linear display space to the scene luminance signal Ys before luminance mapping;
- Rt, Gt, and Bt are the red, green, and blue components of the linear display light signal after the luminance mapping process, respectively;
- BLoffset is the black level of the display device. For details, refer to the description in step 507.
- the processed linear display light signal RtGtBt After obtaining the processed linear display light signal RtGtBt, according to the color space of the actual display device, color space conversion is performed on the linear display light signal RtGtBt, and the RtGtBt signal is converted to the color space of the display device to obtain the display brightness range with the display device.
- the processed video signal that matches the color space; for example, if the color space of the display device is sRGB space, it can be displayed directly without color space conversion; if the color space of the display device is YUV, RtGtBt The signal is converted into a video signal YUV1 in YUV space.
- the luminance signal of the scene light signal is converted into a display light luminance signal, and the linear display light luminance signal is converted into a non-linear space, and the luminance of the display brightness is mapped on the non-linear space.
- the display brightness range of the video signal is reasonably mapped to the brightness range that the display device can display to improve the contrast, brightness, and detail performance of the picture; after the HLG scene signal is converted into a linear display light brightness signal, the display is not directly performed, but instead The linear display brightness signal is converted into a non-linear space, and the brightness mapping is performed in the non-linear space.
- the embodiment of the present application considers the influence of the black level of the display device on the brightness mapping curve, and keeps the low Brightness details in the brightness section. Further, the HDR signal brightness processing method provided in the embodiment of the present application can convert an HDR signal into an SDR signal, thereby improving the compatibility of the SDR display device with the HDR signal.
- FIG. 6, FIG. 9, and FIG. 10 describe the methods in the form of steps. Although the method sequence is shown in the method flowcharts 6, 9, and 10, In some cases, the steps described may be performed in a different order than described here.
- a device for video signal processing includes a brightness obtaining unit 1301, a first conversion unit 1302, a brightness mapping unit 1303, a second conversion unit 1304, and a gain calculation unit. 1305 and a display signal acquisition unit 1306.
- the device may further include a compensation unit 1307 and a color space conversion unit 1308.
- the brightness obtaining unit 1301 is configured to obtain a first linear brightness signal, where the first linear brightness signal is obtained based on a first linear red-green-blue RGB signal corresponding to a video signal to be processed. For details, refer to the description in step 501, and details are not described herein again.
- the brightness obtaining unit 1301 is specifically configured to perform color space conversion on the PQ signal to obtain a first nonlinear RGB signal; according to the PQ electro-optical transfer function, The first non-linear RGB signal is converted into the first linear RGB signal; calculation is performed based on each primary color signal of the first linear RGB signal to obtain the first linear luminance signal.
- the brightness acquisition unit is specifically configured to: perform color space conversion on the HLG signal to obtain a second non-linear RGB signal;
- the second non-linear RGB signal is converted into a second linear RGB signal; calculation is performed based on each primary color signal of the second linear RGB signal to obtain a third linear luminance signal; and the third linear luminance signal is subjected to luminance signal type conversion to The first linear luminance signal is obtained.
- a first conversion unit 1302 configured to convert the first linear luminance signal into a first non-linear luminance signal
- the first conversion unit 1302 may also be used to complete steps 603 and 1004;
- a brightness mapping unit 1303, configured to perform segmented brightness mapping on the first nonlinear brightness signal to obtain a second nonlinear brightness signal;
- the brightness mapping unit 1303 may also be used to complete steps 604 and 1005. It should be understood that the brightness mapping unit may call a look-up table, brightness mapping curve, or brightness mapping formula stored in the memory to complete the brightness mapping of the video signal to be processed.
- the brightness mapping curve please refer to the description of the embodiment corresponding to 600 and FIG. 9.
- the second conversion unit 1304 may also be used to complete steps 605 and 1006;
- a gain calculation unit 1305, configured to calculate a brightness gain of the second linear brightness signal and the first linear brightness signal
- the gain calculation unit 1305 may also be used to complete steps 606 and 1007;
- a display signal obtaining unit 1306, configured to obtain an RGB display signal corresponding to the video signal to be processed based on a product of the brightness gain and the first linear RGB signal;
- a compensation unit 1307 is configured to superimpose a black level on each primary color value of the RGB display signal and increase BlackLevelLift to obtain a processed RGB display signal, where the BlackLevelLift is a minimum value of the display brightness of the display device;
- the display signal acquisition unit 1306 and the compensation unit 1307 may be used together to complete steps 607 and 1008;
- the color space conversion unit 1308 is configured to perform color space conversion on the processed RGB display signal to obtain a target display signal.
- the color format of the target display signal is the same as the color format corresponding to the display device. It should be understood that if the color format corresponding to the display device is RGB, the display can be performed directly, and no color space conversion is required.
- the following is a specific implementation manner of video signal processing provided by the present application.
- the embodiment of the application converts the HDR HLG signal into an SDR signal to adapt to the SDR TV.
- saturation mapping is performed first, and then brightness mapping and color gamut mapping are performed.
- the processing order of saturation mapping, brightness mapping, and color gamut mapping may be reversed. This is not limited.
- Y ⁇ sCbsCrs be a 4: 4: 4 YCbCr non-linear video signal restored by the terminal after AVS2 decoding and reconstruction and chroma upsampling.
- Each component is a 10-bit digitally encoded value.
- Ynorm should clip to [0,1]
- f sm () is a saturation mapping curve, which is calculated according to the brightness mapping curve f tm (), and the calculation steps are:
- L is the input linear brightness
- the unit is nit
- the result of f tm (L) is the linear brightness, the unit is nit;
- e is the normalized HLG signal brightness
- f tmHLG (e) is the normalized HLG signal brightness
- the saturation mapping curve input e, f sm (e) is the saturation mapping gain on the HLG space
- the YiCbiCri signal is a 10-bit limited range digital coded value, where the Yi value should be in the [64,940] interval, and the Cbi, Cri value should be in the [64,960] interval.
- Y ⁇ s Cb s Cr s signal is a 10-bit digital code value limits, obtained through the process R ⁇ s G ⁇ s B ⁇ s nonlinear primary values are floating-point numerical clip to be [0,1] Interval.
- Equation E s s B s represents the signal R s G value of any of a linear primary color component has a value in the interval [0,1];
- E ⁇ s refers R ⁇ s G ⁇ s B ⁇ s signal to any one of The non-linear primary color value of the component.
- the function HLG_OETF -1 () is defined according to ITU BT.2100 as follows:
- the linear brightness Y s is calculated as follows:
- Y s is a real number whose value is in the interval [0,1].
- hmt (x) 0.4064 ⁇ ⁇ 0 (x) + 0.5791 ⁇ ⁇ 1 (x) + ⁇ 0 (x)
- Y t is a real number, and its value should be clipped to the interval [0,200].
- Es represents any component in the R s G s B s signal
- E tm represents any component in the R tm G tm B tm signal.
- the RtGtBt obtained through this process is a floating-point linear primary color value, and the value should be clip to [0,200].
- ⁇ in this embodiment may be 2.2 or 2.4, or other values.
- the value of ⁇ may be selected according to actual conditions, which is not limited in the embodiment of the present application.
- R ⁇ tG ⁇ tB ⁇ t is a non-linear primary color value, and the value is in the interval [0,1].
- the Y ⁇ tCbtCrt signal obtained after this processing is a 10-bit limited range digital coded value, where the Y ⁇ t value should be in the [64,940] interval, and the Cbt, Crt values should be in the [64,960] interval.
- a process flowchart of an HDR terminal technical solution provided in the present application shows a technical framework and a related scope in the entire video end-to-end system.
- the technical framework of the end-to-end system provided by the embodiment of the present application can convert HDR HLG video signals into SDR signals, thereby adapting to SDR TV, and can also convert HDR HLG signals into HDR PQ signals, adapting HDR PQ TV for display; optionally, HDR PQ signals can also be converted to SDR signals for playback on SDR TV; optionally, dynamic metadata of video signals can be discarded, and only static metadata is retained, which is suitable for Play with HDR PQ TV.
- the technique will be a second-generation audio frame video signal and the video coding standard HDR dynamic metadata or static metadata (2 nd Audio Video coding Standard, AVS2) encoded into code AVS2 streamed to the terminal, optionally,
- HDR dynamic metadata or static metadata (2 nd Audio Video coding Standard, AVS2) encoded into code AVS2 streamed to the terminal
- HEVC High Efficiency Video Coding
- the HDR input signal source supported in the embodiment of the present application is an AVS2 code stream received at a terminal.
- an integer YCbCr color difference signal (hereinafter referred to as YCC signal for short, which is not described in this article) in the format of 10bit 4: 2: 0 and related HDR static or dynamic metadata are obtained.
- YCC and HDR signals may be in several formats listed in Table 1:
- the TVs that may be connected to the terminal have very different support capabilities for HDR signals. It is often difficult for older TV models to support new signal formats. To this end, the decoding terminal needs to perform compatible adaptation according to the difference in the TV's ability to support HDR signals.
- the TVs that terminal devices may need to interface with are classified into the following categories, as shown in Table 3:
- the embodiment of the present application performs corresponding signal conversion processing according to different HDR signal formats received by the terminal and different HDR signal support capabilities of the docking TV, so as to achieve compatible adaptation between the signal and the TV.
- YsCbsCrs be a 4: 4: 4 YCbCr non-linear video signal restored by the terminal after AVS2 decoding and reconstruction and chroma upsampling.
- Each component is a 10-bit digitally encoded value.
- the YsCbsCrs signal is a 10-bit limited range digital coded value.
- the R ⁇ sG ⁇ sB ⁇ s obtained through this process is a floating-point non-linear primary color value, and the value should be clip to the [0,1] interval.
- the input HDR HLG video signal is a video signal in YCC space. It can also be said that HLG is a scene light signal.
- the HLG video signal is converted to the RGB color space to obtain a nonlinear scene light signal R ⁇ sG ⁇ sB. ⁇ s.
- the brightness mapping process is completed in the RGB color space. It should be understood that the video signal obtained by the color space conversion is a non-linear video signal. If the input is an HLG signal, the color space conversion is a non-linear scene light signal.
- Equation E s s B s represents a signal of any one of R s G component;
- E ⁇ s refers R ⁇ s G ⁇ s B ⁇ s signal to any one component.
- the function HLG_OETF -1 () is defined according to ITU BT.2100 as follows:
- a first linear display luminance signal is obtained based on the linear video signal.
- the first linear display luminance signal is a linear display luminance signal.
- the linear video signal is a linear scene light signal due to the input HDR HLG signal
- the brightness signal Ys obtained based on the linear scene light signal is a scene brightness signal.
- the linear brightness Ys is calculated as follows:
- the input signal is HDR, HLG signal, and scene light signal
- the Ys calculated in the above steps is the linear scene brightness
- the input of the brightness mapping curve is the display brightness, so it needs to be converted into a display before performing the brightness mapping.
- Lightness signal Yd is the input signal.
- the brightness mapping is performed in a non-linear space, so the brightness mapping input is a non-linear display brightness, so the linear display brightness needs to be converted into a non-linear space to obtain the non-linear display brightness Y dPQ ;
- the signal obtained after the luminance mapping is a non-linear display light luminance signal.
- the non-linear display light luminance signal can be converted into a linear display light luminance signal Yt through a PQ EOTF curve.
- Yt is a real number, and its value is in the interval [0,100].
- ⁇ in this embodiment may be 2.2 or 2.4, or other values.
- the value of ⁇ may be selected according to actual conditions. This embodiment of the present application does not limit this. .
- E s represents any component in the R s G s B s signal
- E tm represents any component in the R tm G tm B tm signal.
- the color gamut mapping may be:
- the function EOTF -1 () may be the inverse function of BT.1886EOTF or the inverse function of BT.2100PQ EOTF curve.
- E ⁇ t (E t / 100) 1 / 2.4 .
- R ⁇ tG ⁇ tB ⁇ t is a floating-point non-linear primary color value, and the value is in the interval [0,1].
- the YtCbtCrt signal obtained after this processing is a 10-bit limited range digitally encoded value.
- the YoCboCro signal is a 10-bit limited range digital coded value.
- This process is suitable for compatible adaptation processing of HDR HLG signals and HDR PQ signals.
- the parameters of each step of the above processing flow are different. It should be understood that, in the embodiment of the present application, the video signal is first subjected to brightness processing, then color gamut processing, and then saturation processing. In an optional case, color gamut processing, brightness processing, and saturation processing may be performed first.
- this embodiment proposes to perform compatible adaptation in the manner listed in Table 4.
- this embodiment also converts the HDR HLG signal to the BT.709 color gamut and sends it to the type 2 TV.
- This part of the processing is a link of HLG signal to SDR signal compatible adaptation processing. Since this processing method has been given a conceptual introduction in the BT.2407 report, this section refers to the content of the ITU report for informational explanation.
- the conversion of BT.2020 wide color gamut signals to BT.709 signals can be achieved by a method based on linear matrix conversion. In addition to hard-clip the output signal, this method is completely the inverse process of ITU standard BT.2087.
- each component signal through a linear transfer function to achieve conversion signal (E R E G E B)
- the conversion function may be an HLG EOTF function.
- BT709 linear RGB signals (ER R G E B ) are to be used in BT709 display devices, and should be converted to BT709 nonlinear RGB signals (E ⁇ R E ⁇ G E ⁇ B ).
- ⁇ in this embodiment may be 2.2 or 2.4, or other values.
- the value of ⁇ may be selected according to actual conditions, which is not limited in the embodiment of the present application.
- the first reference peak brightness Lw from HLG to PQ signal is 1000nit, and the black level Lb is 0.
- the same PQ image as the HLG image can be generated in the color volume within 1000nit using the process shown in Figure 16:
- 1000nit HLG source signal can generate linear scene light signal after HLG OETF inverse function
- the linear scene light signal can generate a linear display light signal through the OOTF function of HLG;
- the linear display light signal can generate 1000nit PQ display light signal through the EOTF inverse function of PQ;
- This processing flow involves the process of converting HDR HLG signals into HDR PQ signals and displaying them on TV.
- YsCbsCrs be a 4: 4: 4 YCbCr non-linear video signal restored by the terminal after AVS2 decoding and reconstruction and chroma upsampling.
- Each component is a 10-bit digitally encoded value.
- the YsCbsCrs signal is a 10-bit limited range digital coded value.
- the R ⁇ sG ⁇ sB ⁇ s obtained through this process is a floating-point non-linear primary color value, and the value should be clip to the [0,1] interval.
- the input HDR HLG video signal is a video signal in YCC space. It can also be said that HLG is a scene light signal.
- the HLG video signal is converted to the RGB color space to obtain a nonlinear scene light signal R ⁇ sG ⁇ sB. ⁇ s.
- the brightness mapping process is completed in the RGB color space. It should be understood that the video signal obtained by the color space conversion is a non-linear video signal. If the input is an HLG signal, the color space conversion is a non-linear scene light signal.
- the non-linear scene light signal R ⁇ sG ⁇ sB ⁇ s is converted into a linear scene light signal RsGsBs.
- the non-linear scene light signal R ⁇ sG ⁇ sB ⁇ s can be based on the HLG photoelectric transfer inverse function. Converted into linear scene light signals RsGsBs.
- HLG_OETF -1 () is defined according to ITU BT.2100 as follows:
- a first linear display luminance signal is obtained based on the linear video signal.
- the first linear display luminance signal is a linear display luminance signal.
- the linear video signal is a linear scene light signal due to the input HDR HLG signal
- the brightness signal Ys obtained based on the linear scene light signal is the scene brightness signal. It needs to be converted into display light before performing brightness mapping. Brightness signal Yd.
- the linear brightness Ys is calculated as follows:
- Es represents any component in the RsGsBs signal
- E tm represents any component in the R tm G tm B tm signal.
- R ⁇ tG ⁇ tB ⁇ t is a floating-point non-linear primary color value, and the value is in the interval [0,1].
- the YtCbtCrt signal obtained after this processing is a 10-bit limited range digitally encoded value.
- the terminal may complete the frame rate adjustment, the bit width adjustment, and the 4: 4: 4 to 4: 2: 2/4: 2 according to the frame rate, bit width, and chroma downsampling method determined by the HDR and PQ TV. : 0 Downsampling and other subsequent processing can transmit the HDR PQ signal generated by the conversion to the HDR PQ TV.
- this embodiment proposes to perform compatible adaptation in the manner listed in Table 5.
- the terminal can complete the frame rate adjustment, bit width adjustment, and 4: 4: 4 to 4: 2: 2/4: 2 according to the frame rate, bit width, and chroma downsampling method determined by the connected SDR TV. Subsequent processing such as 0 downsampling can transmit the SDR signal generated by the conversion to the SDR TV.
- HDR PQ signals have HDMI 2.0A or higher HDMI interface capabilities, and have more mature HDR PQ signal processing capabilities
- this embodiment suggests that when receiving HDR PQ signals, the HDR PQ signal and The static metadata is directly output to the HDR, PQ, and TV through the HDMI interface of HDMI 2.0A and above.
- the TV then completes the subsequent display processing of the HDR and PQ signals.
- the dynamic metadata it cannot be passed to the TV due to the limitation of the HDMI interface, and it should be discarded after decoding.
- the specific method is to continuously broadcast the image to be evaluated and the reference image to the observer according to the test schemes described later, and then leave a certain time interval for the observer to score after the playback, and finally average all the given scores as the
- the evaluation value of the sequence is the evaluation value of the image to be evaluated.
- HDR and HLG signals adapt to SDR and TV
- Test purpose Through comparative tests, it is shown whether the image adaptation conversion from HLG to SDR can provide beneficial image effects when the HDR and HLG signals are sent to the SDR and TV.
- Figure 17 shows a schematic diagram of a test network. among them:
- DUT2 HLG to SDR BT2020
- BenchMark1 HLG watched on SDR BT709
- BenchMark1 HLG watching in HDR, HLG, TV in HLG, BT2020 mode
- HDR HLG signal adapts HDR PQ TV
- FIG. 18 it is a schematic diagram of a networking manner of another test solution according to an embodiment of the present application.
- Benchmark1 HLG watching in SDR and BT709 mode on HDR PQ TV
- Benchmark2 HLG watching on HDR PQ TV in HLG BT2020 mode
- HDR end-to-end system For PQ curve HDR video, the HDR end-to-end system is shown in Figure 19.
- program production PQ curve HDR video and static metadata are obtained, and the HDR video production parameters meet the requirements of GY / T 315-2018 standard.
- HDR pre-processing realizes the extraction of dynamic metadata, obtains HDR video and metadata for encoding and transmission, and encodes and encapsulates the AVS2 before transmitting it on the network.
- the AVS2 decoder decodes the HDR video and metadata.
- HDR video and metadata reconstruction is used to obtain SDR video for display; for HDR display terminals, if the terminal display capability is the same as the HDR video produced and transmitted, the HDR display is directly performed; if the terminal display capability and produced transmission The brightness of the HDR video is different, and then the HDR video and metadata are adapted to be displayed according to the display capability of the terminal.
- HLG curve HDR video when the program production uses the maximum brightness of 1000cd / m2, the HDR end-to-end system is shown in Figure 20. Through program production, HLG curve HDR video is obtained, and the production parameters of HDR video are in compliance with GY / T315-2018 standards. After using AVS2 encoding for HDR video, it is transmitted on the network. At the receiving end, the AVS2 decoder decodes the HDR video and displays it directly on the SDR and HDR terminals.
- HLG curve HDR video when the highest brightness of the program production is not 1000cd / m2, the HDR end-to-end system is shown in Figure 21.
- HLG curve HDR video and static metadata are obtained.
- HDR video production parameters meet the requirements of GY / T315-2018 standard.
- HDR video and static metadata are encoded and encapsulated in AVS2 and transmitted over the network.
- the AVS2 decoder obtains HDR video after decoding
- the AVS2 decoder obtains HDR video and static metadata after decoding.
- SDR display terminals direct display; for HDR display terminals, static metadata can be used to display the adjusted gamma value using the method specified in Table 5 of GY / T315-2018.
- Metadata information metadata_info contains dynamic metadata that can be used to reconstruct HDR and images when combined with the associated transmitted image.
- hdr_characteristics () contains the HDR image signal characteristics, that is, the identification of the HDR image color space, and the identification for adjusting the primary color of the HDR image monitor.
- This variable represents the primary color and the reference white coordinates that conform to the HDR image color space defined in CIE1931.
- This variable represents the primary and reference white coordinates of the main monitor color space of the adjusted HDR image as defined by CIE1931.
- This variable defines the nominal maximum display brightness of the main monitor that adjusts the HDR image, measured in candela per square meter (cd / m2) and rounded to an integer multiple of 50 cd / m2.
- This variable defines the nominal minimum display brightness of the main monitor that adjusts the HDR image. In 0.0001cd / m2.
- hdrDisplayMinLuminance should be less than hdrDisplayMaxLuminance. If the value of the variable is unknown, it is recommended to set it to 0.
- sdr_characteristics contains the characteristics of the SDR image signal, that is, the color space identification of the SDR image, and the nominal maximum and minimum brightness values used to adjust the SDR image main monitor.
- This variable defines the nominal maximum display brightness of the main monitor used to adjust the SDR image, in units of 1cd / m2, and rounded to an integer multiple of 50cd / m2.
- luminance_mapping_variables contains luminance mapping variables used to build the lookup table lutMapY.
- This variable is used in the first step of calculating the signal gain during the reconstruction of the luminance mapping curve, and represents the black level offset eliminated during the chrominance volume reconstruction.
- the value of this variable should be in the range [0,1] and be an integer multiple of (1 ⁇ 255).
- the set-top box learns the display capabilities of the TV.
- the set-top box analyzes the received program stream and obtains the information that the program is SDR program / HDR program, maximum brightness, minimum brightness, etc. If it matches the display capability of the TV, it decodes the audio and video and sends it to the TV for display via HDMI. ; If it does not match the TV ’s actual capabilities, decode the audio and video and obtain the program signal adapted to the TV ’s display capabilities through display adaptation, and send it to the TV via HDMI for display.
- the all-in-one decodes the received program code stream and adjusts and displays it according to the actual display capability of the TV.
- This embodiment of the present application describes a method for transmitting HDR metadata in an HDMI interface.
- the HDMI 2.0a specification specifies how to transmit HDR static metadata in the HDMI interface
- the HDMI 2.1 specification specifies how to transmit HDR dynamic data in the HDMI interface. Metadata.
- the HDMI 2.0a standard adopts the CEA-861.3-2014 specification for HDR static metadata transmission.
- upstream source processing equipment such as a set-top box
- the HDR static metadata can be in the CEA-861.3-2014 interface It is transmitted to a receiver (such as a television) that can receive and process HDR static metadata.
- This appendix explains the HDR static metadata data block "HDR Static Metadata Data Block” (C.2.2, corresponding to CEA-861.3-2014 specification 4.2) and the dynamic range main supervisor information frame "Dynamic MetaMetadata" in CEA-861.3-2014.
- Mastering InfoFrame “(C.2.3, corresponding to 3.2 of CEA-861.3-2014 specification) information transmission and mapping.
- the HDR static metadata data block "HDR Static Metadata Data Block” is used to transmit the receiving end HDR static metadata support capability information to the source device.
- the "ET_2" bit in the data block transmitted from the receiver to the source device is equal to 1, it means that the receiver supports the PQ EOTF curve specified by GY / T315-2018, and the "SM_0" bit is equal to 1, which means that it supports the static elements specified in this article. data.
- the dynamic range master information frame "Dynamic Range and Mastering InfoFrame" is used by the source device to identify and transmit HDR dynamic metadata to the receiving device.
- the source device uses the "EOTF” value of 2 to mark the EOTF of the transport stream as the PQ curve specified by GY / T315-2018, and "Static_MetaData_Descriptor_ID" is 0 to mark that the static metadata of this standard is carried in the user extension information, so that the transmission complies with Static metadata specified in this standard.
- the HDMI 2.1 standard uses the CTA-861-G-2016 specification for HDR dynamic metadata transmission.
- CTA-861-G2016 specifies how to carry HDR metadata in the CE terminal interface.
- the upstream source processing equipment such as a set-top box
- the HDR metadata can be transmitted in the CTA-861-G-2016 interface to a receiver (such as a television) that can receive and process the HDR metadata.
- This appendix explains the HDR dynamic metadata data block "HDR Dynamic Metadata Data Block” (C.3.2, corresponding to 7.5.14 of the CTA-861-G-2016 specification) and HDR dynamic metadata in CTA-861-G-2016.
- Information transmission and mapping of the extended information frame "HDR Dynamic Metadata Extended InfoFrame” (C.3.3, corresponding to the 6.10.1 of the CTA-861-G-2016 specification).
- the HDR dynamic metadata data block "HDR Dynamic Metadata Data Block” is used to transmit the receiving end HDR dynamic metadata support capability information to the source device.
- the value of the supported dynamic metadata type "Supported HDR Dynamic Metadata Type" in the data block transmitted from the receiving end to the source device is 0x0002, it means that the receiving end supports the dynamic metadata specified in this standard.
- the support flag "Support Flags" bytes in the data block represent different HDR transmission modes. Its binary value consisting of 0 to 3 bits is greater than or equal to 1, and the fifth bit is equal to 1, indicating that it supports the HDR transmission mode of this standard.
- HDR dynamic metadata extended information frame "HDR Dynamic Metadata Extended InfoFrame" is used by the source device to identify and transmit the HDR dynamic metadata to the receiving device.
- the source device uses the extended information frame type "Extended InfoFrame Type" value of 0x0002 to mark that this standard's dynamic metadata is carried in the user's extended information, thereby transmitting dynamic metadata that complies with the specifications of this standard.
- An embodiment of the present application further provides a computer-readable storage medium.
- the computer-readable storage medium stores instructions, and when the computer-readable storage medium runs on the computer, the computer causes one or more steps in any one of the foregoing methods to be executed.
- each component module of the above signal processing device is implemented in the form of a software functional unit and sold or used as an independent product, it may be stored in the computer-readable storage medium.
- the embodiments of the present application further provide a computer program product containing instructions.
- the technical solution of the present application is essentially a part that contributes to the existing technology or all or part of the technical solution may be a software product.
- the computer software product is stored in a storage medium, and includes several instructions for causing a computer device (which may be a personal computer, a server, or a network device) or a processor therein to execute the embodiments of the present application. All or part of the steps of the method. Please refer to the relevant description of the memory 302 for the type of the storage medium.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Computer Hardware Design (AREA)
- Processing Of Color Television Signals (AREA)
- Control Of Indicators Other Than Cathode Ray Tubes (AREA)
- Controls And Circuits For Display Device (AREA)
- Image Processing (AREA)
- Transforming Electric Information Into Light Information (AREA)
Abstract
Description
电视机类型 | 色域 | 转换曲线 |
类型1 | BT.709 | 伽马 |
类型2 | BT.2020 | 伽马 |
类型3 | BT.2020 | 支持PQ |
类型4 | BT.2020 | 支持PQ和HLG |
Claims (25)
- 一种视频信号的处理方法,其特征在于,所述方法包括:获得第一线性亮度信号,所述第一线性亮度信号基于待处理视频信号对应的第一线性红绿蓝RGB信号获得;将所述第一线性亮度信号转换为第一非线性亮度信号;对所述第一非线性亮度信号进行分段亮度映射,以得到第二非线性亮度信号;将所述第二非线性亮度信号转换为第二线性亮度信号;计算所述第二线性亮度信号与所述第一线性亮度信号的亮度增益;基于所述亮度增益和所述第一线性RGB信号的乘积,获得所述待处理视频信号对应的RGB显示信号。
- 根据权利要求1所述的方法,其特征在于,所述待处理视频信号为感知量化PQ信号,所述获得第一线性亮度信号,包括:对所述PQ信号进行色彩空间转换,以获得第一非线性RGB信号;根据PQ电光转移函数,将所述第一非线性RGB信号转换为所述第一线性RGB信号;基于所述第一线性RGB信号的各基色信号进行计算,以获得所述第一线性亮度信号。
- 根据权利要求1所述的方法,其特征在于,所述待处理视频信号为混合对数伽玛HLG信号,所述获得第一线性亮度信号,包括:对所述HLG信号进行色彩空间转换,以获得第二非线性RGB信号;根据HLG光电转移逆函数,将所述第二非线性RGB信号转换为第二线性RGB信号;基于所述第二线性RGB信号的各基色信号进行计算,以获得第三线性亮度信号;对所述第三线性亮度信号进行亮度信号类型转换,以获得所述第一线性亮度信号。
- 根据权利要求1至3任一项所述的方法,其特征在于,在所述获得所述待处理视频信号对应的RGB显示信号之后,还包括:对所述RGB显示信号进行色彩空间转换,以获得目标显示信号,其中,所述目标显示信号的色彩格式和显示设备对应的色彩格式相同。
- 根据权利要求4所述的方法,其特征在于,在所述获得所述待处理视频信号对应的RGB显示信号之后,还包括:对所述RGB显示信号的各基色值叠加黑位电平提升BlackLevelLift,以获得处理后的RGB显示信号,所述BlackLevelLift为所述显示设备的显示亮度的最小值;对应的,所述对所述RGB显示信号进行色彩空间转换,包括:对所述处理后的RGB显示信号进行色彩空间转换。
- 根据权利要求4或5任一项所述的方法,其特征在于,所述对所述第一非线性亮度信号进行分段亮度映射,以得到第二非线性亮度信号,包括:确定第一阈值和第二阈值,所述第一阈值小于所述第二阈值;当所述第一非线性亮度信号的亮度值小于或等于所述第一阈值时,所述第二非线性亮度信号的亮度值等于所述第一非线性亮度信号的亮度值;当所述第一非线性亮度信号的亮度值大于所述第一阈值,且小于或等于所述第二阈值时,所述第二非线性亮度信号的亮度值基于以所述第一非线性亮度信号的亮度值 为自变量的拟合曲线获得;当所述第一非线性亮度信号的亮度值大于所述第二阈值时,所述第二非线性亮度信号的亮度值等于所述显示设备对应的最大非线性显示亮度值。
- 根据权利要求6所述的方法,所述拟合曲线通过对所述第一阈值和所述第二阈值进行埃尔米特Hermite插值得到。
- 根据权利要求6至8任一项所述的方法,其特征在于,所述确定第一阈值和第二阈值,包括:根据所述第一非线性亮度信号的显示亮度范围和所述显示设备的显示亮度范围的关系确定所述第一阈值;将所述第一非线性亮度信号的最大亮度值作为所述第二阈值。
- 根据权利要求1至5任一项所述的方法,其特征在于,所述对所述第一非线性亮度信号进行分段亮度映射,以得到第二非线性亮度信号,包括:基于预设的所述第一非线性亮度信号和所述第二非线性亮度信号的亮度值的映射关系,确定与所述第一非线性亮度信号的亮度值相对应的第二非线性亮度信号的亮度值。
- 根据权利要求1至10任一项所述的方法,其特征在于,所述将所述第一线性亮度信号转换为第一非线性亮度信号,包括:根据PQ电光转移逆函数,将所述第一线性亮度信号转换为所述第一非线性亮度信号;对应的,所述将所述第二非线性亮度信号转换为第二线性亮度信号,包括:根据PQ电光转移函数,将所述第二非线性亮度信号转换为所述第二线性亮度信号。
- 一种视频信号处理的装置,其特征在于,所述装置包括:亮度获取单元,用于获得第一线性亮度信号,所述第一线性亮度信号基于待处理视频信号对应的第一线性红绿蓝RGB信号获得;第一转换单元,用于将所述第一线性亮度信号转换为第一非线性亮度信号;亮度映射单元,用于对所述第一非线性亮度信号进行分段亮度映射,以得到第二非线性亮度信号;第二转换单元,用于将所述第二非线性亮度信号转换为第二线性亮度信号;增益计算单元,用于计算所述第二线性亮度信号与所述第一线性亮度信号的亮度增益;显示信号获取单元,用于基于所述亮度增益和所述第一线性RGB信号的乘积,获得所述待处理视频信号对应的RGB显示信号。
- 根据权利要求12所述的装置,其特征在于,所述待处理视频信号为感知量化PQ信号,所述亮度获取单元,具体用于:对所述PQ信号进行色彩空间转换,以获得第一非线性RGB信号;根据PQ电光转移函数,将所述第一非线性RGB信号转换为所述第一线性RGB信号;基于所述第一线性RGB信号的各基色信号进行计算,以获得所述第一线性亮度信号。
- 根据权利要求12所述的装置,其特征在于,所述待处理视频信号为混合对数伽玛HLG信号,所述亮度获取单元,具体用于:对所述HLG信号进行色彩空间转换,以获得第二非线性RGB信号;根据HLG光电转移逆函数,将所述第二非线性RGB信号转换为第二线性RGB信号;基于所述第二线性RGB信号的各基色信号进行计算,以获得第三线性亮度信号;对所述第三线性亮度信号进行亮度信号类型转换,以获得所述第一线性亮度信号。
- 根据权利要求12至14任一项所述的装置,其特征在于,所述装置还包括:色彩空间转换单元,用于:对所述RGB显示信号进行色彩空间转换,以获得目标显示信号,其中,所述目标显示信号的色彩格式和显示设备对应的色彩格式相同。
- 根据权利要求15所述的装置,其特征在于,所述装置还包括:补偿单元,用于:对所述RGB显示信号的各基色值叠加黑位电平提升BlackLevelLift,以获得处理后的RGB显示信号,所述BlackLevelLift为所述显示设备的显示亮度的最小值;对应的,所述色彩空间转换单元,具体用于:对所述处理后的RGB显示信号进行色彩空间转换。
- 根据权利要求12至16任一项所述的装置,其特征在于,所述亮度映射单元,具体用于:确定第一阈值和第二阈值,所述第一阈值小于所述第二阈值;当所述第一非线性亮度信号的亮度值小于或等于所述第一阈值时,所述第二非线性亮度信号的亮度值等于所述第一非线性亮度信号的亮度值;当所述第一非线性亮度信号的亮度值大于所述第一阈值,且小于或等于所述第二阈值时,所述第二非线性亮度信号的亮度值基于以所述第一非线性亮度信号的亮度值为自变量的拟合曲线获得;当所述第一非线性亮度信号的亮度值大于所述第二阈值时,所述第二非线性亮度信号的亮度值等于所述显示设备对应的最大非线性显示亮度值。
- 根据权利要求17所述的装置,其特征在于,所述拟合曲线通过对所述第一阈值和所述第二阈值进行埃尔米特Hermite插值得到。
- 根据权利要求17至19任一项所述的装置,其特征在于,所述亮度映射单元,具体用于:根据所述第一非线性亮度信号的显示亮度范围和所述显示设备的显示亮度范围的关系确定所述第一阈值;将所述第一非线性亮度信号的最大亮度值作为所述第二阈值。
- 根据权利要求12至16任一项所述的装置,其特征在于,所述亮度映射单元,具体用于:基于预设的所述第一非线性亮度信号和所述第二非线性亮度信号的亮度值的映射关系,确定与所述第一非线性亮度信号的亮度值相对应的第二非线性亮度信号的亮度值。
- 根据权利要求12至21任一项所述的装置,其特征在于,所述第一转换单元, 具体用于:根据PQ电光转移逆函数,将所述第一线性亮度信号转换为所述第一非线性亮度信号;对应的,所述第二转换单元,具体用于:根据PQ电光转移函数,将所述第二非线性亮度信号转换为所述第二线性亮度信号。
- 一种视频信号处理的装置,其特征在于,所述装置包括:处理器和存储器;所述处理器,用于调用所述存储器中的软件指令,以执行如权利要求1至11任一项所述的方法。
- 一种计算机可读存储介质,所述计算机可读存储介质中存储有指令,当其在计算机或处理器上运行时,使得所述计算机或处理器执行如权利要求1至11任一项所述的方法。
- 一种包含指令的计算机程序产品,当其在计算机或处理器上运行时,使得所述计算机或处理器执行如权利要求1至11任一项所述的方法。。
Priority Applications (5)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
KR1020217000634A KR102511572B1 (ko) | 2018-07-05 | 2019-06-10 | 비디오 신호 처리 방법 및 장치 |
JP2020568530A JP7115776B2 (ja) | 2018-07-05 | 2019-06-10 | ビデオ信号の処理方法及び装置 |
CN201980025155.1A CN112272947B (zh) | 2018-07-05 | 2019-06-10 | 一种视频信号处理的方法、装置及计算机可读存储介质 |
EP19830987.4A EP3796648A4 (en) | 2018-07-05 | 2019-06-10 | VIDEO SIGNAL PROCESSING METHOD AND DEVICE |
US17/135,568 US11317071B2 (en) | 2018-07-05 | 2020-12-28 | Video signal processing method and apparatus |
Applications Claiming Priority (4)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201810732833.5 | 2018-07-05 | ||
CN201810732833 | 2018-07-05 | ||
CN201810797968.X | 2018-07-19 | ||
CN201810797968.XA CN108900823B (zh) | 2018-07-05 | 2018-07-19 | 一种视频信号处理的方法及装置 |
Related Child Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US17/135,568 Continuation US11317071B2 (en) | 2018-07-05 | 2020-12-28 | Video signal processing method and apparatus |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2020007165A1 true WO2020007165A1 (zh) | 2020-01-09 |
Family
ID=64351196
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/CN2019/090566 WO2020007165A1 (zh) | 2018-07-05 | 2019-06-10 | 一种视频信号处理的方法及装置 |
Country Status (6)
Country | Link |
---|---|
US (1) | US11317071B2 (zh) |
EP (1) | EP3796648A4 (zh) |
JP (1) | JP7115776B2 (zh) |
KR (1) | KR102511572B1 (zh) |
CN (2) | CN108900823B (zh) |
WO (1) | WO2020007165A1 (zh) |
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP4053834A1 (en) * | 2021-03-05 | 2022-09-07 | Canon Kabushiki Kaisha | Image processing apparatus, image capture apparatus, control method, and program |
CN117440118A (zh) * | 2023-10-23 | 2024-01-23 | 镕铭微电子(上海)有限公司 | 图像处理方法、装置、电子设备及存储介质 |
CN117440118B (zh) * | 2023-10-23 | 2024-06-11 | 镕铭微电子(上海)有限公司 | 图像处理方法、装置、电子设备及存储介质 |
Families Citing this family (16)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN108900823B (zh) * | 2018-07-05 | 2019-07-12 | 华为技术有限公司 | 一种视频信号处理的方法及装置 |
CN109616037A (zh) * | 2018-12-18 | 2019-04-12 | Oppo广东移动通信有限公司 | 显示屏的显示校准方法、电子装置及计算机可读存储介质 |
US11405582B2 (en) * | 2019-06-28 | 2022-08-02 | Meta Platforms, Inc. | Preprocessing of high-dynamic-range video using a hybrid lookup table scheme |
CN110691239B (zh) * | 2019-12-11 | 2020-05-19 | 武汉精立电子技术有限公司 | 一种hdr显示模组检测方法、装置及电子设备 |
US11688361B2 (en) * | 2020-06-19 | 2023-06-27 | Apple Inc. | Quantifying performance of a display device |
CN112750086A (zh) * | 2020-08-31 | 2021-05-04 | 腾讯科技(深圳)有限公司 | 一种图像处理方法、装置、电子设备及存储介质 |
US11508331B1 (en) * | 2021-04-28 | 2022-11-22 | Novatek Microelectronics Corp. | Image uniformity compensation device |
CN113676773B (zh) * | 2021-08-16 | 2023-11-14 | 广州虎牙信息科技有限公司 | 一种视频播放方法、系统、装置、计算机设备和存储介质 |
TWI788983B (zh) * | 2021-08-30 | 2023-01-01 | 瑞昱半導體股份有限公司 | 視訊訊號處理裝置與方法 |
US11769464B2 (en) * | 2021-09-02 | 2023-09-26 | Arm Limited | Image processing |
WO2023064426A1 (en) * | 2021-10-13 | 2023-04-20 | Freedom Scientific, Inc. | Apparatus and method for reducing photophobia in electronic screens |
CN114466147B (zh) * | 2021-12-23 | 2024-03-15 | 阿里巴巴(中国)有限公司 | 一种视频亮度的调整方法、装置、电子设备以及存储介质 |
CN115564659B (zh) * | 2022-02-28 | 2024-04-05 | 荣耀终端有限公司 | 视频处理方法和装置 |
CN114898722A (zh) * | 2022-05-25 | 2022-08-12 | 深圳创维-Rgb电子有限公司 | 画面显示方法、装置、设备及计算机可读存储介质 |
CN115063325A (zh) * | 2022-08-17 | 2022-09-16 | 中央广播电视总台 | 一种视频信号处理方法、装置、计算机设备和存储介质 |
CN115762404B (zh) * | 2023-01-10 | 2023-04-07 | 禹创半导体(深圳)有限公司 | Oled调光的dbv重新映射方法、装置和oled显示设备 |
Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20090310015A1 (en) * | 2008-06-16 | 2009-12-17 | El-Mahdy Ahmed Hazem Mohamed R | Transformation of a video image from a high dynamic range image to a low dynamic range image |
CN105516674A (zh) * | 2015-12-24 | 2016-04-20 | 潮州响石数码技术有限公司 | 一种具有hdr显示功能的监视设备 |
CN106780402A (zh) * | 2017-01-16 | 2017-05-31 | 聚龙智瞳科技有限公司 | 基于Bayer格式的图像动态范围扩展方法及装置 |
CN106878694A (zh) * | 2015-12-10 | 2017-06-20 | 瑞昱半导体股份有限公司 | 高动态范围信号处理系统与方法 |
CN107046628A (zh) * | 2017-03-15 | 2017-08-15 | 四川长虹电器股份有限公司 | 一种低亮度8k电视机显示hdr效果的方法 |
CN108900823A (zh) * | 2018-07-05 | 2018-11-27 | 华为技术有限公司 | 一种视频信号处理的方法及装置 |
Family Cites Families (17)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
TWI466093B (zh) * | 2007-06-26 | 2014-12-21 | Apple Inc | 用於視訊播放的管理技術 |
US8285059B2 (en) * | 2008-05-20 | 2012-10-09 | Xerox Corporation | Method for automatic enhancement of images containing snow |
JP5275122B2 (ja) | 2008-05-30 | 2013-08-28 | パナソニック株式会社 | ダイナミックレンジ圧縮装置、ダイナミックレンジ圧縮方法、プログラム、集積回路および撮像装置 |
KR102060604B1 (ko) * | 2013-02-28 | 2019-12-31 | 삼성디스플레이 주식회사 | 휘도 조절부, 이를 포함하는 표시 장치 및 이를 이용한 휘도 조절 방법 |
CN110460744B (zh) | 2014-06-10 | 2022-03-08 | 松下知识产权经营株式会社 | 亮度变换装置和亮度变换方法 |
JP2016058848A (ja) | 2014-09-08 | 2016-04-21 | ソニー株式会社 | 画像処理装置及び画像処理方法 |
US20160309154A1 (en) | 2015-04-17 | 2016-10-20 | Qualcomm Incorporated | Dynamic range adjustment for high dynamic range and wide color gamut video coding |
GB2539917B (en) * | 2015-06-30 | 2021-04-07 | British Broadcasting Corp | Method and apparatus for conversion of HDR signals |
BR112018000910A2 (pt) | 2015-07-16 | 2018-09-11 | Thomson Licensing | método e dispositivo para mapear por tons uma imagem utilizando uma função de ajuste de tom paramétrica |
CN106878964B (zh) * | 2015-12-14 | 2020-07-28 | 中国电信股份有限公司江苏客户运营服务中心 | 基于短信信道的鉴权系统和方法 |
EP3235236B1 (en) | 2015-12-17 | 2018-06-20 | Koninklijke Philips N.V. | Dynamic range coding for images and video |
EP3220349A1 (en) * | 2016-03-16 | 2017-09-20 | Thomson Licensing | Methods, apparatus, and systems for extended high dynamic range ("hdr") hdr to hdr tone mapping |
JP7054851B2 (ja) | 2016-09-09 | 2022-04-15 | パナソニックIpマネジメント株式会社 | 表示装置および信号処理方法 |
KR102554379B1 (ko) | 2016-10-31 | 2023-07-11 | 엘지디스플레이 주식회사 | 하이 다이나믹 레인지 영상 처리 방법 및 영상 처리 모듈과 그를 이용한 표시 장치 |
PL3389278T3 (pl) | 2016-11-22 | 2023-03-13 | Saturn Licensing, Llc | Urządzenie nadawcze, sposób nadawania, urządzenie odbiorcze i sposób odbioru |
CN108090879B (zh) * | 2017-12-12 | 2020-11-10 | 上海顺久电子科技有限公司 | 一种对输入的高动态范围图像进行处理的方法和显示设备 |
CN110691277B (zh) * | 2018-07-05 | 2024-03-05 | 华为技术有限公司 | 视频信号的处理方法及装置 |
-
2018
- 2018-07-19 CN CN201810797968.XA patent/CN108900823B/zh active Active
-
2019
- 2019-06-10 WO PCT/CN2019/090566 patent/WO2020007165A1/zh unknown
- 2019-06-10 JP JP2020568530A patent/JP7115776B2/ja active Active
- 2019-06-10 EP EP19830987.4A patent/EP3796648A4/en active Pending
- 2019-06-10 KR KR1020217000634A patent/KR102511572B1/ko active IP Right Grant
- 2019-06-10 CN CN201980025155.1A patent/CN112272947B/zh active Active
-
2020
- 2020-12-28 US US17/135,568 patent/US11317071B2/en active Active
Patent Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20090310015A1 (en) * | 2008-06-16 | 2009-12-17 | El-Mahdy Ahmed Hazem Mohamed R | Transformation of a video image from a high dynamic range image to a low dynamic range image |
CN106878694A (zh) * | 2015-12-10 | 2017-06-20 | 瑞昱半导体股份有限公司 | 高动态范围信号处理系统与方法 |
CN105516674A (zh) * | 2015-12-24 | 2016-04-20 | 潮州响石数码技术有限公司 | 一种具有hdr显示功能的监视设备 |
CN106780402A (zh) * | 2017-01-16 | 2017-05-31 | 聚龙智瞳科技有限公司 | 基于Bayer格式的图像动态范围扩展方法及装置 |
CN107046628A (zh) * | 2017-03-15 | 2017-08-15 | 四川长虹电器股份有限公司 | 一种低亮度8k电视机显示hdr效果的方法 |
CN108900823A (zh) * | 2018-07-05 | 2018-11-27 | 华为技术有限公司 | 一种视频信号处理的方法及装置 |
Non-Patent Citations (1)
Title |
---|
See also references of EP3796648A4 |
Cited By (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP4053834A1 (en) * | 2021-03-05 | 2022-09-07 | Canon Kabushiki Kaisha | Image processing apparatus, image capture apparatus, control method, and program |
US11770511B2 (en) | 2021-03-05 | 2023-09-26 | Canon Kabushiki Kaisha | Image processing apparatus, image capture apparatus, control method, and computer-readable storage medium |
CN117440118A (zh) * | 2023-10-23 | 2024-01-23 | 镕铭微电子(上海)有限公司 | 图像处理方法、装置、电子设备及存储介质 |
CN117440118B (zh) * | 2023-10-23 | 2024-06-11 | 镕铭微电子(上海)有限公司 | 图像处理方法、装置、电子设备及存储介质 |
Also Published As
Publication number | Publication date |
---|---|
KR102511572B1 (ko) | 2023-03-17 |
EP3796648A1 (en) | 2021-03-24 |
JP7115776B2 (ja) | 2022-08-09 |
CN108900823A (zh) | 2018-11-27 |
EP3796648A4 (en) | 2021-06-02 |
JP2021529339A (ja) | 2021-10-28 |
KR20210019076A (ko) | 2021-02-19 |
CN108900823B (zh) | 2019-07-12 |
US11317071B2 (en) | 2022-04-26 |
CN112272947A (zh) | 2021-01-26 |
US20210152801A1 (en) | 2021-05-20 |
CN112272947B (zh) | 2022-05-13 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2020007165A1 (zh) | 一种视频信号处理的方法及装置 | |
US11183143B2 (en) | Transitioning between video priority and graphics priority | |
JP6234920B2 (ja) | ハイダイナミックレンジ画像信号生成及び処理 | |
US11032579B2 (en) | Method and a device for encoding a high dynamic range picture, corresponding decoding method and decoding device | |
CN107211130B (zh) | 对彩色画面进行编码和解码的方法和装置 | |
WO2020007167A1 (zh) | 视频信号的处理方法及装置 | |
KR102509504B1 (ko) | 코딩 및 디코딩 방법과 대응하는 디바이스 | |
TW201702988A (zh) | 用於高動態範圍色調映射運算子之方法,裝置及系統 | |
KR102385726B1 (ko) | 하이 다이내믹 레인지 화상을 코딩된 비트스트림으로 인코딩/디코딩하기 위한 방법 및 장치 | |
WO2021073304A1 (zh) | 一种图像处理的方法及装置 | |
CN110691227A (zh) | 一种视频信号处理方法及装置 | |
CN115052137A (zh) | 一种视频信号处理方法及装置 | |
KR20160058153A (ko) | 압축된 및 압축해제된 디지털 이미지들 및 비디오의 디스플레이에서 가시적 결함들을 감소시키기 위한 시스템 및 방법 | |
CN117319620B (zh) | 一种hdr预览级现场实时调色方法 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 19830987 Country of ref document: EP Kind code of ref document: A1 |
|
ENP | Entry into the national phase |
Ref document number: 2020568530 Country of ref document: JP Kind code of ref document: A |
|
ENP | Entry into the national phase |
Ref document number: 2019830987 Country of ref document: EP Effective date: 20201216 |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
ENP | Entry into the national phase |
Ref document number: 20217000634 Country of ref document: KR Kind code of ref document: A |