WO2023016038A1 - 视频处理方法、装置、电子设备和存储介质 - Google Patents

视频处理方法、装置、电子设备和存储介质 Download PDF

Info

Publication number
WO2023016038A1
WO2023016038A1 PCT/CN2022/094744 CN2022094744W WO2023016038A1 WO 2023016038 A1 WO2023016038 A1 WO 2023016038A1 CN 2022094744 W CN2022094744 W CN 2022094744W WO 2023016038 A1 WO2023016038 A1 WO 2023016038A1
Authority
WO
WIPO (PCT)
Prior art keywords
video
log
frame rate
lut
camera
Prior art date
Application number
PCT/CN2022/094744
Other languages
English (en)
French (fr)
Inventor
崔瀚涛
邵涛
王梓蓉
张作超
Original Assignee
荣耀终端有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 荣耀终端有限公司 filed Critical 荣耀终端有限公司
Priority to US18/274,343 priority Critical patent/US20240080405A1/en
Priority to EP22855024.0A priority patent/EP4266675A1/en
Publication of WO2023016038A1 publication Critical patent/WO2023016038A1/zh

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N5/00Details of television systems
    • H04N5/222Studio circuitry; Studio devices; Studio equipment
    • H04N5/262Studio circuits, e.g. for mixing, switching-over, change of character of image, other special effects ; Cameras specially adapted for the electronic generation of special effects
    • H04N5/2621Cameras specially adapted for the electronic generation of special effects during image pickup, e.g. digital cameras, camcorders, video cameras having integrated special effects capability
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/60Control of cameras or camera modules
    • H04N23/62Control of parameters via user interfaces
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/60Control of cameras or camera modules
    • H04N23/63Control of cameras or camera modules by using electronic viewfinders
    • H04N23/631Graphical user interfaces [GUI] specially adapted for controlling image capture or setting capture parameters
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/60Control of cameras or camera modules
    • H04N23/667Camera operation mode switching, e.g. between still and video, sport and normal or high- and low-resolution modes
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/80Camera processing pipelines; Components thereof
    • H04N23/81Camera processing pipelines; Components thereof for suppressing or minimising disturbance in the image signal generation
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/80Camera processing pipelines; Components thereof
    • H04N23/84Camera processing pipelines; Components thereof for processing colour signals
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N9/00Details of colour television systems
    • H04N9/64Circuits for processing colour signals
    • H04N9/646Circuits for processing colour signals for image enhancement, e.g. vertical detail restoration, cross-colour elimination, contour correction, chrominance trapping filters
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N5/00Details of television systems
    • H04N5/14Picture signal circuitry for video frequency region
    • H04N5/20Circuitry for controlling amplitude response

Definitions

  • the present application relates to the technical field of video shooting, and in particular to a video processing method, device, electronic equipment and storage medium.
  • a video processing method, device, electronic equipment, and storage medium, which can make videos captured by electronic equipment have different style effects based on the characteristics of LUTs, so as to meet higher color matching requirements.
  • a video processing method including: determining a video style template among a plurality of video style templates, each video style template corresponding to a preset color lookup table LUT; obtaining a video taken by a camera; The logarithmic LOG curve corresponding to the current sensitivity ISO processes the video captured by the camera to obtain the LOG video; when the capture frame rate of the video captured by the camera is the first frame rate, based on the determined video style template corresponding The two-dimensional 2D-LUT processes the LOG video to obtain the video corresponding to the determined video style template; when the capture frame rate of the video captured by the camera is the second frame rate, the 3D 3D corresponding to the determined video style template - The LUT processes the LOG video to obtain a video corresponding to the determined video style template, and the first frame rate is greater than the second frame rate.
  • the LOG video is processed based on the determined two-dimensional 2D-LUT corresponding to the video style template, and the determined The video process corresponding to the video style template is performed in the HSV color space.
  • a 2D-LUT is applied.
  • the LOG video is processed based on the determined 3D 3D-LUT corresponding to the video style template to obtain the determined
  • the process of the video corresponding to the video style template includes: establishing a cube interpolation space based on the LUT; determining the cube to which each pixel in the LOG video belongs in the cube interpolation space, and the cube is divided into six tetrahedrons; The tetrahedron to which the pixel points belong; for the pixel point corresponding to the vertex of the cube, the pixel value is converted into the pixel value after LUT processing; for the pixel point not corresponding to the vertex of the cube, interpolation is performed according to the tetrahedron to which each pixel point belongs, Convert pixel values to LUT-processed pixel values.
  • the LOG video when the capture frame rate of the video captured by the camera is the second frame rate, the LOG video is processed based on the determined 3D 3D-LUT corresponding to the video style template to obtain the determined Before the process of the video corresponding to the video style template, it also includes: converting the LOG video from the LOG video in the RGB color space to the LOG video in the YUV color space; performing YUV denoising processing on the LOG video in the YUV color space to obtain the denoised LOG video. Since the LOG video obtained by processing the video through the LOG curve will introduce noise, it is possible to convert the LOG video to the YUV color space, perform YUV denoising processing, and use an algorithm to reduce noise to improve the video image quality.
  • the LOG video is processed based on the determined 3D 3D-LUT corresponding to the video style template to obtain the determined Before the process of the video corresponding to the video style template, it also includes: converting the denoised LOG video from the LOG video of the YUV color space to the LOG video of the RGB color space; After processing to obtain the video corresponding to the determined video style template, it further includes: converting the RGB color space video corresponding to the determined video style template into a YUV color space video.
  • the video corresponding to the determined video style template is split into two streams, one of which is saved, and the other stream is previewed .
  • the preview video and the final video can have the same visual effect, which is convenient for users to directly preview the video based on the color-graded style.
  • the first frame rate is 120 FPS or 60 FPS
  • the second frame rate is 30 FPS.
  • a video processing device including: a processor and a memory, the memory is used to store at least one instruction, and when the instruction is loaded and executed by the processor, the above video processing method is realized.
  • an electronic device including: a camera; and the above-mentioned video processing device.
  • a computer-readable storage medium In a fourth aspect, a computer-readable storage medium is provided.
  • a computer program is stored in the computer-readable storage medium, and when running on a computer, the computer is made to execute the above video processing method.
  • the video processing method, device, electronic equipment, and storage medium in the embodiments of the present application use the LUT technology in the film industry to process the LOG video based on the LUT corresponding to the determined video style template during the video recording process, so that the recorded
  • the video has the style effect corresponding to the determined video style template, so as to meet the higher color grading requirements and make the recorded video have a movie feel.
  • use 3D-LUT to process the video to achieve more precise color control
  • use 2D-LUT to process the video to achieve limited LUT processing is performed within the frame interval.
  • FIG. 1 is a structural block diagram of an electronic device in an embodiment of the present application
  • FIG. 2 is a flowchart of a video processing method in an embodiment of the present application
  • FIG. 3 is a schematic diagram of a user interface in a movie mode in an embodiment of the present application.
  • Fig. 4 is the schematic diagram of a kind of LOG curve in the embodiment of the present application.
  • Fig. 5 is a kind of specific flow diagram of step 104 in Fig. 2;
  • FIG. 6 is a schematic diagram of the relationship between a cube and a tetrahedron in a cube interpolation space in an embodiment of the present application
  • FIG. 7 is a flowchart of another video processing method in the embodiment of the present application.
  • Fig. 8 is a UV plane schematic diagram
  • FIG. 9 is another structural block diagram of an electronic device in the embodiment of the present application.
  • FIG. 10 is another structural block diagram of an electronic device in the embodiment of the present application.
  • Fig. 11 is a software structural block diagram of an electronic device in the embodiment of the present application.
  • FIG. 12 is a schematic diagram of a user interface in a professional mode in the embodiment of the present application.
  • the electronic device 100 may include a processor 110, a camera 193, a display screen 194, and the like. It can be understood that, the structure illustrated in the embodiment of the present invention does not constitute a specific limitation on the electronic device 100 . In other embodiments of the present application, the electronic device 100 may include more or fewer components than shown in the figure, or combine certain components, or separate certain components, or arrange different components. The illustrated components can be realized in hardware, software or a combination of software and hardware.
  • the processor 110 may include one or more processing units, for example: the processor 110 may include a graphics processing unit (graphics processing unit, GPU), an image signal processor (image signal processor, ISP), a controller, a video codec, Digital signal processor (digital signal processor, DSP), etc. Wherein, different processing units may be independent devices, or may be integrated in one or more processors.
  • the controller can generate an operation control signal according to the instruction opcode and timing signal, and complete the control of fetching and executing the instruction.
  • a memory may also be provided in the processor 110 for storing instructions and data.
  • the electronic device 100 realizes the display function through the GPU, the display screen 194 , and the application processor.
  • the GPU is a microprocessor for image processing, and is connected to the display screen 194 and the application processor. GPUs are used to perform mathematical and geometric calculations for graphics rendering.
  • Processor 110 may include one or more GPUs that execute program instructions to generate or change display information.
  • the electronic device 100 can realize the shooting function through the ISP, the camera 193 , the video codec, the GPU, the display screen 194 and the application processor.
  • the ISP is used for processing the data fed back by the camera 193 .
  • the light is transmitted to the photosensitive element of the camera through the lens, and the light signal is converted into an electrical signal, and the photosensitive element of the camera transmits the electrical signal to the ISP for processing, and converts it into an image visible to the naked eye.
  • ISP can also perform algorithm optimization on image noise, brightness, and skin color.
  • ISP can also optimize the exposure, color temperature and other parameters of the shooting scene.
  • the ISP may be located in the camera 193 .
  • Camera 193 is used to capture still images or video.
  • the object generates an optical image through the lens and projects it to the photosensitive element.
  • the photosensitive element may be a charge coupled device (CCD) or a complementary metal-oxide-semiconductor (CMOS) phototransistor.
  • CMOS complementary metal-oxide-semiconductor
  • the photosensitive element converts the light signal into an electrical signal, and then transmits the electrical signal to the ISP to convert it into a digital image signal.
  • the ISP outputs the digital image signal to the DSP for processing.
  • DSP converts digital image signals into standard RGB, YUV and other image signals.
  • the electronic device 100 may include 1 or N cameras 193 , where N is a positive integer greater than 1.
  • Digital signal processors are used to process digital signals. In addition to digital image signals, they can also process other digital signals. For example, when the electronic device 100 selects a frequency point, the digital signal processor is used to perform Fourier transform on the energy of the frequency point.
  • Video codecs are used to compress or decompress digital video.
  • the electronic device 100 may support one or more video codecs.
  • the electronic device 100 can play or record videos in various encoding formats, for example: moving picture experts group (moving picture experts group, MPEG) 1, MPEG2, MPEG3, MPEG4 and so on.
  • MPEG moving picture experts group
  • the embodiment of the present application provides a video processing method.
  • the video processing method may be executed by a processor 110, specifically an ISP or a combination of an ISP and other processors.
  • the video processing method includes:
  • Step 101 determine a video style template among a plurality of video style templates, each video style template corresponds to a preset color lookup table (Look Up Table, LUT);
  • LUT is a mathematical conversion model.
  • the image data value can be output as another image data value, thereby changing the exposure and color of the picture. Therefore, LUTs corresponding to different video styles can be pre-generated.
  • a video style template can be determined before the electronic device records a video.
  • the video style template can be determined based on the user's choice, or based on artificial intelligence (Artificial Intelligence, AI),
  • AI Artificial Intelligence, AI
  • the video style template is automatically determined according to the scene corresponding to the image captured by the current camera. For example, assuming that the electronic device is a mobile phone, in a possible implementation, as shown in Figure 3, the user operates the mobile phone to enter the shooting interface, and the shooting interface includes movie mode options.
  • the movie mode interface of including multiple video style template options, for example including "A" movie style template, "B” movie style template and “C” movie style template, only one " A "movie style template, understandably, multiple different movie style templates can be displayed side by side in the user interface, and the LUTs corresponding to different movie style templates can be generated based on the corresponding movie color matching style in advance, and the color conversion of the LUT has Corresponding to the style characteristics of the movie, for example, the color matching style of the movie "A" is complementary color.
  • Complementary color refers to the contrast effect of two corresponding colors. Two colors of warm color and cool color are used to emphasize the contrast to enhance the vividness, For outstanding effects, usually two contrasting colors symbolize conflicting behaviors.
  • the LUT corresponding to the "A" movie style template is ready to use After transforming the colormap, the complementary colors are more pronounced to simulate the color scheme of the "A" movie.
  • the mobile phone when the user operates the mobile phone to enter the movie mode, the mobile phone will obtain the picture taken by the current camera, and based on the AI algorithm, determine the scene corresponding to the picture and determine the scene corresponding to the scene.
  • the recommended video style template for example, if it is recognized that the subject of the currently captured picture is a young female character, the corresponding recommended video style template is determined according to the algorithm as the "C" movie style template, and the movie "C" has a young female character as the theme movie, its corresponding LUT can simulate the color matching style of the movie "C”; It is a movie with city streets as the main scene, and its corresponding LUT can simulate the color matching style of the "B" movie. In this way, a video style template matching the current scene can be automatically recommended for the user. Film styles can be pre-extracted to produce LUTs suitable for mobile electronics.
  • Step 102 obtain the video shot by camera, for example, after determining the video style template in step 101, if the user clicks on the shooting option, the mobile phone starts to obtain the video shot by camera;
  • Step 103 process the video captured by the camera through the logarithm (Logarithm, LOG) curve corresponding to the current sensitivity ISO of the camera to obtain the LOG video;
  • Figure 4 illustrates a LOG curve, where the abscissa is a linear signal, represented by a 16-bit code value Code Value, and the ordinate is the LOG signal processed by the LOG curve, represented by a 10-bit code value.
  • the signal input of the camera can be used to encode the information in the dark area to the middle tone (as shown in the steep part of the curve in Figure 4), forming a 10-bit signal output, which conforms to the human eye's LOG sensing rule for light, and maximizes the The dark information is preserved, and the LOG video can use the limited bit depth to maximize the details of shadows and highlights.
  • the ASA in Figure 4 is the sensitivity, and different ASAs correspond to different ISOs, and the two belong to different systems.
  • Step 1041 when the capture frame rate of the video captured by the camera is at the first frame rate, process the LOG video based on the two-dimensional 2D-LUT corresponding to the determined video style template, and obtain the corresponding to the determined video style template video;
  • Step 1042 when the capture frame rate of the video captured by the camera is the second frame rate, process the LOG video based on the three-dimensional 3D-LUT corresponding to the determined video style template to obtain the video corresponding to the determined video style template,
  • the first frame rate is greater than the second frame rate.
  • the LOG video is used as an input, and the LUT corresponding to the video style template determined in step 101 is applied to perform mapping conversion processing on the LOG video image.
  • the output can be the video of the Rec.709 color standard, or the video of the High-Dynamic Range (HDR) 10 standard, that is, the LOG video can be processed through the LUT, Convert video to HDR10 standard.
  • HDR High-Dynamic Range
  • the capture frame rate refers to the readout frame rate of the camera sensor, or the shooting frame rate of the video.
  • the frame rate when the camera is shooting also has an impact on the video effect.
  • the video capture frame rate is greater than the playback frame rate, it is an upgrade.
  • the frame rate of the video playback is 30 frames per second, that is, the frame rate is 30 frames per second.
  • the frame rate when shooting is 60FPS, that is, upgrading, the effect that can be achieved is slow motion, and the frame rate when shooting with the camera in upgrading is suitable for different scene expressions.
  • 60FPS upgraded shooting is suitable for expressing scenes such as panning, slow walking, laughing, applauding, etc.
  • 120FPS upgraded shooting is suitable for expressing scenes such as running, turning around, pulling hair, and throwing flowers. Therefore, during the video recording process, the user can choose to perform slow-motion shooting, or choose other methods to use a higher frame rate to shoot.
  • the faster the processing speed is, for example, the capture frame rate of 60FPS needs to apply LUT to process the video within 15ms, and the capture frame rate of 120FPS needs to apply LUT to process the video within 8ms.
  • the LUT can be 3D-LUT or 2D-LUT.
  • 3D-LUT can achieve more accurate color control.
  • 3D-LUT requires more processing time due to the complexity of the algorithm when applying it.
  • the processing time required for 2D-LUT application is relatively short. Therefore, in the embodiment of the present application, when the capture frame rate of the video captured by the camera is a relatively high first frame rate, the 2D-LUT is applied to achieve LUT processing is performed within a limited frame interval time.
  • the capture frame rate of the video captured by the camera is a lower second frame rate
  • 3D-LUT is applied to achieve LUT processing within a long enough frame interval time to achieve more Precise color control.
  • the embodiment of the present application is not limited to application in upscaling, and the embodiment of the present application is also applicable to other methods that require a higher capture frame rate.
  • Different LUTs are applied to electronic equipment, and related modules in the electronic equipment can be adapted to adapt to different styles of LUTs.
  • the video style template determined in step 101 is a gray tone video style template
  • the gray tone The characteristics of the picture are that the texture in the picture is strong, the saturation is low, there is no more color interference except for the color of the character's skin, and the dark part is cooler. Based on these characteristics, the electronic device can monitor the relevant Adjust the module parameters to keep the texture in the picture, do not do strong denoising and sharpening, properly reduce the saturation of the picture, keep the skin color in the picture true to restore, and adjust the dark part of the picture to cool colors.
  • the LUT technology of the film industry is used to process the LOG video based on the LUT corresponding to the determined video style template, so that the recorded video has the determined video style
  • the style effect corresponding to the template can meet the higher color grading requirements and make the recorded video have a cinematic feel.
  • the video capture frame rate is low, use 3D-LUT to process the video to achieve more precise color control; when the video capture frame rate is high, use 2D-LUT to process the video to achieve limited LUT processing is performed within the frame interval.
  • step 1042 when the capture frame rate of the video captured by the camera is the second frame rate, the 3D 3D-LUT corresponding to the determined video style template is used to pair the LOG
  • the video is processed, and the process of obtaining the video corresponding to the determined video style template includes:
  • Step 10421 establish a cubic interpolation space based on the LUT, and the LUT is a three-dimensional 3D-LUT;
  • 3D-LUT is a color mapping relationship commonly used in the film industry. It can convert any input RGB pixel value into corresponding other RGB pixel values, such as inputting 12bit RGB Video image, output 12bit RGB video image after LUT processing and mapping.
  • Step 10422 determine the cube to which each pixel in the LOG video belongs in the cube interpolation space, and the cube is divided into 6 tetrahedrons;
  • the LOG video is used as the input in the LUT processing process, and each pixel in the LOG video picture is obtained through the pixel point after the LUT processing mapping, which can realize the process of processing the LOG video through the LUT.
  • step 10422 it is necessary to Determine the cube to which each pixel in each input LOG video belongs to in the above cube interpolation space, and the cube is divided into 6 tetrahedrons.
  • Step 10423 determine the tetrahedron to which each pixel in the LOG video belongs
  • Step 10424 for the pixel points corresponding to the vertices of the cube, the pixel value is converted into the pixel value after LUT processing, and for the pixel points not corresponding to the vertices of the cube, interpolation is performed according to the tetrahedron to which each pixel point belongs, and the pixel value is converted into The pixel value after LUT processing.
  • the mapped RGB pixel value can be directly obtained, that is, the pixel value can be directly mapped and converted into The corresponding pixel value, and if the pixel is located between the vertices of the cube, interpolate according to the tetrahedron to which the pixel belongs.
  • the cube has the 0th to 7th vertices, which are respectively represented by numbers 0 to 7 in Figure 6, and the direction from the 0th vertex to the first vertex is blue B
  • the coordinate axis direction of the channel, the direction from the 0th vertex to the 4th vertex is the coordinate axis direction of the red R channel
  • the direction from the 0th vertex to the second vertex is the coordinate axis direction of the green G channel
  • the 2nd vertex and the 3rd vertex are located on the same plane
  • the 1st vertex, the 3rd vertex, the 5th vertex and the 7th vertex are located on the same plane
  • the 4th vertex, the 5th vertex, the 6th vertex and the 7th vertex are located on the same plane
  • the 3rd vertex is located on the same plane.
  • Vertex 0, Vertex 2, Vertex 4 and Vertex 6 are on the same plane; Vertex 0, Vertex 1, Vertex 5 and Vertex 7 form the first tetrahedron, Vertex 0, Vertex 1, Vertex 3
  • the vertex and the 7th vertex form the second tetrahedron, the 0th vertex, the 2nd vertex, the 3rd vertex and the 7th vertex form the third tetrahedron, the 0th vertex, the 4th vertex, the 5th vertex and the 7th vertex form
  • the 0th vertex, the 4th vertex, the 6th vertex and the 7th vertex form the fifth tetrahedron
  • the 0th vertex, the 2nd vertex, the 6th vertex and the 7th vertex form the sixth tetrahedron;
  • the coordinates of the i-th vertex are (Ri, Gi, Bi)
  • the value of i is 0, 1, 2, 3, ..., 7, and the pixel value of the pixel value of the
  • step 1044 for the pixel points that do not correspond to the vertices of the cube, interpolation is performed according to the tetrahedron to which each pixel point belongs, and the process of converting the pixel value into the pixel value processed by the LUT includes:
  • VE(R, G, B) VE(R0, G0, B0)+(delta_valueR_E ⁇ deltaR+delta_valueG_E ⁇ deltaG+delta_valueB_E ⁇ deltaB+(step_size>1))/(step_size);
  • VE(R0, G0, B0) is the E channel pixel value of the 0th vertex (R0, G0, B0) after LUT processing, and E is R, G and B;
  • delta_valueR_E is the difference between the two vertices in the direction of the coordinate axis of the R channel corresponding to the tetrahedron to which the current pixel belongs, and the difference between the pixel values of the E channel after LUT processing.
  • delta_valueB_E is the difference between the pixel values of the E channel after the LUT processing of two vertices in the direction of the coordinate axis of the tetrahedron to which the current pixel belongs corresponding to the B channel;
  • deltaR is the difference between the R value in the current pixel point (R, G, B) and the R0 value in the 0th vertex (R0, G0, B0)
  • deltaG is the G value in the current pixel point (R, G, B).
  • deltaB is the difference between the B value in the current pixel point (R, G, B) and the B0 value in the 0th vertex (R0, G0, B0).
  • step_size is the side length of the cube.
  • step_size>>1 means step_size is shifted to the right by one bit.
  • deltaR, deltaG and deltaB represent the distance between the current pixel point (R, G, B) and the 0th vertex
  • deltaR R-R0
  • deltaG G-G0
  • deltaB B-B0
  • deltaB ⁇ deltaR and deltaR ⁇ deltaG determine that the current pixel belongs to the first tetrahedron; if deltaB ⁇ deltaG and deltaG ⁇ deltaR, determine that the current pixel belongs to the second tetrahedron; if deltaG ⁇ deltaB and deltaB ⁇ deltaR , it is determined that the current pixel point belongs to the third tetrahedron; if deltaR ⁇ deltaB and deltaB ⁇ deltaG, then it is determined that the current pixel point belongs to the fourth tetrahedron; if deltaR ⁇ deltaG and deltaG ⁇ deltaB, then it is determined that the current pixel point belongs to the fourth tetrahedron Five tetrahedrons; if the relationship among deltaR, deltaG, and deltaB does not belong to the above conditions of the first to fifth tetrahedrons, it is determined that the current pixel point belongs to the sixth tetrahedron.
  • delta_valueR_E is the tetrahedron to which the current pixel point belongs
  • delta_valueR_R VR(R5, G5, B5)-VR(R1, G1, B1)
  • delta_valueG_R VR( R7, G7, B7)-VR(R5, G5, B5)
  • delta_valueB_R VR(R1, G1, B1)-VR(R0, G0, B0)
  • VR(R, G, B) VR(R0, G0 , B0)+(delta_valueR_R ⁇ deltaR+delta_valueG_R ⁇ deltaG+delta_valueB_R ⁇ deltaB+
  • the LOG video is processed based on the two-dimensional 2D-LUT corresponding to the determined video style template to obtain The processing of the video corresponding to the determined video style template is performed in the HSV color space.
  • the 2D-LUT is obtained through 3D-LUT simulation in advance, for example, the 3D-LUT is known in advance.
  • LUT, and the corresponding input data and output data of 3D-LUT, the input data and output data belong to the RGB color space, at this time, the input data can be converted from the RGB color space to the HSV color space, and the output data can be converted from the RGB color space
  • the specific model of the 2D-LUT can be obtained through the conversion relationship of the data in the HSV color space, so that the 2D-LUT can be applied in the process of video recording. Since the 2D-LUT can only be based on two Therefore, the hue H and saturation S can be converted, and the lightness V can be ignored to achieve the effect of 2D-LUT simulating 3D-LUT.
  • step 1042 when the capture frame rate of the video captured by the camera is the second frame rate, the 3D 3D-LUT pair corresponding to the determined video style template LOG video is processed, before the process of obtaining the video corresponding to the determined video style template, it also includes: step 105, converting the LOG video to the LOG video of the YUV color space by the LOG video of the RGB color space; step 106, to YUV
  • the LOG video in the color space is subjected to YUV denoising processing to obtain the denoising LOG video, and the LOG video to which the LUT is applied in step 1042 is the LOG video after YUV denoising in step 106 . Since the LOG video obtained in step 103 will introduce noise, after the LOG video can be converted into a YUV color space, YUV denoising processing can be performed, and the noise can be reduced by an algorithm to improve the video image quality.
  • step 1042 when the capture frame rate of the video captured by the camera is the second frame rate, the 3D 3D-LUT pair corresponding to the determined video style template LOG video is processed, and before the process of obtaining the video corresponding to the determined video style template, it also includes: step 107, converting the LOG video after denoising to the LOG video of RGB color space by the LOG video of YUV color space; Above-mentioned step 104, process LOG video based on the LUT corresponding to the determined video style template, and after the process of obtaining the video corresponding to the determined video style template, it also includes: step 108, combining the RGB color space with the determined The video corresponding to the video style template is converted to a video in the YUV color space.
  • the process of processing the LOG video based on the LUT in the step 1042 is realized based on the RGB color space, therefore, before the step 1042, the video in the YUV color space is converted to the video in the RGB color space, and after the step 1042, then Reconvert video in RGB color space to video in YUV color space.
  • YUV also known as YCbCr
  • YCbCr is a color coding method used by the European television system.
  • three-tube color cameras or color CCD cameras are usually used to capture images, and then the obtained color image signals are separated, amplified and corrected to obtain RGB signals, and then the brightness signals Y and Y are obtained through a matrix conversion circuit.
  • Two color-difference signals B-Y (ie U) and R-Y (ie V), and finally the sending end encodes the three signals separately and sends them out on the same channel.
  • This color representation method is the YUV color space.
  • YCbCr is the specific implementation of the YUV model, which is actually a scaled and offset replica of YUV.
  • Y has the same meaning as Y in YUV, and both Cb and Cr refer to color, but they are different in the way of expression.
  • YCbCr is the most widely used member in computer systems, and its application fields are very wide. Both JPEG and MPEG use this format. Generally speaking, YUV mostly refers to YCbCr.
  • the UV plane is shown in Figure 8.
  • RGB and YUV color spaces can be realized by a 3x3 matrix:
  • the electronic device may specifically include a camera 193, an anti-mosaic Demosaic module 21, a deformation module 22, a fusion module 23, a noise processing module 24, and a color correction matrix (Color Correction Matrix, CCM) module 25, global tone mapping (Global Tone Mapping, GTM) module 26, scaling Scaler module 27, YUV denoising module 28 and LUT processing module 29, for example, in the process of video recording, when the video of camera 193 takes When the capture frame rate is the second frame rate, the camera 193 captures a long-exposure frame video image and a short-exposure frame video image, and the exposure time corresponding to the long-exposure frame video image is longer than the exposure time corresponding to the short-exposure frame video image.
  • CCM Color Correction Matrix
  • the frame video image and the short-exposure frame video image are respectively processed by the anti-mosaic module 21, so that the image is converted from the RAW domain to the RGB domain, and then the two video images are respectively processed by the deformation warp module 22, and the alignment is realized by deforming the video image , anti-shake effect, the two-way video images are processed by the fusion module 23 afterwards, and the two kinds of video images are fused into the same one, and the data after the fusion is divided into two paths.
  • the video processing method includes the first video processing flow S1 and the second In the video processing flow S2, after being processed by the fusion module 23, one path enters the first video processing flow S1, and the other path enters the second video processing flow S2.
  • the above step 103 is executed, the process of processing the video captured by the camera through the logarithmic LOG curve, and the process of obtaining the LOG video, and the above step 104, based on the LUT corresponding to the determined video style template to the LOG
  • the video is processed to obtain the video corresponding to the determined video style template.
  • the first video processing flow S1 includes, the video taken by the camera 193 from the fusion module 23 is denoised by the noise processing module 24, and then processed by the CCM module 25 to convert the video into RGB wide color gamut color Space, then carry out above-mentioned step 103 by GTM module 26, process video by LOG curve, obtain LOG video, then carry out scaling processing to video by scaling module 27, then carry out above-mentioned step 106 by YUV denoising module 28, video is carried out YUV denoising, and then perform the above step 1042 through the LUT processing module 29, process the video through the 3D-LUT, and obtain the video corresponding to the determined video style module.
  • the video corresponding to the determined video style template in the first video processing flow S1 is saved as video 1, and a video with style can be obtained.
  • the second video processing flow S2 includes: the video taken by the camera 193 from the fusion module 23 is denoised by the noise processing module 24, and then processed by the CCM module 25 to convert the video into the color space of RGB wide color gamut, Then carry out above-mentioned step 103 by GTM module 26, process video by LOG curve, obtain LOG video, then carry out scaling processing to video by scaling module 27, then carry out above-mentioned step 106 by YUV denoising module 28, carry out YUV to video noise, and then execute the above step 1042 through the LUT processing module 29, and process the video through the 3D-LUT to obtain the video corresponding to the determined video style module. Previewing is performed based on the video corresponding to the determined video style template in the second video processing flow S2.
  • the two video streams are respectively processed in the first video processing flow S1 and the second video processing flow S2.
  • the two video streams are respectively processed in the first video processing flow S1 and the second video processing flow S2.
  • the LOG curve-based processing in step 103 and the 3D-LUT-based processing in step 1042 one video stream is used for encoding and saving, and the other video stream is used for preview.
  • Bayer field Each lens on a digital camera has a light sensor to measure the brightness of the light, but to obtain a full-color image, generally three light sensors are required to obtain the three primary colors of red, green and blue information, and in order to reduce the cost and volume of digital cameras, manufacturers usually use CCD or CMOS image sensors.
  • CCD or CMOS image sensors usually use CCD or CMOS image sensors.
  • the original image output by CMOS image sensors is in Bayer domain RGB format, and a single pixel contains only one color value. To obtain the gray value of the image, it is necessary to interpolate the complete color information of each pixel, and then calculate the gray value of each pixel.
  • the Bayer domain refers to a raw image format inside a digital camera.
  • the Raw domain or Raw format refers to unprocessed images. Further, the Raw image can be understood as that the photosensitive element of the camera such as Complementary Metal Oxide Semiconductor (CMOS) or Charge-coupled Device (Charge-coupled Device, CCD) converts the captured light source signal into digital The raw data of the signal.
  • CMOS Complementary Metal Oxide Semiconductor
  • CCD Charge-coupled Device
  • a RAW file is a record of the original information of the digital camera sensor, while recording some metadata (Metadata, such as ISO (International Organization for Standardization, International Organization for Standardization) settings, shutter speed, aperture value) generated by the camera. , white balance, etc.) files.
  • the Raw domain is a format that has not been processed by the ISP nonlinearly and has not been compressed.
  • the full name of Raw format is RAW Image Format.
  • YUV is a color encoding method that is often used in various video processing components. YUV takes human perception into account when encoding photos or videos, allowing bandwidth reduction for chroma. YUV is a type of compiling true-color color space (color space). Proper nouns such as Y'UV, YUV, YCbCr, and YPbPr can all be called YUV, and they overlap with each other. Among them, "Y” represents the brightness (Luminance or Luma), that is, the grayscale value, "U” and “V” represent the chroma (Chrominance or Chroma), which are used to describe the color and saturation of the image, and are used to specify the color of the pixel .
  • YUV is divided into two formats, one is: packed formats, which store Y, U, and V values into a Macro Pixels array, which is similar to the storage method of RGB.
  • the other is: planar formats, which store the three components of Y, U, and V in different matrices.
  • Planar formats means that each Y component, U component and V component are organized in an independent plane, that is to say, all U components are behind the Y component, and V components are behind all U components.
  • the LOG video is processed based on the 3D-LUT corresponding to the determined video style template to obtain the video image corresponding to the determined video style template.
  • it also includes: saving the LOG video, that is, saving the video 2. That is to say, after the LOG video is obtained in step 103, one LOG video is split and saved directly, that is, a non-style LOG video is saved, that is, video 2, and the other LOG video is saved after being processed based on 3D-LUT in step 1042. That is to save a video corresponding to the video style template, that is, recording 1.
  • two video files are encoded and saved in a temporary folder at the same time. Based on the user's choice, one of the videos is determined as the final video file. If the user If you want to directly record and generate a "B" movie-style video file, use Video 1 as the final video file. If you want to save the original material to improve the ability of post-editing, then use Video 2 as the final video file. It should be noted that the splitting process of saving the LOG video may be performed after the zooming module 27 scales the video and before the LUT processing module 29 executes the process of step 104 above.
  • the LOG video is processed based on the 3D-LUT corresponding to the determined video style template to obtain the video image corresponding to the determined video style template.
  • it also includes: converting the LOG video to a video of the Rec.709 color standard; saving the video of the Rec.709 color standard, that is, video 2. That is to say, after obtaining the LOG video in step 103, divert one LOG video and convert it into a primary color video and save it as video 2. This video has not been color-graded, and the other LOG video is saved after being processed based on 3D-LUT in step 1042.
  • a video corresponding to the video style template that is, save it as video 1.
  • encode and save two video files in a temporary folder at the same time Based on the user's choice, determine one of the videos as the final video file. If the user wants to directly record and generate a "B" movie-style video file, the video corresponding to the video style template, that is, recording 1, will be used as the final video file. If the user wants to save the ungraded video, the ungraded The LOG video is converted to a standard video file and saved as the final video file, that is, recording 2 is the final video file. It should be noted that the splitting process of ungraded video can be performed after the zooming module 27 scales the video and before the LUT processing module 29 executes the process of step 1042 above.
  • the video corresponding to the determined video style template is divided into two streams, one of which is Save it, and preview it on the other way.
  • the electronic device also includes a 2D-LUT processing module 30, 2D-LUT
  • the processing module 30 can be a hardware module in the ISP, and the above-mentioned LUT processing module 29 can be a software module.
  • the camera 193 shoots The long exposure frame video image and the short exposure frame video image are obtained, the exposure time corresponding to the long exposure frame video image is greater than the exposure time corresponding to the short exposure frame video image, and the long exposure frame video image and the short exposure frame video image are respectively passed through anti-mosaic
  • the processing of the module 21 converts the image from the RAW domain to the RGB domain, and then the two channels of video images are processed by the warp warp module 22 respectively to achieve alignment and anti-shake effects through the deformation of the video images, and then the two channels of video images are fused
  • Module 23 is processed, and two kinds of video images are fused into the same one, and the video image after fusion is denoised by noise processing module 24, then processed by CCM module 25, and the video is converted into the color space of RGB wide color gamut, and then passed
  • the GTM module 26 executes the above step 103, processes the video through the LOG curve to obtain
  • the two-dimensional 2D-LUT corresponding to the determined video style template processes the LOG video, obtains the video corresponding to the determined video style template, then zooms the video through the scaling module 27, and divides the processed video image into two streams. One way is saved, and the other way is previewed. Based on the comparison of FIG. 9 and FIG.
  • FIG. 11 is a block diagram of the software structure of the electronic device 100 according to the embodiment of the present application.
  • the layered architecture divides the software into several layers, and each layer has a clear role and division of labor. Layers communicate through software interfaces.
  • the Android system is divided into five layers, which are, from top to bottom, the Application layer, the application framework framework layer, the system library library, the Hardware Abstraction Layer (Hardware Abstraction Layer, HAL) and the kernel layer.
  • the application layer can include applications such as cameras.
  • the application framework layer may include camera application programming interface (Application Programming Interface, API), media recording MediaRecorder and surface view Surfaceview, etc.
  • Media recording is used to record video or image data and make this data accessible to applications.
  • Surface views are used to display preview images.
  • a system library can include multiple function modules. For example: camera service CameraSevice, etc.
  • the hardware abstraction layer is used to provide interface support, for example, including the camera process CameraPipeline for the camera service to call Call.
  • the kernel layer is the layer between hardware and software.
  • the kernel layer includes display drivers, camera drivers, etc.
  • HAL reports the ability information of recording two videos at the same time, and the application layer sends a capture request CaptureRequest, which corresponds to a stream of recording 1, a stream of recording 2 and a preview stream, and creates two streams at the same time.
  • CaptureRequest An example of a media codec mediacodec that receives the encoding of two video streams.
  • HAL calls back three streams according to the above dataflow dataflow. Among them, the preview stream shows that the two video streams are sent to mediacodec respectively.
  • the video recording and video processing method provided in the embodiment of the present application may be represented as multiple functions in two shooting modes, where the two shooting modes may refer to: movie mode and professional mode.
  • the movie mode is a shooting mode related to the theme of the movie.
  • the image displayed by the electronic device 100 can give the user a sense of watching a movie.
  • the electronic device 100 also provides a plurality of video related to the theme of the movie Style templates, users can use these video style templates to obtain tone-adjusted images or videos, and the tone of these images or videos is similar or identical to the tone of the movie.
  • the movie mode can at least provide an interface for the user to trigger the LUT function and the HDR10 function. For specific descriptions about the LUT function and the HDR10 function, please refer to the following embodiments.
  • the electronic device 100 may enter a movie mode in response to a user's operation.
  • the electronic device 100 may detect a user's touch operation on the camera application, and in response to the operation, the electronic device 100 displays a default camera interface of the camera application.
  • the default camera interface can include: preview frame, shooting mode list, gallery shortcut keys, shutter controls, etc. in:
  • the preview frame can be used to display images collected by the camera 193 in real time.
  • the electronic device 100 can refresh the displayed content therein in real time, so that the user can preview the image currently captured by the camera 193 .
  • One or more shooting mode options may be displayed in the shooting mode list.
  • the one or more shooting mode options may include: portrait mode options, video recording mode options, camera mode options, movie mode options, and professional options.
  • the one or more shooting mode options can be represented as text information on the interface, such as "portrait”, “video recording”, “photographing”, “movie”, “professional”.
  • the one or more shooting mode options may also be represented as icons or other forms of interactive elements (interactive element, IE) on the interface.
  • Gallery shortcuts can be used to launch the Gallery application.
  • the gallery application program is an application program for picture management on electronic devices such as smart phones and tablet computers, and may also be called "album".
  • the name of the application program is not limited in this embodiment.
  • the gallery application program can support users to perform various operations on pictures stored on the electronic device 100, such as browsing, editing, deleting, selecting and other operations.
  • the shutter control can be used to listen for user actions that trigger a photo.
  • the electronic device 100 may detect a user operation acting on the shutter control, and in response to the operation, the electronic device 100 may save the image in the preview frame as a picture in the gallery application.
  • the electronic device 100 may also display the thumbnails of the saved images in the gallery shortcut key. That is, users can tap the shutter control to trigger a photo.
  • the shutter control may be a button or other forms of control.
  • the electronic device 100 may detect a user's touch operation on the movie mode option, and in response to the operation, the electronic device displays a user interface as shown in FIG. 3 .
  • the electronic device 100 may turn on the movie mode by default after starting the camera application. Not limited thereto, the electronic device 100 may also enable the movie mode in other ways, for example, the electronic device 100 may also enable the movie mode according to a user's voice command, which is not limited in this embodiment of the present application.
  • the electronic device 100 may detect a user's touch operation on the movie mode option, and in response to the operation, the electronic device displays a user interface as shown in FIG. 3 .
  • the user interface shown in FIG. 3 includes function options, and the function options include HDR10 options, flash options, LUT options, and setting options. These multiple function options can detect the user's touch operation, and in response to the operation, enable or disable the corresponding shooting function, for example, HDR10 function, flash function, LUT function, setting function.
  • the electronic device can enable the LUT function, and the LUT function can change the display effect of the preview image.
  • the LUT function introduces a color lookup table, which is equivalent to a color conversion model, which can output adjusted color values according to the input color values.
  • the color value of the image captured by the camera is equivalent to the input value, and different color values can be correspondingly obtained as an output value after passing through the color conversion model.
  • the image displayed in the preview box is the image adjusted by the color transformation model.
  • the electronic device 100 uses the LUT function to display an image composed of color values adjusted by the color conversion model, so as to achieve the effect of adjusting the tone of the image.
  • the electronic device 100 can provide multiple video style templates, one video style template corresponds to one color conversion model, and different video style templates can bring different display effects to the preview image.
  • these video style templates can be associated with the theme of the movie, and the tone adjustment effect brought by the video style template to the preview image can be close to or the same as the tone in the movie, creating an atmosphere for the user to shoot a movie.
  • the electronic device 100 can determine a video style template among multiple video style templates according to the current preview video image, and the determined video style template can be displayed on the interface, so that the user can understand Currently determined video style templates, for example, a plurality of video style templates including "A" movie style template, "B" movie style template and "C” movie style template, the corresponding LUTs of different movie style templates can be based on the corresponding Generated by the movie color matching style, the color conversion of the LUT has the style characteristics of the corresponding movie. Film styles can be pre-extracted to produce LUTs suitable for mobile electronics. Turning on the LUT function will change the color tone of the preview video screen. As illustrated in FIG. 3 , the electronic device 100 determines and displays the "A" movie style template.
  • the electronic device 100 may select a video style template according to the user's sliding operation. Specifically, when the electronic device 100 detects the user operation of enabling the LUT function and displays the LUT preview window, the electronic device 100 can select the first video style template located in the LUT preview window by default as the video style template selected by the electronic device 100. template. Afterwards, the electronic device 100 can detect the left and right sliding operation of the user acting on the LUT preview window, and move the position of each video style template in the LUT preview window. The first video style template displayed in the preview window is used as the video style template selected by the electronic device 100 .
  • the electronic device 100 in addition to using the video style template to change the display effect of the preview image, can also detect a user operation to start recording a video after adding the video style template, and in response to the operation, the electronic device 100 starts recording Video, so as to obtain the video after adjusting the display effect using the video style template.
  • the electronic device 100 can also detect the user operation of taking a photo. In response to this operation, the electronic device 100 saves the preview image with the video style template added in the preview box as a picture, so as to obtain The style template adjusts the image after the display effect.
  • HDR10 is a high-dynamic range image (High-Dynamic Range, HDR). Compared with ordinary images, HDR can provide more dynamic range and image details, and can better Reflecting the visual effects in the real environment, 10 in HDR10 is 10 bits, and HDR10 can record video with a high dynamic range of 10 bits.
  • the electronic device 100 may detect the user's touch operation on the professional mode option, and enter the professional mode.
  • the functional options that can be included in the user interface are, for example: LOG option, flashlight option, LUT option, and setting option.
  • the user interface also includes parameter adjustment options, such as: measurement Light M option, ISO option, shutter S option, exposure compensation EV option, focus mode AF option and white balance WB option.
  • the electronic device 100 may turn on the professional mode by default after starting the camera application.
  • the electronic device 100 can also enable the professional mode in other ways, for example, the electronic device 100 can also enable the professional mode according to the user's voice command, which is not limited in this embodiment of the present application.
  • the electronic device 100 may detect a user operation on the LOG option by the user, and in response to the operation, the electronic device 100 enables the LOG function.
  • the LOG function can apply the logarithmic function to the exposure curve to preserve the details of the highlights and shadows in the image captured by the camera to the maximum extent, so that the saturation of the final preview image is lower.
  • the video recorded with LOG function is called LOG video.
  • the electronic device 100 can not only record a video with a video style template added through the professional mode, but also add a video style template to the video after recording a video without a video style template, or record a LOG video after enabling the LOG function. Then add a video style template for the LOG video. In this way, the electronic device 100 can not only adjust the display effect of the picture before recording the video, but also adjust the display effect of the recorded video after the video recording is completed, which increases the flexibility and freedom of image adjustment.
  • the embodiment of the present application also provides a video processing device, including: a video style determination module, configured to determine a video style template among a plurality of video style templates, and each video style template corresponds to a preset three-dimensional color lookup table LUT;
  • the video acquisition module is used to obtain the video taken by the camera;
  • the first processing module is used to process the video taken by the camera through the logarithmic LOG curve corresponding to the current sensitivity ISO of the camera to obtain the LOG video;
  • the second processing A module configured to process the LOG video based on the two-dimensional 2D-LUT corresponding to the determined video style template when the capture frame rate of the video captured by the camera is the first frame rate, to obtain the corresponding to the determined video style template Video;
  • the third processing module used for processing the LOG video based on the three-dimensional 3D-LUT corresponding to the determined video style template when the capture frame rate of the video captured by the camera is the second frame rate, to obtain the determined video
  • the first frame rate is greater than
  • the video processing device can apply the above-mentioned video processing method, and the specific process and principle will not be repeated here, wherein, the first processing module can specifically be the GTM module 26 in the above-mentioned embodiment, and the second processing module can be specifically the above-mentioned embodiment In the 2D-LUT processing module 30, the third processing module may specifically be the LUT processing module 29 in the above embodiment.
  • the third processing module is specifically configured to: establish a cube interpolation space based on the LUT, where the LUT is a three-dimensional 3D-LUT; determine the cube to which each pixel in the LOG video belongs in the cube interpolation space, and the cube Divided into 6 tetrahedrons; Determine the tetrahedron to which each pixel in the LOG video belongs; For pixels corresponding to the vertices of the cube, convert the pixel value to the pixel value after LUT processing; For pixels not corresponding to the vertices of the cube , perform interpolation according to the tetrahedron to which each pixel belongs, and convert the pixel value into a pixel value after LUT processing.
  • the second processing module is specifically configured to, when the capture frame rate of the video captured by the camera is at the second frame rate, process the LOG video based on the determined 3D 3D-LUT corresponding to the video style template. processing to obtain a video corresponding to the determined video style template, and the first frame rate is greater than the second frame rate.
  • the video processing device further includes a stream splitting processing module, configured to split the video corresponding to the determined video style template into two streams when the capture frame rate of the video captured by the camera is the first frame rate. One way is saved, and the other way is previewed.
  • the first frame rate is 120 FPS or 60 FPS
  • the second frame rate is 30 FPS.
  • the video processing device further includes: a conversion module, configured to convert the LOG video from the LOG video of the RGB color space to the LOG video of the YUV color space. Between the module 27 and the YUV denoising module 28, the conversion module is not shown in FIG. 9; the YUV denoising module 28 is used to perform YUV denoising processing on the LOG video in the YUV color space to obtain the denoised LOG video.
  • the video processing device further includes: a first conversion module, configured to convert the denoised LOG video from a LOG video in a YUV color space to a LOG video in an RGB color space, and the first conversion module Can be located between the YUV denoising module 28 and the LUT processing module 29 in Fig. 9;
  • the second conversion module is used to convert the video corresponding to the determined video style template of the RGB color space into the video of the YUV color space, the The second conversion module may be located between the LUT processing module 29 in FIG. 9 and the process of saving the video recording 1.
  • the first conversion module and the second conversion module are not shown in FIG. 9 .
  • the video processing device further includes: a first saving module, configured to save the video corresponding to the determined video style template; a second saving module, configured to save the LOG video.
  • the video processing device further includes: a first saving module, used to save the video corresponding to the determined video style template; a backup conversion module, used to convert the LOG video to the Rec.709 color standard video; the second saving module is used to save the video of the Rec.709 color standard.
  • the process of processing the video captured by the camera through the logarithmic LOG curve to obtain the LOG video, and the LUT corresponding to the LOG based on the determined video style template The video is processed to obtain the process of video corresponding to the determined video style template;
  • the video processing method also includes a second video processing flow, and the second video processing flow includes: processing the video taken by the camera through a logarithmic LOG curve, The process of obtaining the LOG video; the process of processing the LOG video based on the LUT corresponding to the determined video style template to obtain the video corresponding to the determined video style template;
  • the video processing method also includes: a first saving module for storing The video corresponding to the determined video style template in the first video processing flow is saved; the preview module is configured to perform preview based on the video corresponding to the determined video style template in the second video processing flow.
  • the cube has the 0th to 7th vertices, the direction from the 0th vertex to the 1st vertex is the coordinate axis direction of the blue B channel, and the direction from the 0th vertex to the 4th vertex is the red R
  • the coordinate axis direction of the channel, the direction from the 0th vertex to the 2nd vertex is the coordinate axis direction of the green G channel
  • the 0th vertex, the 1st vertex, the 2nd vertex and the 3rd vertex are on the same plane
  • the 5th vertex and the 7th vertex are located on the same plane
  • the 4th vertex, the 5th vertex, the 6th vertex and the 7th vertex are located on the same plane
  • the 0th vertex, the 2nd vertex, the 4th vertex and the 6th vertex are located on the same plane
  • each module of the video processing device is only a division of logical functions, and may be fully or partially integrated into one physical entity or physically separated during actual implementation.
  • these modules can all be implemented in the form of software called by the processing element; they can also be implemented in the form of hardware; some modules can also be implemented in the form of software called by the processing element, and some modules can be implemented in the form of hardware.
  • any one of the video style determination module, the video acquisition module, the first processing module, the second processing module, and the third processing module may be a separate processing element, or may be integrated in a video processing device, such as integrated in
  • it can also be stored in the memory of the video processing device in the form of a program, and a certain processing element of the video processing device calls and executes the functions of the above modules.
  • the implementation of other modules is similar.
  • all or part of these modules can be integrated together, and can also be implemented independently.
  • the processing element mentioned here may be an integrated circuit with signal processing capabilities.
  • each step of the above method or each module above can be completed by an integrated logic circuit of hardware in the processor element or an instruction in the form of software.
  • the video style determination module, the video acquisition module, the first processing module, the second processing module and the third processing module may be one or more integrated circuits configured to implement the above method, for example: one or more specific Integrated Circuit (Application Specific Integrated Circuit, ASIC), or, one or more microprocessors (digital signal processor, DSP), or, one or more Field Programmable Gate Array (Field Programmable Gate Array, FPGA), etc.
  • ASIC Application Specific Integrated Circuit
  • DSP digital signal processor
  • FPGA Field Programmable Gate Array
  • the processing element may be a general-purpose processor, such as a central processing unit (Central Processing Unit, CPU) or other processors that can call programs.
  • CPU central processing unit
  • these modules can be integrated together and implemented in the form of a system-on-a-chip (SOC).
  • SOC system-on-a-chip
  • An embodiment of the present application further provides a video processing device, including: a processor and a memory, the memory is used to store at least one instruction, and when the instruction is loaded and executed by the processor, the video processing method in any of the foregoing embodiments is implemented.
  • the video processing apparatus may apply the above-mentioned video processing method, and the specific process and principle will not be repeated here.
  • the number of processors may be one or more, and the processors and memory may be connected through a bus or in other ways.
  • the memory can be used to store non-transitory software programs, non-transitory computer-executable programs and modules, such as program instructions/modules corresponding to the video processing device in the embodiment of the present application.
  • the processor executes various functional applications and data processing by running non-transitory software programs, instructions and modules stored in the memory, that is, implements the method in any of the above method embodiments.
  • the memory may include a program storage area and a data storage area, wherein the program storage area may store an operating system, an application program required by at least one function; and necessary data and the like.
  • the memory may include high-speed random access memory, and may also include non-transitory memory, such as at least one magnetic disk storage device, flash memory device, or other non-transitory solid-state storage devices.
  • an embodiment of the present application further provides an electronic device, including: a camera 193 and the above-mentioned video processing device, where the video processing device includes a processor 110 .
  • the electronic device may be any product or component with a video shooting function such as a mobile phone, a TV, a tablet computer, a watch, a bracelet, and the like.
  • An embodiment of the present application further provides a computer-readable storage medium, in which a computer program is stored, and when running on a computer, the computer is made to execute the video processing method in any of the foregoing embodiments.
  • all or part of them may be implemented by software, hardware, firmware or any combination thereof.
  • software When implemented using software, it may be implemented in whole or in part in the form of a computer program product.
  • the computer program product includes one or more computer instructions. When the computer program instructions are loaded and executed on the computer, the processes or functions according to the present application will be generated in whole or in part.
  • the computer can be a general purpose computer, a special purpose computer, a computer network, or other programmable devices.
  • the computer instructions may be stored in or transmitted from one computer-readable storage medium to another computer-readable storage medium, for example, the computer instructions may be transmitted from a website, computer, server or data center Transmission to another website site, computer, server, or data center by wired (eg, coaxial cable, optical fiber, DSL) or wireless (eg, infrared, wireless, microwave, etc.) means.
  • the computer-readable storage medium may be any available medium that can be accessed by a computer, or a data storage device such as a server or a data center integrated with one or more available media.
  • the available medium may be a magnetic medium (for example, a floppy disk, a hard disk, a magnetic tape), an optical medium (for example, DVD), or a semiconductor medium (for example, a Solid State Disk).
  • "at least one” means one or more, and “multiple” means two or more.
  • “And/or” describes the association relationship of associated objects, indicating that there may be three kinds of relationships, for example, A and/or B may indicate that A exists alone, A and B exist simultaneously, or B exists alone. Among them, A and B can be singular or plural.
  • the character “/” generally indicates that the contextual objects are an “or” relationship.
  • “At least one of the following” and similar expressions refer to any combination of these items, including any combination of single items or plural items.
  • At least one of a, b, and c may represent: a, b, c, a-b, a-c, b-c, or a-b-c, wherein a, b, and c may be single or multiple.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Human Computer Interaction (AREA)
  • Image Processing (AREA)

Abstract

本申请实施例提供一种视频处理方法、装置、电子设备和存储介质,涉及视频拍摄技术领域,可以基于LUT的特性使电子设备所拍摄的视频具有不同的风格效果,以满足更高的调色要求。视频处理方法,包括:在多个视频风格模板中确定一个视频风格模板,每个视频风格模板对应一个LUT;获取通过摄像头拍摄的视频;通过摄像头当前的感光度ISO所对应的对数LOG曲线对通过摄像头拍摄的视频进行处理,得到LOG视频;当摄像头拍摄的视频的捕获帧率为第一帧率时,基于所确定的视频风格模板对应的2D-LUT对LOG视频进行处理;当摄像头拍摄的视频的捕获帧率为第二帧率时,基于所确定的视频风格模板对应的3D-LUT对LOG视频进行处理,第一帧率大于第二帧率。

Description

视频处理方法、装置、电子设备和存储介质
本申请要求于2021年8月12日提交中国专利局、申请号为202110926600.0、申请名称为“视频处理方法、装置、电子设备和存储介质”的中国专利申请的优先权,其全部内容通过引用结合在本申请中。
技术领域
本申请涉及视频拍摄技术领域,特别涉及一种视频处理方法、装置、电子设备和存储介质。
背景技术
随着技术的发展,用户对通过手机等终端所拍摄的视频效果和风格的要求越来越高,然而,目前的手机中拍摄视频所使用的滤镜通常沿用拍照模式下的滤镜原理,经过滤镜处理的视频无法满足较高的调色要求。
发明内容
一种视频处理方法、装置、电子设备和存储介质,可以基于LUT的特性使电子设备所拍摄的视频具有不同的风格效果,以满足更高的调色要求。
第一方面,提供一种视频处理方法,包括:在多个视频风格模板中确定一个视频风格模板,每个视频风格模板对应一个预设的颜色查找表LUT;获取通过摄像头拍摄的视频;通过摄像头当前的感光度ISO所对应的对数LOG曲线对通过摄像头拍摄的视频进行处理,得到LOG视频;当摄像头拍摄的视频的捕获帧率为第一帧率时,基于所确定的视频风格模板对应的二维2D-LUT对LOG视频进行处理,得到与所确定的视频风格模板对应的视频;当摄像头拍摄的视频的捕获帧率为第二帧率时,基于所确定的视频风格模板对应的三维3D-LUT对LOG视频进行处理,得到与所确定的视频风格模板对应的视频,第一帧率大于第二帧率。
在一种可能的实施方式中,当摄像头拍摄的视频的捕获帧率为第一帧率时,基于所确定的视频风格模板对应的二维2D-LUT对LOG视频进行处理,得到与所确定的视频风格模板对应的视频的过程在HSV色彩空间执行。以便于利用HSV色彩空间的特点,应用2D-LUT。
在一种可能的实施方式中,当摄像头拍摄的视频的捕获帧率为第二帧率时,基于所确定的视频风格模板对应的三维3D-LUT对所述LOG视频进行处理,得到与所确定的视频风格模板对应的视频的过程包括:基于LUT建立立方体插值空间;确定LOG视频中每个像素点在立方体插值空间中所属的立方体,立方体中被划分为6个四面体;确定LOG视频中每个像素点所属的四面体;对于对应立方体顶点的像素点,将像素值转换为经过LUT处理后的像素值;对于不对应立方体顶点的像素点,根据每个像素点 所属的四面体进行插值,将像素值转换为经过LUT处理后的像素值。
在一种可能的实施方式中,在当摄像头拍摄的视频的捕获帧率为第二帧率时,基于所确定的视频风格模板对应的三维3D-LUT对LOG视频进行处理,得到与所确定的视频风格模板对应的视频的过程之前,还包括:将LOG视频由RGB色彩空间的LOG视频转换为YUV色彩空间的LOG视频;对YUV色彩空间的LOG视频进行YUV去噪处理,得到去噪后的LOG视频。由于通过LOG曲线对视频进行处理得到的LOG视频会引入噪声,因此可以将LOG视频转换为YUV色彩空间之后,进行YUV去噪处理,通过算法降噪,以改善视频图像质量。
在一种可能的实施方式中,在当摄像头拍摄的视频的捕获帧率为第二帧率时,基于所确定的视频风格模板对应的三维3D-LUT对LOG视频进行处理,得到与所确定的视频风格模板对应的视频的过程之前,还包括:将去噪后的LOG视频由YUV色彩空间的LOG视频转换为RGB色彩空间的LOG视频;在基于所确定的视频风格模板对应的LUT对LOG视频进行处理,得到与所确定的视频风格模板对应的视频的过程之后,还包括:将RGB色彩空间的与所确定的视频风格模板对应的视频转换为YUV色彩空间的视频。
在一种可能的实施方式中,当摄像头拍摄的视频的捕获帧率为第一帧率时,将与所确定的视频风格模板对应的视频分流为两路,其中一路进行保存,另外一路进行预览。可以使预览视频和最终得到的视频具有相同的视觉效果,便于用户直接基于调色后的风格进行视频预览。
在一种可能的实施方式中,第一帧率为120每秒传输帧数FPS或60FPS,第二帧率为30FPS。
第二方面,提供一种视频处理装置,包括:处理器和存储器,存储器用于存储至少一条指令,指令由所述处理器加载并执行时以实现上述的视频处理方法。
第三方面,提供一种电子设备,包括:摄像头;上述的视频处理装置。
第四方面,提供一种计算机可读存储介质,计算机可读存储介质中存储有计算机程序,当其在计算机上运行时,使得计算机执行上述的视频处理方法。
本申请实施例中的视频处理方法、装置、电子设备和存储介质,在视频录制过程中,利用电影行业的LUT技术,基于所确定的视频风格模板对应的LUT对LOG视频进行处理,使所录制的视频具有所确定的视频风格模板对应的风格效果,以满足较高的调色要求,使所录制的视频具有电影感。并且,在视频捕获帧率较低时,使用3D-LUT对视频进行处理,以实现更加精确的色彩控制,在视频捕获帧率较高时,使用2D-LUT对视频进行处理,以实现在有限的帧间隔时间内进行LUT处理。
附图说明
图1为本申请实施例中一种电子设备的结构框图;
图2为本申请实施例中一种视频处理方法的流程图;
图3为本申请实施例中一种电影模式下用户界面的示意图;
图4为本申请实施例中一种LOG曲线的示意图;
图5为图2中步骤104的一种具体流程示意图;
图6为本申请实施例中一种立方体插值空间中立方体和四面体关系的示意图;
图7为本申请实施例中另一种视频处理方法的流程图;
图8为UV平面示意图;
图9为本申请实施例中一种电子设备的另一种结构框图;
图10为本申请实施例中一种电子设备的另一种结构框图;
图11为本申请实施例中一种电子设备的软件结构框图;
图12为本申请实施例中一种专业模式下用户界面的示意图。
具体实施方式
本申请的实施方式部分使用的术语仅用于对本申请的具体实施例进行解释,而非旨在限定本申请。
在介绍本申请实施例之前,首先对本申请实施例所涉及的电子设备进行介绍,如图1所示,电子设备100可以包括处理器110,摄像头193,显示屏194等。可以理解的是,本发明实施例示意的结构并不构成对电子设备100的具体限定。在本申请另一些实施例中,电子设备100可以包括比图示更多或更少的部件,或者组合某些部件,或者拆分某些部件,或者不同的部件布置。图示的部件可以以硬件,软件或软件和硬件的组合实现。
处理器110可以包括一个或多个处理单元,例如:处理器110可以包括图形处理器(graphics processing unit,GPU),图像信号处理器(image signal processor,ISP),控制器,视频编解码器,数字信号处理器(digital signal processor,DSP)等。其中,不同的处理单元可以是独立的器件,也可以集成在一个或多个处理器中。控制器可以根据指令操作码和时序信号,产生操作控制信号,完成取指令和执行指令的控制。处理器110中还可以设置存储器,用于存储指令和数据。
电子设备100通过GPU,显示屏194,以及应用处理器等实现显示功能。GPU为图像处理的微处理器,连接显示屏194和应用处理器。GPU用于执行数学和几何计算,用于图形渲染。处理器110可包括一个或多个GPU,其执行程序指令以生成或改变显示信息。
电子设备100可以通过ISP,摄像头193,视频编解码器,GPU,显示屏194以及应用处理器等实现拍摄功能。
ISP用于处理摄像头193反馈的数据。例如,拍照时,打开快门,光线通过镜头被传递到摄像头感光元件上,光信号转换为电信号,摄像头感光元件将所述电信号传递给ISP处理,转化为肉眼可见的图像。ISP还可以对图像的噪点,亮度,肤色进行算法优化。ISP还可以对拍摄场景的曝光,色温等参数优化。在一些实施例中,ISP可以设置在摄像头193中。
摄像头193用于捕获静态图像或视频。物体通过镜头生成光学图像投射到感光元件。感光元件可以是电荷耦合器件(charge coupled device,CCD)或互补金属氧化物半导体(complementary metal-oxide-semiconductor,CMOS)光电晶体管。感光元件把光信号转换成电信号,之后将电信号传递给ISP转换成数字图像信号。ISP将数字图像信号输出到DSP加工处理。DSP将数字图像信号转换成标准的RGB,YUV等格式的图像信 号。在一些实施例中,电子设备100可以包括1个或N个摄像头193,N为大于1的正整数。
数字信号处理器用于处理数字信号,除了可以处理数字图像信号,还可以处理其他数字信号。例如,当电子设备100在频点选择时,数字信号处理器用于对频点能量进行傅里叶变换等。
视频编解码器用于对数字视频压缩或解压缩。电子设备100可以支持一种或多种视频编解码器。这样,电子设备100可以播放或录制多种编码格式的视频,例如:动态图像专家组(moving picture experts group,MPEG)1,MPEG2,MPEG3,MPEG4等。
如图2所示,本申请实施例提供一种视频处理方法,该视频处理方法的执行主体可以为处理器110,具体可以为ISP或者ISP与其他处理器的组合,该视频处理方法包括:
步骤101、在多个视频风格模板中确定一个视频风格模板,每个视频风格模板对应一个预设的颜色查找表(Look Up Table,LUT);
其中,LUT的本质为数学转换模型,利用LUT可以将图像数据值输出为另外的图像数据值,从而改变画面的曝光与色彩。因此,可以预先生成对应不同视频风格的LUT,在电子设备录制视频之前,首先确定出一个视频风格模板,例如可以基于用户的选择来确定视频风格模板,或者基于人工智能(Artificial Intelligence,AI),根据当前摄像头获取的图像所对应的场景自动确定视频风格模板。例如,假设电子设备为手机,在一种可能的实施方式中,如图3所示,用户操作手机进入拍摄界面,拍摄界面包括电影模式选项,当用户进一步选择电影模式选项进入电影模式,在对应的电影模式界面中,包括多个视频风格模板选项,例如包括《A》电影风格模板、《B》电影风格模板和《C》电影风格模板,图3所示的用户界面中仅显示了一个《A》电影风格模板,可以理解地,用户界面中可以并排显示多个不同的电影风格模板,不同的电影风格模板所对应的LUT可以是预先基于对应电影配色风格所生成的,LUT的颜色转换具有对应电影所具有的风格特点,例如《A》电影的配色风格为互补色,互补色是指两种对应的颜色形成对比效果,以暖色系与冷色系的两种颜色来强调对比度以提升鲜艳、突出的效果,通常两种对比的色彩象征冲突行为,透过外在的互补色彩的呈现来表达角色内心正处于矛盾或是身心交瘁的状态,《A》电影风格模板所对应的LUT即用于将颜色映射转换之后,更明显地呈现互补色,以模拟《A》电影的配色风格。在一种可能的实施方式中,如图3所示,用户操作手机进入电影模式,手机会通过获取当前摄像头所拍摄的画面,并基于AI算法确定画面所对应的场景并确定与该场景对应的推荐的视频风格模板,例如若识别到当前所拍摄的画面主体为年轻女性人物,根据算法确定对应的推荐的视频风格模板为《C》电影风格模板,《C》电影为以年轻女性人物为主题的电影,其对应的LUT可以模拟《C》电影的配色风格;例如若识别到当前所拍摄的画面为城市街道,根据算法确定对应的视频风格模板为《B》电影风格模板,《B》电影为以城市街道为主要场景的电影,其对应的LUT可以模拟《B》电影的配色风格。这样,可以自动为用户推荐符合当前场景的视频风格模板。可以预先从电影风格中提取,产生适合移动电子设备的LUT。
步骤102、获取通过摄像头拍摄的视频,例如,在步骤101确定了视频风格模板 之后,如果用户点击拍摄选项,则手机开始获取通过摄像头拍摄的视频;
步骤103、通过摄像头当前的感光度ISO所对应的对数(Logarithm,LOG)曲线对通过摄像头拍摄的视频进行处理,得到LOG视频;
其中,LOG曲线是基于场景的曲线,不同ISO下LOG曲线略有不同。随着ISO的增加,LOG曲线最大值也在增加。当ISO提高到一定程度,高光处有肩部形状,保持高光不过曝。如图4所示,图4示意了一种LOG曲线,其中横坐标为线性信号,以16比特bit编码值Code Value表示,纵坐标为经过LOG曲线处理后的LOG信号,以10bit编码值表示。通过LOG曲线处理,可以利用摄像头的信号输入,将暗部区间的信息编码到中间调(如图4中曲线陡峭的部分),形成10bit的信号输出,符合人眼对光线LOG感应规则,最大化的保留了暗部信息,LOG视频可以利用有限的bit深度最大化的保留阴影和高光的细节。图4中的ASA即为感光度,不同的ASA即对应不同的ISO,两者属于不同制式。
步骤1041、当摄像头拍摄的视频的捕获capture帧率为第一帧率时,基于所确定的视频风格模板对应的二维2D-LUT对LOG视频进行处理,得到与所确定的视频风格模板对应的视频;
步骤1042、当摄像头拍摄的视频的捕获帧率为第二帧率时,基于所确定的视频风格模板对应的三维3D-LUT对LOG视频进行处理,得到与所确定的视频风格模板对应的视频,第一帧率大于第二帧率。
具体地,在得到LOG视频之后,以LOG视频作为输入,应用在步骤101中所确定的视频风格模板对应的LUT,对LOG视频图像进行映射转换处理,在处理之后,既可以得到与所确定的视频风格模板对应的视频。基于LUT对LOG视频进行处理后输出的可以是Rec.709色彩标准的视频,也可以是高动态范围图像(High-Dynamic Range,HDR)10标准的视频,即可以通过LUT对LOG视频的处理,将视频转换为HDR10标准。
其中,捕获帧率是指摄像头传感器的读出帧率,或者说是视频的拍摄帧率,对于视频录制来说,摄像头拍摄时的帧率对视频效果也有影响,例如,升格是电影拍摄中的一种技术手段,可以实现不同的效果,当视频的捕获帧率大于播放帧率时,即为升格,例如视频播放时的帧率为每秒显示30帧画面,即帧率为30每秒传输帧数(Frames Per Second,FPS),而拍摄时的帧率为60FPS,即为升格,可以实现的效果为慢动作,升格中摄像头拍摄时的帧率适用于不同的场景表达。例如,60FPS的升格拍摄适于表达平移、慢走、欢笑、鼓掌等场景,120FPS的升格拍摄适于表达跑步、转身、撩头发、撒花等场景。因此,在视频录制过程中,用户可以选择进行升格拍摄,或者选择其他方式使用较高帧率拍摄,摄像头拍摄的视频的捕获帧率越快,帧间隔就越短,则要求视频录制过程中的处理速度也要越快,例如,60FPS的捕获帧率,需要应用LUT对视频进行处理的时间在15ms以内,120FPS的捕获帧率,需要应用LUT对视频进行处理的时间在8ms以内。LUT可以为3D-LUT或2D-LUT,相对于2D-LUT,3D-LUT可以实现更加精确的色彩控制,但是,3D-LUT在应用时由于算法的复杂性,需要更多的处理时间,而2D-LUT在应用时所需要的处理时间较短,因此,在本申请实施例中,当摄像头拍摄的视频的捕获帧率为较高的第一帧率时,应用2D-LUT,以实现在有限的 帧间隔时间内进行LUT处理,当摄像头拍摄的视频的捕获帧率为较低的第二帧率时,应用3D-LUT,以实现在足够长的帧间隔时间内进行LUT处理,实现更加精确的色彩控制。需要说明的是,本申请实施例并不仅限于应用在升格中,本申请实施例同样适用其他需要较高捕获帧率的方式。
不同的LUT应用在电子设备上,可以对电子设备中相关的模块进行适配,以适应不同风格的LUT,例如,如果在步骤101中所确定的视频风格模板为灰色调视频风格模板,灰色调画面的特点为使画面中纹理感较强、饱和度较低、除了人物皮肤的颜色,没有更多的颜色干扰、暗部较冷,基于这些特点,电子设备在录制视频的过程中,可以对相关的模块参数进行调整,保持画面中的纹理,不做很强的去噪和锐化,适当降低画面的饱和度,保持画面中的皮肤颜色真实还原,使画面的暗部向冷色调整。
本申请实施例中的视频处理方法,在视频录制过程中,利用电影行业的LUT技术,基于所确定的视频风格模板对应的LUT对LOG视频进行处理,使所录制的视频具有所确定的视频风格模板对应的风格效果,以满足较高的调色要求,使所录制的视频具有电影感。并且,在视频捕获帧率较低时,使用3D-LUT对视频进行处理,以实现更加精确的色彩控制,在视频捕获帧率较高时,使用2D-LUT对视频进行处理,以实现在有限的帧间隔时间内进行LUT处理。
在一种可能的实施方式中,如图5所示,上述步骤1042、当摄像头拍摄的视频的捕获帧率为第二帧率时,基于所确定的视频风格模板对应的三维3D-LUT对LOG视频进行处理,得到与所确定的视频风格模板对应的视频的过程包括:
步骤10421、基于LUT建立立方体插值空间,LUT为三维3D-LUT;
其中,3D-LUT的实现是在RGB域进行的,3D-LUT为电影工业中常用的调色映射关系,可以将任意输入的RGB像素值转换为对应的其他RGB像素值,例如输入12bit的RGB视频图像,经过LUT处理映射之后输出12bit的RGB视频图像。在步骤1041中,将整个RGB色彩空间均匀地分为例如33×33×33的立方体,对应LUT,每个立方体的边长step_size例如为2 (12-5)=2 7
步骤10422、确定LOG视频中每个像素点在立方体插值空间中所属的立方体,立方体中被划分为6个四面体;
其中,LOG视频作为LUT处理过程中的输入,对LOG视频画面中每个像素点得到通过LUT处理映射后的像素点,既可以实现通过LUT对LOG视频进行处理的过程,在步骤10422中,需要确定每个作为输入的LOG视频中每个像素点在上述立方体插值空间中所属的立方体,立方体被划分为6个四面体。
步骤10423、确定LOG视频中每个像素点所属的四面体;
步骤10424、对于对应立方体顶点的像素点,将像素值转换为经过LUT处理后的像素值,对于不对应立方体顶点的像素点,根据每个像素点所属的四面体进行插值,将像素值转换为经过LUT处理后的像素值。
具体地,对于输入的像素点来说,如果像素点位于立方体的顶点,根据顶点的索引以及3D-LUT,可以直接获取映射后的RGB像素值,即可以直接通过LUT将其像素值映射转换为对应的像素值,而如果像素点位于立方体的顶点之间,则根据像素点所属的四面体进行插值。
在一种可能的实施方式中,如图6所示,立方体中具有第0至第7顶点,在图6中分别以数字0~7表示,第0顶点至第1顶点的方向为蓝色B通道的坐标轴方向,第0顶点至第4顶点的方向为红色R通道的坐标轴方向,第0顶点至第2顶点的方向为绿色G通道的坐标轴方向,第0顶点、第1顶点、第2顶点和第3顶点位于同一平面,第1顶点、第3顶点、第5顶点和第7顶点位于同一平面,第4顶点、第5顶点、第6顶点和第7顶点位于同一平面,第0顶点、第2顶点、第4顶点和第6顶点位于同一平面;第0顶点、第1顶点、第5顶点和第7顶点形成第一个四面体,第0顶点、第1顶点、第3顶点和第7顶点形成第二个四面体,第0顶点、第2顶点、第3顶点和第7顶点形成第三个四面体,第0顶点、第4顶点、第5顶点和第7顶点形成第四个四面体,第0顶点、第4顶点、第6顶点和第7顶点形成第五个四面体,第0顶点、第2顶点、第6顶点和第7顶点形成第六个四面体;其中,第i顶点的坐标为(Ri,Gi,Bi),i的取值为0、1、2、3、…、7,第i顶点经过LUT处理后的像素值为VE(Ri,Gi,Bi),其中E取R、G和B;
上述步骤1044中,对于不对应立方体顶点的像素点,根据每个像素点所属的四面体进行插值,将像素值转换为经过LUT处理后的像素值的过程包括:
根据当前像素点(R,G,B)生成经过LUT处理后的E通道像素值VE(R,G,B),E取R、G和B,当前像素点是指输入的LOG视频中的当前待进行插值计算的像素点;
VE(R,G,B)=VE(R0,G0,B0)+(delta_valueR_E×deltaR+delta_valueG_E×deltaG+delta_valueB_E×deltaB+(step_size>>1))/(step_size);
VE(R0,G0,B0)为第0顶点(R0,G0,B0)经过LUT处理后的E通道像素值,E取R、G和B;
delta_valueR_E为当前像素点所属四面体对应R通道的坐标轴方向上的两个顶点经过LUT处理后的E通道像素值之差,delta_valueG_E为当前像素点所属四面体对应G通道的坐标轴方向上的两个顶点经过LUT处理后的E通道像素值之差,delta_valueB_E为当前像素点所属四面体对应B通道的坐标轴方向上的两个顶点经过LUT处理后的E通道像素值之差;
deltaR为当前像素点(R,G,B)中的R值与第0顶点(R0,G0,B0)中的R0值之差,deltaG为当前像素点(R,G,B)中的G值与第0顶点(R0,G0,B0)中的G0值之差,deltaB为当前像素点(R,G,B)中的B值与第0顶点(R0,G0,B0)中的B0值之差;
step_size为立方体的边长。
其中,>>表示右移运算,(step_size>>1)即step_size右移一位。
具体地,例如,对于输入的当前像素点(R,G,B),计算deltaR、deltaG和deltaB,deltaR、deltaG和deltaB表示当前像素点(R,G,B)与第0顶点的距离,deltaR=R-R0,deltaG=G-G0,deltaB=B-B0,在上述步骤1043中,可以根据deltaR、deltaG以及deltaB之间的关系判断当前像素点属于哪个四面体。如果deltaB≥deltaR且deltaR≥deltaG,则确定当前像素点属于第一个四面体;如果deltaB≥deltaG且deltaG≥deltaR,则确定当前像素点属于第二个四面体;如果deltaG≥deltaB且deltaB≥deltaR,则确定 当前像素点属于第三个四面体;如果deltaR≥deltaB且deltaB≥deltaG,则确定当前像素点属于第四个四面体;如果deltaR≥deltaG且deltaG≥deltaB,则确定当前像素点属于第五个四面体;如果deltaR、deltaG以及deltaB之间的关系不属于上述第一~第五个四面体的条件,则确定当前像素点属于第六个四面体。假设当前像素点(R,G,B)属于第一个四面体,该像素点经过LUT处理后的R通道像素值VR(R,G,B)的计算过程中,delta_valueR_E为当前像素点所属四面体对应R通道的坐标轴方向上的两个顶点经过LUT处理后的E通道像素值之差,即delta_valueR_R=VR(R5,G5,B5)-VR(R1,G1,B1),delta_valueG_R=VR(R7,G7,B7)-VR(R5,G5,B5),delta_valueB_R=VR(R1,G1,B1)-VR(R0,G0,B0),VR(R,G,B)=VR(R0,G0,B0)+(delta_valueR_R×deltaR+delta_valueG_R×deltaG+delta_valueB_R×deltaB+(step_size>>1))/(step_size);该像素点经过LUT处理后的G通道像素值VG(R,G,B)的计算过程中,delta_valueG_E为当前像素点所属四面体对应G通道的坐标轴方向上的两个顶点经过LUT处理后的E通道像素值之差,即delta_valueR_G=VR(R5,G5,B5)-VR(R1,G1,B1),delta_valueG_G=VG(R7,G7,B7)-VG(R5,G5,B5),delta_valueB_G=VG(R1,G1,B1)-VG(R0,G0,B0),VG(R,G,B)=VG(R0,G0,B0)+(delta_valueR_G×deltaR+delta_valueG_G×deltaG+delta_valueB_G×deltaB+(step_size>>1))/(step_size);该像素点经过LUT处理后的B通道像素值VG(R,G,B)的计算过程中,delta_valueB_E为当前像素点所属四面体对应B通道的坐标轴方向上的两个顶点经过LUT处理后的E通道像素值之差,即delta_valueR_B=VB(R5,G5,B5)-VB(R1,G1,B1),delta_valueG_B=VB(R7,G7,B7)-VB(R5,G5,B5),delta_valueB_B=VB(R1,G1,B1)-VB(R0,G0,B0),VB(R,G,B)=VB(R0,G0,B0)+(delta_valueR_B×deltaR+delta_valueG_B×deltaG+delta_valueB_B×deltaB+(step_size>>1))/(step_size)。对于当前像素点(R,G,B)属于其他四面体的情况,计算过程类似,区别在于delta_valueR_E的计算,例如对于第二个四面体,delta_valueR_R=VR(R7,G7,B7)-VR(R3,G3,B3),delta_valueG_R=VR(R3,G3,B3)-VR(R1,G1,B1),delta_valueB_R=VR(R1,G1,B1)-VR(R0,G0,B0),基于其他四面体的具体计算过程在此不再赘述。
在一种可能的实施方式中,上述步骤1041、当摄像头拍摄的视频的捕获帧率为第一帧率时,基于所确定的视频风格模板对应的二维2D-LUT对LOG视频进行处理,得到与所确定的视频风格模板对应的视频的过程在HSV色彩空间执行。
具体地,当摄像头拍摄的视频的捕获帧率为第一帧率时,即基于2D-LUT对LOG视频进行处理时,2D-LUT为预先通过3D-LUT模拟得到的,例如,预先知道3D-LUT,以及3D-LUT对应的输入数据和输出数据,输入数据和输出数据均属于RGB色彩空间,此时,可以将输入数据由RGB色彩空间转换至HSV色彩空间,将输出数据由RGB色彩空间转换至HSV色彩空间,通过HSV色彩空间中数据的转换关系,既可以得到2D-LUT的具体模型,以便于在视频录制的过程中,应用该2D-LUT,由于在2D-LUT中只能基于两个变量进行转换,因此可以对色调H和饱和度S进行转换,将明度V忽略,以实现2D-LUT模拟3D-LUT的效果。
在一种可能的实施方式中,如图7所示,在上述步骤1042、当摄像头拍摄的视频 的捕获帧率为第二帧率时,基于所确定的视频风格模板对应的三维3D-LUT对LOG视频进行处理,得到与所确定的视频风格模板对应的视频的过程之前,还包括:步骤105、将LOG视频由RGB色彩空间的LOG视频转换为YUV色彩空间的LOG视频;步骤106、对YUV色彩空间的LOG视频进行YUV去噪处理,得到去噪后的LOG视频,在步骤1042中应用LUT的LOG视频即为在步骤106中经过YUV去噪后的LOG视频。由于步骤103中得到的LOG视频会引入噪声,因此可以将LOG视频转换为YUV色彩空间之后,进行YUV去噪处理,通过算法降噪,以改善视频图像质量。
在一种可能的实施方式中,如图7所示,在上述步骤1042、当摄像头拍摄的视频的捕获帧率为第二帧率时,基于所确定的视频风格模板对应的三维3D-LUT对LOG视频进行处理,得到与所确定的视频风格模板对应的视频的过程之前,还包括:步骤107、将去噪后的LOG视频由YUV色彩空间的LOG视频转换为RGB色彩空间的LOG视频;在上述步骤104、基于所确定的视频风格模板对应的LUT对LOG视频进行处理,得到与所确定的视频风格模板对应的视频的过程之后,还包括:步骤108、将RGB色彩空间的与所确定的视频风格模板对应的视频转换为YUV色彩空间的视频。由于步骤1042中基于LUT对LOG视频进行处理的过程是基于RGB色彩空间实现的,因此,因此,在步骤1042之前先将YUV色彩空间的视频转换为RGB色彩空间的视频,在步骤1042之后,再将RGB色彩空间的视频重新转换为YUV色彩空间的视频。
YUV(亦称YCbCr)是欧洲电视系统采用的一种颜色编码方法。在现代彩色电视系统中,通常采用三管彩色摄像机或彩色CCD摄影机进行取像,然后把取得的彩色图像信号经分色、分别放大校正后得到RGB信号,再经过矩阵变换电路得到亮度信号Y和两个色差信号B-Y(即U)、R-Y(即V),最后发送端将三个信号分别进行编码后用同一信道发送出去。这种色彩表示方法就是YUV颜色空间。YCbCr是YUV模型的具体实现,其实是YUV经过缩放和偏移的翻版。其中Y与YUV中的Y含义一致,Cb和Cr同样都指色彩,只是在表示方法上不同而已。在YUV家族中,YCbCr是在计算机系统中应用最多的成员,其应用领域很广泛,JPEG、MPEG均采用此格式。一般人们所讲的YUV大多是指YCbCr。UV平面如图8所示。
RGB和YUV颜色空间的相互转换可以通过3x3的矩阵实现:
Figure PCTCN2022094744-appb-000001
YUV主要有4种采样格式:YCbCr 4:2:0、YCbCr 4:2:2、YCbCr 4:1:1和YCbCr 4:4:4。
在一种可能的实施方式中,如图9所示,电子设备具体可以包括摄像头193、反马赛克Demosaic模块21、变形模块22、融合模块23、噪声处理模块24、色彩校正矩阵(Color Correction Matrix,CCM)模块25、全局色调映射(Global Tone Mapping,GTM)模块26、缩放Scaler模块27、YUV去噪模块28和LUT处理模块29,例如,在视频录制的过程中,当摄像头193拍摄的视频的捕获帧率为第二帧率时,摄像头193拍摄得到长曝光帧视频图像和短曝光帧视频图像,长曝光帧视频图像所对应的曝光时间大于短曝光帧视频图像所对应的曝光时间,长曝光帧视频图像和短曝光帧视频图像分别通过反马赛克模块21的处理,使图像从RAW域转换为RGB域,之后两路视频 图像分别通过变形warp模块22的处理,通过对视频图像的变形实现对齐、防抖的效果,之后两路视频图像通过融合模块23处理,将两种视频图像融合为同一个,融合之后的数据再分流为两路,视频处理方法包括第一视频处理流程S1和第二视频处理流程S2,通过融合模块23处理之后的其中一路进入第一视频处理流程S1,另外一路进入第二视频处理流程S2。
在第一视频处理流程S1中执行上述步骤103、通过对数LOG曲线对通过摄像头拍摄的视频进行处理,得到LOG视频的过程、以及上述步骤104、基于所确定的视频风格模板对应的LUT对LOG视频进行处理,得到与所确定的视频风格模板对应的视频的过程。
例如,第一视频处理流程S1包括,将来自于融合模块23的通过摄像头193拍摄的视频通过噪声处理模块24进行去噪处理,然后通过CCM模块25处理,将视频转换为RGB广色域的色彩空间,然后通过GTM模块26执行上述步骤103,通过LOG曲线对视频进行处理,得到LOG视频,然后通过缩放模块27对视频进行缩放处理,然后通过YUV去噪模块28执行上述步骤106,对视频进行YUV去噪,然后通过LUT处理模块29执行上述步骤1042,通过3D-LUT对视频进行处理,得到与所确定的视频风格模块对应的视频。在第一视频处理流程S1之后,将第一视频处理流程S1中与所确定的视频风格模板对应的视频进行保存,保存为录像1,即可以得到具有风格的视频。
第二视频处理流程S2包括:将来自于融合模块23的通过摄像头193拍摄的视频通过噪声处理模块24进行去噪处理,然后通过CCM模块25处理,将视频转换为RGB广色域的色彩空间,然后通过GTM模块26执行上述步骤103,通过LOG曲线对视频进行处理,得到LOG视频,然后通过缩放模块27对视频进行缩放处理,然后通过YUV去噪模块28执行上述步骤106,对视频进行YUV去噪,然后通过LUT处理模块29执行上述步骤1042,通过3D-LUT对视频进行处理,得到与所确定的视频风格模块对应的视频。基于第二视频处理流程S2中与所确定的视频风格模板对应的视频进行预览。
也就是说,在视频录制过程中,会基于两路视频流分别在第一视频处理流程S1和第二视频处理流程S2中进行处理,两路视频流分别挂在两套相同的算法,均包含步骤103中的基于LOG曲线的处理和步骤1042中基于3D-LUT的处理,其中一路视频流用于编码保存,另一路视频流用于预览。
以下对RAW和YUV的相关内容进行说明:
拜耳域:数码相机上的每个镜头都带有一个光传感器,用以测量光线的明亮程度,但若要获得一幅全彩图像,一般需要有三个光传感器分别获得红、绿、蓝三基色信息,而为了降低数码相机的成本与体积,生产厂商通常会采用CCD或CMOS图像传感器,通常的,CMOS图像传感器输出的原始图像为拜尔域RGB格式,单个像素点只包含一种颜色值,要得到图像的灰度值,需要先插补完整各像素点的颜色信息,再计算各像素点的灰度值。也就是说拜耳域是指数码相机内部的一种原始图片格式。
Raw域或称Raw格式,是指未经加工图像。进一步地,所述Raw图像可以理解为,就是相机的感光元件比如互补金属氧化物半导体(Complementary Metal  OxideSemiconductor,CMOS)或者电荷耦合器件(Charge-coupled Device,CCD)将捕捉到的光源信号转化为数字信号的原始数据。RAW文件是一种记录了数码相机传感器的原始信息,同时记录了由相机拍摄所产生的一些元数据(Metadata,如感光度ISO(InternationalOrganization for Standardization,国际标准化组织)的设置、快门速度、光圈值、白平衡等)的文件。Raw域是未经ISP非线性处理、也未经压缩的格式。Raw格式的全称是RAW Image Format。
YUV是一种颜色编码方法,常使用在各个视频处理组件中。YUV在对照片或视频编码时,考虑到人类的感知能力,允许降低色度的带宽。YUV是编译true-color颜色空间(color space)的种类,Y'UV、YUV、YCbCr、YPbPr等专有名词都可以称为YUV,彼此有重叠。其中“Y”表示明亮度(Luminance或Luma),也就是灰阶值,“U”和“V”表示色度(Chrominance或Chroma),作用是描述影像色彩及饱和度,用于指定像素的颜色。一般YUV分成两种格式,一种是:紧缩格式(packedformats),将Y、U、V值存储成Macro Pixels数组,和RGB的存放方式类似。另一种是:平面格式(planarformats),将Y、U、V的三个分量分别存放在不同的矩阵中。平面格式(planarformats)是指每Y分量,U分量和V分量都是以独立的平面组织的,也就是说所有的U分量都在Y分量后面,而V分量在所有的U分量后面。
在一种可能的实施方式中,如图9所示,在上述步骤1042、基于所确定的视频风格模板对应的3D-LUT对LOG视频进行处理,得到与所确定的视频风格模板对应的视频的过程之前,还包括:保存LOG视频,即保存录像2。也就是说,在步骤103中得到LOG视频之后,分流一路LOG视频直接保存,即保存一份无风格的LOG视频,即录像2,另外一路LOG视频在步骤1042中基于3D-LUT处理之后保存,即保存一份对应视频风格模板的视频,即录像1,例如在录制过程中,同时编码保存两份视频文件在临时文件夹,基于用户的选择,确定其中一个视频作为最终的视频文件,如果用户希望直接录制生成《B》电影风格的视频文件,则将录像1作为最终的视频文件,如果用户希望保存原始素材,以便于提高后期编辑的能力,则将录像2作为最终的视频文件。需要说明的是,保存LOG视频的分流过程可以在通过缩放模块27对视频进行缩放处理的过程之后,在通过LUT处理模块29执行上述步骤104的过程之前。
在一种可能的实施方式中,如图9所示,在上述步骤1042、基于所确定的视频风格模板对应的3D-LUT对LOG视频进行处理,得到与所确定的视频风格模板对应的视频的过程之前,还包括:将LOG视频转换为Rec.709色彩标准的视频;保存Rec.709色彩标准的视频,即录像2。也就是说,在步骤103中得到LOG视频之后,分流一路LOG视频转换为原色视频并保存为录像2,该视频没有经过调色,另外一路LOG视频在步骤1042中基于3D-LUT处理之后保存,即保存一份对应视频风格模板的视频,即保存为录像1,例如在录制过程中,同时编码保存两份视频文件在临时文件夹,基于用户的选择,确定其中一个视频作为最终的视频文件,如果用户希望直接录制生成《B》电影风格的视频文件,则将对应视频风格模板的视频,即将录像1作为最终的视频文件,如果用户希望保存未经调色的视频,则将未经调色的LOG视频转换为标准视频文件并保存为最终的视频文件,即将录像2作为最终的视频文件。需要说明的是,未经调色的视频分流过程可以在通过缩放模块27对视频进行缩放处理的过程之后,在通过 LUT处理模块29执行上述步骤1042的过程之前。
在一种可能的实施方式中,如图10所示,当摄像头193拍摄的视频的捕获帧率为第一帧率时,将与所确定的视频风格模板对应的视频分流为两路,其中一路进行保存,另外一路进行预览。电子设备除了上述摄像头193、Demosaic模块21、变形模块22、融合模块23、噪声处理模块24、CCM模块25、GTM模块26、缩放模块27之外,还包括2D-LUT处理模块30,2D-LUT处理模块30可以为ISP中的硬件模块,上述LUT处理模块29可以为软件模块,例如,在视频录制的过程中,当摄像头193拍摄的视频的捕获帧率为第二帧率时,摄像头193拍摄得到长曝光帧视频图像和短曝光帧视频图像,长曝光帧视频图像所对应的曝光时间大于短曝光帧视频图像所对应的曝光时间,长曝光帧视频图像和短曝光帧视频图像分别通过反马赛克模块21的处理,使图像从RAW域转换为RGB域,之后两路视频图像分别通过变形warp模块22的处理,通过对视频图像的变形实现对齐、防抖的效果,之后两路视频图像通过融合模块23处理,将两种视频图像融合为同一个,融合之后的视频图像通过噪声处理模块24进行去噪处理,然后通过CCM模块25处理,将视频转换为RGB广色域的色彩空间,然后通过GTM模块26执行上述步骤103,通过LOG曲线对视频进行处理,得到LOG视频,然后通过2D-LUT处理模块30执行上述步骤1041、当摄像头拍摄的视频的捕获帧率为第一帧率时,基于所确定的视频风格模板对应的二维2D-LUT对LOG视频进行处理,得到与所确定的视频风格模板对应的视频,然后通过缩放模块27对视频进行缩放处理,处理之后的视频图像分流为两路,其中一路进行保存,另外一路进行预览。基于图9和图10的对比可知,当摄像头193拍摄的视频的捕获帧率为第一帧率时,由于捕获帧率较低,因此可以使用算法更加复杂的3D-LUT对视频进行处理,以实现更好的色彩控制,并且可以使用两路流分别在各模块中进行处理,其中一路进行保存,另外一路进行预览,以使预览画面的效果更好,更加接近所保存的视频。而当摄像头193拍摄的视频的捕获帧率为第二帧率时,由于捕获帧率较高,为了保证在较高帧率下视频处理过程能够完成,可以使用算法较为简单的2D-LUT对视频进行处理,并且仅使用一路流在各模块中进行处理,在处理完成之后再分流为两路,一路进行保存,另外一路进行预览。
以下结合软件架构对本申请实施例进行说明,本申请实施例以分层架构的Android系统为例,示例性说明电子设备100的软件结构。图11是本申请实施例的电子设备100的软件结构框图。
分层架构将软件分成若干个层,每一层都有清晰的角色和分工。层与层之间通过软件接口通信。在一些实施例中,将Android系统分为五层,从上至下分别为应用程序Application层、应用程序框架framework层、系统库library、硬件抽象层(Hardware Abstraction Layer,HAL)以及内核层。
应用程序层可以包括相机等应用程序。
应用程序框架层可以包括相机应用程序编程接口(Application Programming Interface,API)、媒体录制MediaRecorder和表面视图Surfaceview等。媒体录制用来录制视频或图片数据,并使这些数据可以被应用程序访问。表面视图用来显示预览画面。
系统库可以包括多个功能模块。例如:相机服务CameraSevice等。
硬件抽象层用于提供接口支持,例如包括相机流程CameraPipeline以供相机服务调用Call。
内核层是硬件和软件之间的层。内核层包含显示驱动,摄像头驱动等。
结合捕获视频的一种具体场景,HAL上报同时录制两段视频的能力信息,应用程序层下发捕获请求CaptureRequest,请求对应一个录像1的流、一个录像2的流和一个预览流,同时创建两个媒体编解码器mediacodec示例,接收两个录像流的编码。HAL按照上述的数据流dataflow,回调三路流。其中,预览流送显示,两个录像流分别送mediacodec。
本申请实施例提供的录像视频处理方法可以表现为两种拍摄模式下的多个功能,其中这两种拍摄模式可以是指:电影模式、专业模式。
电影模式是一种与电影主题相关的拍摄模式,在该模式下,电子设备100显示的图像能够从感官上给用户一种观看电影的效果,电子设备100还提供多个与电影主题相关的视频风格模板,用户可以利用这些视频风格模板获得色调调整后的图像或视频,这些图像或视频的色调与电影的色调类似或相同。在本申请以下实施例中,电影模式至少可提供用户触发LUT功能、HDR10功能的接口。具体关于LUT功能、HDR10功能的描述可以参见以下实施例。
例如,假设电子设备100为手机,在一种可能的实施方式中,如图3所示,电子设备可以响应用户的操作进入电影模式。例如,电子设备100可以检测到用户作用于相机应用程序的触控操作,响应于该操作,电子设备100显示相机应用程序的默认拍照界面。默认拍照界面可包括:预览框、拍摄模式列表、图库快捷键、快门控件等。其中:
预览框可用于显示摄像头193实时采集的图像。电子设备100可以实时刷新其中的显示内容,以便于用户预览摄像头193当前采集的图像。
拍摄模式列表中可以显示有一个或多个拍摄模式选项。这一个或多个拍摄模式选项可以包括:人像模式选项、录像模式选项、拍照模式选项、电影模式选项、专业选项。这一个或多个拍摄模式选项在界面上可以表现为文字信息,例如“人像”、“录像”、“拍照”、“电影”、“专业”。不限于此,这一个或多个拍摄模式选项在界面上还可以表现为图标或者其他形式的交互元素(interactive element,IE)。
图库快捷键可用于开启图库应用程序。图库应用程序是智能手机、平板电脑等电子设备上的一款图片管理的应用程序,又可以称为“相册”,本实施例对该应用程序的名称不做限制。图库应用程序可以支持用户对存储于电子设备100上的图片进行各种操作,例如浏览、编辑、删除、选择等操作。
快门控件可用于监听触发拍照的用户操作。电子设备100可以检测到作用于快门控件的用户操作,响应于该操作,电子设备100可以将预览框中的图像保存为图库应用程序中的图片。另外,电子设备100还可以在图库快捷键中显示所保存的图像的缩略图。也即是说,用户可以点击快门控件来触发拍照。其中,快门控件可以是按钮或者其他形式的控件。
电子设备100可以检测到用户作用于电影模式选项的触控操作,响应于该操作, 电子设备显示如图3所示的用户界面。
在一些实施例中,电子设备100可以在启动相机应用程序后默认开启电影模式。不限于此,电子设备100还可以通过其他方式开启电影模式,例如电子设备100还可以根据用户的语音指令开启电影模式,本申请实施例对此不作限制。
电子设备100可以检测到用户作用于电影模式选项的触控操作,响应于该操作,电子设备显示如图3所示的用户界面。
如图3示出的用户界面中包括功能选项,功能选项包括HDR10选项、闪光灯选项、LUT选项、设置选项。这多个功能选项都可以检测到用户的触控操作,并响应于该操作,开启或关闭对应的拍摄功能,例如,HDR10功能、闪光灯功能、LUT功能、设置功能。
电子设备可以开启LUT功能,该LUT功能可以改变预览图像的显示效果。实质上,LUT功能引入了颜色查找表,颜色查找表相当于一个颜色转换模型,该颜色转换模型能够根据输入的色彩值,输出调整后的色彩值。摄像头采集的图像的色彩值相当于输入值,不同的色彩值经过颜色转换模型后,都可以对应得到一个输出值。最终,显示在预览框中的图像即为经过颜色转换模型调整后的图像。电子设备100利用该LUT功能,显示经过颜色转换模型调整后的色彩值组成的图像,达到调整图像色调的效果。开启LUT功能之后,电子设备100可以提供多个视频风格模板,一个视频风格模板对应一个颜色转换模型,不同的视频风格模板可以给预览图像带来不同的显示效果。并且,这些视频风格模板可以与电影主题相关联,视频风格模板给预览图像带来的色调调整效果可以和电影中的色调接近或相同,为用户营造拍摄电影的氛围感。
另外,在电子设备100开启LUT功能之后,电子设备100可以根据当前预览视频画面,在多个视频风格模板中确定一个视频风格模板,所确定的视频风格模板可以显示在界面中,以便于用户了解当前所确定的视频风格模板,例如多个视频风格模板包括《A》电影风格模板、《B》电影风格模板和《C》电影风格模板,不同的电影风格模板所对应的LUT可以是预先基于对应电影配色风格所生成的,LUT的颜色转换具有对应电影所具有的风格特点。可以预先从电影风格中提取,产生适合移动电子设备的LUT。LUT功能的开启会改变预览视频画面的色调。如图3中示意的,电子设备100确定《A》电影风格模板并进行显示。
在一些实施例中,电子设备100可以根据用户的滑动操作来选择视频风格模板。具体地,当电子设备100检测到用户开启LUT功能的用户操作,显示LUT预览窗口之后,电子设备100可以默认选择位于LUT预览窗口中的第一个视频风格模板,作为电子设备100选中的视频风格模板。之后,电子设备100可以检测到用户作用于LUT预览窗口的左右滑动操作,移动LUT预览窗口中各视频风格模板的位置,当电子设备100不再检测到用户的滑动操作时,电子设备100将LUT预览窗口中显示的第一个视频风格模板作为电子设备100选中的视频风格模板。
在一些实施例中,电子设备100除了可以使用视频风格模板改变预览图像的显示效果,还可以在添加视频风格模板之后,检测到开始录制视频的用户操作,响应于该操作,电子设备100开始录制视频,从而获得使用视频风格模板调整显示效果后的视频。另外,在录制视频的过程中,电子设备100还可以检测到拍摄照片的用户操作, 响应于该操作,电子设备100将预览框中添加了视频风格模板的预览图像保存成图片,从而获得使用视频风格模板调整显示效果后的图像。
电子设备可以开启HDR10功能,HDR10模式中,HDR即为高动态范围图像(High-Dynamic Range,HDR),相比于普通的图像,HDR可以提供更多的动态范围和图像细节,能够更好地反映出真实环境中的视觉效果,HDR10中的10即为10比特,HDR10可以以10位高动态范围录制视频。
电子设备100可以检测到用户作用于专业模式选项的触控操作,进入专业模式。如图12所示,电子设备处于专业模式时,用户界面中可以包括的功能选项例如为:LOG选项、闪光灯选项、LUT选项、设置选项,另外,用户界面还包括参数调节选项,例如为:测光M选项、ISO选项、快门S选项、曝光补偿EV选项、对焦方式AF选项和白平衡WB选项。
在一些实施例中,电子设备100可以在启动相机应用程序后默认开启专业模式。不限于此,电子设备100还可以通过其他方式开启专业模式,例如电子设备100还可以根据用户的语音指令开启专业模式,本申请实施例对此不作限制。
电子设备100可以检测到用户作用于LOG选项的用户操作,响应于该操作,电子设备100开启LOG功能。其中,LOG功能能够将对数函数应用到曝光曲线上,最大限度地保留摄像头采集的图像中,高光和阴影部分的细节,使最终呈现出来的预览图像的饱和度较低。其中,使用LOG功能录制的视频称为LOG视频。
电子设备100通过专业模式除了可以录制添加了视频风格模板的视频,还可以在录制未添加视频风格模板的视频后,为该视频添加视频风格模板,或者,在开启LOG功能后,录制LOG视频,之后再为该LOG视频添加视频风格模板。这样,电子设备100不仅可以在录制视频的之前调整画面的显示效果,还可以在视频录制完成之后,调整录制的视频的显示效果,增加了图像调整的灵活性和自由度。
本申请实施例还提供一种视频处理装置,包括:视频风格确定模块,用于在多个视频风格模板中确定一个视频风格模板,每个视频风格模板对应一个预设的三维颜色查找表LUT;视频获取模块,用于获取通过摄像头拍摄的视频;第一处理模块,用于通过摄像头当前的感光度ISO所对应的对数LOG曲线对通过摄像头拍摄的视频进行处理,得到LOG视频;第二处理模块,用于当摄像头拍摄的视频的捕获帧率为第一帧率时,基于所确定的视频风格模板对应的二维2D-LUT对LOG视频进行处理,得到与所确定的视频风格模板对应的视频;第三处理模块,用于当摄像头拍摄的视频的捕获帧率为第二帧率时,基于所确定的视频风格模板对应的三维3D-LUT对LOG视频进行处理,得到与所确定的视频风格模板对应的视频,第一帧率大于所述第二帧率。
该视频处理装置可以应用上述的视频处理方法,具体过程和原理在此不再赘述,其中,第一处理模块具体可以为上述实施例中的GTM模块26,第二处理模块具体可以为上述实施例中的2D-LUT处理模块30,第三处理模块具体可以为上述实施例中的LUT处理模块29。
在一种可能的实施方式中,第三处理模块具体用于:基于LUT建立立方体插值空间,LUT为三维3D-LUT;确定LOG视频中每个像素点在立方体插值空间中所属的立方体,立方体中被划分为6个四面体;确定LOG视频中每个像素点所属的四面体; 对于对应立方体顶点的像素点,将像素值转换为经过LUT处理后的像素值;对于不对应立方体顶点的像素点,根据每个像素点所属的四面体进行插值,将像素值转换为经过LUT处理后的像素值。
在一种可能的实施方式中,第二处理模块具体用于,当摄像头拍摄的视频的捕获帧率为第二帧率时,基于所确定的视频风格模板对应的三维3D-LUT对LOG视频进行处理,得到与所确定的视频风格模板对应的视频,第一帧率大于第二帧率。
在一种可能的实施方式中,视频处理装置还包括分流处理模块,用于当摄像头拍摄的视频的捕获帧率为第一帧率时,将与所确定的视频风格模板对应的视频分流为两路,其中一路进行保存,另外一路进行预览。
在一种可能的实施方式中,第一帧率为120每秒传输帧数FPS或60FPS,第二帧率为30FPS。
在一种可能的实施方式中,视频处理装置还包括:转换模块,用于将LOG视频由RGB色彩空间的LOG视频转换为YUV色彩空间的LOG视频,该转换模块具体可以位于图9中的缩放模块27和YUV去噪模块28之间,图9中未示出转换模块;YUV去噪模块28,用于对YUV色彩空间的LOG视频进行YUV去噪处理,得到去噪后的LOG视频。
在一种可能的实施方式中,视频处理装置还包括:第一转换模块,用于将去噪后的LOG视频由YUV色彩空间的LOG视频转换为RGB色彩空间的LOG视频,该第一转换模块可以位于图9中的YUV去噪模块28和LUT处理模块29之间;第二转换模块,用于将RGB色彩空间的与所确定的视频风格模板对应的视频转换为YUV色彩空间的视频,该第二转换模块可以位于图9中的LUT处理模块29和保存录像1的过程之间,图9中未示出第一转换模块和第二转换模块。
在一种可能的实施方式中,视频处理装置还包括:第一保存模块,用于保存与所确定的视频风格模板对应的视频;第二保存模块,用于保存LOG视频。
在一种可能的实施方式中,视频处理装置还包括:第一保存模块,用于保存与所确定的视频风格模板对应的视频;备份转换模块,用于将LOG视频转换为Rec.709色彩标准的视频;第二保存模块,用于保存Rec.709色彩标准的视频。
在一种可能的实施方式中,在第一视频处理流程中执行通过对数LOG曲线对通过摄像头拍摄的视频进行处理,得到LOG视频的过程、以及基于所确定的视频风格模板对应的LUT对LOG视频进行处理,得到与所确定的视频风格模板对应的视频的过程;视频处理方法还包括第二视频处理流程,第二视频处理流程包括:通过对数LOG曲线对通过摄像头拍摄的视频进行处理,得到LOG视频的过程;基于所确定的视频风格模板对应的LUT对LOG视频进行处理,得到与所确定的视频风格模板对应的视频的过程;视频处理方法还包括:第一保存模块,用于将第一视频处理流程中与所确定的视频风格模板对应的视频进行保存;预览模块,用于基于第二视频处理流程中与所确定的视频风格模板对应的视频进行预览。
在一种可能的实施方式中,立方体中具有第0至第7顶点,第0顶点至第1顶点的方向为蓝色B通道的坐标轴方向,第0顶点至第4顶点的方向为红色R通道的坐标轴方向,第0顶点至第2顶点的方向为绿色G通道的坐标轴方向,第0顶点、第1顶 点、第2顶点和第3顶点位于同一平面,第1顶点、第3顶点、第5顶点和第7顶点位于同一平面,第4顶点、第5顶点、第6顶点和第7顶点位于同一平面,第0顶点、第2顶点、第4顶点和第6顶点位于同一平面;第0顶点、第1顶点、第5顶点和第7顶点形成第一个四面体,第0顶点、第1顶点、第3顶点和第7顶点形成第二个四面体,第0顶点、第2顶点、第3顶点和第7顶点形成第三个四面体,第0顶点、第4顶点、第5顶点和第7顶点形成第四个四面体,第0顶点、第4顶点、第6顶点和第7顶点形成第五个四面体,第0顶点、第2顶点、第6顶点和第7顶点形成第六个四面体;第一处理模块具体用于,根据当前像素点(R,G,B)生成经过所述LUT处理后的E通道像素值VE(R,G,B),E取R、G和B;VE(R,G,B)=VE(R0,G0,B0)+(delta_valueR_E×deltaR+delta_valueG_E×deltaG+delta_valueB_E×deltaB+(step_size>>1))/(step_size);VE(R0,G0,B0)为第0顶点(R0,G0,B0)经过LUT处理后的E通道像素值,E取R、G和B;delta_valueR为当前像素点所属四面体对应R通道的坐标轴方向上的两个顶点经过所述LUT处理后的R通道像素值之差,delta_valueG为当前像素点所属四面体对应G通道的坐标轴方向上的两个顶点经过所述LUT处理后的G通道像素值之差,delta_valueB为当前像素点所属四面体对应B通道的坐标轴方向上的两个顶点经过所述LUT处理后的B通道像素值之差;deltaR为当前像素点(R,G,B)中的R值与第0顶点(R0,G0,B0)中的R0值之差,deltaG为当前像素点(R,G,B)中的G值与第0顶点(R0,G0,B0)中的G0值之差,deltaB为当前像素点(R,G,B)中的B值与第0顶点(R0,G0,B0)中的B0值之差;step_size为立方体的边长。
应理解以上视频处理装置的各个模块的划分仅仅是一种逻辑功能的划分,实际实现时可以全部或部分集成到一个物理实体上,也可以物理上分开。且这些模块可以全部以软件通过处理元件调用的形式实现;也可以全部以硬件的形式实现;还可以部分模块以软件通过处理元件调用的形式实现,部分模块通过硬件的形式实现。例如,视频风格确定模块、视频获取模块、第一处理模块、第二处理模块和第三处理模块中的任意一者可以为单独设立的处理元件,也可以集成在视频处理装置中,例如集成在视频处理装置的某一个芯片中实现,此外,也可以以程序的形式存储于视频处理装置的存储器中,由视频处理装置的某一个处理元件调用并执行以上各个模块的功能。其它模块的实现与之类似。此外这些模块全部或部分可以集成在一起,也可以独立实现。这里所述的处理元件可以是一种集成电路,具有信号的处理能力。在实现过程中,上述方法的各步骤或以上各个模块可以通过处理器元件中的硬件的集成逻辑电路或者软件形式的指令完成。
例如,视频风格确定模块、视频获取模块、第一处理模块、第二处理模块和第三处理模块这些模块可以是被配置成实施以上方法的一个或多个集成电路,例如:一个或多个特定集成电路(Application Specific Integrated Circuit,ASIC),或,一个或多个微处理器(digital singnal processor,DSP),或,一个或者多个现场可编程门阵列(Field Programmable Gate Array,FPGA)等。再如,当以上某个模块通过处理元件调度程序的形式实现时,该处理元件可以是通用处理器,例如中央处理器(Central Processing Unit,CPU)或其它可以调用程序的处理器。再如,这些模块可以集成在一 起,以片上系统(system-on-a-chip,SOC)的形式实现。
本申请实施例还提供一种视频处理装置,包括:处理器和存储器,存储器用于存储至少一条指令,指令由处理器加载并执行时以实现上述任意实施例中的视频处理方法。
该视频处理装置可以应用上述的视频处理方法,具体过程和原理在此不再赘述。
处理器的数量可以为一个或多个,处理器和存储器可以通过总线或者其他方式连接。存储器作为一种非暂态计算机可读存储介质,可用于存储非暂态软件程序、非暂态计算机可执行程序以及模块,如本申请实施例中的视频处理装置对应的程序指令/模块。处理器通过运行存储在存储器中的非暂态软件程序、指令以及模块,从而执行各种功能应用以及数据处理,即实现上述任意方法实施例中的方法。存储器可以包括存储程序区和存储数据区,其中,存储程序区可存储操作系统、至少一个功能所需要的应用程序;以及必要数据等。此外,存储器可以包括高速随机存取存储器,还可以包括非暂态存储器,例如至少一个磁盘存储器件、闪存器件、或其他非暂态固态存储器件。
如图1所示,本申请实施例还提供一种电子设备,包括:摄像头193和上述的视频处理装置,视频处理装置包括处理器110。
视频处理装置的具体原理和工作过程与上述实施例相同,在此不再赘述。该电子设备可以是例如手机、电视、平板电脑、手表、手环等任何具有视频拍摄功能的产品或部件。
本申请实施例还提供一种计算机可读存储介质,计算机可读存储介质中存储有计算机程序,当其在计算机上运行时,使得计算机执行上述任意实施例中的视频处理方法。
在上述实施例中,可以全部或部分地通过软件、硬件、固件或者其任意组合来实现。当使用软件实现时,可以全部或部分地以计算机程序产品的形式实现。所述计算机程序产品包括一个或多个计算机指令。在计算机上加载和执行所述计算机程序指令时,全部或部分地产生按照本申请所述的流程或功能。所述计算机可以是通用计算机、专用计算机、计算机网络、或者其他可编程装置。所述计算机指令可以存储在计算机可读存储介质中,或者从一个计算机可读存储介质向另一个计算机可读存储介质传输,例如,所述计算机指令可以从一个网站站点、计算机、服务器或数据中心通过有线(例如同轴电缆、光纤、数字用户线)或无线(例如红外、无线、微波等)方式向另一个网站站点、计算机、服务器或数据中心进行传输。所述计算机可读存储介质可以是计算机能够存取的任何可用介质或者是包含一个或多个可用介质集成的服务器、数据中心等数据存储设备。所述可用介质可以是磁性介质,(例如,软盘、硬盘、磁带)、光介质(例如,DVD)、或者半导体介质(例如固态硬盘Solid State Disk)等。
本申请实施例中,“至少一个”是指一个或者多个,“多个”是指两个或两个以上。“和/或”,描述关联对象的关联关系,表示可以存在三种关系,例如,A和/或B,可以表示单独存在A、同时存在A和B、单独存在B的情况。其中A,B可以是单数或者复数。字符“/”一般表示前后关联对象是一种“或”的关系。“以下至少一项”及其类似表达,是指的这些项中的任意组合,包括单项或复数项的任意组合。例如, a,b和c中的至少一项可以表示:a,b,c,a-b,a-c,b-c,或a-b-c,其中a,b,c可以是单个,也可以是多个。
以上仅为本申请的优选实施例而已,并不用于限制本申请,对于本领域的技术人员来说,本申请可以有各种更改和变化。凡在本申请的精神和原则之内,所作的任何修改、等同替换、改进等,均应包含在本申请的保护范围之内。

Claims (10)

  1. 一种视频处理方法,其特征在于,包括:
    在多个视频风格模板中确定一个视频风格模板,每个视频风格模板对应一个预设的颜色查找表LUT;
    获取通过摄像头拍摄的视频;
    通过所述摄像头当前的感光度ISO所对应的对数LOG曲线对所述通过摄像头拍摄的视频进行处理,得到LOG视频;
    当所述摄像头拍摄的视频的捕获帧率为第一帧率时,基于所确定的视频风格模板对应的二维2D-LUT对所述LOG视频进行处理,得到与所确定的视频风格模板对应的视频;
    当所述摄像头拍摄的视频的捕获帧率为第二帧率时,基于所确定的视频风格模板对应的三维3D-LUT对所述LOG视频进行处理,得到与所确定的视频风格模板对应的视频,所述第一帧率大于所述第二帧率。
  2. 根据权利要求1所述的视频处理方法,其特征在于,
    当所述摄像头拍摄的视频的捕获帧率为第一帧率时,基于所确定的视频风格模板对应的二维2D-LUT对所述LOG视频进行处理,得到与所确定的视频风格模板对应的视频的过程在HSV色彩空间执行。
  3. 根据权利要求1所述的视频处理方法,其特征在于,
    当所述摄像头拍摄的视频的捕获帧率为第二帧率时,基于所确定的视频风格模板对应的三维3D-LUT对所述LOG视频进行处理,得到与所确定的视频风格模板对应的视频的过程包括:
    基于所述LUT建立立方体插值空间;
    确定所述LOG视频中每个像素点在所述立方体插值空间中所属的立方体,所述立方体中被划分为6个四面体;
    确定所述LOG视频中每个像素点所属的四面体;
    对于对应立方体顶点的像素点,将像素值转换为经过所述LUT处理后的像素值;
    对于不对应立方体顶点的像素点,根据每个像素点所属的四面体进行插值,将像素值转换为经过所述LUT处理后的像素值。
  4. 根据权利要求1所述的视频处理方法,其特征在于,
    在所述当所述摄像头拍摄的视频的捕获帧率为第二帧率时,基于所确定的视频风格模板对应的三维3D-LUT对所述LOG视频进行处理,得到与所确定的视频风格模板对应的视频的过程之前,还包括:
    将所述LOG视频由RGB色彩空间的LOG视频转换为YUV色彩空间的LOG视频;
    对所述YUV色彩空间的LOG视频进行YUV去噪处理,得到去噪后的LOG视频。
  5. 根据权利要求4所述的视频处理方法,其特征在于,
    在所述当所述摄像头拍摄的视频的捕获帧率为第二帧率时,基于所确定的视频风格模板对应的三维3D-LUT对所述LOG视频进行处理,得到与所确定的视频风格模板 对应的视频的过程之前,还包括:
    将所述去噪后的LOG视频由YUV色彩空间的LOG视频转换为RGB色彩空间的LOG视频;
    在所述基于所确定的视频风格模板对应的LUT对所述LOG视频进行处理,得到与所确定的视频风格模板对应的视频的过程之后,还包括:
    将RGB色彩空间的与所确定的视频风格模板对应的视频转换为YUV色彩空间的视频。
  6. 根据权利要求1所述的视频处理方法,其特征在于,还包括:
    当所述摄像头拍摄的视频的捕获帧率为第一帧率时,将与所确定的视频风格模板对应的视频分流为两路,其中一路进行保存,另外一路进行预览。
  7. 根据权利要求1所述的视频处理方法,其特征在于,
    所述第一帧率为120每秒传输帧数FPS或60FPS,所述第二帧率为30FPS。
  8. 一种视频处理装置,其特征在于,包括:
    处理器和存储器,所述存储器用于存储至少一条指令,所述指令由所述处理器加载并执行时以实现如权利要求1至7中任意一项所述的视频处理方法。
  9. 一种电子设备,其特征在于,包括:
    摄像头;
    如权利要求8所述的视频处理装置。
  10. 一种计算机可读存储介质,其特征在于,所述计算机可读存储介质中存储有计算机程序,当其在计算机上运行时,使得计算机执行如权利要求1至7中任意一项所述的视频处理方法。
PCT/CN2022/094744 2021-08-12 2022-05-24 视频处理方法、装置、电子设备和存储介质 WO2023016038A1 (zh)

Priority Applications (2)

Application Number Priority Date Filing Date Title
US18/274,343 US20240080405A1 (en) 2021-08-12 2022-05-24 Video processing method and apparatus, electronic device, and storage medium
EP22855024.0A EP4266675A1 (en) 2021-08-12 2022-05-24 Video processing method and apparatus, electronic device, and storage medium

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN202110926600.0A CN115706764B (zh) 2021-08-12 2021-08-12 视频处理方法、装置、电子设备和存储介质
CN202110926600.0 2021-08-12

Publications (1)

Publication Number Publication Date
WO2023016038A1 true WO2023016038A1 (zh) 2023-02-16

Family

ID=85180980

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2022/094744 WO2023016038A1 (zh) 2021-08-12 2022-05-24 视频处理方法、装置、电子设备和存储介质

Country Status (4)

Country Link
US (1) US20240080405A1 (zh)
EP (1) EP4266675A1 (zh)
CN (1) CN115706764B (zh)
WO (1) WO2023016038A1 (zh)

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101867685A (zh) * 2010-06-25 2010-10-20 北京理工大学 基于颜色查找表的双波段视频快速自然感彩色融合方法
US20110058793A1 (en) * 2009-09-10 2011-03-10 Greg Mullins Video Format for Digital Video Recorder
CN107077828A (zh) * 2014-11-25 2017-08-18 英特尔公司 对颜色查找表的大小进行压缩
CN111009027A (zh) * 2019-12-27 2020-04-14 南京奥视威电子科技股份有限公司 一种利用3d lut实现色彩校正的方法和显示设备
CN111510698A (zh) * 2020-04-23 2020-08-07 惠州Tcl移动通信有限公司 图像处理方法、装置、存储介质及移动终端
CN111751982A (zh) * 2019-03-29 2020-10-09 成都理想境界科技有限公司 一种扫描显示方法及装置

Family Cites Families (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR101544843B1 (ko) * 2008-07-28 2015-08-18 삼성디스플레이 주식회사 표시 장치와 그 구동 방법
US10462439B2 (en) * 2015-10-02 2019-10-29 Vid Scale, Inc. Color correction with a lookup table
CN111050211B (zh) * 2019-12-13 2021-10-26 广州酷狗计算机科技有限公司 视频处理方法、装置及存储介质

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20110058793A1 (en) * 2009-09-10 2011-03-10 Greg Mullins Video Format for Digital Video Recorder
CN101867685A (zh) * 2010-06-25 2010-10-20 北京理工大学 基于颜色查找表的双波段视频快速自然感彩色融合方法
CN107077828A (zh) * 2014-11-25 2017-08-18 英特尔公司 对颜色查找表的大小进行压缩
CN111751982A (zh) * 2019-03-29 2020-10-09 成都理想境界科技有限公司 一种扫描显示方法及装置
CN111009027A (zh) * 2019-12-27 2020-04-14 南京奥视威电子科技股份有限公司 一种利用3d lut实现色彩校正的方法和显示设备
CN111510698A (zh) * 2020-04-23 2020-08-07 惠州Tcl移动通信有限公司 图像处理方法、装置、存储介质及移动终端

Also Published As

Publication number Publication date
EP4266675A1 (en) 2023-10-25
CN115706764B (zh) 2023-09-19
CN115706764A (zh) 2023-02-17
US20240080405A1 (en) 2024-03-07

Similar Documents

Publication Publication Date Title
WO2023016039A1 (zh) 视频处理方法、装置、电子设备和存储介质
WO2023016035A1 (zh) 视频处理方法、装置、电子设备和存储介质
WO2023016037A1 (zh) 视频处理方法、装置、电子设备和存储介质
US11138765B2 (en) Non-linear color correction
CN113824914B (zh) 视频处理方法、装置、电子设备和存储介质
US10600170B2 (en) Method and device for producing a digital image
CN114449199B (zh) 视频处理方法、装置、电子设备和存储介质
WO2023016040A1 (zh) 视频处理方法、装置、电子设备和存储介质
WO2023016044A1 (zh) 视频处理方法、装置、电子设备和存储介质
WO2023016038A1 (zh) 视频处理方法、装置、电子设备和存储介质
WO2023016041A1 (zh) 视频处理方法、装置、电子设备和存储介质
WO2023016042A1 (zh) 视频处理方法、装置、电子设备和存储介质
WO2023016043A1 (zh) 视频处理方法、装置、电子设备和存储介质
US20230215051A1 (en) Systems, apparatus, and methods for color space representation
CN115706853A (zh) 视频处理方法、装置、电子设备和存储介质

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 22855024

Country of ref document: EP

Kind code of ref document: A1

ENP Entry into the national phase

Ref document number: 2022855024

Country of ref document: EP

Effective date: 20230717

WWE Wipo information: entry into national phase

Ref document number: 18274343

Country of ref document: US

NENP Non-entry into the national phase

Ref country code: DE