WO2023016039A1 - Procédé et appareil de traitement vidéo, dispositif électronique et support de stockage - Google Patents

Procédé et appareil de traitement vidéo, dispositif électronique et support de stockage Download PDF

Info

Publication number
WO2023016039A1
WO2023016039A1 PCT/CN2022/094754 CN2022094754W WO2023016039A1 WO 2023016039 A1 WO2023016039 A1 WO 2023016039A1 CN 2022094754 W CN2022094754 W CN 2022094754W WO 2023016039 A1 WO2023016039 A1 WO 2023016039A1
Authority
WO
WIPO (PCT)
Prior art keywords
video
frame rate
determined
mode
camera
Prior art date
Application number
PCT/CN2022/094754
Other languages
English (en)
Chinese (zh)
Inventor
崔瀚涛
张东
朱登奎
王燕东
郭永利
Original Assignee
荣耀终端有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 荣耀终端有限公司 filed Critical 荣耀终端有限公司
Publication of WO2023016039A1 publication Critical patent/WO2023016039A1/fr

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N5/00Details of television systems
    • H04N5/76Television signal recording
    • H04N5/91Television signal processing therefor

Definitions

  • the present application relates to the technical field of video shooting, and in particular to a video processing method, device, electronic equipment and storage medium.
  • a video processing method, device, electronic equipment, and storage medium, which can make videos captured by electronic equipment have different style effects based on the characteristics of LUTs, so as to meet higher color matching requirements.
  • a video processing method including: determining a slow-motion mode among multiple slow-motion modes, the determined slow-motion mode corresponds to a capture frame rate and an encoding frame rate, and the capture frame rate is greater than the encoding frame rate;
  • a video style template is determined among the multiple video style templates of the determined upgrade mode, and each video style template corresponds to a preset two-dimensional color lookup table 2D-LUT; based on the determined capture frame rate corresponding to the upgrade mode, it is obtained through the camera The captured video; the logarithmic LOG curve corresponding to the current sensitivity ISO of the camera is used to process the video captured by the camera to obtain the LOG video; the LOG video is processed based on the 2D-LUT corresponding to the determined video style template to obtain The video corresponding to the determined video style template; encoding and saving the video corresponding to the determined video style template based on the determined encoding frame rate corresponding to the upscaling mode.
  • the multiple upscaling modes include a first upscaling mode and a second upscaling mode, the first upscaling mode corresponds to the first capture frame rate and the first encoding frame rate, and the second upscaling mode corresponds to the second capture frame rate and the second encoding frame rate, the first capturing frame rate is less than the second capturing frame rate, the first encoding frame rate is equal to the second encoding frame rate, and the second capturing frame rate is greater than the second encoding frame rate.
  • the video captured by the camera is processed through the logarithmic LOG curve corresponding to the current sensitivity ISO of the camera to obtain the LOG video
  • it also includes: performing electronic anti-shake processing on the video captured by the camera; if the determined upscaling mode is the first upscaling mode, the logarithmic LOG curve corresponding to the current sensitivity ISO of the camera is used for the video captured by the camera.
  • the process of processing to obtain the LOG video is as follows: the video after electronic anti-shake processing is processed through the logarithmic LOG curve corresponding to the current sensitivity ISO of the camera to obtain the LOG video.
  • the electronic anti-shake function is omitted, so as to realize video content in a limited frame interval time.
  • an electronic anti-shake function is added to achieve better video processing effects.
  • the process of encoding and saving the video corresponding to the determined video style template based on the determined encoding frame rate corresponding to the upscaling mode includes:
  • the video stream is divided into two streams, one of which is encoded and saved based on the encoding frame rate corresponding to the determined upscaling mode, and the other stream is previewed.
  • the preview video and the final video can have the same visual effect, which is convenient for users to directly preview the video based on the color-graded style.
  • a video processing device including: a processor and a memory, the memory is used to store at least one instruction, and when the instruction is loaded and executed by the processor, the above video processing method is implemented.
  • an electronic device including: a camera; and the above-mentioned video processing device.
  • a computer-readable storage medium In a fourth aspect, a computer-readable storage medium is provided.
  • a computer program is stored in the computer-readable storage medium, and when running on a computer, the computer is made to execute the above video processing method.
  • the video processing method, device, electronic equipment, and storage medium in the embodiments of the present application use the LUT technology in the film industry to process the LOG video based on the LUT corresponding to the determined video style template during the video recording process, so that the recorded
  • the video has the style effect corresponding to the determined video style template, so as to meet the higher color grading requirements and make the recorded video have a movie feel.
  • a video with an upscaling effect can be obtained in a relatively simple way.
  • 2D-LUT is used to process the video to achieve limited LUT processing is performed within the frame interval.
  • FIG. 1 is a structural block diagram of an electronic device in an embodiment of the present application
  • FIG. 2 is a flowchart of a video processing method in an embodiment of the present application
  • FIG. 3 is a schematic diagram of a user interface in a movie mode in an embodiment of the present application.
  • FIG. 4 is a schematic diagram of a video recording interface in an embodiment of the present application.
  • Fig. 5 is the schematic diagram of a kind of LOG curve in the embodiment of the present application.
  • FIG. 6 is a schematic diagram of a comparison of the playing time of captured video and encoded video file in the embodiment of the present application.
  • FIG. 7 is a flowchart of a video processing method in an embodiment of the present application.
  • FIG. 8 is a structural block diagram corresponding to the execution process in the second upgrade mode in the embodiment of the present application.
  • FIG. 9 is a structural block diagram corresponding to an execution process in the first upgrade mode in the embodiment of the present application.
  • FIG. 10 is a software structural block diagram of an electronic device in an embodiment of the present application.
  • FIG. 11 is a schematic diagram of a user interface in a professional mode in an embodiment of the present application.
  • the electronic device 100 may include a processor 110, a camera 193, a display screen 194, and the like. It can be understood that, the structure illustrated in the embodiment of the present invention does not constitute a specific limitation on the electronic device 100 . In other embodiments of the present application, the electronic device 100 may include more or fewer components than shown in the figure, or combine certain components, or separate certain components, or arrange different components. The illustrated components can be realized in hardware, software or a combination of software and hardware.
  • the processor 110 may include one or more processing units, for example: the processor 110 may include a graphics processing unit (graphics processing unit, GPU), an image signal processor (image signal processor, ISP), a controller, a video codec, Digital signal processor (digital signal processor, DSP), etc. Wherein, different processing units may be independent devices, or may be integrated in one or more processors.
  • the controller can generate an operation control signal according to the instruction opcode and timing signal, and complete the control of fetching and executing the instruction.
  • a memory may also be provided in the processor 110 for storing instructions and data.
  • the electronic device 100 realizes the display function through the GPU, the display screen 194 , and the application processor.
  • the GPU is a microprocessor for image processing, and is connected to the display screen 194 and the application processor. GPUs are used to perform mathematical and geometric calculations for graphics rendering.
  • Processor 110 may include one or more GPUs that execute program instructions to generate or change display information.
  • the electronic device 100 can realize the shooting function through the ISP, the camera 193 , the video codec, the GPU, the display screen 194 and the application processor.
  • the ISP is used for processing the data fed back by the camera 193 .
  • the light is transmitted to the photosensitive element of the camera through the lens, and the light signal is converted into an electrical signal, and the photosensitive element of the camera transmits the electrical signal to the ISP for processing, and converts it into an image visible to the naked eye.
  • ISP can also perform algorithm optimization on image noise, brightness, and skin color.
  • ISP can also optimize the exposure, color temperature and other parameters of the shooting scene.
  • the ISP may be located in the camera 193 .
  • Camera 193 is used to capture still images or video.
  • the object generates an optical image through the lens and projects it to the photosensitive element.
  • the photosensitive element may be a charge coupled device (CCD) or a complementary metal-oxide-semiconductor (CMOS) phototransistor.
  • CMOS complementary metal-oxide-semiconductor
  • the photosensitive element converts the light signal into an electrical signal, and then transmits the electrical signal to the ISP to convert it into a digital image signal.
  • the ISP outputs the digital image signal to the DSP for processing.
  • DSP converts digital image signals into standard RGB, YUV and other image signals.
  • the electronic device 100 may include 1 or N cameras 193 , where N is a positive integer greater than 1.
  • Digital signal processors are used to process digital signals. In addition to digital image signals, they can also process other digital signals. For example, when the electronic device 100 selects a frequency point, the digital signal processor is used to perform Fourier transform on the energy of the frequency point.
  • Video codecs are used to compress or decompress digital video.
  • the electronic device 100 may support one or more video codecs.
  • the electronic device 100 can play or record videos in various encoding formats, for example: moving picture experts group (moving picture experts group, MPEG) 1, MPEG2, MPEG3, MPEG4 and so on.
  • MPEG moving picture experts group
  • the embodiment of the present application provides a video processing method.
  • the video processing method may be executed by a processor 110, specifically an ISP or a combination of an ISP and other processors.
  • the video processing method includes:
  • Step 100 determine a step-up mode in a plurality of step-up modes, the determined step-up mode corresponds to a capture frame rate and an encoding encoder frame rate, and the capture frame rate is greater than the encoding frame rate;
  • the capture frame rate refers to the readout frame rate of the camera sensor, and the encoding frame rate affects the default playback frame rate of the video.
  • the capture frame rate is greater than the encoding frame rate, a slow motion effect is generated.
  • the video capture frame rate is greater than the playback frame rate, it is upgraded. Upscaling is a technical means in movie shooting, which can achieve different effects. For example, when the video playback frame rate is displayed at 30 frames per second, that is The frame rate is 30 frames per second (Frames Per Second, FPS), and the frame rate during shooting is 60FPS, which is upgraded.
  • the frame rate of the camera during the upgrade is suitable for different scene expressions.
  • 60FPS upgraded shooting is suitable for expressing scenes such as panning, slow walking, laughing, applauding, etc.
  • 120FPS upgraded shooting is suitable for expressing scenes such as running, turning around, pulling hair, and throwing flowers.
  • Boost mode may be entered based on user selection.
  • Step 101 determine a video style template among the plurality of video style templates in the determined upgrade mode, each video style template corresponds to a preset two-dimensional color look-up table (2D-Look Up Table, 2D-LUT);
  • LUT is a mathematical conversion model. Using LUT, one image data value can be output as another image data value, thereby changing the exposure and color of the picture. Therefore, LUTs corresponding to different video styles can be pre-generated.
  • a video style template can be determined before the electronic device records a video.
  • the video style template can be determined based on the user's choice, or based on artificial intelligence (Artificial Intelligence, AI),
  • AI Artificial Intelligence
  • the video style template is automatically determined according to the scene corresponding to the image captured by the current camera. For example, assuming that the electronic device is a mobile phone, in a possible implementation, as shown in Figure 3, the user operates the mobile phone to enter the shooting interface, and the shooting interface includes movie mode options.
  • the movie mode interface of including multiple video style template options, for example including "A" movie style template, "B” movie style template and “C” movie style template, only one " A "movie style template, understandably, multiple different movie style templates can be displayed side by side in the user interface, and the LUTs corresponding to different movie style templates can be generated based on the corresponding movie color matching style in advance, and the color conversion of the LUT has Corresponding to the style characteristics of the movie, for example, the color matching style of the movie "A" is complementary color.
  • Complementary color refers to the contrast effect of two corresponding colors. Two colors of warm color and cool color are used to emphasize the contrast to enhance the vividness, For outstanding effects, usually two contrasting colors symbolize conflicting behaviors.
  • the LUT corresponding to the "A" movie style template is ready to use After transforming the colormap, the complementary colors are more pronounced to simulate the color scheme of the "A" movie.
  • the mobile phone when the user operates the mobile phone to enter the movie mode, the mobile phone will obtain the picture taken by the current camera, and based on the AI algorithm, determine the scene corresponding to the picture and determine the scene corresponding to the scene.
  • the recommended video style template for example, if it is recognized that the subject of the currently captured picture is a young female character, the corresponding recommended video style template is determined according to the algorithm as the "C" movie style template, and the movie "C" has a young female character as the theme movie, its corresponding LUT can simulate the color matching style of the movie "C”; It is a movie with city streets as the main scene, and its corresponding LUT can simulate the color matching style of the "B" movie. In this way, a video style template matching the current scene can be automatically recommended for the user. Film styles can be pre-extracted to produce LUTs suitable for mobile electronics.
  • the frame rate of the video captured by the camera is faster, that is, the frame interval is shorter, and the processing speed in the video recording process is required to be faster.
  • the capture frame rate of 60FPS needs to apply LUT to the video
  • the processing time is within 15ms
  • the capture frame rate of 120FPS requires the application of LUT to process the video within 8ms. Therefore, for the non-upscaled mode, the LUT can be a 3D-LUT, while for the upscaled mode, a 2D-LUT needs to be used. Compared with the 2D-LUT, the 3D-LUT can achieve more accurate color control.
  • Step 102 based on the determined capture frame rate corresponding to the upscaling mode, the video captured by the camera is acquired;
  • the mobile phone After determining the upscaling mode and video style template, if the user clicks on the shooting option, the mobile phone starts to acquire the video captured by the camera based on the determined capture frequency corresponding to the upscaling mode.
  • the capture frame rate is an integer multiple of the encoding frame rate to avoid the judder effect, as shown in Figure 4.
  • the number of "X” can be used to indicate the frame rate in the slow-motion mode Multiples, assuming that the encoding frame rate is 30FPS, two “X” indicate that the capture frame rate is 60FPS, that is, the capture frame rate is twice the encoding frame rate, and four “X” indicate that the capture frame rate is 120FPS, that is, the capture frame rate is Four times the encoding frame rate, an "X” indicates that the capture frame rate is 30FPS, and it is the non-upgraded mode at this time.
  • the embodiment of this application only involves the relevant content of the upgraded mode, and does not introduce the relevant content of the non-upgraded mode;
  • Step 103 process the video captured by the camera through the logarithm (Logarithm, LOG) curve corresponding to the current sensitivity ISO of the camera to obtain the LOG video;
  • Figure 5 illustrates a LOG curve, where the abscissa is a linear signal, represented by a 16-bit code value Code Value, and the ordinate is the LOG signal processed by the LOG curve, represented by a 10-bit code value.
  • the signal input of the camera can be used to encode the information in the dark area to the middle tone (as shown in the steep part of the curve in Figure 5), forming a 10-bit signal output, which conforms to the human eye's LOG sensing rule for light, and maximizes the The dark information is preserved, and the LOG video can use the limited bit depth to maximize the details of shadows and highlights.
  • the ASA in Figure 5 is the sensitivity, and different ASAs correspond to different ISOs, and the two belong to different systems.
  • Step 104 process the LOG video based on the two-dimensional 2D-LUT corresponding to the determined video style template, and obtain the video corresponding to the determined video style template;
  • the LOG video is used as an input, and the LUT corresponding to the video style template determined in step 101 is applied to perform mapping conversion processing on the LOG video image.
  • the output can be the video of the Rec.709 color standard, or the video of the High-Dynamic Range (HDR) 10 standard, that is, the LOG video can be processed through the LUT, Convert video to HDR10 standard.
  • HDR High-Dynamic Range
  • Different LUTs are applied to electronic equipment, and related modules in the electronic equipment can be adapted to adapt to different styles of LUTs.
  • the video style template determined in step 101 is a gray tone video style template
  • the gray tone The characteristics of the picture are that the texture in the picture is strong, the saturation is low, there is no more color interference except for the color of the character's skin, and the dark part is cooler. Based on these characteristics, the electronic device can monitor the relevant Adjust the module parameters to keep the texture in the picture, do not do strong denoising and sharpening, properly reduce the saturation of the picture, keep the skin color in the picture true to restore, and adjust the dark part of the picture to cool colors.
  • Step 105 Based on the encoding frame rate corresponding to the determined upscaling mode, the video corresponding to the determined video style template is encoded and stored. After encoding and saving, the video with the upscaling effect can be directly obtained, that is, the video with the slow motion effect. By converting the video separately, a video with an upscaling effect, that is, a video with a slow motion effect can be obtained in a relatively simple way. For example, as shown in Figure 6, assuming that the capture frame rate is 120FPS and the encoding frame rate is 30FPS, in The total video capture time at this frame rate is 1 minute, and the playback time of the encoded video file at this frame rate is 4 minutes.
  • the LUT technology of the film industry is used to process the LOG video based on the LUT corresponding to the determined video style template, so that the recorded video has the determined video style
  • the style effect corresponding to the template can meet the higher color grading requirements and make the recorded video have a cinematic feel.
  • a video with an upscaling effect can be obtained in a relatively simple way.
  • 2D-LUT is used to process the video to achieve limited LUT processing is performed within the frame interval.
  • the multiple upscaling modes include a first upscaling mode and a second upscaling mode, the first upscaling mode corresponds to the first capture frame rate and the first encoding frame rate, and the second upscaling mode corresponds to the second capture frame rate and the second encoding frame rate, the first capturing frame rate is less than the second capturing frame rate, the first encoding frame rate is equal to the second encoding frame rate, and the second capturing frame rate is greater than the second encoding frame rate.
  • the user can select the required upscaling mode based on different scenes. Different upscaling modes correspond to different upscaling frame rates. In different upscaling modes, the encoding frame rate is the same, but the capture frame rate is different.
  • the logarithmic LOG curve corresponding to the current sensitivity ISO of the camera passes through the The video captured by the camera is processed, and before the process of obtaining the LOG video, it also includes: step 106, performing electronic image stabilization (Electric Image Stabilization, EIS) processing on the video captured by the camera; if the determined upgrade mode is the first upgrade mode , above step 103, process the video taken by the camera through the logarithmic LOG curve corresponding to the current sensitivity ISO of the camera, and the process of obtaining the LOG video is: through the logarithmic LOG corresponding to the current sensitivity ISO of the camera The curve processes the video after the electronic anti-shake processing to obtain the LOG video.
  • EIS Electronic Image Stabilization
  • the electronic device may specifically include a camera 193, an anti-mosaic Demosaic module 21, a deformation module 22, a fusion module 23, a noise processing module 24, a color correction matrix (Color Correction Matrix, CCM) module 25, a global tone mapping (Global Tone Mapping , GTM) module 26, scaling Scaler module 27, YUV denoising module 28, 2D-LUT processing module 30 and electronic anti-shake module 31, for example, if the determined slow-motion mode is the second slow-motion mode, in the process of video recording , the camera 193 shoots video at the second capture frame rate, and the camera 193 shoots a long-exposure frame video image and a short-exposure frame video image, and the exposure time corresponding to the long-exposure frame video image is greater than the exposure time corresponding to the short-exposure frame video image, The long-exposure frame video image and the short-exposure frame video image are respectively processed by the anti-mosaic module 21, so that the image is converted from the RAW domain to the RGB domain, and
  • the two-way video image electronic anti-shake module 31 performs electronic anti-shake processing, and then processes through the fusion module 23 to fuse the two video images into the same one, and the video image after fusion passes through the noise processing module 24
  • Carry out denoising processing then process through the CCM module 25, convert the video into the color space of RGB wide color gamut, then execute the above-mentioned step 103 through the GTM module 26, process the video through the LOG curve, obtain the LOG video, and then pass 2D-
  • the LUT processing module 30 executes the above step 104, processes the LOG video based on the two-dimensional 2D-LUT corresponding to the determined video style template, obtains the video corresponding to the determined video style template, and then zooms the video through the scaling module 27 Processing, the processed video image is split into two channels, one of which is saved, and the other is previewed.
  • the camera 193 shoots a video at the first capture frame rate, and the camera 193 captures a long exposure frame video image and a short exposure frame.
  • the image fusion is the same, the video image after the fusion is denoised by the noise processing module 24, then processed by the CCM module 25, the video is converted into the color space of RGB wide color gamut, and then the above-mentioned steps 103 are performed by the GTM module 26,
  • the video is processed through the LOG curve to obtain the LOG video, and then the 2D-LUT processing module 30 executes the above step 104, and the LOG video is processed based on the two-dimensional 2D-LUT corresponding to the determined video style template to obtain the determined LOG video.
  • the video corresponding to the video style template is then scaled by the scaling module 27, and the processed video image is divided into two streams, one of which is saved, and the other is previewed.
  • the electronic anti-shake function is omitted to realize the limited frame interval time
  • the content is used for video processing.
  • an electronic anti-shake function is added to achieve better video processing effects.
  • Bayer field Each lens on a digital camera has a light sensor to measure the brightness of the light, but to obtain a full-color image, generally three light sensors are required to obtain the three primary colors of red, green and blue information, and in order to reduce the cost and volume of digital cameras, manufacturers usually use CCD or CMOS image sensors.
  • CCD or CMOS image sensors usually use CCD or CMOS image sensors.
  • the original image output by CMOS image sensors is in Bayer domain RGB format, and a single pixel contains only one color value. To obtain the gray value of the image, it is necessary to interpolate the complete color information of each pixel, and then calculate the gray value of each pixel.
  • the Bayer domain refers to a raw image format inside a digital camera.
  • the Raw domain or Raw format refers to unprocessed images. Further, the Raw image can be understood as that the photosensitive element of the camera such as Complementary Metal Oxide Semiconductor (Complementary Metal Oxide Semiconductor, CMOS) or Charge-coupled Device (Charge-coupled Device, CCD) converts the captured light source signal into digital The raw data of the signal.
  • CMOS Complementary Metal Oxide Semiconductor
  • CCD Charge-coupled Device
  • a RAW file is a record of the original information of the digital camera sensor, while recording some metadata (Metadata, such as ISO (International Organization for Standardization, International Organization for Standardization) settings, shutter speed, aperture value) generated by the camera. , white balance, etc.) files.
  • the Raw domain is a format that has not been processed by the ISP nonlinearly and has not been compressed.
  • the full name of Raw format is RAW Image Format.
  • YUV is a color encoding method that is often used in various video processing components. YUV takes human perception into account when encoding photos or videos, allowing bandwidth reduction for chroma. YUV is a type of compiling true-color color space (color space). Proper nouns such as Y'UV, YUV, YCbCr, and YPbPr can all be called YUV, and they overlap with each other. Among them, "Y” represents the brightness (Luminance or Luma), that is, the grayscale value, "U” and “V” represent the chroma (Chrominance or Chroma), which are used to describe the color and saturation of the image, and are used to specify the color of the pixel .
  • YUV is divided into two formats, one is: packed formats, which store Y, U, and V values into a Macro Pixels array, which is similar to the storage method of RGB.
  • the other is: planar formats, which store the three components of Y, U, and V in different matrices.
  • Planar formats means that each Y component, U component and V component are organized in an independent plane, that is to say, all U components are behind the Y component, and V components are behind all U components.
  • the first capture frame rate is 60 FPS
  • the second capture frame rate is 120 FPS
  • the first encoding frame rate and the second encoding frame rate are 30 FPS.
  • the above step 104 based on the two-dimensional 2D-LUT corresponding to the determined video style template, processes the LOG video, and the process of obtaining the video corresponding to the determined video style template is in the HSV color space implement.
  • the 2D-LUT is obtained through 3D-LUT simulation in advance, for example, the 3D-LUT is known in advance, and the input data and output data corresponding to the 3D-LUT, the input data and the The output data all belong to the RGB color space.
  • the input data can be converted from the RGB color space to the HSV color space, and the output data can be converted from the RGB color space to the HSV color space.
  • the process of encoding and saving the video corresponding to the determined video style template based on the determined encoding frame rate corresponding to the upscaling mode in the above step 105 includes : Split the video corresponding to the determined video style template into two streams, one of which is encoded and saved based on the encoding frame rate corresponding to the determined upscaling mode, and the other stream is previewed.
  • the 2D-LUT with a relatively simple algorithm can be used to process the video, and only one stream is used for processing in each module. Processing, after the processing is completed, it will be divided into two channels, one for saving, and the other for previewing.
  • FIG. 10 is a block diagram of the software structure of the electronic device 100 according to the embodiment of the present application.
  • the layered architecture divides the software into several layers, and each layer has a clear role and division of labor. Layers communicate through software interfaces.
  • the Android system is divided into five layers, which are, from top to bottom, the Application layer, the application framework framework layer, the system library library, the Hardware Abstraction Layer (Hardware Abstraction Layer, HAL) and the kernel layer.
  • the application layer can include applications such as cameras.
  • the application framework layer may include camera application programming interface (Application Programming Interface, API), media recording MediaRecorder and surface view Surfaceview, etc.
  • Media recording is used to record video or image data and make this data accessible to applications.
  • Surface views are used to display preview images.
  • a system library can include multiple function modules. For example: camera service CameraSevice, etc.
  • the hardware abstraction layer is used to provide interface support, for example, including the camera process CameraPipeline for the camera service to call Call.
  • the kernel layer is the layer between hardware and software.
  • the kernel layer includes display drivers, camera drivers, etc.
  • HAL reports the capability information of recording two videos at the same time, and the application layer sends a capture request CaptureRequest to capture at a frame rate of 120FPS.
  • the request corresponds to a video stream and a preview stream, and simultaneously creates two An example of a media codec, mediacodec, that receives 30FPS encoding.
  • HAL calls back two streams according to the dataflow mentioned above. Among them, the preview stream is sent to display, and the video stream is sent to mediacodec. It should be noted that when switching between different upscaling modes, due to the different capture frame rates, the outflow methods of the camera sensor are different, so a restart is required to switch.
  • the video recording and video processing method provided in the embodiment of the present application may be represented as multiple functions in two shooting modes, where the two shooting modes may refer to: movie mode and professional mode.
  • the movie mode is a shooting mode related to the theme of the movie.
  • the image displayed by the electronic device 100 can give the user a sense of watching a movie.
  • the electronic device 100 also provides a plurality of video related to the theme of the movie Style templates, users can use these video style templates to obtain tone-adjusted images or videos, and the tone of these images or videos is similar or identical to the tone of the movie.
  • the movie mode can at least provide an interface for the user to trigger the LUT function and the HDR10 function. For specific descriptions about the LUT function and the HDR10 function, please refer to the following embodiments.
  • the electronic device 100 may enter a movie mode in response to a user's operation.
  • the electronic device 100 may detect a user's touch operation on the camera application, and in response to the operation, the electronic device 100 displays a default camera interface of the camera application.
  • the default camera interface can include: preview frame, shooting mode list, gallery shortcut keys, shutter controls, etc. in:
  • the preview frame can be used to display images collected by the camera 193 in real time.
  • the electronic device 100 can refresh the displayed content therein in real time, so that the user can preview the image currently captured by the camera 193 .
  • One or more shooting mode options may be displayed in the shooting mode list.
  • the one or more shooting mode options may include: portrait mode options, video recording mode options, camera mode options, movie mode options, and professional options.
  • the one or more shooting mode options can be represented as text information on the interface, such as "portrait”, “video recording”, “photographing”, “movie”, “professional”.
  • the one or more shooting mode options may also be represented as icons or other forms of interactive elements (interactive element, IE) on the interface.
  • Gallery shortcuts can be used to launch the Gallery application.
  • the gallery application program is an application program for picture management on electronic devices such as smart phones and tablet computers, and may also be called "album".
  • the name of the application program is not limited in this embodiment.
  • the gallery application program can support users to perform various operations on pictures stored on the electronic device 100, such as browsing, editing, deleting, selecting and other operations.
  • the shutter control can be used to listen for user actions that trigger a photo.
  • the electronic device 100 may detect a user operation acting on the shutter control, and in response to the operation, the electronic device 100 may save the image in the preview frame as a picture in the gallery application.
  • the electronic device 100 may also display the thumbnails of the saved images in the gallery shortcut key. That is, users can tap the shutter control to trigger a photo.
  • the shutter control may be a button or other forms of control.
  • the electronic device 100 may detect a user's touch operation on the movie mode option, and in response to the operation, the electronic device displays a user interface as shown in FIG. 3 .
  • the electronic device 100 may turn on the movie mode by default after starting the camera application. Not limited thereto, the electronic device 100 may also enable the movie mode in other ways, for example, the electronic device 100 may also enable the movie mode according to a user's voice command, which is not limited in this embodiment of the present application.
  • the electronic device 100 may detect a user's touch operation on the movie mode option, and in response to the operation, the electronic device displays a user interface as shown in FIG. 3 .
  • the user interface shown in FIG. 3 includes function options, and the function options include HDR10 options, flash options, LUT options, and setting options. These multiple function options can detect the user's touch operation, and in response to the operation, enable or disable the corresponding shooting function, for example, HDR10 function, flash function, LUT function, setting function.
  • the electronic device can enable the LUT function, and the LUT function can change the display effect of the preview image.
  • the LUT function introduces a color lookup table, which is equivalent to a color conversion model, which can output adjusted color values according to the input color values.
  • the color value of the image captured by the camera is equivalent to the input value, and different color values can be correspondingly obtained as an output value after passing through the color conversion model.
  • the image displayed in the preview box is the image adjusted by the color transformation model.
  • the electronic device 100 uses the LUT function to display an image composed of color values adjusted by the color conversion model, so as to achieve the effect of adjusting the tone of the image.
  • the electronic device 100 can provide multiple video style templates, one video style template corresponds to one color conversion model, and different video style templates can bring different display effects to the preview image.
  • these video style templates can be associated with the theme of the movie, and the tone adjustment effect brought by the video style template to the preview image can be close to or the same as the tone in the movie, creating an atmosphere for the user to shoot a movie.
  • the electronic device 100 can determine a video style template among multiple video style templates according to the current preview video image, and the determined video style template can be displayed on the interface, so that the user can understand Currently determined video style templates, for example, a plurality of video style templates including "A" movie style template, "B" movie style template and "C” movie style template, the corresponding LUTs of different movie style templates can be based on the corresponding Generated by the movie color matching style, the color conversion of the LUT has the style characteristics of the corresponding movie. Film styles can be pre-extracted to produce LUTs suitable for mobile electronics. Turning on the LUT function will change the color tone of the preview video screen. As illustrated in FIG. 3 , the electronic device 100 determines and displays the "A" movie style template.
  • the electronic device 100 may select a video style template according to the user's sliding operation. Specifically, when the electronic device 100 detects the user operation of enabling the LUT function and displays the LUT preview window, the electronic device 100 can select the first video style template located in the LUT preview window by default as the video style template selected by the electronic device 100. template. Afterwards, the electronic device 100 can detect the left and right sliding operation of the user acting on the LUT preview window, and move the position of each video style template in the LUT preview window. The first video style template displayed in the preview window is used as the video style template selected by the electronic device 100 .
  • the electronic device 100 in addition to using the video style template to change the display effect of the preview image, can also detect a user operation to start recording a video after adding the video style template, and in response to the operation, the electronic device 100 starts recording Video, so as to obtain the video after adjusting the display effect using the video style template.
  • the electronic device 100 can also detect the user operation of taking a photo, and in response to this operation, the electronic device 100 saves the preview image with the video style template added in the preview frame as a picture, so as to obtain the user's operation of using the video
  • the style template adjusts the image after the display effect.
  • HDR10 is a high-dynamic range image (High-Dynamic Range, HDR). Compared with ordinary images, HDR can provide more dynamic range and image details, and can better Reflecting the visual effects in the real environment, 10 in HDR10 is 10 bits, and HDR10 can record video with a high dynamic range of 10 bits.
  • the electronic device 100 may detect the user's touch operation on the professional mode option, and enter the professional mode.
  • the functional options that can be included in the user interface are, for example: LOG option, flashlight option, LUT option, and setting option.
  • the user interface also includes parameter adjustment options, such as: measurement Light M option, ISO option, shutter S option, exposure compensation EV option, focus mode AF option and white balance WB option.
  • the electronic device 100 may turn on the professional mode by default after starting the camera application.
  • the electronic device 100 can also enable the professional mode in other ways, for example, the electronic device 100 can also enable the professional mode according to the user's voice command, which is not limited in this embodiment of the present application.
  • the electronic device 100 may detect a user operation on the LOG option by the user, and in response to the operation, the electronic device 100 enables the LOG function.
  • the LOG function can apply the logarithmic function to the exposure curve to preserve the details of the highlights and shadows in the image captured by the camera to the maximum extent, so that the saturation of the final preview image is lower.
  • the video recorded with LOG function is called LOG video.
  • the electronic device 100 can not only record a video with a video style template added through the professional mode, but also add a video style template to the video after recording a video without a video style template, or record a LOG video after enabling the LOG function. Then add a video style template for the LOG video. In this way, the electronic device 100 can not only adjust the display effect of the picture before recording the video, but also adjust the display effect of the recorded video after the video recording is completed, which increases the flexibility and freedom of image adjustment.
  • the embodiment of the present application also provides a video processing device, including: a step-up mode determination module, configured to determine a step-up mode among a plurality of step-up modes, the determined step-up mode corresponds to a capture frame rate and an encoding frame rate, and the capture frame The rate is greater than the encoding frame rate; the video style determination module is used to determine a video style template among the plurality of video style templates in the determined upgrade mode, and each video style template corresponds to a preset two-dimensional color lookup table 2D-LUT ;
  • the video acquisition module is used to obtain the video taken by the camera based on the captured frame rate corresponding to the determined step-up mode;
  • the LOG processing module is used to pass through the camera according to the logarithmic LOG curve corresponding to the current sensitivity ISO of the camera
  • the captured video is processed to obtain the LOG video; the 2D-LUT processing module is used to process the LOG video based on the 2D-LUT corresponding to the determined video style template to obtain
  • the video processing device may apply the above-mentioned video processing method, and the specific process and principle will not be repeated here.
  • the LOG processing module may specifically be the GTM module 26 in the above-mentioned embodiment.
  • the multiple upscaling modes include a first upscaling mode and a second upscaling mode, the first upscaling mode corresponds to the first capture frame rate and the first encoding frame rate, and the second upscaling mode corresponds to the second capture frame rate and the second encoding frame rate, the first capturing frame rate is less than the second capturing frame rate, the first encoding frame rate is equal to the second encoding frame rate, and the second capturing frame rate is greater than the second encoding frame rate.
  • the video processing device further includes: an electronic anti-shake module 31, configured to perform electronic anti-shake processing on the video captured by the camera.
  • the LOG processing module is specifically used to process the video after electronic anti-shake processing through the logarithmic LOG curve corresponding to the current sensitivity ISO of the camera to obtain the LOG video.
  • the first capture frame rate is 60 FPS
  • the second capture frame rate is 120 FPS
  • the first encoding frame rate and the second encoding frame rate are 30 FPS.
  • the first frame rate is 120 FPS or 60 FPS
  • the second frame rate is 30 FPS.
  • the above step 104 based on the two-dimensional 2D-LUT corresponding to the determined video style template, processes the LOG video, and the process of obtaining the video corresponding to the determined video style template is in the HSV color space implement.
  • the encoding module is specifically configured to split the video corresponding to the determined video style template into two streams, one of which is encoded and saved based on the encoding frame rate corresponding to the determined upgrade mode, and the other stream is to preview.
  • each module of the video processing device is only a division of logical functions, and may be fully or partially integrated into one physical entity or physically separated during actual implementation.
  • these modules can all be implemented in the form of software called by the processing element; they can also be implemented in the form of hardware; some modules can also be implemented in the form of software called by the processing element, and some modules can be implemented in the form of hardware.
  • any one of the upgrade mode determination module, the video style determination module, the video acquisition module, the LOG processing module, the 2D-LUT processing module and the encoding module can be a separate processing element, or can be integrated in the video processing device, For example, it can be integrated into a certain chip of the video processing device.
  • each step of the above method or each module above can be completed by an integrated logic circuit of hardware in the processor element or an instruction in the form of software.
  • the upgrade mode determination module, video style determination module, video acquisition module, LOG processing module, 2D-LUT processing module and encoding module can be one or more integrated circuits configured to implement the above method, for example: one or Multiple specific integrated circuits (Application Specific Integrated Circuit, ASIC), or, one or more microprocessors (digital signal processor, DSP), or, one or more Field Programmable Gate Arrays (Field Programmable Gate Array, FPGA) wait.
  • ASIC Application Specific Integrated Circuit
  • DSP digital signal processor
  • FPGA Field Programmable Gate Array
  • the processing element may be a general-purpose processor, such as a central processing unit (Central Processing Unit, CPU) or other processors that can call programs.
  • these modules can be integrated together and implemented in the form of a system-on-a-chip (SOC).
  • SOC system-on-a-chip
  • An embodiment of the present application further provides a video processing device, including: a processor and a memory, the memory is used to store at least one instruction, and when the instruction is loaded and executed by the processor, the video processing method in any of the foregoing embodiments is implemented.
  • the video processing apparatus may apply the above-mentioned video processing method, and the specific process and principle will not be repeated here.
  • the number of processors may be one or more, and the processor and the memory may be connected through a bus or in other ways.
  • the memory can be used to store non-transitory software programs, non-transitory computer-executable programs and modules, such as program instructions/modules corresponding to the video processing device in the embodiment of the present application.
  • the processor executes various functional applications and data processing by running non-transitory software programs, instructions and modules stored in the memory, that is, implements the method in any of the above method embodiments.
  • the memory may include a program storage area and a data storage area, wherein the program storage area may store an operating system, an application program required by at least one function; and necessary data and the like.
  • the memory may include high-speed random access memory, and may also include non-transitory memory, such as at least one magnetic disk storage device, flash memory device, or other non-transitory solid-state storage devices.
  • an embodiment of the present application further provides an electronic device, including: a camera 193 and the above-mentioned video processing device, where the video processing device includes a processor 110 .
  • the electronic device may be any product or component with a video shooting function such as a mobile phone, a TV, a tablet computer, a watch, a bracelet, and the like.
  • An embodiment of the present application further provides a computer-readable storage medium, in which a computer program is stored, and when running on a computer, the computer is made to execute the video processing method in any of the foregoing embodiments.
  • all or part of them may be implemented by software, hardware, firmware or any combination thereof.
  • software When implemented using software, it may be implemented in whole or in part in the form of a computer program product.
  • the computer program product includes one or more computer instructions. When the computer program instructions are loaded and executed on the computer, the processes or functions according to the present application will be generated in whole or in part.
  • the computer can be a general purpose computer, a special purpose computer, a computer network, or other programmable devices.
  • the computer instructions may be stored in or transmitted from one computer-readable storage medium to another computer-readable storage medium, for example, the computer instructions may be transmitted from a website, computer, server or data center Transmission to another website site, computer, server, or data center by wired (eg, coaxial cable, optical fiber, DSL) or wireless (eg, infrared, wireless, microwave, etc.) means.
  • the computer-readable storage medium may be any available medium that can be accessed by a computer, or a data storage device such as a server or a data center integrated with one or more available media.
  • the available medium may be a magnetic medium (for example, a floppy disk, a hard disk, a magnetic tape), an optical medium (for example, DVD), or a semiconductor medium (for example, a Solid State Disk).
  • "at least one” means one or more, and “multiple” means two or more.
  • “And/or” describes the association relationship of associated objects, indicating that there may be three kinds of relationships, for example, A and/or B may indicate that A exists alone, A and B exist simultaneously, or B exists alone. Among them, A and B can be singular or plural.
  • the character “/” generally indicates that the contextual objects are an “or” relationship.
  • “At least one of the following” and similar expressions refer to any combination of these items, including any combination of single items or plural items.
  • At least one of a, b, and c may represent: a, b, c, a-b, a-c, b-c, or a-b-c, wherein a, b, and c may be single or multiple.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Studio Devices (AREA)
  • Image Processing (AREA)

Abstract

Les modes de réalisation de la présente invention concernent un procédé et un appareil de traitement vidéo, un dispositif électronique et un support de stockage relevant du domaine technique de la capture vidéo et permettant de conférer différents effets de style à des vidéos capturées par un dispositif électronique sur la base des caractéristiques d'une table de consultation (LUT) de façon à satisfaire de plus grandes exigences de teintes. Le procédé de traitement vidéo comprend les étapes consistant à : déterminer un mode accéléré parmi une pluralité de modes accélérés, la fréquence d'images de capture étant supérieure à la fréquence d'images codées ; déterminer un modèle de style vidéo parmi une pluralité de modèles de style vidéo du mode accéléré déterminé ; obtenir une vidéo capturée par une caméra sur la base de la fréquence d'images de capture déterminée correspondant au mode accéléré déterminé ; traiter la vidéo capturée par la caméra au moyen d'une courbe logarithmique correspondant à une valeur ISO de photosensibilité actuelle de la caméra ; traiter la vidéo logarithmique sur la base d'une LUT 2D correspondant au modèle de style vidéo déterminé ; puis coder et stocker la vidéo correspondant au modèle de style vidéo déterminé sur la base d'une fréquence d'images de codage correspondant au mode accéléré déterminé.
PCT/CN2022/094754 2021-08-12 2022-05-24 Procédé et appareil de traitement vidéo, dispositif électronique et support de stockage WO2023016039A1 (fr)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN202110926817.1A CN113824913A (zh) 2021-08-12 2021-08-12 视频处理方法、装置、电子设备和存储介质
CN202110926817.1 2021-08-12

Publications (1)

Publication Number Publication Date
WO2023016039A1 true WO2023016039A1 (fr) 2023-02-16

Family

ID=78913138

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2022/094754 WO2023016039A1 (fr) 2021-08-12 2022-05-24 Procédé et appareil de traitement vidéo, dispositif électronique et support de stockage

Country Status (2)

Country Link
CN (2) CN115242992B (fr)
WO (1) WO2023016039A1 (fr)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN117478929A (zh) * 2023-12-28 2024-01-30 昆明中经网络有限公司 一种基于ai大模型的新媒体精品影像处理系统

Families Citing this family (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113810642B (zh) * 2021-08-12 2023-02-28 荣耀终端有限公司 视频处理方法、装置、电子设备和存储介质
CN115242992B (zh) * 2021-08-12 2023-08-18 荣耀终端有限公司 视频处理方法、装置、电子设备和存储介质
US20240155254A1 (en) * 2021-12-31 2024-05-09 Honor Device Co., Ltd. Image Processing Method and Related Electronic Device
CN114520874B (zh) * 2022-01-28 2023-11-24 西安维沃软件技术有限公司 视频处理方法、装置及电子设备
CN116805992B (zh) * 2022-03-15 2024-04-19 荣耀终端有限公司 生成对数曲线的方法、设备和存储介质
CN114598834A (zh) * 2022-05-10 2022-06-07 中国铁塔股份有限公司 一种视频处理方法、装置、电子设备及可读存储介质

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107077828A (zh) * 2014-11-25 2017-08-18 英特尔公司 对颜色查找表的大小进行压缩
CN108616696A (zh) * 2018-07-19 2018-10-02 北京微播视界科技有限公司 一种视频拍摄方法、装置、终端设备及存储介质
US20200007718A1 (en) * 2018-06-29 2020-01-02 Ati Technologies Ulc Method and apparatus for nonlinear interpolation color conversion using look up tables
CN111510698A (zh) * 2020-04-23 2020-08-07 惠州Tcl移动通信有限公司 图像处理方法、装置、存储介质及移动终端
CN113824913A (zh) * 2021-08-12 2021-12-21 荣耀终端有限公司 视频处理方法、装置、电子设备和存储介质

Family Cites Families (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2015122734A (ja) * 2013-11-25 2015-07-02 パナソニックIpマネジメント株式会社 撮像装置及び撮像方法
KR102449872B1 (ko) * 2015-12-18 2022-09-30 삼성전자주식회사 촬영 디바이스 및 그 제어 방법
CN108600646B (zh) * 2018-07-25 2020-09-08 张维良 一种用于多图像采集设备影视拍摄器材的控制方法
CN110636375B (zh) * 2019-11-11 2022-03-11 RealMe重庆移动通信有限公司 视频流处理方法、装置、终端设备及计算机可读存储介质
CN113067994B (zh) * 2021-03-31 2022-08-19 联想(北京)有限公司 一种视频录制方法及电子设备

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107077828A (zh) * 2014-11-25 2017-08-18 英特尔公司 对颜色查找表的大小进行压缩
US20200007718A1 (en) * 2018-06-29 2020-01-02 Ati Technologies Ulc Method and apparatus for nonlinear interpolation color conversion using look up tables
CN108616696A (zh) * 2018-07-19 2018-10-02 北京微播视界科技有限公司 一种视频拍摄方法、装置、终端设备及存储介质
CN111510698A (zh) * 2020-04-23 2020-08-07 惠州Tcl移动通信有限公司 图像处理方法、装置、存储介质及移动终端
CN113824913A (zh) * 2021-08-12 2021-12-21 荣耀终端有限公司 视频处理方法、装置、电子设备和存储介质

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN117478929A (zh) * 2023-12-28 2024-01-30 昆明中经网络有限公司 一种基于ai大模型的新媒体精品影像处理系统
CN117478929B (zh) * 2023-12-28 2024-03-08 昆明中经网络有限公司 一种基于ai大模型的新媒体精品影像处理系统

Also Published As

Publication number Publication date
CN115242992B (zh) 2023-08-18
CN113824913A (zh) 2021-12-21
CN115242992A (zh) 2022-10-25

Similar Documents

Publication Publication Date Title
WO2023016039A1 (fr) Procédé et appareil de traitement vidéo, dispositif électronique et support de stockage
WO2023016035A1 (fr) Procédé et appareil de traitement vidéo, dispositif électronique et support de stockage
WO2023016037A1 (fr) Procédé et appareil de traitement vidéo, dispositif électronique et support de stockage
CN113824914B (zh) 视频处理方法、装置、电子设备和存储介质
WO2023160295A1 (fr) Procédé et appareil de traitement vidéo
CN114449199B (zh) 视频处理方法、装置、电子设备和存储介质
WO2023016040A1 (fr) Procédé et appareil de traitement vidéo, dispositif électronique et support de stockage
WO2023016044A1 (fr) Procédé et appareil de traitement vidéo, dispositif électronique et support de stockage
WO2023016038A1 (fr) Procédé et appareil de traitement vidéo, dispositif électronique et support de stockage
WO2023016042A1 (fr) Procédé et appareil de traitement vidéo, dispositif électronique et support de stockage
WO2023016043A1 (fr) Procédé et appareil de traitement vidéo, dispositif électronique et support de stockage
WO2023016041A1 (fr) Procédé et appareil de traitement vidéo, dispositif électronique et support de stockage
US20240236492A9 (en) Video Processing Method and Apparatus, Electronic Device, and Storage Medium
CN115706853A (zh) 视频处理方法、装置、电子设备和存储介质
US20230215051A1 (en) Systems, apparatus, and methods for color space representation

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 22855025

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE