WO2022262537A1 - Procédé de traitement de transition pour données vidéo et dispositif électronique - Google Patents

Procédé de traitement de transition pour données vidéo et dispositif électronique Download PDF

Info

Publication number
WO2022262537A1
WO2022262537A1 PCT/CN2022/094793 CN2022094793W WO2022262537A1 WO 2022262537 A1 WO2022262537 A1 WO 2022262537A1 CN 2022094793 W CN2022094793 W CN 2022094793W WO 2022262537 A1 WO2022262537 A1 WO 2022262537A1
Authority
WO
WIPO (PCT)
Prior art keywords
interface
electronic device
video
video data
frame
Prior art date
Application number
PCT/CN2022/094793
Other languages
English (en)
Chinese (zh)
Inventor
牛思月
Original Assignee
荣耀终端有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 荣耀终端有限公司 filed Critical 荣耀终端有限公司
Priority to US18/257,018 priority Critical patent/US20240106967A1/en
Priority to EP22824018.0A priority patent/EP4240011A4/fr
Publication of WO2022262537A1 publication Critical patent/WO2022262537A1/fr

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N5/00Details of television systems
    • H04N5/222Studio circuitry; Studio devices; Studio equipment
    • H04N5/262Studio circuits, e.g. for mixing, switching-over, change of character of image, other special effects ; Cameras specially adapted for the electronic generation of special effects
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N5/00Details of television systems
    • H04N5/76Television signal recording
    • H04N5/91Television signal processing therefor
    • H04N5/915Television signal processing therefor for field- or frame-skip recording or reproducing
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T3/00Geometric image transformations in the plane of the image
    • G06T3/40Scaling of whole images or parts thereof, e.g. expanding or contracting
    • G06T3/4038Image mosaicing, e.g. composing plane images from plane sub-images
    • GPHYSICS
    • G11INFORMATION STORAGE
    • G11BINFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
    • G11B27/00Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
    • G11B27/02Editing, e.g. varying the order of information signals recorded on, or reproduced from, record carriers
    • G11B27/031Electronic editing of digitised analogue information signals, e.g. audio or video signals
    • GPHYSICS
    • G11INFORMATION STORAGE
    • G11BINFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
    • G11B27/00Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
    • G11B27/02Editing, e.g. varying the order of information signals recorded on, or reproduced from, record carriers
    • G11B27/031Electronic editing of digitised analogue information signals, e.g. audio or video signals
    • G11B27/036Insert-editing
    • GPHYSICS
    • G11INFORMATION STORAGE
    • G11BINFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
    • G11B27/00Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
    • G11B27/10Indexing; Addressing; Timing or synchronising; Measuring tape travel
    • G11B27/34Indicating arrangements 
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/44Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs
    • H04N21/44016Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs involving splicing one content stream with another content stream, e.g. for substituting a video clip
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/83Generation or processing of protective or descriptive data associated with content; Content structuring
    • H04N21/845Structuring of content, e.g. decomposing content into time segments
    • H04N21/8456Structuring of content, e.g. decomposing content into time segments by decomposing the content in the time domain, e.g. in time segments
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/60Control of cameras or camera modules
    • H04N23/62Control of parameters via user interfaces
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/60Control of cameras or camera modules
    • H04N23/63Control of cameras or camera modules by using electronic viewfinders
    • H04N23/631Graphical user interfaces [GUI] specially adapted for controlling image capture or setting capture parameters
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/60Control of cameras or camera modules
    • H04N23/667Camera operation mode switching, e.g. between still and video, sport and normal or high- and low-resolution modes

Definitions

  • the present application relates to the field of terminal technologies, and in particular to a video data transition processing method and electronic equipment.
  • electronic devices such as mobile phones and tablet computers are generally equipped with multiple cameras, such as front cameras, rear cameras, wide-angle cameras, and the like. Multiple cameras are convenient for users to use electronic equipment to shoot video works.
  • Embodiments of the present application provide a video data transition processing method and an electronic device, which are used to improve the efficiency of human-computer interaction in video editing.
  • an embodiment of the present application provides a video data transition processing method, the method is applied to an electronic device, and the method includes: the electronic device displays a first interface; wherein the first interface includes The first thumbnail image of the first video data; the first video data includes a first transition special effect; the first transition special effect is superimposed on the first video frame of multiple consecutive frames in the first video data; The electronic device receives a user's first operation on the first thumbnail; the electronic device displays a second interface in response to the first operation; the second interface is a video editing interface for the first video data ; The second interface includes a one-key blockbuster control; after receiving the user's second operation on the one-key blockbuster control, the electronic device displays a third interface; the third interface is used to display the second video data ; The second video data includes the video frame of the first video data, a multi-frame replacement frame, the first music, and a second transition effect corresponding to the first music; the second transition effect is superimposed on the On the multi-frame replacement frame, the replacement frame
  • the electronic device can automatically edit and process the first video data to obtain the second video data in response to the user's operation on the one-key blockbuster control.
  • the first music is configured in the second video data.
  • the electronic device can replace the first transition effect with a second transition effect that matches the first music.
  • the second video The transition effects appearing in the data can be matched with the first music, which improves the degree of adaptation between music and content in the second video data, improves user satisfaction with the second video data, and reduces the possibility of rework.
  • the entire operation of triggering the creation of the second video data is simple, which effectively improves the efficiency of human-computer interaction for creating video data.
  • the method before the electronic device displays the first interface, the method further includes: the electronic device displays a fourth interface; the fourth interface is a viewfinder preview interface provided by a camera application; the The fourth interface includes a first control indicating to start shooting video; the electronic device receives a third operation of the first control by the user; the electronic device displays a fifth interface in response to the third operation, and starts recording The first video data; wherein, the fifth interface is a video recording interface in the first lens mode; the fifth interface includes a second control indicating to switch lens modes; when recording to the first video data At the first time point, the electronic device displays a sixth interface in response to the user's fourth operation on the second control, and determines that the video frame corresponding to the first time point is the first video frame; wherein , the sixth interface is a video recording interface in the second lens mode; the sixth interface includes a third control indicating to stop shooting; the electronic device receives a fifth operation of the third control by the user; the The electronic device displaying the first interface includes: the electronic device responding
  • the first video data may be a video captured by the electronic device in a normal mode.
  • the electronic device receives an operation indicating to switch the lens mode at the first time point, not only can directly switch the lens mode, but also can determine the video frame corresponding to the first time point in the first video data as the first In this way, when editing and processing the first video data, the video frames affected by the lens switching can be processed, the viewability of the second video data can be improved, and the human-computer interaction efficiency of editing video can be improved.
  • the method further includes: the electronic device superimposes the first video frame on the first video frame. Transition effects.
  • the video clips collected before and after the switching of the lens mode are connected by the first transition special effect, so as to improve the video frequency due to the switching of the lens mode.
  • the problem of incoherence caused by the content can improve the quality of the film.
  • the method before the electronic device displays the first interface, the method further includes: the electronic device displays a main interface; the main interface includes an icon of a gallery application; The sixth operation of the icon of the gallery application; the electronic device displaying the first interface includes: the electronic device responding to the sixth operation, displaying the first interface, the first interface being the gallery The application interface provided by the application.
  • the first video data may also be video data already stored in the gallery, that is, it may be a video shot by other devices, or a video that has been created once.
  • the user can process a variety of video materials through the electronic device, and the operations required for processing are simple, and the efficiency of human-computer interaction in video creation is improved.
  • the method before the electronic device displays the third interface, the method further includes: in response to the second operation, the electronic device determines a first Effect template; the first effect template includes the first music; the electronic device deletes the first video frame in the first video data; the electronic device deletes the first video frame in the first video data
  • the second video frame freezes to obtain the replacement frame used to replace the first video frame; the second video frame is the video frame of the adjacent previous frame of the first video frame, or is the first video frame A video frame next to a video frame; the electronic device superimposes the second transition effect on the replacement frame.
  • the first effect template corresponds to the first style; determining the first effect template from a plurality of preconfigured effect templates includes: the electronic device using a preset artificial intelligence model It is determined that the first video data matches the first style; the electronic device determines the first effect template from the effect templates belonging to the first style; or, the electronic device determines the first effect template from multiple preset Among the effect templates set, the first shooting template is randomly determined.
  • the method before the electronic device displays the third interface, the method further includes: determining, by the electronic device, the second transition special effect corresponding to the first music.
  • the electronic device determining the second transition effect corresponding to the first music includes: determining, by the electronic device, the second transition effect corresponding to the first music from various preset transition effects The second transition effect with associated identification between a piece of music.
  • the electronic device determining the second transition effect corresponding to the first music includes: determining, by the electronic device, from various preset transition effects based on matching weights The second transition effect; wherein, each of the preset transition effects corresponds to a matching weight, and the matching weight is the degree of adaptation between the first music and the preset transition effects The quantization ratio parameter of .
  • the second transition effect is relatively related to the first music, and the type is also random, so that while ensuring that the transition effect is adapted to the first music, the diversification of the transition effect is increased, The viewability of the second video data is improved.
  • the second video data further includes: a third transition effect; the third transition effect is added to the video frame corresponding to the second time point in the first video data ;
  • the third transition effect is one of multiple preset transition effects; the multiple preset transition effects include the second transition effect.
  • an electronic device provided by an embodiment of the present application, includes one or more processors and a memory; the memory is coupled to the processor, and the memory is used to store computer program codes, the computer program codes include computer instructions, When one or more processors execute computer instructions, the one or more processors are configured to: display a first interface; wherein, the first interface includes a first thumbnail image of the first video data; the first interface A piece of video data includes a first transition effect; the first transition effect is superimposed on a plurality of consecutive first video frames in the first video data; receiving a first user operation on the first thumbnail; In response to the first operation, a second interface is displayed; the second interface is a video editing interface for the first video data; the second interface includes a one-key blockbuster control; After the second operation of the block control, a third interface is displayed; the third interface is used to display the second video data; the second video data includes video frames of the first video data, multi-frame replacement frames, first Music and a second transition effect corresponding to the first music
  • the one or more processors are further configured to: display a fourth interface; the fourth interface is a viewfinder preview interface provided by the camera application; the first interface The fourth interface includes a first control indicating to start shooting video; receiving a third operation of the first control by the user; in response to the third operation, displaying a fifth interface and starting to record the first video data; wherein, The fifth interface is a video recording interface in the first lens mode; the fifth interface includes a second control indicating to switch the lens mode; when recording to the first time point of the first video data, in response to the user The fourth operation on the second control is to display a sixth interface, and determine that the video frame corresponding to the first time point is the first video frame; wherein, the sixth interface is the second camera mode A video recording interface; the sixth interface includes a third control indicating to stop shooting; receiving a fifth user operation on the third control;
  • the one or more processors are further configured to: display the first interface in response to the fifth operation; the first interface is also a framing preview interface provided by the camera application.
  • the one or more processors are further configured to: superimpose the first video frame on the first video frame field effects.
  • the one or more processors are further configured to: display a main interface; the main interface includes an icon of a gallery application;
  • the sixth operation of icons: displaying a first interface includes: the electronic device responding to the sixth operation and displaying the first interface, where the first interface is an application interface provided by the gallery application.
  • the one or more processors are further configured to: determine the first effect from a plurality of preconfigured effect templates in response to the second operation template; the first effect template includes the first music; the first video frame in the first video data is deleted; the second video frame in the first video data is frozen to obtain the The replacement frame that replaces the first video frame; the second video frame is a video frame adjacent to the first video frame, or a frame adjacent to the first video frame A video frame; superimposing the second transition effect on the replacement frame.
  • the one or more processors are further configured to: use a preset artificial intelligence model to determine that the first video data matches the first style;
  • the first effect template is determined from among the effect templates; or, the first shooting template is randomly determined from a plurality of preset effect templates.
  • the one or more processors are further configured to: determine the second transition effect corresponding to the first music.
  • the one or more processors are further configured to: determine the second transition that has an associated identifier with the first music from a variety of preset transition effects special effects.
  • the one or more processors are further configured to: determine the second transition effect from various preset transition effects based on matching weights; wherein, each of the The preset transition effect corresponds to a matching weight, and the matching weight is a quantitative ratio parameter of the degree of adaptation between the first music and the preset transition effect.
  • the second video data further includes: a third transition effect; the third transition effect is added to the video frame corresponding to the second time point in the first video data ;
  • the third transition effect is one of multiple preset transition effects; the multiple preset transition effects include the second transition effect.
  • a computer storage medium provided by an embodiment of the present application includes computer instructions.
  • the computer instructions When the computer instructions are run on the electronic device, the electronic device executes the above-mentioned first aspect and its possible embodiments. method.
  • the present application provides a computer program product, which, when the computer program product runs on the above-mentioned electronic device, causes the electronic device to execute the method described in the above-mentioned first aspect and its possible embodiments.
  • FIG. 1 is a schematic structural diagram of an electronic device provided in an embodiment of the present application.
  • FIG. 2 is one of the example diagrams of the display interface provided by the embodiment of the present application.
  • Fig. 3 is the second example diagram of the display interface provided by the embodiment of the present application.
  • Figure 4 conceptually shows an example of the impact on video data 1 after the lens mode is switched from the front-to-back mode to the picture-in-picture mode;
  • Figure 5 conceptually shows an example diagram of processing video data 1 after the lens mode is switched from the front-to-back mode to the picture-in-picture mode;
  • Fig. 6 is the third example diagram of the display interface provided by the embodiment of the present application.
  • FIG. 7A conceptually shows one of the example diagrams of processing video data 1 in the scene where the lens mode is switched from front-to-back mode to back-to-back mode;
  • FIG. 7B conceptually shows the second example of processing video data 1 in the scene where the lens mode is switched from front-to-back mode to back-to-back mode;
  • FIG. 8 is the fourth example diagram of the display interface provided by the embodiment of the present application.
  • Fig. 9 is the fifth exemplary diagram of the display interface provided by the embodiment of the present application.
  • Fig. 10 is the sixth example diagram of the display interface provided by the embodiment of the present application.
  • Fig. 11 is the seventh example diagram of the display interface provided by the embodiment of the present application.
  • FIG. 12 is a flow chart of the steps of the video data transition processing method provided by the embodiment of the present application.
  • Fig. 13 conceptually shows an example diagram of replacing the original transition effect in the video data 1;
  • Fig. 14 is the eighth example diagram of the display interface provided by the embodiment of the present application.
  • FIG. 15 is a schematic diagram of a chip system provided by an embodiment of the present application.
  • first and second are used for descriptive purposes only, and cannot be understood as indicating or implying relative importance or implicitly specifying the quantity of indicated technical features. Thus, a feature defined as “first” and “second” may explicitly or implicitly include one or more of these features. In the description of this embodiment, unless otherwise specified, “plurality” means two or more.
  • An embodiment of the present application provides a video data transition processing method, which can be applied to an electronic device with multiple cameras.
  • the electronic device can automatically process video data in response to user operations, such as adding transition effects, configuring video music, and the like.
  • user operations such as adding transition effects, configuring video music, and the like.
  • it reduces the operational complexity of editing video data and improves the efficiency of human-computer interaction in video creation.
  • the electronic device in the embodiment of the present application may be a mobile phone, a tablet computer, a smart watch, a desktop, a laptop, a handheld computer, a notebook computer, an ultra-mobile personal computer (ultra-mobile personal computer, UMPC), a netbook , and cellular phones, personal digital assistants (personal digital assistant, PDA), augmented reality (augmented reality, AR) ⁇ virtual reality (virtual reality, VR) equipment, etc.
  • PDA personal digital assistant
  • augmented reality augmented reality, AR
  • VR virtual reality
  • the specific form is not particularly limited.
  • FIG. 1 is a schematic structural diagram of an electronic device 100 provided in an embodiment of the present application.
  • the electronic device 100 may include: a processor 110, an external memory interface 120, an internal memory 121, a universal serial bus (universal serial bus, USB) interface 130, a charging management module 140, a power management module 141, a battery 142, antenna 1, antenna 2, mobile communication module 150, wireless communication module 160, audio module 170, speaker 170A, receiver 170B, microphone 170C, earphone interface 170D, sensor module 180, button 190, motor 191, indicator 192, camera 193, a display screen 194, and a subscriber identification module (subscriber identification module, SIM) card interface 195, etc.
  • SIM subscriber identification module
  • the above-mentioned sensor module 180 may include sensors such as pressure sensor, gyroscope sensor, air pressure sensor, magnetic sensor, acceleration sensor, distance sensor, proximity light sensor, fingerprint sensor, temperature sensor, touch sensor, ambient light sensor and bone conduction sensor.
  • sensors such as pressure sensor, gyroscope sensor, air pressure sensor, magnetic sensor, acceleration sensor, distance sensor, proximity light sensor, fingerprint sensor, temperature sensor, touch sensor, ambient light sensor and bone conduction sensor.
  • the structure shown in this embodiment does not constitute a specific limitation on the electronic device 100 .
  • the electronic device 100 may include more or fewer components than shown, or combine certain components, or separate certain components, or arrange different components.
  • the illustrated components can be realized in hardware, software or a combination of software and hardware.
  • the processor 110 may include one or more processing units, for example: the processor 110 may include an application processor (application processor, AP), a modem processor, a graphics processing unit (graphics processing unit, GPU), an image signal processor (image signal processor, ISP), controller, memory, video codec, digital signal processor (digital signal processor, DSP), baseband processor, and/or neural network processor (neural-network processing unit, NPU) Wait. Wherein, different processing units may be independent devices, or may be integrated in one or more processors.
  • application processor application processor, AP
  • modem processor graphics processing unit
  • GPU graphics processing unit
  • image signal processor image signal processor
  • ISP image signal processor
  • controller memory
  • video codec digital signal processor
  • DSP digital signal processor
  • baseband processor baseband processor
  • neural network processor neural-network processing unit
  • the controller may be the nerve center and command center of the electronic device 100 .
  • the controller can generate an operation control signal according to the instruction opcode and timing signal, and complete the control of fetching and executing the instruction.
  • a memory may also be provided in the processor 110 for storing instructions and data.
  • the memory in processor 110 is a cache memory.
  • the memory may hold instructions or data that the processor 110 has just used or recycled. If the processor 110 needs to use the instruction or data again, it can be called directly from the memory. Repeated access is avoided, and the waiting time of the processor 110 is reduced, thus improving the efficiency of the system.
  • processor 110 may include one or more interfaces.
  • the interface may include an integrated circuit (inter-integrated circuit, I2C) interface, an integrated circuit built-in audio (inter-integrated circuit sound, I2S) interface, a pulse code modulation (pulse code modulation, PCM) interface, a universal asynchronous transmitter (universal asynchronous receiver/transmitter, UART) interface, mobile industry processor interface (mobile industry processor interface, MIPI), general-purpose input and output (general-purpose input/output, GPIO) interface, subscriber identity module (subscriber identity module, SIM) interface, and /or universal serial bus (universal serial bus, USB) interface, etc.
  • I2C integrated circuit
  • I2S integrated circuit built-in audio
  • PCM pulse code modulation
  • PCM pulse code modulation
  • UART universal asynchronous transmitter
  • MIPI mobile industry processor interface
  • GPIO general-purpose input and output
  • subscriber identity module subscriber identity module
  • SIM subscriber identity module
  • USB universal serial bus
  • the interface connection relationship between the modules shown in this embodiment is only for schematic illustration, and does not constitute a structural limitation of the electronic device 100 .
  • the electronic device 100 may also adopt different interface connection methods in the above embodiments, or a combination of multiple interface connection methods.
  • the electronic device 100 realizes the display function through the GPU, the display screen 194 , and the application processor.
  • the GPU is a microprocessor for image processing, and is connected to the display screen 194 and the application processor. GPUs are used to perform mathematical and geometric calculations for graphics rendering.
  • Processor 110 may include one or more GPUs that execute program instructions to generate or change display information.
  • the display screen 194 is used to display images, videos and the like.
  • the display screen 194 includes a display panel.
  • the display panel can be a liquid crystal display (LCD), an organic light-emitting diode (OLED), an active matrix organic light emitting diode or an active matrix organic light emitting diode (active-matrix organic light emitting diode, AMOLED), flexible light-emitting diode (flex light-emitting diode, FLED), Miniled, MicroLed, Micro-oLed, quantum dot light emitting diodes (quantum dot light emitting diodes, QLED), etc.
  • LCD liquid crystal display
  • OLED organic light-emitting diode
  • AMOLED active matrix organic light emitting diode
  • FLED flexible light-emitting diode
  • Miniled MicroLed, Micro-oLed
  • quantum dot light emitting diodes quantum dot light emitting diodes (quantum dot light emitting diodes, QLED),
  • the electronic device 100 can realize the shooting function through the ISP, the camera 193 , the video codec, the GPU, the display screen 194 and the application processor.
  • the ISP is used for processing the data fed back by the camera 193 .
  • the light is transmitted to the photosensitive element of the camera through the lens, and the light signal is converted into an electrical signal, and the photosensitive element of the camera transmits the electrical signal to the ISP for processing, and converts it into an image visible to the naked eye.
  • ISP can also perform algorithm optimization on image noise, brightness, and skin color.
  • ISP can also optimize the exposure, color temperature and other parameters of the shooting scene.
  • the ISP may be located in the camera 293 .
  • Camera 193 is used to capture still images or video.
  • the object generates an optical image through the lens and projects it to the photosensitive element.
  • the photosensitive element may be a charge coupled device (CCD) or a complementary metal-oxide-semiconductor (CMOS) phototransistor.
  • CMOS complementary metal-oxide-semiconductor
  • the photosensitive element converts the light signal into an electrical signal, and then transmits the electrical signal to the ISP to convert it into a digital image signal.
  • the ISP outputs the digital image signal to the DSP for processing.
  • DSP converts digital image signals into standard RGB, YUV and other image signals.
  • the electronic device 100 may include N cameras 193 , where N is a positive integer greater than 1.
  • the aforementioned N cameras 193 may include: one or more front cameras and one or more rear cameras.
  • the mobile phone includes at least one front camera.
  • the front camera is configured on the front side of the mobile phone, for example, the front camera 201 shown in (a) of FIG. 2 .
  • the phone includes at least one rear camera.
  • the rear camera is arranged on the back side of the mobile phone. This way, the front and rear cameras face different directions.
  • the electronic device may enable at least one of the N cameras 139 to take pictures and generate corresponding photos or videos.
  • a front camera of the electronic device 100 is used alone for shooting.
  • a single rear camera of the electronic device 100 is used for shooting.
  • two front-facing cameras are enabled for shooting at the same time.
  • a front-facing camera and a rear-facing camera are enabled for shooting at the same time.
  • enabling a single camera 139 for shooting may be referred to as enabling a single-camera mode, such as a proactive mode (also known as a single front mode) and a rear-camera mode (also known as a single rear mode).
  • enabling multiple cameras 139 to shoot at the same time can be collectively referred to as enabling a multi-camera mode, such as front-to-front mode, front-to-back mode, back-to-back mode, and picture-in-picture mode.
  • the electronic device may render and combine image frames collected by the front-facing camera and the rear-facing camera.
  • the above rendering merging may be splicing image frames collected by different cameras.
  • the image frames collected by different cameras can be spliced up and down.
  • the image frames collected by different cameras can be spliced left and right.
  • the image frames collected by one camera may be embedded in the image frames collected by another camera. Then, it is coded to generate a photo.
  • the front camera collects a video stream and caches it.
  • the rear camera captures a video stream and caches it.
  • the electronic device 100 renders and merges the two buffered video streams frame by frame, that is, renders and merges the video frames whose acquisition time points are the same or match in the two video streams. After that, encoding is performed to generate a video file.
  • Digital signal processors are used to process digital signals. In addition to digital image signals, they can also process other digital signals. For example, when the electronic device 100 selects a frequency point, the digital signal processor is used to perform Fourier transform on the energy of the frequency point.
  • Video codecs are used to compress or decompress digital video.
  • the electronic device 100 may support one or more video codecs.
  • the electronic device 100 can play or record videos in various encoding formats, for example: moving picture experts group (moving picture experts group, MPEG) 1, MPEG2, MPEG3, MPEG4 and so on.
  • MPEG moving picture experts group
  • the NPU is a neural-network (NN) computing processor.
  • NN neural-network
  • Applications such as intelligent cognition of the electronic device 100 can be realized through the NPU, such as image recognition, face recognition, speech recognition, text understanding, and the like.
  • the audio module 170 is used to convert digital audio information into analog audio signal output, and is also used to convert analog audio input into digital audio signal.
  • the audio module 170 may also be used to encode and decode audio signals.
  • the audio module 170 may be set in the processor 110 , or some functional modules of the audio module 170 may be set in the processor 110 .
  • Speaker 170A also referred to as a "horn", is used to convert audio electrical signals into sound signals. In this way, the electronic device 100 can play audio data, such as video music and the like.
  • the pressure sensor is used to sense the pressure signal and convert the pressure signal into an electrical signal.
  • a pressure sensor may be located on the display screen 194 .
  • the gyro sensor can be used to determine the motion posture of the electronic device 100 . When the electronic device 100 is stationary, the magnitude and direction of gravity can be detected. It can also be used to recognize the posture of the electronic device 100, and be applied to applications such as horizontal and vertical screen switching.
  • Touch sensor also known as "touch panel”.
  • the touch sensor can be arranged on the display screen 194, and the touch sensor 180K and the display screen 194 form a touch screen, also called “touch screen”.
  • the touch sensor is used to detect a touch operation on or near it. The touch sensor can pass the detected touch operation to the application processor to determine the type of touch event.
  • the shot video can be edited by operating the electronic device, such as configuring video music, adding animation special effects, adding transition special effects, and the like.
  • the video after the second creation will be more vivid and rich, and it also conforms to the user's creative intention.
  • adding transition effects can not only make the transition of video content more natural, but also enrich the content presented in the video.
  • an embodiment of the present application provides a method for processing transitions of video data.
  • the method in the embodiment of the present application will be described below by taking the above-mentioned electronic device 100 as a mobile phone as an example.
  • the mobile phone includes a main interface.
  • the above-mentioned main interface is also called desktop 202 .
  • the main interface may be a user interface displayed after the mobile phone is unlocked.
  • the aforementioned main interface may include icons of installed application programs (Application, APP), such as the icon 203 of the camera APP.
  • Application Application
  • the mobile phone can receive the user's operation on the main interface, and start the APP indicated by the operation.
  • the mobile phone may receive a user's operation on the icon 203, such as a click operation, and start the camera APP in response to the operation.
  • an application interface provided by the camera APP may be displayed.
  • a viewfinder interface for performing a shooting function is displayed, that is, an interface 204 as shown in (b) of FIG. 2 .
  • the user can switch between different functional modes of the camera APP in the interface 204, such as a portrait functional mode, a photographing functional mode, a video recording functional mode, and a multi-lens video recording functional mode. That is, the mobile phone can receive the user's operation 1 in the interface 204, and the operation 1 is used to instruct the camera APP to switch between different function modes.
  • the interface 204 includes controls corresponding to multiple functional modes of the camera APP, such as a portrait control, a camera control, a video control, a multi-mirror video control, and the like.
  • the photographing control is in a selected state, which is used to prompt the user that the current viewfinder interface is used to execute the photographing function.
  • the mobile phone can receive the user's operation on any one of the portrait control, video control, and multi-mirror video control, and determine the switched function mode based on the operation, and display the viewfinder interface before executing the function mode. For example, when the mobile phone receives a user's operation on the portrait control, such as a click operation, the viewfinder interface before performing the portrait shooting function may be displayed. At the same time, the portrait control is in the selected state. When the mobile phone receives the user's operation on the recording control, for example, when the operation is clicked, the viewfinder interface before performing the recording function can be displayed. At the same time, the recording control is in the selected state.
  • the viewfinder interface before performing the multi-mirror video function can be displayed, that is, as shown in Figure 2 In the interface 205 shown in (c), at the same time, the multi-mirror video control is in the selected state.
  • the above-mentioned interface 205 is an example of the fourth interface, and in some other embodiments, the viewfinder interface before performing the conventional video recording function may also be called the fourth interface.
  • both the recording function and the multi-camera recording function can record video data, and the difference between them is that when the recording starts, the enabled lens modes are different.
  • the mobile phone Under the video recording function, the mobile phone can respond to the user's operation and enable single-camera mode such as single-front mode or single-rear mode to shoot video.
  • the mobile phone Under the multi-camera video recording function, the mobile phone can respond to the user's operation and enable multi-camera lens modes such as front-to-back mode, back-to-back mode, or picture-in-picture to shoot videos.
  • the method provided by the embodiment of the present application is not only applicable to the video data captured under the video recording function, but also applicable to the video data captured under the multi-mirror video recording function, and the realization principle is the same. In the subsequent embodiments, the method provided in the embodiments of the present application is mainly introduced by taking the multi-camera video recording function as an example.
  • the viewfinder interface (that is, interface 205 ) displayed by the mobile phone includes a plurality of viewfinder frames, such as viewfinder frame 206 and viewfinder frame 207 .
  • the arrangement position relationship of the viewfinder frame 206 and the viewfinder frame 207 is related to the posture of the mobile phone. For example, in a scene where the gyroscope sensor of the mobile phone recognizes that the mobile phone is in a vertical screen state, the viewfinder frame 206 and the viewfinder frame 207 are arranged up and down. In a scenario where the gyro sensor of the mobile phone recognizes that the mobile phone is in a landscape orientation, the viewfinder frame 206 and the viewfinder frame 207 are arranged left and right.
  • the viewfinder frame 206 and the viewfinder frame 207 respectively correspond to cameras.
  • the viewfinder frame 206 corresponds to the camera 1 (eg, the rear camera a), so the viewfinder frame 206 can be used to display the video stream uploaded by the camera 1 .
  • the viewfinder frame 207 corresponds to the camera 2 (for example, a front camera), so that the viewfinder frame 207 can be used to display the video stream uploaded by the camera 2 .
  • the camera corresponding to each viewfinder frame eg, viewfinder frame 206 and viewfinder frame 207
  • the camera corresponding to each viewfinder frame can be adjusted according to the user's operation. After the camera corresponding to each viewfinder frame changes, it means that the lens mode used by the mobile phone also changes accordingly.
  • the mobile phone can receive the user's operation 2, and the operation 2 is used to trigger the mobile phone to directly start video shooting without selecting any special effects.
  • This operation 2 may also be referred to as the third operation.
  • the interface 205 includes a control 208 for instructing to start shooting, that is, a first control.
  • the mobile phone receives the user's third operation on the control 208.
  • a viewfinder interface that is recording a video can be displayed, such as the fifth interface, for example, the interface 209 shown in (d) in FIG. 2 .
  • the interface 209 is a recording and framing interface in the first lens mode, for example, it is a recording and framing interface corresponding to the front and rear modes.
  • Interface 209 also includes frame 206 and frame 207 .
  • the interface 209 of the mobile phone can display the video streams collected in real time by the front camera and the rear camera a.
  • the mobile phone can also render and merge the video streams captured by the front camera and the rear camera a, and then encode, generate and save video data.
  • the video frame of the video data will gradually increase.
  • the mobile phone may receive the user's operation 3 in the interface 209, and the above operation 3 may be an operation indicating to switch the camera mode.
  • the mobile phone can enable different cameras or combinations of different cameras to collect video streams, so that users can create videos with various scenes and rich content.
  • the interface 209 may include a control for instructing to switch the lens mode, which is also called a second control, such as the control 301 shown in (a) of FIG. 3 .
  • the icon of the control 301 is used to indicate the currently enabled lens mode.
  • the window 302 lists multiple optional lens modes, such as front-to-back mode, rear-to-back mode, picture-in-picture mode, single-back mode, single-back mode, and the like. In window 302, the front and back mode is selected.
  • the mobile phone may receive a user's selection operation of the rear-to-rear mode, the picture-in-picture mode, the single-rear mode or the single-rear mode, and switch the used lens mode in response to the selection operation.
  • the above-mentioned operation for realizing switching the lens mode may be referred to as a fourth operation.
  • the mobile phone may switch the lens mode to the picture-in-picture mode. That is, as (c) in FIG. 3 , the mobile phone can switch the display interface 303 .
  • the interface 303 is an example of the sixth interface. It can be understood that the sixth interface refers to the corresponding video recording interface after switching the camera mode (also called the second camera mode).
  • the interface 303 also includes a viewing frame 206 and a viewing frame 207 . Wherein, the viewfinder frame 206 will continue to be used to display the video stream uploaded by the rear camera a, and the viewfinder frame 207 will also continue to be used to display the video stream uploaded by the front camera.
  • the viewfinder frame 207 shrinks. Instead, viewfinder frame 206 increases. In addition, the viewfinder frame 207 is superimposed on the viewfinder frame 206 .
  • the camera parameters of the front camera and rear camera a are also adjusted. After adjusting the camera parameters, the video streams captured by the front camera and rear camera a may not be uploaded in time. This will cause a pause segment to appear in the captured video data 1 . That is, as shown in (c) in FIG. 3 , the viewfinder frame 206 and the viewfinder frame 207 of the interface 303 will appear a short black screen.
  • FIG. 4 conceptually shows the impact on the video data 1 after the lens mode is switched (from the front-to-back mode to the picture-in-picture mode) during the video shooting process of the mobile phone.
  • the relative time axis of video data 1 between time 00:00 and 00:05, both the front camera and rear camera a can upload video streams normally, and at the same time, the mobile phone can also upload video streams based on the front camera and rear camera a.
  • the uploaded video stream is normally encoded to generate continuous multi-frame video frames, also called video segment 1.
  • the relative time axis is the time axis created based on the video data 1 .
  • the time 00:00 of the relative time axis corresponds to the acquisition time of the first video frame (also referred to as the first frame) of the video data 1 .
  • the mobile phone After time 00:05, the mobile phone detects an operation indicating to switch to the picture-in-picture mode, and this time 00:05 is an example of the first time point. This operation can temporarily affect the video stream return of the front camera and rear camera a, so the problem of missing video frames occurs between time 00:05 and time 00:06. After time 00:06, the front camera and the rear camera a return to normal, and they can be collected and encoded to obtain continuous video frames, which can also be called video segment 2.
  • the mobile phone may insert multiple frames to replace frame 1 between time 00:05 and time 00:06, so as to obtain a coherent flow of video data 1 .
  • the substitute frame 1 inserted after time 00:05 (that is, the first time point) may also be referred to as a video frame corresponding to the first time point.
  • the mobile phone may freeze the video frame 401 (the last frame in the video segment 1) to obtain the substitute frame 1 .
  • the screen content displayed in the substitute frame 1 is the same as that of the video frame 401 .
  • the mobile phone cancels the frame freeze of the video frame 401 .
  • the mobile phone may insert a preview configured image frame after the video frame 401, for example, a black image frame or a white image frame, until the camera receives the image again. After the returned video stream, stop inserting alternate frame 1. Understandably, the inserted image frames may also be collectively referred to as a substitute frame 1 .
  • the mobile phone may mark the video frame 401 . In this way, after the shooting of the video data 1 is completed, the mobile phone automatically inserts multiple frames to replace the frame 1 after the video frame 401 .
  • the substitute frame 1 may be the same image frame as the video frame 401, or may be a pre-configured white image frame or black image frame.
  • the mobile phone may receive the user's selection operation of the back-to-back mode. In this way, the mobile phone can switch the used lens mode to the rear mode in response to the selection operation. That is, as (b) in FIG. 6 , the mobile phone can switch the display interface 601 .
  • the interface 601 also includes a viewfinder frame 206 and a viewfinder frame 207 . Wherein, the viewfinder frame 206 is used to display the video stream uploaded by the rear camera a, and the viewfinder frame 207 is used to display the video stream uploaded by the rear camera b.
  • the lens mode can also be set to Switch to back-to-back mode.
  • the mobile phone activates the rear camera b, and turns off the front camera, so that the rear camera b corresponds to the viewfinder frame 207 superior.
  • the viewfinder frame 206 of the mobile phone continues to display the video stream uploaded by the rear camera a, while the viewfinder frame 207 displays the video stream uploaded by the rear camera b.
  • the video stream returned by the rear camera a is not affected, as shown in (b) in FIG. 6 , the image display of the viewfinder frame 206 is not affected during the lens switching.
  • the picture display of the viewfinder frame 207 will be affected, as shown in (b) in FIG. 6 , a black screen appears briefly in the viewfinder frame 207 .
  • the interface 601 is also an example of the sixth interface.
  • FIG. 7A conceptually shows an example diagram of generating video data 1 in the case of lens mode switching (switching from front-to-back mode to back-to-back mode) during the process of shooting video by the mobile phone.
  • the mobile phone may mark the video frame with dots after instructing to activate the rear camera b. After it is determined that the rear-facing camera b has returned the video stream normally, stop marking. In this way, as shown in (b) in FIG. 7A , the mobile phone can remove the video frame corresponding to the marked position, so that video clip 1 and video clip 2 can be obtained. Then, in order to ensure the continuity of the video data, as shown in (c) in Figure 7A, the mobile phone can also add a substitute frame 1 between the video segment 1 and the video segment 2, that is, corresponding to the first time point video frame, thus, video data 1 is obtained.
  • the last captured video frame can be frozen, for example, the video frame 701 in FIG. 7B is frame-freeze , to obtain the substitute frame 1, that is, the video frame corresponding to the first time point.
  • the frame freeze for the video frame 701 is cancelled, and the video data 1 is generated by normal encoding according to the video streams uploaded by the rear camera a and the rear camera b.
  • the addition of the substitute frame 1 mentioned in the above embodiment may be after the video shooting is completed, or during the video shooting process.
  • switching the lens mode will not interrupt the normal shooting of the video.
  • the mobile phone can display the interface 601 and shoot a video. In this way, the number of video frames corresponding to the video data 1 will continue to increase.
  • the mobile phone is enumerated to switch the lens mode from the front-to-back mode to the picture-in-picture mode, and from the front-to-back mode to the rear-to-back mode.
  • similar problems also exist in the switching between other lens modes, which can also be solved by inserting substitute frames as described in the previous example, and will not be repeated here.
  • the mobile phone may also receive an operation 4 in the interface 601 instructing the user to stop shooting. Then, the mobile phone may stop video shooting in response to the operation 4 .
  • the interface 601 further includes a control for instructing to pause shooting, which may also be called a third control, such as control 801 .
  • the mobile phone may receive a fifth operation of the user on the control 801, for example, a click operation. And in response to the click operation on the control 801, continue shooting is stopped, and the captured video data 1 is saved.
  • the interface 205 is displayed again. After the video data 1 (that is, the first video data) is captured and saved, the displayed interface 205 may also be referred to as the first interface. In this scenario, the first interface is actually a viewfinder preview interface provided by the camera application.
  • the mobile phone may exit the operation of the camera APP according to the user's instruction, such as an upward sliding operation on the interface 601, stop continuing to shoot, and save the captured video data 1 .
  • the mobile phone can also display the main interface again.
  • the mobile phone can display the captured video data 1 according to the user's operation 4 , which is convenient for the user to view or edit the video data 1 .
  • the interface 205 includes a thumbnail of the video data 1, such as an icon 802, which may also be called a first thumbnail.
  • the mobile phone may receive the user's operation on the icon 802, and in response to the operation, display a video editing interface, such as the interface 803 shown in (c) in FIG. 8 .
  • the interface 803 is used to display the video data 1 and may also be called the second interface.
  • the mobile phone can exit the operation of the camera APP according to the user's instruction, such as swipe up, and display the main interface of the mobile phone, that is, the desktop 202 again.
  • the desktop 202 also includes an icon of a gallery APP.
  • the mobile phone can receive the sixth operation for the icon 901 of the gallery APP, such as a click operation, and in response to this operation, display the image provided by the gallery APP.
  • the application interface is an interface 902 shown in (b) of FIG. 9 .
  • the mobile phone can directly receive the click operation on the icon 901 of the Gallery APP, and in response to this operation, display the application interface provided by the Gallery APP, as shown in (b) of Figure 9 ) interface 902 shown.
  • the interface 902 will display thumbnails of various picture resources and video resources, and these picture resources or video resources may be captured and stored by the mobile phone, for example, the thumbnail 903 of video data 1 (also referred to as the first thumbnail), or it can be a thumbnail of an image, video, etc. downloaded from the Internet, or it can also be a thumbnail of an image, video, etc. synchronized to the cloud.
  • the interface 902 for displaying video resource thumbnails may also be referred to as a first interface. In this scenario, the first interface is actually an application interface provided by the gallery application.
  • the mobile phone may receive a user's selection operation on any video thumbnail in the interface 902, and in response to the user's selection operation on any video thumbnail, the mobile phone may display a corresponding video editing interface.
  • the mobile phone can display the interface 803 shown in (c) in FIG. Can be called the second interface.
  • the mobile phone can play the corresponding video according to the user's operation on the video editing interface.
  • the interface 803 includes controls for instructing to play video data 1 , such as control 1001 .
  • the mobile phone receives the user's click operation on the control 1001 , the mobile phone plays the video data 1 in the interface 803 .
  • the substitute frame 1 is added to the video data 1, the picture may be still during the playback process.
  • the mobile phone adds a substitute frame to the video data 1, it can also superimpose a transition effect on multiple frames of the substitute frame 1, such as called transition effect 1 or the first transition effect.
  • the substitute frame 1 actually superimposed with the first transition effect may also be referred to as the first video frame.
  • transition effect 1 can be any type of transition effect pre-specified in the mobile phone, such as left shift transition, right shift transition, rotation transition, dissolve transition, blur transition, melting transition , black field transition, white field transition, zoom in transition, zoom out transition, up transition and down transition, etc.
  • an instruction manual is required.
  • the left transition and right transition are only applicable to the scene of vertical screen video shooting, while the up transition and down transition are only applicable to the scene of horizontal screen video shooting .
  • the upper transition effect 1 may be a transition randomly determined from the above-mentioned multiple types of transition effects.
  • the transition effect 1 can better connect the video clips before and after the replacement frame 1, that is, the transition between the video clip 1 and the video clip 2 can be better, and the viewing experience of the user can be improved. , but also increase the quality of video shooting.
  • the mobile phone can also mark the substitute frame 1 on which the transition effect is actually superimposed, so that the mobile phone can identify the position where the transition is added.
  • the mobile phone can automatically perform secondary creation on the video data 1 according to the user's operations on the video editing interface, such as configuring video music, adding transition effects, and the like.
  • the interface 803 also includes a control for instructing to edit video data 1 , such as a one-click blockbuster control 1101 .
  • the mobile phone receives the user's operation on the one-key blockbuster control 1101 , such as the second operation, the mobile phone can automatically edit the video data 1 .
  • the mobile phone can automatically edit the video data 1, which may include the following steps:
  • the mobile phone determines an effect template matching the video data 1.
  • effect templates of multiple styles can be pre-configured in the mobile phone, and each effect template corresponds to a piece of video music.
  • the effect template also corresponds to filters, special effects, transitions, and stickers.
  • the mobile phone can use an artificial intelligence model to analyze the picture content of the video data 1, and determine an effect template matching the video data 1, that is, the first effect template, also called the effect template 1.
  • the artificial intelligence model of the mobile phone searches for similar videos based on the picture content of the video data 1 . And get video music for similar videos. In this way, the corresponding effect template 1 is determined according to the acquired video music.
  • the artificial intelligence model of the mobile phone searches for similar videos based on the screen content of the video data 1 . Multiple effect templates belonging to the same style are determined according to style names of similar videos. Then, an effect template 1 is randomly determined from the determined plurality of effect templates. In some other embodiments, the mobile phone may also randomly determine an effect template as the effect template 1 .
  • the mobile phone processes the video data 1 according to the effect template 1.
  • the mobile phone can adjust the volume of the original audio track of the video data 1 to zero, and then add the video music of the shooting template 1 (that is, the first music) to the video data 1, so that the video music and the video data 1 The video screen fits.
  • the volume of the original audio track can also be adjusted to other decibel values according to the user's operation.
  • the mobile phone may add the filter, special effect, transition, and sticker corresponding to the effect template 1 to the video data 1 .
  • the mobile phone in the process of adding transition effects, can not only add new transition effects in video data 1, but also replace the original transition effects in video data 1 (for example, the transition effects superimposed on the replacement frame 1 1).
  • transition effects with higher adaptability are more suitable for the style of the effect template and the corresponding video music.
  • a transition effect with a lower degree of adaptation is relatively less suitable for the style of the effect template and the corresponding video music.
  • the mobile phone may first identify whether there is a transition effect 1 in the video data 1 . For example, it can be detected by detecting whether there is a mark in the video data 1 . If the mark is identified, the marked video frame (that is, the substitute frame 1 superimposed with the transition effect 1 ) is deleted. In this way, the video data 1 is divided into a video segment 1 and a video segment 2 . Then, the mobile phone generates multiple substitute frames 2 and multiple substitute frames 3 . Wherein, the substitute frame 2 may be the same as the last frame of the video clip 1, and the substitute frame 3 may be the same as the first frame of the video clip 2. The last frame of video clip 1 and the first frame of video clip 2 may be collectively referred to as the second video frame.
  • the substitute frame 2 may be an image frame obtained after the mobile phone freezes the last frame of the video segment 1 .
  • the above-mentioned substitute frame 2 may be an image frame obtained after the mobile phone freezes the first frame of the video segment 2 .
  • the total number of replacement frames 2 and 3 is the same as the number of deleted video frames, ensuring that the length of the final video data 1 is not affected. Then, the mobile phone determines the transition effect 2 according to the degree of adaptation between the effect template 1 and various transition effects, and superimposes the transition effect 2 between the replacement frame 2 and the replacement frame 3 to realize the connection of the video clip 1 and video clip 2.
  • the degree of adaptation between the effect template and the transition effect can be quantified as a matching weight.
  • the mobile phone can combine the matching weights between the effect template 1 and each transition effect, and randomly select the transition effect 2 that matches the effect template 1 from multiple types of transition effects. Understandably, a transition effect with a higher matching weight is relatively more likely to be selected as the transition effect 2 that matches the effect template 1 . The transition effect with lower matching weight is relatively more difficult to be selected as the transition effect 2 that matches the effect template 1.
  • the matching weights of effect templates and transition effects can be preconfigured. Exemplarily, as shown in Table 1 below:
  • the above table 1 exemplifies the corresponding relationship between different effect templates and video music, style, and matching weights of different transition effects.
  • the percentage value corresponding to each transition effect in the table is the matching weight between the transition effect and the effect template.
  • the style corresponding to effect template 1 is also called the first style.
  • the matching weight between the effect template and the dissolve transition is 50%, that is, the dissolve transition has a 50% probability of being selected as a matching transition effect.
  • the matching weight between the effect template and the blur transition is 0%, that is, the blur transition will not be selected as a matching transition effect.
  • the matching weight between this effect template and the melting transition is 0%, that is, the melting transition will not be selected as a matching transition effect.
  • the matching weight between the effect template and the upward transition is 50%, that is, in the scene where the mobile phone needs to process the horizontal screen video data 1, the upward transition has a 50% probability of being selected as the matching transition effect .
  • the matching weight between the effect template and the down transition is 50%.
  • the down transition has a 50% probability of being selected as the matching transition effect.
  • the matching weight between this effect template and the left-shift transition is 50%, that is, in the case of the vertical screen video data 1 that the mobile phone needs to process, the left-shift transition has a 50% probability of being selected as the matching transition effect .
  • the matching weight between the effect template and the right-shift transition is 50%.
  • the right-shift transition has a 50% probability of being selected as the matching transition effect.
  • the matching weight between the effect template and the black transition is 90%, that is, the black transition has a 90% probability of being selected as a matching transition effect.
  • the matching weight between the effect template and the white transition is 90%, that is, the white transition has a 90% probability of being selected as a matching transition effect.
  • the matching weight between the effect template and the zoom-in transition is 90%, that is, the zoom-in transition has a 90% probability of being selected as a matching transition effect.
  • the matching weight between the effect template and the zoom-out transition is 90%, that is, the zoom-out transition has a 90% probability of being selected as a matching transition effect.
  • the matching weight between the effect template and the rotation transition is 30%, that is, the rotation transition has a 30% probability of being selected as a matching transition effect.
  • the mobile phone can randomly select a transition effect 2 to replace the transition effect 1 by using the matching weights corresponding to each transition effect.
  • This selection method not only has high flexibility, but also can ensure that there is a high probability of a high correlation between the selected transition effect 2 and the effect template 1.
  • the mobile phone adds a new transition effect scene in the video data 1, it can also use the matching weights corresponding to each transition effect to randomly determine the transition effect 3 (also called the third transition effect 3) that matches the effect template 1. special effect), and add the transition effect 3 to the video data 1, for example, to the video frame corresponding to the second time point in the video data 1.
  • the video frame corresponding to the second time point may be: in the video data 1, video frames located before and after the second time point. In this way, the style that can be presented by the processed video data 1 can be closer to the expected effect of the effect model.
  • the effect template 1 may also have associated identifiers with one or more transition effects.
  • the transition effects with the associated identifiers may be preferentially selected as transition effects. field effects2.
  • the mobile phone after processing the video data 1 with the matching effect template 1, can also change the effect template used when processing the video data 1 or separately change the used video music according to the user's operation.
  • the mobile phone may display a preview interface, such as interface 1401 , which is called a third interface.
  • the third interface is used to display video data 2, also called second video data, where video data 2 is a video obtained after processing effect templates on the basis of video data 1.
  • the second video data includes video frames in video data 1, substitute frames (eg, substitute frame 2 and substitute frame 3), video music corresponding to effect template 1 (eg, referred to as the first music) and transition effects 2 (Also known as the second transition effect).
  • the mobile phone can receive the user's operation 6 on the interface 1401 , for example, the operation of clicking the style control 1402 on the interface 1401 .
  • an interface 1403 is displayed.
  • the interface 1403 is a guide interface guiding the user to select an effect template.
  • the interface 1403 includes multiple template windows indicating different effect templates. For example, window 1404, window 1405, window 1406, and window 1407.
  • the above-mentioned window 1404 is used to indicate the effect template named Hello Summer
  • window 1405 is used to indicate the effect template named sunny
  • window 1406 is used to indicate the effect template named HAPPY
  • window 1407 is used to indicate the effect template named Xiaomeimei effect template.
  • the template window of the hello summer effect template is selected, indicating that effect template 1 is the hello summer effect template.
  • the mobile phone can determine the effect template 2 selected by the user according to the user's operations on other template windows. For example, after the mobile phone receives the user's click operation on the window 1405, the preview window 1408 in the interface 1403 may display a sample of the sunny effect template. In this way, if the mobile phone receives the user's operation on the control 1409 in the interface 1403, it can determine that the sunny effect template is the selected effect template 2. Afterwards, the mobile phone can use the effect template 2 to process the original video data 1 to obtain the video data 1 conforming to the style of the effect template 2, and as shown in (c) in FIG. 14, the interface 1401 is displayed again. Video data 1 processed based on effect template 2.
  • the mobile phone receives the user's operation on the control 1409 when it does not receive the user's operation indicating to select another effect template.
  • the mobile phone can determine that the user indicates to use the effect template 1 to reprocess the original video data 1 .
  • the matching weights between each transition effect and the effect template 1 can still be used to re-determine the transition effect 2 and the transition effect 3 randomly, and use them to reprocess the original video data 1 .
  • the rerandomized transition effects 2 and 3 may be different from the transition effects determined by using the effect template 1 for the first time. In this way, the visual effect of the video data 1 obtained after reprocessing will also be different, which improves the diversity of one-click filming.
  • the mobile phone may receive the user's operation 7 on the interface 1401 , for example, the operation of clicking the music control 1410 on the interface 1401 . And in response to the operation 7, different video music is replaced.
  • the video music to be replaced can be the same style of music corresponding to the effect template 2, or a random piece of music, which is not limited.
  • the interface 1401 also includes a control 1411 indicating confirmation.
  • the mobile phone receives the user's operation on the control 1411, such as after the click operation, saves the processed video data 1, for example, the video data 1 processed based on the effect template is also called video data2.
  • the mobile phone can also display a video editing interface corresponding to the video data 2, such as the interface 1412 shown in (d) in FIG. 14 .
  • the interface 1412 can display video data 2 .
  • the mobile phone can play the video data 2 according to the user's operation on the interface 1412 .
  • the interface 1401 may also include a control indicating to undo the effect template, for example, the control 1413 shown in (a) of FIG. 14 .
  • the mobile phone may receive the user's click operation on the control 1413, delete the video data 1 processed based on the effect template, and display the interface 803 again.
  • the interface 803 still includes the one-key blockbuster control 1101 . If the mobile phone receives the user's operation on the one-key blockbuster 1101 again, it can determine a matching effect template again, and use the newly determined effect template to process the video data 1 again.
  • the processing process can refer to the foregoing embodiments, and will not be repeated here. repeat.
  • the mobile phone can make the effect templates determined by the one-click blockbuster two adjacent times different, increasing the diversity of the second creation of video data.
  • the embodiment of the present application also provides an electronic device, and the electronic device may include: a memory and one or more processors.
  • the memory is coupled to the processor.
  • the memory is used to store computer program code comprising computer instructions.
  • the electronic device can be made to perform various steps performed by the mobile phone in the foregoing embodiments.
  • the electronic device includes, but is not limited to, the aforementioned memory and one or more processors.
  • the structure of the electronic device may refer to the structure of the mobile phone shown in FIG. 1 .
  • the chip system includes at least one processor 2201 and at least one interface circuit 2202 .
  • the processor 2201 may be a processor in the aforementioned electronic device.
  • the processor 2201 and the interface circuit 2202 may be interconnected through wires.
  • the processor 2201 can receive and execute computer instructions from the memory of the above-mentioned electronic device through the interface circuit 2202 .
  • the electronic device can be made to perform various steps performed by the mobile phone in the above-mentioned embodiments.
  • the chip system may also include other discrete devices, which is not specifically limited in this embodiment of the present application.
  • Each functional unit in each embodiment of the embodiment of the present application may be integrated into one processing unit, or each unit may physically exist separately, or two or more units may be integrated into one unit.
  • the above-mentioned integrated units can be implemented in the form of hardware or in the form of software functional units.
  • the integrated unit is realized in the form of a software function unit and sold or used as an independent product, it can be stored in a computer-readable storage medium.
  • the technical solution of the embodiment of the present application is essentially or the part that contributes to the prior art or all or part of the technical solution can be embodied in the form of a software product, and the computer software product is stored in a storage
  • the medium includes several instructions to enable a computer device (which may be a personal computer, server, or network device, etc.) or a processor to execute all or part of the steps of the methods described in the various embodiments of the present application.
  • the aforementioned storage medium includes: flash memory, removable hard disk, read-only memory, random access memory, magnetic disk or optical disk, and other various media that can store program codes.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Human Computer Interaction (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Studio Devices (AREA)
  • Television Signal Processing For Recording (AREA)

Abstract

La présente demande concerne le domaine technique des terminaux et fournit un procédé de traitement de transition pour des données vidéo et un dispositif électronique, destinés à être utilisés dans la résolution du problème de faible efficacité d'interaction homme-ordinateur pendant l'édition vidéo. La solution spécifique consiste à : afficher une première interface, la première interface comprenant une première vignette de premières données vidéo, les premières données vidéo comprenant un premier effet spécial de transition et le premier effet spécial de transition étant superposé sur de multiples premières trames vidéo consécutives dans les premières données vidéo ; recevoir une première opération d'un utilisateur pour la première vignette ; afficher une deuxième interface en réponse à la première opération, la deuxième interface comprenant une commande de superproduction à une seule touche ; et lors de la réception d'une seconde opération de l'utilisateur pour la commande de superproduction à une seule touche, afficher une troisième interface, la troisième interface étant utilisée pour afficher des secondes données vidéo, les secondes données vidéo comprenant des trames vidéo des premières données vidéo, de multiples trames alternatives, une première musique et un second effet spécial de transition correspondant à la première musique, le second effet spécial de transition étant superposé sur les multiples trames alternatives et les trames alternatives étant utilisées pour remplacer les premières trames vidéo dans les premières données vidéo.
PCT/CN2022/094793 2021-06-16 2022-05-24 Procédé de traitement de transition pour données vidéo et dispositif électronique WO2022262537A1 (fr)

Priority Applications (2)

Application Number Priority Date Filing Date Title
US18/257,018 US20240106967A1 (en) 2021-06-16 2022-05-24 Video Data Transition Processing Method and Electronic Device
EP22824018.0A EP4240011A4 (fr) 2021-06-16 2022-05-24 Procédé de traitement de transition pour données vidéo et dispositif électronique

Applications Claiming Priority (6)

Application Number Priority Date Filing Date Title
CN202110676709.3 2021-06-16
CN202110676709 2021-06-16
CN202111439351.9 2021-11-29
CN202111439351 2021-11-29
CN202210056943.0A CN115484424B (zh) 2021-06-16 2022-01-18 一种视频数据的转场处理方法及电子设备
CN202210056943.0 2022-01-18

Publications (1)

Publication Number Publication Date
WO2022262537A1 true WO2022262537A1 (fr) 2022-12-22

Family

ID=84420486

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2022/094793 WO2022262537A1 (fr) 2021-06-16 2022-05-24 Procédé de traitement de transition pour données vidéo et dispositif électronique

Country Status (4)

Country Link
US (1) US20240106967A1 (fr)
EP (1) EP4240011A4 (fr)
CN (1) CN115484424B (fr)
WO (1) WO2022262537A1 (fr)

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2006279968A (ja) * 2006-04-05 2006-10-12 Hitachi Ltd 映像アクセス装置及び映像アクセスプログラムを記録した記録媒体
CN104184960A (zh) * 2014-08-19 2014-12-03 厦门美图之家科技有限公司 一种对视频文件进行特效处理的方法
JP2015219817A (ja) * 2014-05-20 2015-12-07 オリンパス株式会社 表示装置、表示方法、およびプログラム
WO2016124095A1 (fr) * 2015-02-04 2016-08-11 腾讯科技(深圳)有限公司 Procédé, appareil et terminal de génération de vidéo
CN107888988A (zh) * 2017-11-17 2018-04-06 广东小天才科技有限公司 一种视频剪辑方法及电子设备
US20180176481A1 (en) * 2016-12-21 2018-06-21 Samsung Electronics Co., Ltd. Method for producing media file and electronic device thereof
CN111866404A (zh) * 2019-04-25 2020-10-30 华为技术有限公司 一种视频编辑方法及电子设备

Family Cites Families (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110072070B (zh) * 2019-03-18 2021-03-23 华为技术有限公司 一种多路录像方法及设备、介质
CN111835986B (zh) * 2020-07-09 2021-08-24 腾讯科技(深圳)有限公司 视频编辑处理方法、装置及电子设备

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2006279968A (ja) * 2006-04-05 2006-10-12 Hitachi Ltd 映像アクセス装置及び映像アクセスプログラムを記録した記録媒体
JP2015219817A (ja) * 2014-05-20 2015-12-07 オリンパス株式会社 表示装置、表示方法、およびプログラム
CN104184960A (zh) * 2014-08-19 2014-12-03 厦门美图之家科技有限公司 一种对视频文件进行特效处理的方法
WO2016124095A1 (fr) * 2015-02-04 2016-08-11 腾讯科技(深圳)有限公司 Procédé, appareil et terminal de génération de vidéo
US20180176481A1 (en) * 2016-12-21 2018-06-21 Samsung Electronics Co., Ltd. Method for producing media file and electronic device thereof
CN107888988A (zh) * 2017-11-17 2018-04-06 广东小天才科技有限公司 一种视频剪辑方法及电子设备
CN111866404A (zh) * 2019-04-25 2020-10-30 华为技术有限公司 一种视频编辑方法及电子设备

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
See also references of EP4240011A4

Also Published As

Publication number Publication date
CN115484424A (zh) 2022-12-16
CN115484424B (zh) 2024-07-02
EP4240011A1 (fr) 2023-09-06
EP4240011A4 (fr) 2024-04-24
US20240106967A1 (en) 2024-03-28

Similar Documents

Publication Publication Date Title
CN113747085B (zh) 拍摄视频的方法和装置
WO2021223500A1 (fr) Procédé et dispositif photographique
CN113099146B (zh) 一种视频生成方法、装置及相关设备
WO2022252660A1 (fr) Procédé de captation vidéo et dispositif électronique
US20230043815A1 (en) Image Processing Method and Electronic Device
CN108513069B (zh) 图像处理方法、装置、存储介质及电子设备
CN113705665B (zh) 图像变换网络模型的训练方法和电子设备
EP4258632A1 (fr) Procédé de traitement vidéo et dispositif associé
EP4273684A1 (fr) Procédé de photographie et dispositif électronique
WO2023035921A1 (fr) Procédé d'instantané d'image dans un enregistrement vidéo, et dispositif électronique
CN115689963A (zh) 一种图像处理方法及电子设备
CN114520886A (zh) 一种慢动作录像方法及设备
WO2022083325A1 (fr) Procédé de prévisualisation photographique, dispositif électronique, et support de stockage
CN108259767B (zh) 图像处理方法、装置、存储介质及电子设备
WO2024109224A1 (fr) Procédé de recommandation de mode photographique
CN115484423A (zh) 一种转场特效添加方法及电子设备
CN108495038B (zh) 图像处理方法、装置、存储介质及电子设备
WO2022262537A1 (fr) Procédé de traitement de transition pour données vidéo et dispositif électronique
WO2023036007A1 (fr) Procédé d'acquisition d'image et dispositif électronique
CN115484400B (zh) 一种视频数据处理方法及电子设备
EP4290874A1 (fr) Procédé de traitement vidéo et dispositif électronique
CN114285963B (zh) 多镜头视频录制方法及相关设备
CN115623319B (zh) 一种拍摄方法及电子设备
CN116055863B (zh) 一种相机的光学图像稳定装置的控制方法及电子设备
CN115484425A (zh) 一种转场特效的确定方法及电子设备

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 22824018

Country of ref document: EP

Kind code of ref document: A1

ENP Entry into the national phase

Ref document number: 2022824018

Country of ref document: EP

Effective date: 20230530

WWE Wipo information: entry into national phase

Ref document number: 18257018

Country of ref document: US

NENP Non-entry into the national phase

Ref country code: DE