WO2024078238A1 - Procédé de commande d'enregistrement vidéo, dispositif électronique et support - Google Patents

Procédé de commande d'enregistrement vidéo, dispositif électronique et support Download PDF

Info

Publication number
WO2024078238A1
WO2024078238A1 PCT/CN2023/118317 CN2023118317W WO2024078238A1 WO 2024078238 A1 WO2024078238 A1 WO 2024078238A1 CN 2023118317 W CN2023118317 W CN 2023118317W WO 2024078238 A1 WO2024078238 A1 WO 2024078238A1
Authority
WO
WIPO (PCT)
Prior art keywords
image
recording
content
camera
instruction
Prior art date
Application number
PCT/CN2023/118317
Other languages
English (en)
Chinese (zh)
Inventor
卞超
Original Assignee
华为技术有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 华为技术有限公司 filed Critical 华为技术有限公司
Publication of WO2024078238A1 publication Critical patent/WO2024078238A1/fr

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/44Local feature extraction by analysis of parts of the pattern, e.g. by detecting edges, contours, loops, corners, strokes or intersections; Connectivity analysis, e.g. of connected components
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/74Image or video pattern matching; Proximity measures in feature spaces
    • G06V10/75Organisation of the matching processes, e.g. simultaneous or sequential comparisons of image or video features; Coarse-fine approaches, e.g. multi-scale approaches; using context analysis; Selection of dictionaries
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/20Movements or behaviour, e.g. gesture recognition
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M1/00Substation equipment, e.g. for use by subscribers
    • H04M1/72Mobile telephones; Cordless telephones, i.e. devices for establishing wireless links to base stations without route selection
    • H04M1/724User interfaces specially adapted for cordless or mobile telephones
    • H04M1/72403User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality
    • H04M1/7243User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality with interactive means for internal management of messages
    • H04M1/72439User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality with interactive means for internal management of messages for image or video messaging
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M1/00Substation equipment, e.g. for use by subscribers
    • H04M1/72Mobile telephones; Cordless telephones, i.e. devices for establishing wireless links to base stations without route selection
    • H04M1/724User interfaces specially adapted for cordless or mobile telephones
    • H04M1/72448User interfaces specially adapted for cordless or mobile telephones with means for adapting the functionality of the device according to specific conditions
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N5/00Details of television systems
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N5/00Details of television systems
    • H04N5/76Television signal recording
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N7/00Television systems

Definitions

  • the present application relates to the field of recording technology, and in particular to a video recording control method, electronic equipment and medium.
  • the present application provides a video recording control method, electronic device and medium.
  • the present application provides a video recording control method, which is applied to an electronic device, and the method includes: in response to a first operation of a user to start recording, performing recording at least through a camera, and the recorded content includes at least video stream content; identifying image instructions input by the user through the camera in the video stream content, and the image instructions are used to achieve control over the recording; wherein, identifying image instructions input by the user through the camera in the video stream content includes: identifying at least a first image block of at least one image frame in the video stream content, identifying an image instruction that matches a characteristic behavior in the first image block, and determining a first time interval where the image instruction is located based on at least one image frame where the image instruction is located.
  • the first time interval where the image instruction is located can be determined, so that it is convenient to determine the second time interval including the first time interval based on the first time interval where the image instruction is located, and obtain the corresponding splicing content after deleting the second time interval in the recorded content, so as to realize recording control through image instructions during the recording process, meet the control requirements in video recording, and obtain the first recording file without image instructions, so as to facilitate the recording of scenes such as single-person recording and live broadcast, and improve the user experience.
  • the image instructions used for recording control can include image instructions in multiple forms, so that users can use different instruction forms to control the recording process in different scenes.
  • the recognition of image instructions can be realized by recognizing partial image blocks of image frames in the video stream content, effectively improving the recognition efficiency.
  • the characteristic behavior may be the image instruction content corresponding to each operation stored in the electronic device, which is used to match the image input by the user during the recording process to identify the image instruction input by the user during the recording process.
  • image blocks of an image frame in the video stream content may be identified, or all image blocks of an image frame in the video stream content may be identified.
  • An image block may be an image region of a set size.
  • determining the first time interval where the image instruction is located based on at least one image frame where the image instruction is located includes: at least one image frame includes a first image frame and a second image frame, the first image frame is the first image frame where the image instruction matches the characteristic behavior in the first image block, the second image frame is the last image frame where the image instruction matches the characteristic behavior in the first image block, the start time of the first time interval is the moment where the first image frame is located, and the end time of the first time interval is the moment where the second image frame is located.
  • the electronic device can compare each image frame in the recording process with each image frame in the image instruction content (or characteristic behavior) corresponding to the operation in the image resource library.
  • image frame from the beginning to the end of an image instruction that matches the image frame from the beginning to the end of any image instruction content stored in the resource library, that is, it is completely consistent, then it is confirmed that the corresponding image instruction is recognized, and it can be Execute the operation corresponding to the image instruction.
  • the first image frame in the image command input by the user that can match the image command content of any image command content in the resource library, that is, the first image frame can have its corresponding time point marked as the starting time point of the first time interval in which the image command is located
  • the last image frame that can match the image command content in the resource library, that is, the second image frame can have its corresponding time point marked as the ending time point of the first time interval of the image command.
  • the method before determining the image instruction that matches the characteristic behavior in the first image block, the method also includes: identifying a second image block of at least one image frame in the video stream content, and switching from the second image block to the first image block when no image instruction is identified.
  • the first image block is larger than the second image block, and the first image block includes the second image block.
  • the second image block is located at the center of the image frame, and the first image block extends in all directions based on the second image block.
  • the second image block may be an image block located at the center of the image frame, that is, for example, it may be the first center image block mentioned in the embodiment of the present application, and the first image block may be an image block of a set size extending from the image block at the center of the image frame to the surrounding areas, for example, it may be the second center image block mentioned in the embodiment of the present application.
  • the core information of the image generally only occupies a local area of the entire image, and has a higher probability of occupying the central area.
  • the images will generally match. Therefore, in the embodiment of the present application, a local comparison method in which the center position image block is gradually expanded is used to match the image. This can improve the matching efficiency while ensuring the matching accuracy.
  • the first image block is adjacent to the second image block.
  • the first image block may be an image block adjacent to the right side of the second image block, or an image block adjacent to the left side, top side, bottom side, or diagonal direction of the second image block.
  • the next image block arranged after the second image block may be determined according to a preset switching sequence, and each image block may be identified and matched according to the switching sequence.
  • the first image block is the next image block arranged after the second image block according to a switching sequence
  • the switching sequence includes preset positions of image blocks of different sequences.
  • the core information of the image generally only occupies a local area of the entire image, and when the core information matches, the images will generally match. Therefore, the local comparison method of comparing each image block in sequence according to the switching order is adopted in this application to match the image, which can improve the matching efficiency while ensuring the matching accuracy.
  • the method when the image instruction is matched with the characteristic behavior in the first image block, the method further includes: identifying a first pixel in the first image block, a display parameter difference between the first pixel and adjacent pixels being higher than a first threshold, and ignoring matching of the first pixel.
  • the reason for the mismatch in image matching results may be the existence of bad pixels (or noise points), rather than the image mismatch in the actual sense. Therefore, in the present application, pixel points whose display parameter difference with adjacent pixels is higher than the first threshold are identified, that is, bad pixels with obvious jumps, and the bad pixels are not matched, which can effectively ensure the accuracy of image matching.
  • the method when the image instruction is matched with the characteristic behavior in the first image block, the method also includes: identifying the unmatched second pixel and third pixel in the first image block, and ignoring the matching of the second pixel and the third pixel when the position difference between the second pixel and the third pixel exceeds a second threshold.
  • the two unmatched pixels when the position difference between any two unmatched pixels is too large, the two unmatched pixels can be considered as insignificant and negligible points, and the negligible points can be not matched, which can effectively ensure the accuracy of image matching.
  • the characteristic behavior matching the image instruction includes schematic content or body movements appearing in the first image block.
  • the schematic content includes specific text information or image information appearing in the first image block.
  • the recording of the video can be controlled by the schematic content, so that when the user is unable to input limb or facial images, the recording of the video can be controlled by specific text or images, such as text or images displayed on cardboard, to meet the multi-scenario requirements of video recording control.
  • the body movement includes specific gesture information or facial information appearing in the first image block.
  • performing recording through at least one camera includes: performing recording through a first camera; and performing recording through a second camera in response to an image instruction for switching cameras input by a user through the first camera.
  • the first camera is a front camera of the electronic device, and the second camera is a rear camera of the electronic device; or the first camera is a rear camera of the electronic device, and the second camera is a front camera of the electronic device.
  • the method further includes: identifying the user in the video stream content through the first camera. Image command input from the second camera.
  • the characteristic behavior matching the image instruction for switching the camera includes that a recorded image is rotated in the first image block due to the user turning the electronic device.
  • the electronic device when the electronic device detects that the recording screen is rotating, it can be judged that the user has the intention to switch the camera, and it is determined that the image instruction to switch the camera is detected, and the electronic device can be controlled to switch the camera.
  • the user does not need to input body movements or gestures, but only needs to turn the mobile phone to rotate the recording screen to control the electronic device to switch the camera, which meets the recording control needs of multiple scenes and improves the user experience.
  • the method further includes: generating a first recording file according to the recorded content, the first recording file including the spliced content after deleting the recorded content corresponding to the second time interval, and the second time interval is determined based on the first time interval where the image instruction is located.
  • determining the second time interval based on the first time interval where the image instruction is located includes: determining the third time interval according to the waveform similarity of the audio stream content before and after the second time interval; and determining the first time interval according to the third time interval.
  • the time interval where the image instruction is located can be optimized to obtain a second time interval.
  • the spliced content is generated after deleting the recorded content corresponding to the second time interval. This can further ensure that the recorded content will not mutate and effectively ensure that the generated recording file does not include the image instruction content.
  • the method further includes: generating a second recording file according to the recorded content, wherein the second recording file is marked with a start time and an end time of the second time interval.
  • the method also includes: generating a third recording file set based on the recorded content, the third recording file set including at least one first-category recording segment file corresponding to the recorded content in the second time interval and at least one second-category recording segment file of the recorded content outside the second time interval.
  • a third set of recorded files can be generated based on the recorded files marked with the start time and the end time of the second time interval, and the electronic device can store the first type of recorded clip files and the corresponding operations in the image instruction material library for subsequent recognition and matching of image instructions.
  • the electronic device can store the second type of recorded clip files, which can facilitate users to view them separately or perform editing processing such as splicing and synthesis.
  • the present application provides an electronic device, comprising: a memory for storing instructions executed by one or more processors of the electronic device, and a processor, which is one of the one or more processors of the electronic device, for executing the video recording control method mentioned in the present application.
  • the present application provides a readable storage medium having instructions stored thereon.
  • the instructions When the instructions are executed on an electronic device, the electronic device executes the video recording control method claimed in the present application.
  • the present application provides a computer program product, including: execution instructions, the execution instructions are stored in a readable storage medium, at least one processor of an electronic device can read the execution instructions from the readable storage medium, and at least one processor executes the execution instructions so that the electronic device executes the video recording control method mentioned in the present application.
  • FIG1 is a schematic diagram showing a hardware structure of an electronic device according to some embodiments of the present application.
  • FIG2a shows a schematic diagram of microphone distribution according to some embodiments of the present application.
  • FIG2b is a schematic diagram showing a sound receiving range of a microphone according to some embodiments of the present application.
  • 2c-2e respectively show a schematic diagram of the distribution of a rear camera according to some embodiments of the present application
  • FIG2f shows a schematic diagram of the distribution of front cameras according to some embodiments of the present application.
  • FIG2g is a schematic diagram showing a shooting range of a front camera and a rear camera according to some embodiments of the present application.
  • FIG3a shows a schematic diagram of a software structure of an electronic device according to some embodiments of the present application.
  • FIG3 b shows a functional schematic diagram of an image/video acquisition module according to some embodiments of the present application.
  • FIG3c shows a functional schematic diagram of an image/video acquisition module according to some embodiments of the present application.
  • FIG3d is a schematic diagram showing a method of performing independent control of a split flow in an image/video acquisition module according to some embodiments of the present application.
  • FIG3e shows a schematic diagram of a flow splitting control according to some embodiments of the present application.
  • FIG3f shows a functional schematic diagram of an operation control module according to some embodiments of the present application.
  • FIG3g shows a functional schematic diagram of an image/video recognition module according to some embodiments of the present application.
  • FIG3h shows a functional schematic diagram of an audio-video synchronization module according to some embodiments of the present application.
  • FIG3i is a schematic diagram showing a synchronous control method according to some embodiments of the present application.
  • FIG3j is a functional schematic diagram of a shooting result generating module according to some embodiments of the present application.
  • FIGS. 4a-4c are schematic diagrams showing schematic contents according to some embodiments of the present application.
  • FIG5 is a schematic diagram showing a flow chart of a video recording control method according to some embodiments of the present application.
  • FIG6 shows a schematic diagram of starting recording according to some embodiments of the present application.
  • FIG7 shows a schematic diagram of starting recording according to some embodiments of the present application.
  • FIG8 is a schematic diagram showing an audio optimization method according to some embodiments of the present application.
  • FIGS. 9a-9e are schematic diagrams showing recorded scenes according to some embodiments of the present application.
  • FIG10a is a schematic diagram showing the composition of body movements according to some embodiments of the present application.
  • FIG10b is a schematic diagram showing a partial gesture image according to some embodiments of the present application.
  • FIG11 is a schematic diagram showing the composition of the schematic content according to some embodiments of the present application.
  • FIG12 is a schematic diagram showing the composition of an image instruction matching method according to some embodiments of the present application.
  • 13a-13b are schematic diagrams showing a process of image instruction matching according to some embodiments of the present application.
  • 13c-13e are schematic diagrams showing matching situations of image instructions according to some embodiments of the present application.
  • FIG13f shows a schematic diagram of image frame matching according to some embodiments of the present application.
  • FIG13g is a schematic diagram showing a bad pixel in an image frame according to some embodiments of the present application.
  • FIG13h is a schematic diagram showing the composition of a difference point secondary analysis method according to some embodiments of the present application.
  • FIG13i is a schematic diagram showing a neglected point in an image frame according to some embodiments of the present application.
  • FIG14a is a schematic diagram showing a composition of a local image comparison method according to some embodiments of the present application.
  • FIG14b is a schematic diagram showing an arbitrary position movement comparison method according to some embodiments of the present application.
  • FIG14c is a schematic diagram showing an arbitrary position enlargement comparison method according to some embodiments of the present application.
  • FIG14d is a schematic diagram showing a fixed position alignment method according to some embodiments of the present application.
  • FIG15a shows a schematic flow chart of a method for determining a starting time point after a range is expanded according to some embodiments of the present application
  • FIG15b is a schematic flow chart of a method for determining a starting time point after a range is expanded according to some embodiments of the present application;
  • FIG16 is a schematic diagram showing a recording control system according to some embodiments of the present application.
  • FIG17 is a schematic diagram showing an implementation of a recording control method according to some embodiments of the present application.
  • the illustrative embodiments of the present application include, but are not limited to, a video recording control method, an electronic device, and a medium.
  • the electronic device 100 in the embodiment of the present application can be called a user equipment (UE), a terminal, etc.
  • the electronic device 100 can be a tablet computer (portable android device, PAD), a personal digital assistant (personal digital assistant, PDA), a handheld device with wireless communication function, a computing device, a vehicle-mounted device or a wearable device, etc.
  • the form of the terminal device is not specifically limited in the embodiment of the present application.
  • the electronic device 100 may include a processor 110, an external memory interface 120, an internal memory 121, a universal serial bus (USB) interface 130, a charging management module 140, a power management module 141, a battery 142, an antenna 1, an antenna 2, a mobile communication module 150, a wireless communication module 160, an audio module 170, a speaker 170A, a receiver 170B, a microphone 170C, an earphone interface 170D, a sensor module 180, a button 190, a motor 191, an indicator 192, a camera 193, a display screen 194, and a subscriber identification module (SIM) card interface 195, etc.
  • SIM subscriber identification module
  • the sensor module 180 may include a pressure sensor 180A, a gyroscope sensor 180B, an air pressure sensor 180C, a magnetic sensor 180D, an acceleration sensor 180E, a distance sensor 180F, a proximity light sensor 180G, a fingerprint sensor 180H, a temperature sensor 180J, a touch sensor 180K, an ambient light sensor 180L, a bone conduction sensor 180M, etc.
  • the structure illustrated in the embodiment of the present invention does not constitute a specific limitation on the electronic device 100.
  • the electronic device 100 may include more or fewer components than shown in the figure, or combine some components, or split some components, or arrange the components differently.
  • the components shown in the figure may be implemented in hardware, software, or a combination of software and hardware.
  • the processor 110 may include one or more processing units, for example, the processor 110 may include an application processor (AP), a modem processor, a graphics processor (GPU), an image signal processor (ISP), a controller, a video codec, a digital signal processor (DSP), a baseband processor, and/or a neural-network processing unit (NPU), etc.
  • AP application processor
  • GPU graphics processor
  • ISP image signal processor
  • DSP digital signal processor
  • NPU neural-network processing unit
  • Different processing units may be independent devices or integrated in one or more processors.
  • the charging management module 140 is used to receive charging input from a charger.
  • the charger may be a wireless charger or a wired charger.
  • the charging management module 140 may receive charging input from a wired charger through the USB interface 130.
  • the charging management module 140 may receive wireless charging input through a wireless charging coil of the electronic device 100. While the charging management module 140 is charging the battery 142, it may also power the electronic device through the power management module 141.
  • the power management module 141 is used to connect the battery 142, the charging management module 140 and the processor 110.
  • the power management module 141 receives input from the battery 142 and/or the charging management module 140, and supplies power to the processor 110, the internal memory 121, the display screen 194, the camera 193, and the wireless communication module 160.
  • the power management module 141 can also be used to monitor parameters such as battery capacity, battery cycle number, battery health status (leakage, impedance), etc.
  • the power management module 141 can also be set in the processor 110.
  • the power management module 141 and the charging management module 140 can also be set in the same device.
  • the wireless communication function of the electronic device 100 can be implemented through the antenna 1, the antenna 2, the mobile communication module 150, the wireless communication module 160, the modem processor and the baseband processor.
  • Antenna 1 and antenna 2 are used to transmit and receive electromagnetic wave signals.
  • Each antenna in electronic device 100 can be used to cover a single or multiple communication frequency bands. Different antennas can also be reused to improve the utilization of antennas.
  • antenna 1 can be reused as a diversity antenna for a wireless local area network.
  • the antenna can be used in combination with a tuning switch.
  • the mobile communication module 150 can provide solutions for wireless communications including 2G/3G/4G/5G, etc., applied to the electronic device 100.
  • the mobile communication module 150 may include at least one filter, a switch, a power amplifier, a low noise amplifier (LNA), etc.
  • the mobile communication module 150 may receive electromagnetic waves from the antenna 1, and perform filtering, amplification, and other processing on the received electromagnetic waves, and transmit them to the modulation and demodulation processor for demodulation.
  • the mobile communication module 150 may also amplify the signal modulated by the modulation and demodulation processor, and convert it into electromagnetic waves for radiation through the antenna 1.
  • at least some of the functional modules of the mobile communication module 150 may be arranged in the processor 110.
  • at least some of the functional modules of the mobile communication module 150 may be arranged in the same device as at least some of the modules of the processor 110.
  • the wireless communication module 160 can provide wireless communication solutions including wireless local area networks (WLAN) (such as wireless fidelity (Wi-Fi) network), bluetooth (BT), global navigation satellite system (GNSS), frequency modulation (FM), near field communication (NFC), infrared (IR) and the like applied to the electronic device 100.
  • WLAN wireless local area networks
  • BT wireless fidelity
  • GNSS global navigation satellite system
  • FM frequency modulation
  • NFC near field communication
  • IR infrared
  • the wireless communication module 160 can be one or more devices integrating at least one communication processing module.
  • the wireless communication module 160 receives electromagnetic waves via the antenna 2, modulates and filters the electromagnetic wave signals, and sends the processed signals to the processor 110.
  • the wireless communication module 160 can also receive the signal to be sent from the processor 110, modulate the frequency, amplify it, and convert it into electromagnetic waves for radiation through the antenna 2.
  • the electronic device 100 implements the display function through a GPU, a display screen 194, and an application processor.
  • the GPU is a microprocessor for image processing, which connects the display screen 194 and the application processor.
  • the GPU is used to perform mathematical and geometric calculations for graphics rendering.
  • the processor 110 may include one or more GPUs that execute program instructions to generate or change display information.
  • the display screen 194 is used to display images, videos, etc.
  • the display screen 194 includes a display panel.
  • the display panel can be a liquid crystal display (LCD), an organic light-emitting diode (OLED), an active-matrix organic light-emitting diode or an active-matrix organic light-emitting diode (AMOLED), a flexible light-emitting diode (FLED), Miniled, MicroLed, Micro-oLed, quantum dot light-emitting diodes (QLED), etc.
  • the electronic device 100 may include 1 or N display screens 194, where N is a positive integer greater than 1.
  • the external memory interface 120 can be used to connect an external memory card, such as a Micro SD card, to expand the storage capacity of the electronic device 100.
  • the external memory card communicates with the processor 110 through the external memory interface 120 to implement a data storage function. For example, files such as music and videos can be stored in the external memory card.
  • the internal memory 121 may be used to store computer executable program codes, which may include instructions.
  • 121 may include a program storage area and a data storage area.
  • the program storage area may store an operating system, an application required for at least one function (such as a sound playback function, an image playback function, etc.), etc.
  • the data storage area may store data created during the use of the electronic device 100 (such as audio data, a phone book, etc.), etc.
  • the internal memory 121 may include a high-speed random access memory, and may also include a non-volatile memory, such as at least one disk storage device, a flash memory device, a universal flash storage (UFS), etc.
  • the processor 110 executes various functional applications and data processing of the electronic device 100 by running instructions stored in the internal memory 121 and/or instructions stored in a memory provided in the processor.
  • the electronic device 100 can implement audio functions such as music playing and recording through the audio module 170, the speaker 170A, the receiver 170B, the microphone 170C, the headphone jack 170D, and the application processor.
  • the audio module 170 is used to convert digital audio information into analog audio signal output, and is also used to convert analog audio input into digital audio signals.
  • the audio module 170 can also be used to encode and decode audio signals.
  • the audio module 170 can be arranged in the processor 110, or some functional modules of the audio module 170 can be arranged in the processor 110.
  • the speaker 170A also called a "speaker" is used to convert an audio electrical signal into a sound signal.
  • the electronic device 100 can listen to music or listen to a hands-free call through the speaker 170A.
  • the receiver 170B also called a "earpiece" is used to convert audio electrical signals into sound signals.
  • the voice can be received by placing the receiver 170B close to the human ear.
  • Microphone 170C also called “microphone” or “microphone” is used to convert sound signals into electrical signals.
  • the user can make a sound by approaching the microphone 170C with his mouth, and the sound signal is input into the microphone 170C.
  • the electronic device 100 can collect sound signals, reduce noise, identify the sound source, and realize directional recording functions through the microphone.
  • the number of microphones 170C can be one or more.
  • microphone 170C includes microphone A, microphone B, and microphone C
  • the arrangement can be as shown in Figure 2a
  • microphone A can be located on the top of the mobile phone
  • microphone B can be located on the bottom of the mobile phone
  • microphone C can be located on the back of the mobile phone.
  • the sound receiving range of the three microphones can be shown in Figure 2b, the main sound receiving range of microphone A is the middle and upper part, which can be used for the scene of front and rear recording; microphone B mainly receives the sound in the middle and lower part, which can be used for the scene of front and rear recording; microphone C mainly receives the sound in the rear part, which can be used for the scene of rear recording.
  • the electronic device has only one microphone 170C, the acquired single-channel audio stream content can be backed up to achieve the acquisition of multiple copies of the audio stream content. It is understood that in some embodiments, when the electronic device has multiple microphones, the electronic device can directly acquire multiple copies of the audio stream content. In some embodiments, the electronic device has multiple microphones, and multiple copies of an audio stream acquired by one of the microphones can also be used.
  • the earphone interface 170D is used to connect a wired earphone.
  • the earphone interface 170D may be the USB interface 130, or may be a 3.5 mm open mobile terminal platform (OMTP) standard interface or a cellular telecommunications industry association of the USA (CTIA) standard interface.
  • OMTP open mobile terminal platform
  • CTIA cellular telecommunications industry association of the USA
  • the electronic device 100 can implement a recording function through an ISP, a camera 193, a video codec, a GPU, a display screen 194, and an application processor.
  • ISP is used to process the data fed back by camera 193. For example, when taking a photo, the shutter is opened, and the light is transmitted to the camera photosensitive element through the lens. The light signal is converted into an electrical signal, and the camera photosensitive element transmits the electrical signal to ISP for processing and converts it into an image visible to the naked eye. ISP can also perform algorithm optimization on the noise, brightness, and skin color of the image. ISP can also optimize the exposure, color temperature and other parameters of the recorded scene. In some embodiments, ISP can be set in camera 193.
  • the camera 193 is used to capture still images or videos.
  • the object generates an optical image through the lens and projects it onto the photosensitive element.
  • the photosensitive element can be a charge coupled device (CCD) or a complementary metal oxide semiconductor (CMOS) phototransistor.
  • CMOS complementary metal oxide semiconductor
  • the photosensitive element converts the optical signal into an electrical signal, and then passes the electrical signal to the ISP to be converted into a digital image signal.
  • the ISP outputs the digital image signal to the DSP for processing.
  • the DSP converts the digital image signal into an image signal in a standard RGB, YUV or other format.
  • the electronic device 100 may include 1 or N cameras 193, where N is a positive integer greater than 1.
  • the number of cameras can be one or more, and the arrangement of the cameras can be various.
  • the electronic device 100 can also include any number of front cameras.
  • FIG. 2f there can be three front cameras, including front camera a, front camera b, and front camera 7. Head c.
  • the recording ranges of the three rear cameras in FIG2d can be as shown in FIG2g, where the main recording range of the top rear camera 1 is the upper middle part, the main range of the rear camera 2 is the middle part, and the main range of the rear camera 3 is the lower part.
  • the recording ranges of the three front cameras in FIG2f can be as shown in FIG2g, where the main range of the front camera a is the left part, the main range of the front camera b is the middle part, and the main range of the front camera c is the right part.
  • the video stream content can be backed up to obtain multiple copies of the video stream content.
  • the electronic device has multiple cameras, and the electronic device can directly obtain multiple copies of the video stream content.
  • the electronic device has multiple cameras, and multiple copies of one video stream obtained by one of the cameras can also be used.
  • the settings of the camera or microphone in the embodiments of the present application are all examples, and the camera or microphone can be set in any way according to actual needs.
  • the software architecture of the electronic device 100 includes, from top to bottom, an application layer, an application framework layer, a hardware abstraction layer, and a kernel layer.
  • the application layer may include a series of application packages.
  • the application package may include a camera application, wherein the camera application may include a shooting mode control module, a voice acquisition module, an image/video acquisition module, an operation control module, a voice control module, an audio-visual synchronization module, and a shooting result generation module.
  • the shooting mode control module is used to control the recording mode based on user instructions, such as video recording mode, photo taking mode, dual view mode, etc.
  • the image/video acquisition module can be used to establish an image/video resource library, and during the recording process, obtain the video or image captured by the camera, and perform the shunting control of a single-channel video stream and the multi-channel control of multiple-channel video streams.
  • the image/video acquisition module can be the acquisition module in the camera mentioned in this application, wherein the image/video acquisition module performs the shunting control of a single-channel video stream and the multi-channel control of multiple-channel video streams in a manner similar to the control of the audio stream content in the voice acquisition module, which will not be repeated here.
  • the image/video resource library may include an image/video system library, a custom image/video control library, and an image control instruction material library.
  • the image/video system library is used for system preset image instruction content and corresponding operations
  • the custom image/video control library is used to store user-defined image instruction content and corresponding operations.
  • the image control instruction material library is used to store material clips and corresponding operations corresponding to the user's image instructions, and the system preset image instruction content, user-defined image instruction content, and image instruction corresponding material clips can be used as comparison data for control instructions.
  • the image/video acquisition module can be used to obtain a single-channel video captured by a single camera or multiple video streams captured by multiple cameras. It can be understood that in the embodiment of the present application, if the electronic device has only one camera, the image/video acquisition module needs to perform logical processing on the single-channel video acquired by the single camera and perform diversion, that is, the acquired single-channel video stream content can be backed up to achieve the acquisition of multiple video stream contents. Each video stream content can be subsequently operated on as needed.
  • FIG3d the main method of diversion is shown in FIG3d, which can include comprehensive diversion and intelligent diversion: among them, comprehensive diversion refers to copying the entire video file into multiple copies, and then processing them independently; intelligent diversion is to copy the effective part into multiple copies, and the effective part is mainly the recording content with the image instruction part.
  • overall diversion can be to copy the complete video stream A recorded in real time, in which case the original video stream content A can be retained, and the copied video stream content is used for image control processing, such as marking the time period corresponding to the image instruction, and removing the time period content corresponding to the image instruction.
  • Intelligent diversion can be to copy the effective parts B1, B2 and B3 of the real-time recorded video stream B. In this case, the original video stream content B can be retained, and the copied video stream content B1, B2 and B3 are used for image control processing.
  • the image/video acquisition module does not need to copy the video stream content, and several video stream contents among the multiple video stream contents acquired by multiple cameras can be set for image processing, and several video stream contents retain the original files.
  • the electronic device has three cameras, and the acquired video stream content includes video stream content A, video stream content B, and video stream content C. Then, the video stream contents A and B can be not processed, the original files can be retained, and the video stream content C can be set for image control processing, etc.
  • the video stream content A can be not processed, the original file can be retained, the video stream content B can be marked for generating multiple material clips, and the video stream content C can be marked to obtain a cropped recording file.
  • the image/video acquisition module can be used to send the acquired video to the operation control module.
  • the voice acquisition module can be used to establish a voice resource library, and during the recording process, obtain the audio collected by the microphone, and perform shunting control of a single audio stream and multi-channel control of multiple audio streams.
  • the voice resource library can include a voice system library, a custom voice control library, and a voice control instruction material library.
  • the voice system library is used to store the system preset voice instruction content and corresponding operations.
  • the custom voice control library is used to store the user-defined voice instruction content and corresponding operations.
  • the voice control instruction material library is used to store the material clips and corresponding operations corresponding to the user's voice instructions.
  • the voice acquisition module can be used to obtain single-channel audio collected by a single microphone or multi-channel audio stream content collected by multiple microphones. It can be understood that in the embodiment of the present application, if the electronic device has only one microphone, the voice acquisition module needs to perform logical processing on the single-channel audio obtained by the single microphone and perform diversion, that is, the obtained single-channel audio stream content can be backed up to achieve the acquisition of multiple audio stream contents. Each audio stream content can be subsequently operated on as needed.
  • the main methods of diversion can include comprehensive diversion and intelligent diversion: among them, comprehensive diversion refers to copying the entire audio file into multiple copies, and then processing them independently; intelligent diversion is to copy the effective part into multiple copies, and the effective part may include voice with voice command part.
  • the operation control module can be used to identify image instructions during the recording process and control the electronic device to perform corresponding operations.
  • the operation control module mainly includes an image/video recognition module and a system operation module.
  • the image/video recognition module can be used for the fuzzy recognition and precise recognition mentioned later in the embodiments of the present application, and the system operation module is used to execute the operations corresponding to the image instructions recognized by the voice image/video recognition module.
  • the execution mode may include intelligent execution and interactive execution.
  • Intelligent execution is to directly execute the operation corresponding to the image instruction after the image instruction is recognized; interactive execution means to display inquiry information after the image instruction is recognized, or to send inquiry information through voice to confirm whether the image instruction recognition is correct, such as popping up a pop-up window "Do you need to switch the camera?"
  • the user confirmation operation is detected, such as clicking a control representing confirmation, or giving an image or voice instruction representing confirmation, it is determined that the image instruction recognition is correct and the operation corresponding to the image instruction is executed.
  • the image/video recognition module can be used to obtain the correspondence between the system preset image instruction content and the operation, and to obtain the correspondence between the user-defined image instruction content and the operation.
  • the audio and video synchronization module is used to perform video processing and audio processing, that is, to perform synchronization marking or synchronization control on the time period corresponding to the control instruction in the video stream content and the audio stream content, and obtain the marked audio stream content and video stream content.
  • video processing and audio processing that is, to perform synchronization marking or synchronization control on the time period corresponding to the control instruction in the video stream content and the audio stream content, and obtain the marked audio stream content and video stream content.
  • FIG3i There are two ways of synchronization control: timely synchronization and overall synchronization.
  • Real-time synchronization means that during the recording process, after receiving each image instruction, in addition to marking the corresponding time point of the image instruction in the video stream, the image instruction time point in the audio stream content is also synchronized.
  • Overall synchronization means that after detecting the recording end instruction, the image instruction time point in the audio stream content is synchronized.
  • the shooting result generation module is used to generate a recording file based on the marked audio stream content and video stream content when the recording end instruction is detected.
  • the shooting result generation module is shown in Figure 3j, which can be used to match the marked points, remove the recorded content of the second time interval, and generate a first recording file.
  • the first recording file can be the spliced content after cutting out the content corresponding to the second time interval in the recorded content.
  • the marked points There are two ways to match the marked points. One is to match the marked video stream content with the marked audio stream content. For example, you can compare the time points in the marked video stream content and the time points in the audio stream content to see if they are consistent. If they are consistent, it proves that the match is successful. The second is to match the marked video stream content with other unprocessed (unmarked) video stream content. For example, you can use the image frames of the start and end time points in the marked video stream content, or the overall image frames between the start and end time points, to compare with the image frames of the corresponding start and end time points in the unprocessed video stream content, or the overall image frames between the start and end time points. If the image frames are consistent, it can be considered that the match is successful.
  • the content corresponding to the second time interval marked in the audio stream content and the video stream content may be removed.
  • the shooting result generation module may also be used to generate an uncropped original recording file based on the recording content, and to generate a plurality of cropped material segment files based on the recording content, which will be described in detail below.
  • the application layer can also include applications such as gallery, calendar, call, map, navigation, WLAN, Bluetooth, music, video, short message, etc.
  • the application framework layer provides application programming interface (API) and programming framework for the applications in the application layer.
  • API application programming interface
  • the application framework layer includes some predefined functions.
  • the application framework layer can include camera API, camera service and camera framework.
  • the machine frame may include a video processing module and an audio processing module.
  • the application framework layer may include a window manager, a content provider, a view system, a telephony manager, a resource manager, a notification manager, and the like.
  • the window manager is used to manage window programs.
  • the window manager can obtain the display screen size, determine whether there is a status bar, lock the screen, capture the screen, etc.
  • Content providers are used to store and retrieve data and make it accessible to applications.
  • the data may include videos, images, audio, calls made and received, browsing history and bookmarks, phone books, etc.
  • the view system includes visual controls, such as controls for displaying text, controls for displaying images, etc.
  • the view system can be used to build applications.
  • a display interface can be composed of one or more views.
  • a display interface including a text notification icon can include a view for displaying text and a view for displaying images.
  • the phone manager is used to provide communication functions for electronic devices, such as the management of call status (including answering, hanging up, etc.).
  • the resource manager provides various resources for applications, such as localized strings, icons, images, layout files, video files, and so on.
  • the notification manager enables applications to display notification information in the status bar. It can be used to convey notification-type messages and can disappear automatically after a short stay without user interaction. For example, the notification manager is used to notify download completion, message reminders, etc.
  • the notification manager can also be a notification that appears in the system top status bar in the form of a chart or scroll bar text, such as notifications of applications running in the background, or a notification that appears on the screen in the form of a dialog window. For example, a text message is displayed in the status bar, a prompt sound is emitted, an electronic device vibrates, an indicator light flashes, etc.
  • HAL Hardware Abstraction Layer It is an abstraction and encapsulation of hardware devices, providing a unified access interface for Android or Hongmeng system on different hardware devices. Different hardware manufacturers follow the HAL standard to implement their own hardware control logic, but developers do not need to care about the differences between different hardware devices, they only need to access the hardware according to the standard interface provided by HAL.
  • the kernel layer is the layer between hardware and software.
  • the kernel layer contains at least display driver, camera driver, audio driver, sensor driver, camera driver, and microphone (MIC) driver.
  • the MIC driver is used to drive the microphone in the hardware to obtain audio.
  • the camera driver is used to drive the camera to process image signals and obtain video.
  • a video recording control method is provided in an embodiment of the present application, which is used in an electronic device.
  • the method includes: obtaining the recorded content during the recording process, which may include video stream content, or video stream content and audio stream content, identifying the image instructions in the video stream content, and when it is determined that the image instructions in the video stream content match the image instruction content (or characteristic behavior) corresponding to the target operation stored in the electronic device, executing the image instructions, and marking the second time interval in the recorded content based on the first time interval where the image instructions are located, obtaining the marked recorded content, and when the end instruction is obtained, obtaining the first recorded file based on the marked recorded content.
  • the first recorded file can be the spliced content after deleting the content corresponding to the second time interval in the recorded content.
  • the image instruction content corresponding to the target operation may include schematic content.
  • the schematic content may include specific text information, and the specific text information may include text content that is consistent with the name of the operation or the key information.
  • the schematic content corresponding to the switch lens operation may be the text "switch lens".
  • the specific text information may also include fixed mode content set by the user.
  • the schematic content corresponding to the switch lens operation may be various forms of images corresponding to the number "1".
  • the schematic content may also include specific image information, and the specific image information may be an identifier representing a specific meaning.
  • the schematic content corresponding to the lens switching operation may be an identifier representing lens switching.
  • the image commands may not exist in the final recorded content, which is convenient for remote control in any way during the recording process, thereby facilitating the recording of scenes such as single-person recording and live broadcast, and improving the user experience.
  • the commands for recording control include commands in multiple forms, which is convenient for users to control the recording process in different scenes using different command forms.
  • FIG5 shows a flow chart of a video recording control method in the embodiment of the present application, wherein the video recording control method can be executed by the electronic device.
  • the video recording control method may include:
  • a recording start command is detected and video recording is started.
  • the user's recording start instruction may refer to the recording start instruction triggered by the user clicking the recording start control in the electronic device, or it may be a remote control instruction such as a voice instruction or image instruction corresponding to the start recording issued by the user.
  • the electronic device 100 when the user clicks the recording control 001 in the camera application in the electronic device 100, the electronic device 100 can detect the user's recording start instruction and perform video recording. It can be understood that in some embodiments, the electronic device can also start recording when the user clicks the recording control 001 in other applications, such as a chat application, or triggers the start recording function in other ways. For another example, as shown in FIG7 , when the user displays the "fist" image corresponding to the recording start operation, the electronic device can also detect the user's recording start instruction and perform video recording.
  • the recording start instruction in the embodiment of the present application can be any instruction that can trigger the start of recording.
  • the image acquisition module can control the electronic device to perform video recording when a recording start instruction is detected.
  • the electronic device can perform recording through a camera to obtain video stream content, or can perform recording through a camera and a microphone together to obtain video stream content and audio stream content respectively.
  • the electronic device can detect and recognize image instructions input by the user through the camera in the video stream content.
  • the electronic device can recognize the image command by detecting whether the image command in the recording process matches the image command content (or characteristic behavior) corresponding to the target operation stored in the electronic device, and can execute the target operation when the image command is recognized to control the video recording.
  • the method of identifying and matching the image command is described in detail later.
  • the image instruction content corresponding to the target operation may include schematic content, and the schematic content may include specific text information.
  • the specific text information may be understood as text content that is consistent with the name of the operation or the key information.
  • the schematic content corresponding to the switch lens operation may be the text "switch lens".
  • the specific text information may also include fixed mode content customized by the user.
  • the schematic content corresponding to the switch lens operation may be various forms of images corresponding to the number "1".
  • the schematic content may also include specific image information, and the specific image information may be an identifier representing a specific meaning.
  • the schematic content corresponding to the lens switching operation may be an identifier representing lens switching.
  • the image instruction content corresponding to the target operation may also include specific body information or specific facial information, wherein the specific body information may include specific gesture information.
  • the above target operation may include any operation such as switching cameras, adjusting focal length, etc.
  • the image instruction content corresponding to the camera switching operation may also include the rotation of the recorded image caused by turning the direction of the electronic device in the video stream content acquired by the camera.
  • the shooting direction of the mobile phone camera is the first shooting direction
  • the user wants to take a selfie that is, wants to switch to the front camera
  • the user can rotate the mobile phone to a first set angle to achieve the rotation of the recorded image, for example, the mobile phone is turned from the first shooting direction to the second shooting direction that has a first set angle with the first shooting direction.
  • the mobile phone When the mobile phone detects that the recorded image of the rear camera rotates due to the mobile phone being turned from the first shooting direction to the second shooting direction that has a first set angle with the first shooting direction, it is determined that an image instruction to switch the camera is detected, and at this time, the mobile phone is controlled to switch the shooting mode of the rear camera to the shooting mode of the front camera.
  • the user can also rotate the mobile phone to the first set angle to achieve the rotation of the recorded picture, for example, the mobile phone is turned from the third shooting direction to the fourth shooting direction that has the first set angle with the third shooting direction.
  • the mobile phone When the mobile phone detects that the recorded picture of the front camera rotates due to the rotation of the mobile phone from the third shooting direction to the fourth shooting direction that has the first set angle with the third shooting direction, it is determined that the image instruction for switching the camera is detected, and at this time, the mobile phone is controlled to switch the shooting mode of the front camera to the shooting mode of the rear camera.
  • the electronic device may perform fuzzy recognition on the image command, that is, recognize the approximate content or key information of the image, for example, when the image command contains a key part of the text content that is consistent with the stored image command content, or the image similarity reaches a set similarity, it is determined that the image command is recognized, and the operation corresponding to the image command is obtained.
  • the electronic device may recognize the image accurately, that is, when the image instruction is completely consistent with the stored image instruction content, it is determined that the image instruction is recognized and the operation corresponding to the image instruction is obtained.
  • the electronic device will only record the "switch lens” text when it is recorded.
  • the electronic device confirms that a "switch lens” image instruction is detected by using text or a two-finger close-together gesture, and executes the operation corresponding to the "switch lens” image instruction to switch the lens.
  • the electronic device may also display an inquiry message after obtaining the recognition result, or send an inquiry message through voice, so as to confirm whether the image command recognition is correct, such as popping up a pop-up window "Do you need to switch the camera?"
  • a user confirmation operation is detected, such as clicking a control representing confirmation, or giving an image command representing confirmation, etc., it is determined that the image command recognition is correct, and the operation corresponding to the image command is executed.
  • the recognition method when the recognition method is fuzzy recognition, there may be a situation where the recognized image instruction corresponds to multiple operations. At this time, a query message may be displayed to confirm the user's intention. If the user does not make a selection within a set time, the first operation set by the system may be selected by default for execution.
  • the image/video acquisition module can be used to detect image instructions during the recording process.
  • the electronic device after detecting the image instruction, the electronic device can execute the operation corresponding to the image instruction.
  • the method of marking the second time interval in the recorded content based on the first time interval in which the image instruction is located may include:
  • the electronic device can directly use the start time point and end time point of the first time interval in which the detected image instruction is located as the start time point and end time point of the second time interval in the video stream content, and mark the start time point and end time point corresponding to the second time interval in the video stream content.
  • the method of marking the second time interval in the recorded content based on the first time interval in which the image instruction is located may include:
  • the electronic device can directly use the start time point and end time point of the first time interval in which the detected image instruction is located as the start time point and end time point of the second time interval in the audio stream content and the video stream content, and mark the start time point and end time point corresponding to the second time interval in the audio stream content and the video stream content.
  • the start and end times of the first time interval in which the image instructions in the video stream content and the audio stream content are located can be optimized to obtain the optimized start and end times, and the optimized start and end times can be used as the start and end times of the second time interval.
  • the optimized start time point and end time point can be obtained by the audio optimization method, that is, the start time point and the end time point of the first time interval where the image instruction is located in the audio stream content are compared to expand the marking range, and then the start time point and the end time point after the expanded marking range are determined (that is, the start time point and the end time point corresponding to the third time interval are determined), and the start time point and the end time point after the expanded marking range are transparently transmitted to the video stream content for reverse comparison to determine whether there is an obvious conflict in the video, that is, the similarity of any two adjacent frames in each image frame of the expanded part of the video stream content (the interval in the third time interval except the first time interval) is compared to see whether they are both greater than the fifth threshold.
  • the start time point and the end time point after the expanded marking range are used as the start time point and the end time point of the second time interval. If there is less than or equal to the fifth threshold, the start time point and the end time point of the first time interval where the image instruction is located are used as the start time point and the end time point of the second time interval in the audio stream and video stream content.
  • a first recording file is generated based on the marked recording content.
  • the content corresponding to the second time interval in the marked video stream content and the marked audio stream content can be cropped, and the remaining content after cropping can be spliced to generate a first recording file.
  • a complete original recording file can also be generated based on the unmarked recording content.
  • a second recording file marked with the start time point and the end time point of the second time interval can also be generated to facilitate subsequent processing.
  • multiple material clips i.e., a third recording file set
  • the multiple material clips may include image instruction clips, i.e., the first type of recording clip files; and may also include recording content clips, i.e., the second type of recording clip files.
  • the image instruction clips include the recorded content corresponding to the time period corresponding to the image instruction, and the recording content clips are the remaining clip contents in the recorded video after cutting out the time period corresponding to the image instruction.
  • the electronic device can store image instruction-like segments and corresponding operations in an image instruction material library for subsequent image instruction recognition and matching.
  • the electronic device can store recorded content-like segments, which can facilitate users to view them separately or perform editing processing such as splicing and synthesis.
  • the start time point and the end time point in the marked audio stream content and the video stream content can also be matched before cutting.
  • the marked video stream content can be matched with the marked audio stream content or the unmarked video stream content.
  • the matching method may include: comparing the time points in the marked video stream content and the time points in the audio stream content to see if they are consistent, and if they are consistent, the match is successful.
  • the image frames at the start and end time points in the marked video stream content are compared with the image frames at the corresponding start and end time points in the unprocessed video stream content, and the entire image frame between the start and end time points in the marked video stream content is compared with the entire image frame between the start and end time points in the unprocessed video stream content. If the image frames are consistent, the match is considered successful.
  • the image instructions for recording control can include multiple forms of image instructions, which is convenient for users to use different instruction forms to control the recording process in different scenes.
  • FIG9a the user turns on the rear mode for recording, and FIG9a shows the recording screen of the rear camera.
  • FIG9b the user shows the text content of "switch camera” corresponding to the camera switching operation, for example, a paper with the text of "switch camera”, and the electronic device records the image of the text of "switch camera", then performs the corresponding camera switching operation, for example, converting the rear recording mode to the front recording mode.
  • the time period corresponding to the image instruction of "switch camera” in the video stream content and the audio stream content is 00:03-00:05
  • the 00:03-00:05 time period in one of the video stream content and one of the audio stream content is marked.
  • FIG9c shows the screen recorded in the front recording mode after executing the operation corresponding to the image instruction.
  • the time period corresponding to the image instruction in the video stream content and the audio stream content mentioned in this scenario can refer to the time period corresponding to the optimized second time interval.
  • the user shows the "ok" gesture action corresponding to the stop recording operation
  • the electronic device detects the image instruction corresponding to the "stop recording”
  • the electronic device detects the image instruction corresponding to the "stop recording”
  • the electronic device detects the image instruction corresponding to the "stop recording”
  • the electronic device detects the image instruction corresponding to the "stop recording”
  • the electronic device detects the image instruction corresponding to the "stop recording”
  • the corresponding operation such as stopping recording
  • an image/video resource library may be stored in the electronic device, and the image/video resource library may include an image/video system library and a custom image/video control library.
  • the image/video system library is used for system preset image instruction content and corresponding operations
  • the custom image/video control library is used to store user-defined image instruction content and corresponding operations
  • each image/video resource library may limit the recognition mode to fuzzy recognition and precise recognition.
  • the image instruction content corresponding to the operation in the embodiment of the present application can be any executable content customized by the user or preset by the system. It can be understood that the image instruction content referred to in the embodiment of the present application is not limited to a static image or a frame of image, but can also be a continuous multi-frame image instruction content within a period of time.
  • the image instruction content corresponding to the operation may include body movements, as shown in FIG10a, and the body movements may include specific body information and specific facial information.
  • the specific body information may include specific gesture information
  • the specific gesture information may be a specific gesture image, as shown in FIG10b
  • the specific gesture image may include gesture images such as [OK], [Like], [Yeah], [Love], [Fist], and [Love You].
  • Different gesture images may correspond to different operations, for example, [OK] represents "stop recording", and [Like] represents "zoom in focus", etc.
  • the facial information may be a facial image, wherein the facial image may include facial images such as [Nod], [Shake Head], and [Smiley Face].
  • Different facial images may correspond to different operations, for example, the operation corresponding to [Smiley Face] may be "snap shot”, the operation corresponding to [Nod] may be “stop recording”, and the operation corresponding to [Shake Head] may be "switch lens”, etc.
  • specific body information may include specific gesture information, and the specific gesture information may be a specific gesture image, as shown in FIG10b, and the specific gesture image may include gesture images such as [OK], [Like], [Fist], and [Love You].
  • Different gesture images may correspond to different operations, for example, [OK] represents “stop recording", and [Like] represents "zoom in focus", etc.
  • the body information may also include specific human posture information.
  • the specific human posture information may include posture images such as [turning in circles], [squatting], [large character posture], and [jumping]. Different posture images may correspond to different operations. For example, the operation corresponding to [turning in circles] may be "switch camera”, and the operation corresponding to [jumping] may be "enlarge focal length”, etc.
  • the image instruction content corresponding to the operation may also include schematic content, wherein:
  • the schematic content may include specific text information and specific image information.
  • the specific text information may include text content that is consistent with the name of the operation or with the key information.
  • the schematic content corresponding to the switch lens operation may be the text "switch lens".
  • the specific text information may also include fixed mode content that is customized by the user.
  • the schematic content corresponding to the switch lens operation may be various forms of images corresponding to the number "1".
  • the specific image information may be a logo indicating a specific meaning.
  • the schematic content corresponding to the lens switching operation may be a logo indicating lens switching.
  • Table 1 shows a correspondence table between some image instruction contents and operations in an embodiment of the present application.
  • the image instruction content corresponding to the lens switching operation stored in one of the image/video resource libraries in the electronic device can be the text "switch lens" or a two-finger gesture
  • the recognition method can be fuzzy recognition (wherein, in the option column in Table 1, whether there is a one-to-one correspondence is recorded. If yes is selected, the recognition method is limited to precise recognition, and if no is selected, the recognition method is limited to fuzzy recognition).
  • the image instruction content corresponding to the zoom focus operation can be the text "zoom focus” or a five-finger spread gesture, and the recognition method can be fuzzy recognition.
  • the image instruction content corresponding to the switch to the front camera lens operation can be the text "switch to front" or a check mark gesture, and the recognition method can be fuzzy recognition.
  • the matching method may include: an image frame comparison method and an image local comparison method.
  • file A represents the video stream content recorded in real time
  • file B represents the video stream content used for image control processing; it can be understood that file A can be kept without marking or other processing for subsequent comparison or other needs, and file B can be used for marking or cropping the time points corresponding to the control instructions.
  • image 1 in file B can represent the image instruction corresponding to the identified "switch camera" operation
  • image 2 can represent the image instruction corresponding to the "stop recording” operation.
  • Image 1 has a starting time point and an ending time point
  • image 2 also has a starting time point and an ending time point.
  • the process of image instruction recognition and matching is shown in Figure 13b: the electronic device will compare each image frame in the B file with each image frame in the image instruction content (or called feature behavior) corresponding to the operation in the image resource library.
  • image instruction content or called feature behavior
  • the recognition is confirmed to be successful.
  • the first image frame in the real-time captured image that can match any image instruction content in the resource library, such as the first image frame can be marked as the starting time point of the first time interval where the image instruction is located, and the last image frame that can match the image instruction content in the resource library, such as the second image frame, can be marked as the ending time point of the first time interval.
  • the image frames are completely consistent that is, the image frames between the starting time point and the ending time point are consistent, it can be proved that the recognition is successful.
  • the recognition is successful, further recognition can be performed in file A.
  • the marked start time point and end time point can be transparently transmitted to file A, that is, the consistent start time point and end time point can be determined in file A, and the corresponding start time point and end time point in file A can be transferred to file A.
  • the image frames between the end time points are compared with the image frames of the corresponding image instruction content in the resource library. If they match, it proves that the image recognition is successful. If they are inconsistent, it proves that the recognition has failed.
  • the recognition is successful only when both recognitions are successful. As long as there is one recognition failure, it can be considered that the image instruction recognition this time has failed. After the recognition fails, all the marks generated in this recognition process need to be cleared. In this way, the accuracy of image recognition can be effectively improved through the above-mentioned circular two-way recognition method.
  • each pixel of the image frame in the B file is compared with all the pixels of the corresponding image frame in the corresponding image instruction content in the image resource library.
  • the reason for the image matching result being mismatched may be the presence of bad pixels (or noise points) as shown in Figure 13g, rather than the image instruction mismatch in the actual sense, resulting in an erroneous matching result.
  • an embodiment of the present application provides a difference point secondary analysis method for comparing the above image frames, as shown in FIG13h , the difference point secondary analysis method includes a bad point removal method and a difference point ignoring method.
  • the bad pixel removal method refers to first analyzing whether there are bad pixels in the image before comparing two frames of images. When bad pixels exist, the bad pixels are removed before comparing the image frames.
  • a method for determining whether there are bad pixels may include: determining whether there is an obvious jump between any pixel and surrounding pixels. For example, it may be determined whether the display parameter difference (such as hue value) between each pixel and adjacent pixels is higher than a first threshold. If it is higher than the first threshold, it is determined that there is an obvious jump between the pixel and surrounding pixels, and the pixel is determined to be a bad pixel.
  • the display parameter difference such as hue value
  • the method of determining whether there is a bad pixel may include: comparing the similarity between each pixel and a bad pixel in a theoretical sense, and when the similarity exceeds a third threshold (eg, 95%), the pixel may be considered to be a bad pixel.
  • a third threshold eg, 95%)
  • the difference point ignoring method means that when the similarity obtained after comparing two frames of images is greater than the fourth threshold value (for example, 99%), and the position difference between any two unmatched pixel points in the two frames of images is greater than the second threshold value, then these different pixel points can be considered as insignificant ignored points, and no matching is required.
  • the two frames of images are determined to be matched images. For example, as shown in FIG. 13i, the two different pixel points in the two frames of images are far apart, and the position difference is greater than the second threshold value, then both pixel points can be considered as insignificant ignored points.
  • the local image comparison method mentioned in the embodiment of the present application is described below. It can be understood that the local comparison method in the embodiment of the present application is basically the same as the above-mentioned image frame comparison method, the difference is that in the above-mentioned image frame comparison method, the electronic device will compare each complete image frame in the B file with each complete image frame in the image resource library, while in the local comparison method, a specific part of the image frame in the B file is compared with a specific part of the image frame in the image resource library.
  • the local image comparison method may include an arbitrary position moving comparison method, an arbitrary position enlarging comparison method, and a fixed position comparison method.
  • Figure 14b shows a schematic diagram of an arbitrary position moving comparison method in an embodiment of the present application.
  • the arbitrary position moving comparison method is to move and traverse multiple image blocks in each frame of the image according to a certain trajectory or a certain switching order (for example, traversing from left to right in sequence).
  • the second image block in the upper left corner of the image frame can be matched with the image block of the corresponding image frame in the image resource library. If the match is unsuccessful, the first image block adjacent to the right side of the second image block is matched with the image block of the corresponding image frame in the image resource library.
  • the third image block adjacent to the right side of the first image block is matched with the image block of the corresponding image frame in the image resource library.
  • any image block in the image frame in the video stream content used for image control processing is successfully matched with the image block of the corresponding image frame in the image resource library, it proves that the frame image matches successfully.
  • the traversal is completed, there is still no image block that can be matched, which proves that the frame image matches unsuccessfully.
  • the center position enlargement comparison method is to use the first center image block of the set specification of the center position of the image frame as a reference, gradually expand the range of the image block, and perform comparison and matching. Specifically, first, the first central image block of the current image frame to be matched in the video stream content for image control processing (the image block in the leftmost picture in FIG. 14c) is compared with the image block of the corresponding image frame in the image resource library.
  • the match is successful, it proves that the current image frame to be matched in the video stream content and the corresponding image frame in the image resource library are matched successfully; if it is unsuccessful, the first central image block is expanded by a set range, for example, the second central image block (the image block in the middle picture in FIG. 14c) is obtained, and the second central image block is matched.
  • the matching is terminated, and it is determined that the current image frame to be matched in the video stream content and the corresponding image frame in the image resource library are matched successfully; when the second central image block and the corresponding image block in the image resource library are matched unsuccessfully, the second central image block is expanded by a set range, for example, the third central image block (the image block in the rightmost picture in FIG. 14c) is obtained, and the matching is continued. If the range is expanded to the entire image and the match is still unsuccessful, it is determined that the current image frame to be matched in the video stream content and the corresponding image frame in the image resource library are matched unsuccessfully. It can be understood that the second central image The block is larger than the first central image block and includes the first central image block. The second central image block extends in all directions based on the central image block.
  • Figure 14d shows a schematic diagram of a fixed position comparison method in an embodiment of the present application, wherein the fixed position comparison method is to set an image block at a fixed position, and the fixed position is set by an application or a user.
  • the image block at the fixed position may be an image block at a middle position or an image block at a corner position.
  • the number of image blocks at the fixed position may be one or more, and the present application does not limit this.
  • the local comparison method is used for comparison, which can improve the matching efficiency while ensuring the matching accuracy.
  • the marking of the start time point and the end time point of the image instruction can be marked in any feasible form in addition to the above-mentioned method of [Image 1 Start].
  • the marking of the start time point and the end time point of the image instruction can be as shown in Table 2:
  • the real image subscript method can be used, for example, when the image instruction is a switching lens text image, the start time point of the switching lens text image instruction is subscripted with 0, and the end time point is subscripted with 1;
  • the real image pairing method can also be used, for example, when the image instruction content is to enlarge the focal length value, a mark can be subscripted at the start time point of the enlarged focal length value image instruction, and the same mark can be subscripted at the end time point, so that the mark that appears first defaults to the start time point, and the mark that appears later defaults to the end time point;
  • the operation subscript method can also be used, for example, for the switching lens image instruction, the start time point corresponding to the actual operation corresponding to the switching lens image instruction can be subscripted with 0, and the end time point can be subscripted with 1;
  • the operation pairing marking method can also be used
  • Table 2 Image instruction marking table
  • FIG15a shows a method for determining the start time point after the expanded range.
  • the method for determining the start time point after the expanded range may include:
  • go to 1504 use the time point corresponding to the (n+1)th frame waveform in the audio stream content as the corresponding starting time point after the image instruction in the audio stream content is expanded.
  • the starting time point or when it is determined that the similarity between one frame waveform and the waveform corresponding to the first starting time point is greater than the sixth threshold, but the number of frames currently compared has reached the set first number, the time point corresponding to the frame waveform is used as the starting time point after the image instruction is expanded.
  • the time point corresponding to the waveform of the n+1th frame in the audio stream content is used as the corresponding starting time point after the image instruction in the audio stream content is expanded.
  • the time point corresponding to the n-th frame waveform in the audio stream content is used as the starting time point after the range is expanded.
  • FIG. 15b shows another method for determining the start time point after the range is expanded in an embodiment of the present application.
  • the method for determining the start time point after the range is expanded may include:
  • go to 1604 use the time point corresponding to the (n+1)th frame waveform in the audio stream content as the corresponding starting time point after the image instruction in the audio stream content is expanded.
  • the similarity between the previous frame waveform of the first starting time point and the waveform corresponding to the first starting time point, the similarity between the second frame waveform before the first starting time point and the previous frame waveform of the first starting time point, the similarity between the third frame waveform before the first starting time point and the second frame waveform before the first starting time point, and other similarities between each frame waveform and the next frame waveform of each frame waveform is greater than a sixth threshold, until it is determined that the similarity between one of the frame waveforms and the next frame waveform of the frame waveform is less than or equal to the sixth threshold, and the time point corresponding to the next frame waveform of the frame waveform is used as the starting time point after the expansion range of the image instruction, or when it is determined that the similarity between one of the frame waveforms and the next frame waveform of the frame waveform is greater than the sixth threshold, but the number of frames currently compared has reached the set first number, the time point corresponding to the frame
  • the time point corresponding to the waveform of the n+1th frame in the audio stream content is used as the corresponding starting time point after the image instruction in the audio stream content is expanded.
  • the time point corresponding to the n-th frame waveform in the audio stream content is used as the starting time point after the range is expanded.
  • the method for determining the end time point after the expansion of the image instruction can be similar to the method for determining the start time point, except that the end time point is determined by comparing the waveform after the first end time point with the waveform corresponding to the first end time point.
  • different recording files can be generated based on the recorded content.
  • a second time interval can be marked in the recorded content based on the first time interval where the image instruction is located, so as to crop the content corresponding to the second time interval and generate a first recording file that does not include the image instruction.
  • the first recording file is the spliced content after deleting the recording content corresponding to the second time interval from the recorded content.
  • a complete recording file i.e., a second recording file
  • the second time interval can also be marked in the recorded content based on the first time interval where the image instruction is located, so as to generate multiple material clips, i.e., a third recording file set.
  • the multiple material clips may include image instruction clips, i.e., first type of recording clip files; and may also include recording content clips, i.e., second type of recording clip files.
  • the image instruction clips include the recording content corresponding to the second time interval, and the recording content clips are the remaining clip contents after cutting out the corresponding recording content of the second time interval from the recording content.
  • the electronic device can store the image instruction-like segments and corresponding operations in the corresponding control instruction material library for subsequent image instruction recognition and matching.
  • the electronic device can store the recorded content-like segments, which can facilitate users to view them separately or perform editing processing such as splicing and synthesis.
  • obtaining multiple material clips includes: when the recorded content is video stream content, splitting the video stream content based on time points marked in the video stream content to obtain multiple video stream material clips; when the recorded content includes audio stream content and video stream content, splitting the audio stream content and/or video stream content based on time points marked in the audio stream content to obtain multiple audio stream material clips and/or video stream material clips, or splitting the video stream content and/or audio stream content based on time points marked in the video stream content to obtain multiple video stream material clips and/or audio stream material clips, and generating corresponding recording clips based on the correspondence between the corresponding audio stream material clips and the video stream material clips.
  • the electronic device when it detects an image instruction of "switch camera” during the recording process and determines the start time point and end time point corresponding to the image instruction of "switch camera” in the video stream content and the audio stream content, it can generate a first recording segment based on the audio stream content and the video stream content before the start time point of the "switch camera” image instruction, generate a second recording segment based on the audio stream content and the video stream content after the end time point, and generate a third recording segment based on the audio stream content and the video stream content in the time period corresponding to the "switch camera” image instruction.
  • the embodiment of the present application also provides a recording control system, as shown in FIG16 , which may include:
  • An image acquisition module in response to a first operation of starting recording by a user, performs recording at least through a camera, and the recorded content at least includes video stream content;
  • An image control module which can be used to identify image instructions input by the user through the camera in the video stream content, and the image instructions are used to control the recording;
  • the audio-visual synchronization module is used to mark the second time interval in the recorded content at the first time interval where the image instruction in the recorded content is located; and obtain the marked recorded content.
  • the shooting result generating module is used to generate at least a first recording file based on the recording content when a recording end instruction is detected.
  • the recording control method provided in the embodiment of the present application can be directly developed and implemented on the application side, or it can be constructed separately in the form of capability integration.
  • the capability integrated on the application side in the electronic device system can be provided in the form of AAR and JAR packages, which can be updated regardless of the version update, or it can be provided in the form of binary capability packages to the capabilities of all components in the electronic device system, which can be updated regardless of the version update.
  • the capabilities can also be provided to all components in the electronic device system through the interface of the framework layer of the version in the electronic device system, which can be updated with the system upgrade.
  • the various embodiments disclosed in the present application may be implemented in hardware, software, firmware, or a combination of these implementation methods.
  • the embodiments of the present application may be implemented as a computer program or program code executed on a programmable system, the programmable system comprising at least one processor, a storage system (including volatile and non-volatile memory and/or storage elements), at least one input device, and at least one output device.
  • Program code can be applied to input instructions to perform the functions described in this application and generate output information.
  • the output information can be applied to one or more output devices in a known manner.
  • a processing system includes any system having a processor such as, for example, a digital signal processor (DSP), a microcontroller, an application specific integrated circuit (ASIC), or a microprocessor.
  • DSP digital signal processor
  • ASIC application specific integrated circuit
  • Program code can be implemented with high-level programming language or object-oriented programming language to communicate with the processing system.
  • program code can also be implemented with assembly language or machine language.
  • the mechanism described in this application is not limited to the scope of any specific programming language. In either case, the language can be a compiled language or an interpreted language.
  • the disclosed embodiments may be implemented in hardware, firmware, software, or any combination thereof.
  • the disclosed embodiments may also be implemented as instructions carried or stored on one or more temporary or non-temporary machine-readable (e.g., computer-readable) storage media, which may be read and executed by one or more processors.
  • instructions may be distributed over a network or through other computer-readable media.
  • machine-readable media may include any mechanism for storing or transmitting information in a machine (e.g., computer) readable form, including, but not limited to, floppy disks, optical disks, optical disks, read-only memories (CD-ROMs), magneto-optical disks, read-only memories (ROMs), random access memories (RAMs), erasable programmable read-only memories (EPROMs), electrically erasable programmable read-only memories (EEPROMs), magnetic or optical cards, flash memory, or a tangible machine-readable memory for transmitting information (e.g., carrier waves, infrared signals, digital signals, etc.) using the Internet in electrical, optical, acoustic, or other forms of propagation signals. Therefore, machine-readable media include any type of machine-readable media suitable for storing or transmitting electronic instructions or information in a machine (e.g., computer) readable form.
  • a machine-readable media include any type of machine-readable media suitable for storing or transmitting electronic instructions or information in a machine
  • a logical unit/module can be a physical unit/module, or a part of a physical unit/module, or can be implemented as a combination of multiple physical units/modules.
  • the physical implementation method of these logical units/modules themselves is not the most important.
  • the combination of functions implemented by these logical units/modules is the key to solving the technical problems proposed by the present application.
  • the above-mentioned device embodiments of the present application do not introduce units/modules that are not closely related to solving the technical problems proposed by the present application, which does not mean that there are no other units/modules in the above-mentioned device embodiments.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Theoretical Computer Science (AREA)
  • Signal Processing (AREA)
  • Human Computer Interaction (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • General Engineering & Computer Science (AREA)
  • Health & Medical Sciences (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • General Health & Medical Sciences (AREA)
  • Evolutionary Computation (AREA)
  • Databases & Information Systems (AREA)
  • Medical Informatics (AREA)
  • Software Systems (AREA)
  • Psychiatry (AREA)
  • Social Psychology (AREA)
  • Business, Economics & Management (AREA)
  • General Business, Economics & Management (AREA)
  • Computing Systems (AREA)
  • Artificial Intelligence (AREA)
  • Television Signal Processing For Recording (AREA)
  • Signal Processing For Digital Recording And Reproducing (AREA)
  • Management Or Editing Of Information On Record Carriers (AREA)

Abstract

La présente demande concerne le domaine technique de l'enregistrement. Sont divulgués un procédé de commande d'enregistrement vidéo, un dispositif électronique et un support. Le procédé comprend : en réponse à une première opération effectuée par un utilisateur pour démarrer l'enregistrement, l'exécution d'un enregistrement au moins au moyen d'une caméra, le contenu enregistré comprenant au moins un contenu de flux vidéo; et la reconnaissance, à partir du contenu de flux vidéo, d'une instruction d'image entrée par l'utilisateur au moyen de la caméra, l'instruction d'image étant utilisée pour mettre en œuvre une commande d'enregistrement. L'étape de reconnaissance, à partir du contenu de flux vidéo, d'une instruction d'image entrée par l'utilisateur au moyen de la caméra comprend : la reconnaissance d'au moins un premier bloc d'image d'au moins une trame d'images dans le contenu de flux vidéo; la reconnaissance, à partir du premier bloc d'image, d'une instruction d'image qui correspond à un comportement de caractéristique; et la détermination, selon ladite au moins une trame d'images où se situe l'instruction d'image, d'un premier intervalle de temps où se situe l'instruction d'image. Sur la base de la solution, une instruction d'image peut être reconnue pendant l'enregistrement et une commande d'enregistrement peut être exécutée pendant l'enregistrement au moyen de l'instruction d'image, répondant ainsi aux exigences de commande d'un scénario d'enregistrement vidéo.
PCT/CN2023/118317 2022-10-11 2023-09-12 Procédé de commande d'enregistrement vidéo, dispositif électronique et support WO2024078238A1 (fr)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN202211243144.0 2022-10-11
CN202211243144.0A CN117880413A (zh) 2022-10-11 2022-10-11 一种视频录制控制方法、电子设备及介质

Publications (1)

Publication Number Publication Date
WO2024078238A1 true WO2024078238A1 (fr) 2024-04-18

Family

ID=90595415

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2023/118317 WO2024078238A1 (fr) 2022-10-11 2023-09-12 Procédé de commande d'enregistrement vidéo, dispositif électronique et support

Country Status (2)

Country Link
CN (1) CN117880413A (fr)
WO (1) WO2024078238A1 (fr)

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103167230A (zh) * 2011-12-17 2013-06-19 富泰华工业(深圳)有限公司 电子设备及其根据手势控制拍照的方法
CN106506968A (zh) * 2016-11-29 2017-03-15 广东欧珀移动通信有限公司 控制方法、控制装置、电子装置
KR20210038446A (ko) * 2020-02-14 2021-04-07 베이징 바이두 넷컴 사이언스 테크놀로지 컴퍼니 리미티드 제스처를 기반으로 전자기기를 제어하는 방법 및 장치
CN114637439A (zh) * 2022-03-24 2022-06-17 海信视像科技股份有限公司 显示设备和手势轨迹识别方法

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103167230A (zh) * 2011-12-17 2013-06-19 富泰华工业(深圳)有限公司 电子设备及其根据手势控制拍照的方法
CN106506968A (zh) * 2016-11-29 2017-03-15 广东欧珀移动通信有限公司 控制方法、控制装置、电子装置
KR20210038446A (ko) * 2020-02-14 2021-04-07 베이징 바이두 넷컴 사이언스 테크놀로지 컴퍼니 리미티드 제스처를 기반으로 전자기기를 제어하는 방법 및 장치
CN114637439A (zh) * 2022-03-24 2022-06-17 海信视像科技股份有限公司 显示设备和手势轨迹识别方法

Also Published As

Publication number Publication date
CN117880413A (zh) 2024-04-12

Similar Documents

Publication Publication Date Title
US20240168624A1 (en) Screen capture method and related device
JP7326476B2 (ja) スクリーンショット方法及び電子装置
WO2021078284A1 (fr) Procédé de continuation de contenu et dispositif électronique
WO2021129198A1 (fr) Procédé de photographie dans un scénario à longue distance focale, et terminal
WO2021057673A1 (fr) Procédé d'affichage d'image et dispositif électronique
WO2022179405A1 (fr) Procédé d'affichage de projection d'écran et dispositif électronique
CN114185503B (zh) 多屏交互的系统、方法、装置和介质
WO2022143883A1 (fr) Procédé et système de photographie, et dispositif électronique
CN114115674B (zh) 录音和文档内容的定位方法、电子设备及存储介质
WO2022042326A1 (fr) Procédé de commande d'affichage et appareil associé
WO2023241209A9 (fr) Procédé et appareil de configuration de papier peint de bureau, dispositif électronique et support de stockage lisible
WO2023045712A1 (fr) Procédé de traitement d'anomalie de duplication d'écran et dispositif électronique
WO2024012011A1 (fr) Procédé d'enregistrement vidéo et dispositif électronique
WO2024078238A1 (fr) Procédé de commande d'enregistrement vidéo, dispositif électronique et support
WO2023020012A1 (fr) Procédé de communication de données entre dispositifs, dispositif, support de stockage et produit de programme
WO2024078236A1 (fr) Procédé de commande d'enregistrement, dispositif électronique et support
CN114115770B (zh) 显示控制的方法及相关装置
CN118259861A (en) Screen-throwing display method and electronic equipment
CN114911377A (zh) 显示方法和电子设备
CN117714849A (zh) 一种图像拍摄方法及相关设备
CN117201713A (zh) 音频处理方法及电子设备
CN114757955A (zh) 一种目标跟踪方法及电子设备

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 23876443

Country of ref document: EP

Kind code of ref document: A1