CN114390191A - Video recording method, electronic device and storage medium - Google Patents

Video recording method, electronic device and storage medium Download PDF

Info

Publication number
CN114390191A
CN114390191A CN202011142554.7A CN202011142554A CN114390191A CN 114390191 A CN114390191 A CN 114390191A CN 202011142554 A CN202011142554 A CN 202011142554A CN 114390191 A CN114390191 A CN 114390191A
Authority
CN
China
Prior art keywords
video recording
follow
subject
main body
mobile phone
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202011142554.7A
Other languages
Chinese (zh)
Other versions
CN114390191B (en
Inventor
陈彬
朱聪超
李远友
胡斌
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Huawei Technologies Co Ltd
Original Assignee
Huawei Technologies Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Huawei Technologies Co Ltd filed Critical Huawei Technologies Co Ltd
Priority to CN202011142554.7A priority Critical patent/CN114390191B/en
Publication of CN114390191A publication Critical patent/CN114390191A/en
Application granted granted Critical
Publication of CN114390191B publication Critical patent/CN114390191B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/60Control of cameras or camera modules
    • H04N23/67Focus control based on electronic image sensor signals
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/60Control of cameras or camera modules
    • H04N23/63Control of cameras or camera modules by using electronic viewfinders
    • H04N23/631Graphical user interfaces [GUI] specially adapted for controlling image capture or setting capture parameters
    • H04N23/632Graphical user interfaces [GUI] specially adapted for controlling image capture or setting capture parameters for displaying or modifying preview images prior to image capturing, e.g. variety of image resolutions or capturing parameters
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N5/00Details of television systems
    • H04N5/76Television signal recording

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Human Computer Interaction (AREA)
  • Studio Devices (AREA)

Abstract

The application discloses a video recording method, electronic equipment and a storage medium, and relates to the field of electronic equipment. The method can be applied to terminal equipment with a video recording function, and comprises the following steps: displaying an initial video frame including a first subject; in response to the selection operation of the first subject, determining the first subject as a follow-up shooting object; in the video recording process after the first main body is determined to be the follow-shot object, the proportion of the area occupied by the first main body in each frame of video recording picture to the video recording picture is adjusted to be a first proportion. By the video recording method, a user can be assisted to shoot a better and more stable effect when the user follows a motion subject, and a smoother video recording picture is obtained.

Description

Video recording method, electronic device and storage medium
Technical Field
The embodiment of the application relates to the field of electronic equipment, in particular to a video recording method, electronic equipment and a storage medium.
Background
Users often have a need to follow a particular subject, such as: a user may follow-up a moving subject (e.g., a person running, a walking car, etc.) using a cell phone. In many follow-up scenes, the subject of the follow-up may move to a distance close to or far from the lens of the mobile phone. In this case, in order to continue the video recording screen to the subject to be photographed, it is often necessary for the user to move the mobile phone with the hand of the user along with the subject to be photographed.
However, the video recording screen captured by the user moving with the mobile phone in hand along with the subject is not smooth and has poor screen stability.
Disclosure of Invention
The embodiment of the application provides a video recording method, electronic equipment and a storage medium, which can assist a user in shooting a better and more stable effect when the user shoots a motion subject, so that a smoother video recording picture can be obtained.
In a first aspect, an embodiment of the present application provides a video recording method, which is applied to a terminal device with a video recording function, where the method includes: displaying an initial video frame including a first subject; in response to the selection operation of the first subject, determining the first subject as a follow-up shooting object; in the video recording process after the first main body is determined to be the follow-shot object, the proportion of the area occupied by the first main body in each frame of video recording picture to the video recording picture is adjusted to be a first proportion.
For example, the first ratio may be 40%, 50%, etc. In the video recording method, the terminal equipment adjusts the proportion of the area occupied by the first main body in each frame of video recording picture and the video recording picture to be a first proportion in the video recording process after the first main body is determined to be a follow-shot object, so that a better and more stable effect can be shot when an auxiliary user moves the main body with the follow-shot object, and a smoother video recording picture can be obtained.
In one possible design, the method further includes: in the video recording process after the first main body is determined to be the follow-up shooting object, acquiring coordinate information of the first main body in the ith frame image, and adjusting a focusing focus when the (i + 1) th frame image is shot according to the coordinate information of the first main body in the ith frame image; wherein i is an integer greater than 0.
In the design, in the video recording process after the first main body is determined to be the follow-up shooting object, the focusing focus when the (i + 1) th frame image is shot is adjusted according to the coordinate information of the first main body in the (i) th frame image, so that the problem of fuzzy video recording pictures caused by the fact that the focusing focus is not on the first main body when the first main body moves in the video recording process can be avoided. For example, the coordinate information of the first subject at the current frame may be sent to an auto-focusing module of the terminal device. The automatic focusing module can adjust the position of a focusing focus to the first main body according to the coordinate information of the first main body in the current frame, and follow-up shooting and recording are carried out on the first main body, so that a video picture can be continuously focused to the first main body.
Optionally, the method further comprises: and if the coordinate information of the first main body in the ith frame image is not acquired, adjusting the focusing focus when the (i + 1) th frame image is shot according to the coordinate information of the first main body in the (i-1) th frame image.
In one possible design, the method further includes: in the video recording process before the first main body is determined to be the follow-up shooting object, coordinate information of a central point of a j frame image is obtained, and a focusing focus when the j +1 frame image is shot is adjusted to the central point according to the coordinate information of the central point of the j frame image; wherein j is an integer greater than 0.
The design can be applied to a scene in which the user has not performed selection operation on the first main body. In the scene, the terminal device can adjust the focusing focus when shooting the next frame image according to the coordinate information of the central point of the current frame image.
In one possible design, the method further includes: and in the video recording process after the first main body is determined to be the follow-shot object, displaying prompt information in a first area of a video recording picture when the video recording picture corresponding to each frame of picture is detected to be located at the edge of the picture, wherein the prompt information is used for prompting a user to move the terminal equipment to a first direction.
In the design, when the first main body is detected to be located at the edge of the picture, the prompt information is displayed in the first area of the video picture to prompt a user to move the terminal device to the first direction, and the situation that the first main body cannot be shot along with the first main body due to the fact that the first main body moves to the outside of the picture can be avoided.
In one possible design, the hint information is a direction identifier; the method further comprises the following steps: and displaying moving reference lines on the upper side and the lower side of the direction mark, wherein the moving reference lines are used for prompting a user that the direction mark is controlled not to exceed the range of the moving reference lines when the user moves the terminal equipment.
Illustratively, the direction indicator may be an arrow for indicating the first direction. The mobile reference lines are displayed on the upper side and the lower side of the direction mark, so that when a user is prompted to move the terminal equipment, the direction mark is controlled not to exceed the range of the mobile reference lines, and a shot video picture can be more stable.
In one possible design, the initial recording frame further includes a second main body; the method further comprises the following steps: in the video recording process after the first subject is determined to be the follow-up shooting object, responding to a first selection operation on a second subject, and switching the follow-up shooting object from the first subject to the second subject; in the video recording process after the follow-shot object is switched from the first main body to the second main body, the proportion of the area occupied by the second main body in each frame of video recording picture to the video recording picture is adjusted to be the first proportion.
In the design, the user can also switch the follow shooting object in the video recording process. Such as: in the process of performing the follow-up image recording on the first subject, the follow-up object may be switched from the first subject to the second subject.
In one possible design, the initial video frame further includes at least one second main body; the method further comprises the following steps: in the video recording process after the first subject is determined to be the follow-up shooting object, responding to a second selection operation on at least one second subject, and determining the at least one second subject to be the follow-up shooting object; in the video recording process after at least one second main body is also determined as a follow-up shooting object, the proportion of the area occupied by the first main body and the at least one second main body in each video recording picture together to the video recording picture is adjusted to be a first proportion.
In the design, the user can also add a follow shot object in the video recording process. Such as: the second main body can be added as a follow-up object in the process of performing follow-up shooting and recording on the first main body, and the terminal device performs follow-up shooting and recording on the first main body and the second main body simultaneously.
In a second aspect, an embodiment of the present application provides an apparatus, which may be applied to the terminal device with a video recording function in the foregoing first aspect, and is configured to implement the method applied to the terminal device in the foregoing first aspect. The functions of the device can be realized by hardware, and can also be realized by executing corresponding software by hardware. The hardware or software includes one or more modules or units corresponding to the above functions, for example, a display unit, a processing unit, and the like.
The display unit is used for displaying an initial video picture comprising a first main body. The processing unit is used for responding to the selection operation of the first subject and determining the first subject as a follow-up shooting object; in the video recording process after the first main body is determined to be the follow-shot object, the proportion of the area occupied by the first main body in each frame of video recording picture to the video recording picture is adjusted to be a first proportion.
In a possible design, the processing unit is further configured to, in a video recording process after the first subject is determined to be a follow-up shooting object, acquire coordinate information of the first subject in an ith frame image, and adjust a focus when the i +1 th frame image is shot according to the coordinate information of the first subject in the ith frame image; wherein i is an integer greater than 0.
Optionally, the processing unit is further configured to, if coordinate information of the first subject in the i-th frame image is not acquired, adjust a focus point when the i + 1-th frame image is captured according to the coordinate information of the first subject in the i-1-th frame image.
In a possible design, the processing unit is further configured to, in a video recording process before the first subject is determined to be the follow-up subject, obtain coordinate information of a center point of a j-th frame image, and adjust a focus point when the j + 1-th frame image is shot to the center point according to the coordinate information of the center point of the j-th frame image; wherein j is an integer greater than 0.
In a possible design, the processing unit is further configured to, in a video recording process after the first main body is determined to be the follow-up object, display, by the display unit, prompt information for prompting the user to move the terminal device in the first direction in a first area of a video recording picture when the video recording picture corresponding to each frame of picture is detected to be located at a picture edge.
In one possible design, the hint information is a direction identifier; the processing unit is further used for displaying moving reference lines on the upper side and the lower side of the direction mark through the display unit, and the moving reference lines are used for prompting a user to control the direction mark not to exceed the range of the moving reference lines when the user moves the terminal device.
In one possible design, the initial recording frame further includes a second main body; the processing unit is further configured to, in a video recording process after the first subject is determined to be the follow-shot object, respond to a first selection operation on the second subject, and switch the follow-shot object from the first subject to the second subject; in the video recording process after the follow-shot object is switched from the first main body to the second main body, the proportion of the area occupied by the second main body in each frame of video recording picture to the video recording picture is adjusted to be the first proportion.
In one possible design, the initial video frame further includes at least one second main body; the processing unit is further configured to, in a video recording process after the first subject is determined to be the follow-up shooting object, respond to a second selection operation on at least one second subject, and determine the at least one second subject as the follow-up shooting object; in the video recording process after at least one second main body is also determined as a follow-up shooting object, the proportion of the area occupied by the first main body and the at least one second main body in each video recording picture together to the video recording picture is adjusted to be a first proportion.
In a third aspect, an embodiment of the present application provides an electronic device, where the electronic device may be the terminal device with the video recording function. The electronic device includes: a processor, a memory for storing processor-executable instructions; the processor is configured to execute the instructions, such that the electronic device implements the method according to the first aspect.
In a fourth aspect, embodiments of the present application provide a computer-readable storage medium having stored thereon computer program instructions; the computer program instructions, when executed by an electronic device, cause the electronic device to implement a method as described in the first aspect.
The beneficial effects of the second to fourth aspects can be referred to the description of the first aspect, and are not repeated herein.
In a fifth aspect, the present application provides a computer program product, which includes computer readable code, when the computer readable code runs in an electronic device, causes the electronic device to implement the method of the first aspect.
It should be appreciated that the description of technical features, solutions, benefits, or similar language in this application does not imply that all of the features and advantages may be realized in any single embodiment. Rather, it is to be understood that the description of a feature or advantage is intended to include the specific features, aspects or advantages in at least one embodiment. Therefore, the descriptions of technical features, technical solutions or advantages in the present specification do not necessarily refer to the same embodiment. Furthermore, the technical features, technical solutions and advantages described in the present embodiments may also be combined in any suitable manner. One skilled in the relevant art will recognize that an embodiment may be practiced without one or more of the specific features, aspects, or advantages of a particular embodiment. In other instances, additional features and advantages may be recognized in certain embodiments that may not be present in all embodiments.
Drawings
Fig. 1 shows a schematic structural diagram of an electronic device provided in an embodiment of the present application;
FIG. 2 is a schematic diagram illustrating a shooting scene provided by an embodiment of the present application;
FIG. 3 is a schematic diagram illustrating a display screen of a mobile phone according to an embodiment of the present application;
FIG. 4 is a schematic diagram illustrating a method for determining a follow-shot object according to an embodiment of the present application;
FIG. 5 is a schematic diagram illustrating another example of determining a follow-up shot object according to an embodiment of the present disclosure;
FIG. 6 is a schematic diagram illustrating another example of determining a follow-shot object provided by an embodiment of the present application;
fig. 7 is a schematic diagram illustrating another display screen of a mobile phone provided in an embodiment of the present application;
FIG. 8 is a schematic diagram illustrating a heel-and-shoot principle provided by an embodiment of the present application;
FIG. 9 is a schematic diagram illustrating another shooting scene provided by an embodiment of the present application;
FIG. 10 is a schematic diagram illustrating a display screen of a mobile phone according to an embodiment of the present application;
fig. 11 is a schematic diagram illustrating a display screen of a mobile phone provided in an embodiment of the present application;
FIG. 12 is a diagram illustrating a relationship between a first region and a second region provided in an embodiment of the present application;
fig. 13 is a schematic diagram illustrating a display screen of a mobile phone provided in an embodiment of the present application;
fig. 14 is a schematic diagram illustrating a display screen of a mobile phone provided in an embodiment of the present application;
fig. 15 is a schematic diagram illustrating a switching follow-shot object according to an embodiment of the present application;
fig. 16 is a schematic diagram illustrating another switching follow-shot object provided in an embodiment of the present application;
FIG. 17 is a schematic diagram illustrating an example of adding a follow shot object according to the present disclosure;
fig. 18 shows a schematic structural diagram of an apparatus provided in an embodiment of the present application.
Detailed Description
When a common user records a video, the user often needs to follow a specific subject. Illustratively, the subject that the user needs to follow may be a playing child, or a running person, or an actor performing on a stage, or a football or athlete in a football match, etc. In the following shooting scenario of the foregoing example, the subject of the following shooting may move or change the form from time to time, such as: sitting down, standing up, exercising, side, back, half, whole body, etc.
At present, to the scene of the aforesaid with clapping the main part, can continuously follow clapping the main part for making the video picture, clap out better more stable effect, often need the user to adjust video recording equipment initiative when shooing. For example, a user may record a favorite player on a mobile phone while watching a game. In the video recording process, the user expects that the video recording picture can continuously follow the player and have better effect. The photographer (i.e., the user) needs to adjust the position of the mobile phone or adjust the size of the view angle by holding the mobile phone with the hand following the movement of the athlete, such as: the athlete can keep in the video picture all the time by rotating and zooming in, and the athlete keeps close to the same proportional size in the whole video picture. However, the picture taken by the user to adjust the video recording device during the video recording process is not smooth, and the effect is not ideal.
In this background art, the embodiment of the present application provides a video recording method, which can assist a user in shooting a better and more stable effect when the user is shooting a motion subject, so as to obtain a smoother video recording picture. In the method, when the user opens the mobile phone to record the video, the mobile phone may determine the first subject as the follow-up shooting object in response to a selection operation (e.g., a click operation) of the user on the first subject in the current mobile phone screen. Then, in the process of recording, the mobile phone can automatically adjust the proportion of the area occupied by the first main body in each frame of recording picture to the same or nearly the same proportion, such as: a first ratio. Illustratively, the first ratio may be 40%, 50%, etc. By means of the video recording method, the first main body is photographed with the following, and a smoother video recording picture can be obtained through shooting.
Alternatively, the first subject may be a person, an animal, or an object, without limitation. In the embodiments of the present application, the first subject will be described as an example.
The following takes application of the method provided by the embodiment of the present application to a mobile phone as an example, and the embodiment of the present application is exemplarily described with reference to the drawings.
Although the embodiment of the present application is applied to a mobile phone as an example, the video recording method provided in the embodiment of the present application is described. However, it should be understood that the video recording method provided in the embodiment of the present application may also be applied to other shooting devices such as a digital camera, a single lens reflex/micro single camera, a motion video camera, a pan-tilt camera, and an unmanned aerial vehicle, and the present application does not limit the shooting devices to which the video recording method can be applied.
In the description of the present application, "at least one" means one or more, "a plurality" means two or more. The words "first", "second", etc. are used merely to distinguish one element from another, and are not intended to particularly limit one feature. "and/or" is used to describe the association relationship of the associated objects, meaning that three relationships may exist. For example, a and/or B, may represent: a exists alone, A and B exist simultaneously, and B exists alone. The character "/" generally indicates that the former and latter associated objects are in an "or" relationship.
Fig. 1 shows a schematic structural diagram of an electronic device provided in an embodiment of the present application, where the electronic device may be a mobile phone. As shown in fig. 1, the mobile phone may include a processor 110, an external memory interface 120, an internal memory 121, a Universal Serial Bus (USB) interface 130, a charging management module 140, a power management module 141, a battery 142, an antenna 1, an antenna 2, a wireless communication module 160, an audio module 170, a speaker 170A, a receiver 170B, a microphone 170C, an earphone interface 170D, a sensor module 180, a button 190, a motor 191, an indicator 192, a camera 193, a display screen 194, and the like.
Processor 110 may include one or more processing units, such as: the processor 110 may include an Application Processor (AP), a modem processor, a Graphics Processing Unit (GPU), an Image Signal Processor (ISP), a controller, a memory, a video codec, a Digital Signal Processor (DSP), a baseband processor, and/or a neural-Network Processing Unit (NPU), etc. The different processing units may be separate devices or may be integrated into one or more processors. The controller can be the neural center and the command center of the mobile phone. The controller can generate an operation control signal according to the instruction operation code and the timing signal to complete the control of instruction fetching and instruction execution.
A memory may also be provided in processor 110 for storing instructions and data. In some embodiments, the memory in the processor 110 is a cache memory. The memory may hold instructions or data that have just been used or recycled by the processor 110. If the processor 110 needs to reuse the instruction or data, it can be called directly from the memory. Avoiding repeated accesses reduces the latency of the processor 110, thereby increasing the efficiency of the system.
In some embodiments, processor 110 may include one or more interfaces. The interface may include an integrated circuit (I2C) interface, an integrated circuit built-in audio (I2S) interface, a Pulse Code Modulation (PCM) interface, a universal asynchronous receiver/transmitter (UART) interface, a Mobile Industry Processor Interface (MIPI), a general-purpose input/output (GPIO) interface, a SIM interface, and/or a USB interface, etc.
The external memory interface 120 may be used to connect an external memory card, such as a Micro SD card, to extend the storage capability of the mobile phone. The external memory card communicates with the processor 110 through the external memory interface 120 to implement a data storage function. For example, files such as music, video, etc. are saved in an external memory card.
The internal memory 121 may be used to store computer-executable program code, which includes instructions. The processor 110 executes various functional applications of the cellular phone and data processing by executing instructions stored in the internal memory 121. The internal memory 121 may include a program storage area and a data storage area. The storage program area may store an operating system, an application program (such as a sound playing function, an image playing function, etc.) required by at least one function, and the like. The data storage area can store data (such as audio data, a phone book and the like) created in the use process of the mobile phone. In addition, the internal memory 121 may include a high-speed random access memory, and may further include a nonvolatile memory, such as at least one magnetic disk storage device, a flash memory device, a universal flash memory (UFS), and the like.
The charging management module 140 is configured to receive charging input from a charger. The charging management module 140 may also supply power to the mobile phone through the power management module 141 while charging the battery 142. The power management module 141 is used to connect the battery 142, the charging management module 140, and the processor 110. The power management module 141 may also receive input from the battery 142 to power the phone.
Optionally, the mobile phone may further include a mobile communication module 150, a Subscriber Identity Module (SIM) card interface 195, and the like. The wireless communication function of the mobile phone can be realized by the antenna 1, the antenna 2, the mobile communication module 150, the wireless communication module 160, the modem processor, the baseband processor, and the like. The antennas 1 and 2 are used for transmitting and receiving electromagnetic wave signals. Each antenna in the handset may be used to cover a single or multiple communication bands. Different antennas can also be multiplexed to improve the utilization of the antennas. For example: the antenna 1 may be multiplexed as a diversity antenna of a wireless local area network. In other embodiments, the antenna may be used in conjunction with a tuning switch.
The mobile phone can implement audio functions through the audio module 170, the speaker 170A, the receiver 170B, the microphone 170C, the earphone interface 170D, and the application processor. Such as music playing, recording, etc.
The sensor module 180 may include a pressure sensor 180A, a gyro sensor 180B, an air pressure sensor 180C, a magnetic sensor 180D, an acceleration sensor 180E, a distance sensor 180F, a proximity light sensor 180G, a fingerprint sensor 180H, a temperature sensor 180J, a touch sensor 180K, an ambient light sensor 180L, a bone conduction sensor 180M, and the like.
The mobile phone realizes the display function through the GPU, the display screen 194, the application processor and the like. The GPU is a microprocessor for image processing, and is connected to the display screen 194 and an application processor. The processor 110 may include one or more GPUs that execute program instructions to generate or alter display information.
The display screen 194 is used to display images, video, and the like. The display screen 194 includes a display panel. The display panel may adopt a Liquid Crystal Display (LCD), an organic light-emitting diode (OLED), an active-matrix organic light-emitting diode (active-matrix organic light-emitting diode, AMOLED), a flexible light-emitting diode (FLED), a miniature, a Micro-oeld, a quantum dot light-emitting diode (QLED), and the like. In some embodiments, the cell phone may include 1 or N display screens 194, with N being a positive integer greater than 1.
The mobile phone can realize shooting function through the ISP, the camera 193, the video codec, the GPU, the display screen 194, the application processor and the like. In some embodiments, the handset may include 1 or N cameras 193, which may include ultra-wide angle cameras. N is a positive integer greater than 1.
It is to be understood that the structure shown in fig. 1 does not constitute a specific limitation of the cellular phone. In some embodiments, the handset may also include more or fewer components than shown in fig. 1, or combine certain components, or split certain components, or a different arrangement of components, etc. Alternatively still, some of the components shown in FIG. 1 may be implemented in hardware, software, or a combination of software and hardware.
In addition, when the video recording method provided in the embodiment of the present application is applied to other shooting devices such as a digital camera, a single camera/micro single camera, a motion video camera, a pan-tilt camera, and an unmanned aerial vehicle, the specific structure of these other shooting devices may also be referred to as shown in fig. 1. Illustratively, other shooting devices may have additional or fewer components than the structure shown in fig. 1, and are not described in detail here.
Fig. 2 shows a schematic diagram of a shooting scene provided in an embodiment of the present application. As shown in fig. 2, the shooting scene includes two persons, a and B, and a background. When the user opens the mobile phone camera to shoot the scene, the mobile phone can acquire and display the picture of the scene shown in fig. 1. For example, fig. 3 shows a schematic diagram of a display screen of a mobile phone provided in an embodiment of the present application. As shown in fig. 3, the screen displayed by the mobile phone is the same as that shown in fig. 2. In this embodiment, the picture shown in fig. 3 may be referred to as an initial video picture, and the camera that acquires the initial video picture may be an ultra-wide-angle camera.
If the user needs to follow a in the scene shown in fig. 1, a selection operation may be performed on a mobile phone based on the display screen shown in fig. 2. The mobile phone can respond to the selection operation of the user on the A, and the A is determined as the follow shooting object. In the embodiment of the present application, a may also be referred to as a first body or a principal angle.
Alternatively, the initial video recording screen may be a preview screen before the mobile phone starts video recording, or may be a video recording screen after the mobile phone starts video recording. In other words, the user may select a on the display interface of the preview screen before the mobile phone starts recording, or may select a on the display interface of a certain frame screen during the recording after the mobile phone has started recording, which is not limited herein.
The following describes an exemplary manner in which the user performs a selection operation on a on the mobile phone to determine a as the follow shot object.
In one possible design, the user may perform a selection operation on a on the mobile phone by clicking an area where a is located in a display of the mobile phone. For example, fig. 4 shows a schematic diagram for determining a follow-shot object according to an embodiment of the present application. As shown in fig. 4, when the user determines that a is the subject of the follow-up shot, the user may click the area where a is located on the display screen of the mobile phone. The mobile phone can respond to the click operation of the user on the area where the A is located, and the A is determined as the follow shooting object.
For example, the mobile phone may acquire coordinate information of a in the initial video recording screen, appearance characteristics of a, and the like in response to a click operation performed by the user on the area where a is located.
Optionally, as shown in fig. 4, when the user clicks on the area where a is located, the mobile phone may present a dashed box in the area where a is located, so as to prompt the user that the selected follow-up object is a. Alternatively, in other embodiments, the dashed box may not be shown, and is not limited herein.
In another possible design, the user may perform a selection operation on a on the mobile phone by performing a frame selection operation or a sliding operation on an area where a is located in a display screen of the mobile phone. For example, fig. 5 shows a schematic diagram of another method for determining a follow-shot object according to an embodiment of the present application. As shown in fig. 5, when the user determines that a is the object of the follow shot, the user may slide a finger from the upper left corner to the lower right corner of the area where a is located on the display screen of the mobile phone, and select a through the sliding operation. The mobile phone can respond to the sliding operation of the user on the area where the A is located, and the A is determined as the follow shooting object. Or in other embodiments, the user may also slide the finger from the lower right corner to the upper left corner of the area where a is located, so as to select a, which is not limited herein.
Optionally, as shown in fig. 5, when the user performs a sliding operation on the area where a is located, the mobile phone may also present a dashed-line frame in the area where a is located, so as to prompt the user that the selected follow-up object is a. And the dotted line frame can gradually increase from the initial position of the sliding along with the sliding of the finger of the user, so that the frame selection of the area where the A is located is realized. Therefore, in the design, the selection operation of the user on the mobile phone for a may be referred to as a frame selection operation. Of course, in other embodiments, the dashed box may not be shown, and is not limited herein.
In yet another possible design, when the mobile phone captures a picture of the scene shown in fig. 1 for display, a selection mark may be simultaneously displayed on each subject that can be determined as a subject of follow-up shooting in the display picture. The user's selection of a on the mobile phone may be a click operation on the selection mark. For example, fig. 6 shows a schematic diagram of another method for determining a follow-shot object provided in an embodiment of the present application. As shown in fig. 6, when the mobile phone displays the scene shown in fig. 1, a "cross mark" may be displayed on a and B in the display screen at the same time as a selection mark, and when the user determines that a is the object to be followed, the "cross mark" on a may be clicked on the display screen of the mobile phone. The mobile phone can respond to the click operation of the user on the cross mark on the A, and the A is determined as the follow shot object.
In the design, the mobile phone can determine which follow-up objects allowed to be selected by the user are included in the display picture according to a preset target detection algorithm. For example, a portrait detection algorithm may be configured in the mobile phone, and the mobile phone may determine, through the portrait detection algorithm, which persons exist in the display frame that may be determined as the follow-up subject. The embodiment of the application does not limit the way how the mobile phone determines the subject which can be determined as the follow-up subject in the display picture.
Alternatively, the "cross mark" shown in fig. 6 is only one implementation of the selection mark in the present design, and in other implementations, the selection mark may also be a dashed frame, "circle mark" or the like, and is not limited herein.
According to any design mode, after the mobile phone responds to the selection operation of the user on the A, the A is determined as the follow shooting object, the A is adjusted to be approximately the same proportion size in the whole video recording picture in the video recording process, and the follow shooting of the A is achieved. In some embodiments, in the video recording process, an indication mark may be further displayed in the area where the object a is located in the display interface of the mobile phone, so as to prompt the user that the current follow-shot object is the object a. For example, fig. 7 shows a schematic diagram of another display screen of a mobile phone provided in an embodiment of the present application. As shown in fig. 7, after the mobile phone determines that a is the tracking object in response to the click operation performed by the user on the area where a is located, in the process of recording, a solid frame may also be presented in the area where a is located in the display interface of the mobile phone, so as to prompt the user that the current tracking object is a.
Or, in some other embodiments, in the video recording process, another indication mark may be presented in the area where the object a is located in the display interface of the mobile phone, so as to prompt the user that the current follow-up object is the object a. For example, the indicator mark may also be a "cross mark", "circle mark", or the like.
In another embodiment, after the mobile phone determines that a is the object to be followed by shooting in response to the selection operation of the user on a, during the video recording process, no indication mark may be displayed, and only the object to be followed by shooting a may be used, which is not limited herein.
Although the scene shown in the above description includes two persons, a and B, the embodiments of the present application are also applicable to a scene with only a person, or a scene with more persons.
The following is a brief description of a specific implementation manner of adjusting a to the same or nearly the same scale size in the whole video recording screen in the video recording process after the mobile phone determines a as the follow-up object in response to the selection operation of the user on a.
Fig. 8 shows a schematic view of the heel-and-shoot principle provided by the embodiment of the present application. Referring to fig. 8, in the embodiment of the present application, a specific implementation manner of adjusting a in the whole video recording picture to be close to the same scale size in the video recording process of the mobile phone is as follows. Alternatively, the following steps may be performed by a processor of the handset. For example, the module for cropping the sensor raw image in step 1, the anti-shake module, the subject detection module, the tracking module, the smoothing module, and the like may be an algorithm or a program on the processor. The processor may include: ISP, NPU, GPU, etc.
1) And acquiring an original sensing (sensor raw) image, cutting the sensor raw image to obtain an anti-shake input image OMNQ, and outputting the anti-shake input image OMNQ to the anti-shake module. Exemplarily, the resolution of the sensor raw map may be 5408 × 3680; the resolution of the anti-shake input map OMNQ may be 4608 × 3072.
The sensor raw image is raw data of a CMOS or CCD image sensor of the mobile phone converting a captured light source (photon) signal into a digital signal.
2) The anti-shake module performs torsion (warp) transformation on four vertices O, M, N, Q of the anti-shake input graph OMNQ according to information of a multi-frame gyroscope (gyro) to obtain graphs O1, M1, N1, and Q1, and further determines a central point D1 (for example: d1 may be the intersection of intersecting lines of four vertices O1, M1, N1, Q1). Then, the detection input map with the resolution of 3840 × 2160 is cut out from the anti-shake input map OMNQ by centering on the point D1, and is output to the main body detection module.
3) The main body detection module can detect the positions of all people in the detection input graph by adopting a detection algorithm and outputs the positions to the tracking algorithm module. For example, for the scenario shown in fig. 2 described above, the subject detection module may detect coordinate information of a and B in the detection input map.
Alternatively, to increase the detection speed, the subject detection module first needs to down-sample the detection input map to 224 × 224 small maps; then, the feature maps are sent to a detection model for forward propagation to obtain feature maps; and finally, carrying out post-processing on feature maps to obtain the detection confidence coefficient and the position coordinate of all people, such as: detection confidence and location coordinates of a and B.
That is, for example, the input of the detection algorithm may be a small graph of 224 × 224 obtained by the above-mentioned down-sampling, such as: the human body is a BGR map in a vertical state. Alternatively, a rotation buffer (buffer) is required when the screen is erected. The output of the detection algorithm is feature maps in multiple scales (scales). The post-processing refers to comprehensively performing post-processing on feature maps of multiple scales output by the NPU to obtain the detection confidence and the position coordinates of all people.
4) the tracking module obtains location information of all people in the detection input graph (for example: which may be detection box information for each person), the following two processing scenarios may be included.
Scene 1: non-overlapping (overlap) scenarios. In a non-overlay scene, there may be only a person in the picture; alternatively, there may be multiple people including A, but others (e.g., B, C) are far from A and do not overlap in the cell phone camera view (A and B are consistent with a non-overlapping scene as shown in FIG. 2). At this time, the tracking module may calculate an euclidean distance between the detected coordinate information of each person in the current frame and the coordinate information of a in the previous frame, and determine that the coordinate information with the smallest euclidean distance from the coordinate information of a in the previous frame is the coordinate information of a in the current frame, thereby determining the position of a in the current frame. The image of the current frame is the anti-shake input map OMNQ. The coordinate information refers to the coordinates of the pixel points, and may be an area coordinate or a point coordinate.
Assuming that the current frame is the ith frame (i is an integer greater than 1), the previous frame is the (i-1) th frame; in scene 1, the coordinate information of A in the (i-1) th frame comprises two cases; case 1: the coordinate information of the A in the (i-1) th frame is obtained by combining the coordinate information of the A in the (i-2) th frame by the tracking module according to the mode of the step 4). Case 2: and the coordinate information of the A in the (i-1) th frame is obtained when the mobile phone determines the A as the follow-up shooting object in response to the selection operation of the user on the A.
Scene 2: overlapping (overlay) scenes. In the overlap scene, a plurality of people including A exist in the picture, and other people (such as B, C) and A are overlapped in the view angle of the mobile phone camera. For example, fig. 9 shows a schematic diagram of another shooting scene provided in the embodiment of the present application. As shown in fig. 9, when a moves to the front of B in the shooting scene shown in fig. 2, a and B overlap in the viewing angle of the camera of the mobile phone, and different parts may overlap at different time periods in the actual scene as a and B move. At this time, the tracking module may introduce a (re-identification, ReID) module to determine the position of a in the current frame according to the image appearance feature. Specifically, the ReID module may extract appearance features of each person in the current frame, and then compare the appearance features with features in the feature library of a to determine coordinate information of a in the current frame. Such as: the cosine distance between the appearance characteristic of each person in the current frame and the characteristic in the characteristic library of A can be calculated, and the person with the minimum cosine distance is A, so that the coordinate information of A can be obtained. The feature library of a may be obtained by extracting several frames in the video recording process and caching the extracted frames.
Alternatively, the manner of determining the position of a in the scene 1 is also applicable to the scene 2, and the manner of determining the position of a in the scene 2 is also applicable to the scene 1, which is not particularly limited herein.
In the step 4), the tracking module determines the position of the A in the current frame by combining the detection information of the historical frame, so that the problems of false detection, missed detection and the like can be effectively suppressed.
Optionally, when the position of a in the current frame is determined in step 4), the determination may be performed by combining the detection information of more historical frames, and if the requirement is met between a certain person in the current frame and the multiple historical frames, the person is determined to be a. In this way, the mismatch problem can be avoided to some extent.
After the tracking module in step 4) obtains the coordinate information of the current frame of the A, the coordinate information of the current frame of the A can be output to the smoothing module.
5) And the smoothing module outputs the coordinate of the central point H of the A under the coordinate system of the anti-shake input image OMNQ to the rear end for warp conversion and crop (crop) according to the coordinate information of the A in the current frame and combined with the historical frame coordinate information of the A through quasi-static position locking and a multi-frame smoothing algorithm, and finally outputs the coordinate to the previewing and video recording module. Illustratively, the final output map may refer to the output map shown in fig. 8 with a size of 1920 x 1080. The center point of the output graph is H.
It should be noted that the size of the final output graph may be a preset size, such as: 1920 x 1080, the present application does not limit the size of the output graph. For example, in some embodiments, the preset size may be related to the display resolution of the mobile phone, or may be customized by the user.
And after the steps 1) to 5) are carried out, the output image finally output to the preview and video module is the video picture corresponding to the current frame. In the video recording method, after the user selects A, the processing modes of the steps 1) to 5) are adopted for each frame of image, so that the proportion of A in the finally obtained video in the whole video recording picture is nearly the same or the same. The picture of the video shot by the video method is smoother.
For example, fig. 10 shows a schematic diagram of a display screen of a mobile phone provided in an embodiment of the present application. Assuming that for the scene shown in fig. 2, after the mobile phone determines a as the follow-up object in response to the selection operation of the user on a, if a moves to a position farther from the mobile phone during the video recording process, at this time, the video recording picture shot by the mobile phone according to the video recording method may be as shown in fig. 10, and the proportional size of a in the whole video recording picture may be approximately the same as or the same as that in fig. 7.
Alternatively, fig. 11 shows a schematic diagram of a display screen of a mobile phone provided in an embodiment of the present application. Assuming that for the scene shown in fig. 2, after the mobile phone determines a as the follow-up object in response to the selection operation of the user on a, if a moves to a position closer to the mobile phone during the video recording process, then the video recording picture shot by the mobile phone according to the video recording method may be as shown in fig. 11, and the scale size of a in the whole video recording picture may be approximately the same as or the same as that in fig. 7 and 10.
In a possible design, the smoothing module may send the coordinate information of the current frame a to an Auto Focus (AF) module of the mobile phone. The AF module can adjust the position of a focusing focus to A according to the coordinate information of A in the current frame, and follow the A to record images, so that the video picture can be continuously focused to A. The design can avoid the problem of blurred video recording pictures caused by the fact that a focusing focus is not on A when A moves in the video recording process.
In another possible design, the smoothing module may also send a part of the coordinate information of a in the current frame to an AF module of the mobile phone, so that the AF module adjusts the position of the focus. For example, assuming that an area where a is located in the recording picture of the current frame is the first area, the smoothing module may send the coordinate information of an area 20% of the center of the first area to the AF module, so that the AF module adjusts the position of the focus point. Wherein, the central 20% area of the first area may be referred to as the second area.
Taking the video recording picture of the current frame as the picture shown in fig. 7 as an example, fig. 12 shows a schematic relationship diagram of the first region and the second region provided in the embodiment of the present application. As shown in FIG. 12, the first region may be a solid-line region in the recording picture of the current frame, similar to that shown in FIG. 7. The second region may be a region in the center of the first region, and the second region accounts for 20% of the first region. Of course, 20% is only an exemplary illustration, and the proportion of the second region to the first region may be 30%, 50%, etc. in the design, and is not limited herein.
In another possible design, the smoothing module may also send the coordinate information of the center point H of a to the AF module of the mobile phone according to the coordinate information of a in the current frame, so that the AF module adjusts the position of the focus. The center point H of a can refer to the foregoing embodiments, and will not be described herein.
Optionally, the smoothing module described above sends coordinate information to the AF module, so that the scene for the AF module to adjust the position of the focus is achieved on the premise that the mobile phone has determined a as the object to be followed by shooting in response to the selection operation of the user on a. In this embodiment of the application, when the user does not perform the selection operation of a (which may be referred to as an initial state), the smoothing module may output the coordinate information of the center point D1 determined by the anti-shake module in step 2) of the embodiment shown in fig. 8 to the AF module by default, so that the AF module adjusts the position of the focus point.
For example, the mobile phone may obtain coordinate information of a center point of a jth frame image in a video recording process before determining that the first subject is a follow-up shooting object, and adjust a focusing focus when shooting a (j + 1) th frame image to the center point according to the coordinate information of the center point of the jth frame image; wherein j is an integer greater than 0.
In some possible scenarios, for example, when a is in overlap with other people to a high degree, or a exits the screen for a short time, the tracking module may not be able to determine the coordinate information of a in the current frame. For such a scene, the smoothing module may send coordinate information of a in the previous frame, or coordinate information of a second area in the previous frame, or coordinate information of a center point H in the previous frame to the AF module, for the AF module to adjust the position of the focus.
Optionally, in this embodiment of the application, the smoothing module may further send coordinate information of the central point H of the current frame a in a sensor raw image coordinate system to the sensor, so as to guide a dynamic crop of a next frame sensor. When the sensor performs dynamic crop, boundary protection can be performed according to coordinate information of the central point H under a sensor raw image coordinate system, for example: it can be guaranteed that the 4608 × 3072 anti-shake input map OMNQ, centered on the center point H, is within the range of the sensor raw map.
Optionally, in this embodiment of the application, if a is located at an edge of a frame, when the smoothing module generates an output graph with a size of 1920 × 1080, the smoothing module may also perform boundary protection according to a central point H of a, so as to ensure that a boundary is not cut when a back end performs crop. Such as: it can be guaranteed that the output graph with 1920 x 1080 size is within the range of the detection input graph by taking the central point H as the center.
In some embodiments, if a is located at the edge of the screen, the mobile phone may further give a prompt message on the display interface to prompt the user to move the mobile phone, so that a can be located in the screen. For example, the prompt may be a direction indicator for prompting the user to move the mobile phone in a first direction. Exemplarily, fig. 13 shows a schematic diagram of a display screen of a mobile phone provided in an embodiment of the present application. As shown in fig. 13, in the embodiment of the present application, after the mobile phone determines that a is the object to be followed by shooting in response to the selection operation performed by the user on a, in the process of recording, if it is detected that the position of a in the display screen moves to the edge of the right screen, a rightward arrow may be displayed at the edge of the right screen as the prompt information, which prompts the user to move the mobile phone rightward so that a can be located in the screen.
It should be noted that the arrows shown in fig. 13 are only one implementation of the prompt message, and in other embodiments, the prompt message may also be a text prompt or other symbols, which is not limited herein. In addition, the display area of the prompt message may be the edge of the screen as shown in fig. 13, or may be other areas (e.g., the middle area) on the screen, and is not limited as well.
Fig. 14 shows a schematic diagram of a display screen of a mobile phone provided in an embodiment of the present application. Optionally, as shown in fig. 13, in the embodiment of the present application, when the mobile phone displays the prompt message (taking an arrow as an example) at the edge of the screen, the user is prompted to move the mobile phone, and at the same time, a movement reference line may also be displayed around the prompt message, for example: the dotted lines on both sides of the arrow in fig. 14. If the mobile phone shakes up and down (relative to the horizontal plane) in the moving process, the arrow displayed by the mobile phone moves up and down in the screen. When the user moves the mobile phone, the direction and the amplitude of the shaking of the mobile phone can be judged according to the up-down movement of the arrow. Dotted lines on the upper side and the lower side of the arrow can be used for prompting a user to adjust when the mobile phone is moved, and the arrow is controlled to move within the range of the dotted lines on the upper side and the lower side of the arrow, so that a shot video picture is more stable.
Alternatively, when the embodiments of the present application are applied to some specific scenarios, such as: when the mobile phone is only used for following shooting of a person, if the mobile phone detects that only one person exists in the shot picture, for example, the mobile phone can detect that only one person exists in the current frame image by adopting a portrait detection algorithm, and the mobile phone can also default to track the person under the condition that the user does not manually select.
The foregoing embodiments of the present application mainly describe a specific implementation scheme that, in response to a selection operation performed by a user on a first main body (e.g., a), a mobile phone determines the first main body as a follow-up shooting object and performs follow-up shooting and video recording on the first main body. Optionally, on the basis of the foregoing embodiment, the embodiment of the present application may further switch the tracking object in the video recording process. Such as: during the process of tracking and recording the first subject, the mobile phone can also respond to the selection operation of the user on the second subject (such as B) to switch the tracking object from the first subject to the second subject.
In the following, with reference to the scenario shown in fig. 2, an example will be described in which the tracking target is switched from a to B during the tracking recording of a by the user.
Assume that the mobile phone determines a as a tracking object in response to a user's selection operation on a according to the method described in the foregoing embodiment, and performs tracking recording on a. In this case, if the user desires to switch the tracking target from a to B, the selection operation for B may be performed on the mobile phone. The mobile phone can respond to the selection operation of the user on B, switch the following shooting object from A to B, and then follow-shoot and record the image of B according to the same method as the method for follow-shoot and record the image of A in the previous embodiment.
Similar to the foregoing operation of selecting a by the user on the mobile phone, in a possible design, the operation of selecting B by the user on the mobile phone may be a click operation on an area where B is located in a display screen of the mobile phone in the process of performing the shooting and recording on a by the mobile phone. For example, fig. 15 shows a schematic diagram of switching a follow shot object according to an embodiment of the present application. As shown in fig. 15, when the user desires to switch the tracking object from a to B, the user can directly perform a click operation on the area where B is located in the display screen of the mobile phone. The mobile phone can respond to the click operation of the user on the area where the B is located, and the follow-shot object is switched from the A to the B.
Or, in some other possible designs, in the process of performing the following shooting and recording on the a by the mobile phone, the user may perform the selection operation on the B on the mobile phone, or perform the frame selection operation or the sliding operation on the area where the B is located in the display screen of the mobile phone. Alternatively, a click operation or the like may be performed on the selection mark displayed on B on the display screen. The selection method is similar to the selection operation of the user on the mobile phone a, and reference may be made to the foregoing embodiment, which is not repeated herein.
In still another alternative design, for the above-mentioned solid line frame presented in the area where a is located in the display interface of the mobile phone in the process of performing the follow-up shooting and recording of a shown in fig. 7, and for prompting the user of the scene with the current follow-up shooting object being a, the user may further slide the solid line frame presented in the area where a is located to the area where B is located on the mobile phone. The mobile phone can respond to the sliding operation and switch the follow shooting object from A to B. For example, fig. 16 shows a schematic diagram of another switching follow-up object provided in an embodiment of the present application. The process of the user sliding the solid line box presented in the area of a to the area of B on the mobile phone can be referred to fig. 16.
In addition, it should be further noted that other steps after the user selects a in the foregoing embodiment of the present application are also applicable to the scheme after the follow-shot object is switched to B, and are not described herein again.
Optionally, in some embodiments of the present application, during the process of performing the follow-up shooting and image recording on the first subject (e.g., a), the user may further select to add at least one follow-up shooting object, and the mobile phone may perform the follow-up shooting and image recording on the first subject and the added at least one follow-up shooting object simultaneously in response to an operation performed by the user to add the follow-up shooting object.
The following is an example of adding a second body (e.g., B) selected by the user.
Fig. 17 shows a schematic diagram of adding a follow shot object according to an embodiment of the present application. As shown in fig. 17, during the process of performing the follow-up shooting and recording on a, a solid frame appears in the area where a is located in the display interface of the mobile phone. When the user desires to add B as the follow shot object as well, the solid line frame can be dragged and enlarged from the edge of the solid line frame with a finger so that the solid line frame selects a and B simultaneously. The mobile phone can respond to the frame selection operation of the solid frame controlled by the user on the B, determine the B as the follow-shot object, and simultaneously carry out follow-shot image recording on the A and the B.
It is understood that the above-mentioned box selection operation of the user control solid box pair B shown in fig. 17 is only an exemplary illustration, and in other embodiments, the operation of adding the follow shot object performed by the user may be implemented in other manners. For example, the mobile phone may provide an add-function control on the display interface, when the user clicks the add-function control, the display interface of the mobile phone may enter the interface of the add-follow object, at this time, the user may perform a selection operation on another main body (e.g., B) displayed in the interface, and the selection operation refers to the foregoing embodiment. The mobile phone can respond to the selection operation of the user on the B in the interface of the added follow-up shooting object, determine the B as the follow-up shooting object, and simultaneously carry out follow-up shooting and image recording on the A and the B. The specific implementation manner of the operation of adding the follow shooting object performed by the user is not limited in the application.
In the process of performing the follow-up shooting and image recording on the image a, the mobile phone responds to the operation of adding B as the follow-up shooting object performed by the user, determines B as the follow-up shooting object, and then performs the follow-up shooting and image recording on the image a and the image B at the same time. The difference is that when the mobile phone simultaneously performs the following shooting and image recording on the a and the B, the tracking module in the step 4) needs to obtain the coordinate information of the a at the current frame and the coordinate information of the B at the current frame, and outputs the coordinate information of the a at the current frame and the coordinate information of the B at the current frame to the smoothing module. In step 5), the smoothing module outputs the coordinate of the common central point H1 of the a and B under the coordinate system of the anti-shake input diagram to the rear end for warp transformation and cropping (crop) according to the coordinate information of the a in the current frame and the coordinate information of the B in the current frame, by combining the coordinate information of the a and B in the historical frames and by a quasi-stationary position locking and multi-frame smoothing algorithm, and finally outputs the coordinate to the preview and video module. The center point of the final output graph is H1. Wherein, the common center point H1 of a and B may be determined by: the smoothing module first determines a detection box which can select both A and B boxes, and then determines the center point of the detection box as the common center point H1 of A and B.
The above description shows that, in the process of performing the follow-up shooting and image recording on the first subject (for example, a), the user selects to add at least one follow-up shooting object, and the mobile phone can respond to the operation of adding the follow-up shooting object performed by the user, and simultaneously perform the follow-up shooting and image recording on the first subject and the added at least one follow-up shooting object. In other embodiments, the user may also select multiple subjects as follow-up subjects in an initial state (when no follow-up subject is selected). The mobile phone can respond to the selection operation of the user on the plurality of subjects, determine the plurality of subjects as the follow-up shooting objects, and simultaneously carry out follow-up shooting and image recording on the plurality of subjects.
For example, similar to the manner described above in fig. 5 in which the user selects a as the follow shot object, the user may select a common area of a and B on the mobile phone, such as: and sliding the finger from the upper left corner to the lower right corner of the area where the A and the B are jointly located in the display picture of the mobile phone. The mobile phone can respond to the sliding operation of the user on the area where the A and the B are located together, and both the A and the B are determined as the follow shooting objects. However, here too, the specific implementation of the operation of selecting a plurality of subjects as the follow-up subjects by the user is not limited.
The mobile phone may determine the plurality of subjects as the following subjects in response to the selection operation of the user on the plurality of subjects, and the specific implementation of simultaneously performing the following shooting and image recording on the plurality of subjects may refer to the scheme for simultaneously performing the following shooting and image recording on the first subject and the added at least one following shooting subject by the mobile phone, which is not described herein again.
In the foregoing embodiments of the present application, an application of the video recording method in a video recording scene is described, but it should be noted that the video recording method provided in the embodiments of the present application may also be applied to a photo taking scene. For example, the user may perform a selection operation on the first subject at the photo preview interface, and the mobile phone may determine the first subject as the subject to be followed in response to the selection operation on the first subject by the user. Then, when the user clicks a photo button (the button may be an entity button or a virtual button displayed on the screen) of the mobile phone, the mobile phone may obtain a photo picture of the first main body in a manner similar to that in the video recording method described in the foregoing embodiment, and adjust the ratio of the area occupied by the first main body in the photo picture to a first ratio. That is, in the shooting scene, the processing flow or principle for each frame of image is similar to that in the aforementioned video recording scene, and is not repeated here.
Corresponding to the method described in the foregoing embodiment, an apparatus may be applied to a terminal device with a video recording function, and is configured to implement the foregoing video recording method. The functions of the device can be realized by hardware, and can also be realized by executing corresponding software by hardware. The hardware or software includes one or more modules or units corresponding to the above-described functions. For example, fig. 18 shows a schematic structural diagram of an apparatus provided in an embodiment of the present application. As shown in fig. 18, the apparatus may include a display unit 1801, a processing unit 1802.
Among them, the display unit 1801 is used for displaying an initial video recording screen including the first main body. A processing unit 1802 configured to determine, in response to a selection operation on a first subject, that the first subject is a follow-up subject; in the video recording process after the first main body is determined to be the follow-shot object, the proportion of the area occupied by the first main body in each frame of video recording picture to the video recording picture is adjusted to be a first proportion.
In a possible design, the processing unit 1802 is further configured to, during a video recording process after the first subject is determined to be a follow-up shooting object, acquire coordinate information of the first subject in an ith frame image, and adjust a focus point when the i +1 th frame image is shot according to the coordinate information of the first subject in the ith frame image; wherein i is an integer greater than 0.
Optionally, the processing unit 1802 is further configured to, if the coordinate information of the first subject in the i-th frame image is not acquired, adjust the focus point when the i + 1-th frame image is captured according to the coordinate information of the first subject in the i-1-th frame image.
In a possible design, the processing unit 1802 is further configured to, in a video recording process before determining that the first subject is a follow-up subject, acquire coordinate information of a center point of a j-th frame image, and adjust a focus point when the j + 1-th frame image is captured to the center point according to the coordinate information of the center point of the j-th frame image; wherein j is an integer greater than 0.
In a possible design, the processing unit 1802 is further configured to, in a video recording process after the first subject is determined to be the object to be followed, display, by the display unit 1801, a prompt message for prompting the user to move the terminal device in the first direction on a video recording screen corresponding to each frame of image when it is detected that the first subject is located at a screen edge.
In one possible design, the hint information is a direction identifier; the processing unit 1802 is further configured to display moving reference lines on upper and lower sides of the direction indicator through the display unit 1801, where the moving reference lines are used to prompt a user to control the direction indicator not to exceed a range of the moving reference lines when the user moves the terminal device.
In one possible design, the initial recording frame further includes a second main body; processing unit 1802, further configured to, in a video recording process after determining that the first subject is the follow-up subject, switch the follow-up subject from the first subject to the second subject in response to a first selection operation on the second subject; in the video recording process after the follow-shot object is switched from the first main body to the second main body, the proportion of the area occupied by the second main body in each frame of video recording picture to the video recording picture is adjusted to be the first proportion.
In one possible design, the initial video frame further includes at least one second main body; processing unit 1802, further configured to, in a video recording process after determining that the first subject is the follow-up subject, in response to a second selection operation on at least one second subject, determine at least one second subject as a follow-up subject as well; in the video recording process after at least one second main body is also determined as a follow-up shooting object, the proportion of the area occupied by the first main body and the at least one second main body in each video recording picture together to the video recording picture is adjusted to be a first proportion.
Alternatively, the processing unit 1802 may specifically include a module for cropping a sensor raw image, an anti-shake module, a main body detection module, a tracking module, a smoothing module, and the like.
It should be understood that the division of units or modules (hereinafter referred to as units) in the above apparatus is only a division of logical functions, and may be wholly or partially integrated into one physical entity or physically separated in actual implementation. And the units in the device can be realized in the form of software called by the processing element; or may be implemented entirely in hardware; part of the units can also be realized in the form of software called by a processing element, and part of the units can be realized in the form of hardware.
For example, each unit may be a processing element separately set up, or may be implemented by being integrated into a chip of the apparatus, or may be stored in a memory in the form of a program, and a function of the unit may be called and executed by a processing element of the apparatus. In addition, all or part of the units can be integrated together or can be independently realized. The processing element described herein, which may also be referred to as a processor, may be an integrated circuit having signal processing capabilities. In the implementation process, the steps of the method or the units above may be implemented by integrated logic circuits of hardware in a processor element or in a form called by software through the processor element.
In one example, the units in the above apparatus may be one or more integrated circuits configured to implement the above method, such as: one or more ASICs, or one or more DSPs, or one or more FPGAs, or a combination of at least two of these integrated circuit forms.
As another example, when a unit in a device may be implemented in the form of a processing element scheduler, the processing element may be a general purpose processor, such as a CPU or other processor capable of invoking programs. As another example, these units may be integrated together and implemented in the form of a system-on-a-chip (SOC).
In one implementation, the means for implementing the respective corresponding steps of the above method by the above apparatus may be implemented in the form of a processing element scheduler. For example, the apparatus may include a processing element and a memory element, the processing element calling a program stored by the memory element to perform the method described in the above method embodiments. The memory elements may be memory elements on the same chip as the processing elements, i.e. on-chip memory elements.
In another implementation, the program for performing the above method may be in a memory element on a different chip than the processing element, i.e. an off-chip memory element. At this time, the processing element calls or loads a program from the off-chip storage element onto the on-chip storage element to call and execute the method described in the above method embodiment.
For example, the embodiments of the present application may also provide an apparatus, such as: an electronic device may include: a processor, a memory for storing instructions executable by the processor. The processor is configured to execute the above instructions, so that the electronic device implements the method according to the foregoing embodiments. For example, the electronic device may be the terminal device with a video recording function described in the foregoing embodiment. The memory may be located within the electronic device or external to the electronic device. And the processor includes one or more.
In another implementation, the unit of the apparatus for implementing the steps of the above method may be configured as one or more processing elements, which may be disposed on the terminal with the voice control function, where the processing element may be an integrated circuit, for example: one or more ASICs, or one or more DSPs, or one or more FPGAs, or a combination of these types of integrated circuits. These integrated circuits may be integrated together to form a chip.
For example, an embodiment of the present application further provides a chip, where the chip may be applied to the terminal device with the video recording function. The chip includes one or more interface circuits and one or more processors; the interface circuit and the processor are interconnected through a line; the processor receives and executes computer instructions from the memory of the electronic device through the interface circuitry to implement the methods described in the method embodiments above.
Embodiments of the present application further provide a computer program product, which includes computer readable code, when the computer readable code is run in an electronic device, the electronic device is caused to implement the method described in the foregoing embodiments.
Through the above description of the embodiments, it is clear to those skilled in the art that, for convenience and simplicity of description, the foregoing division of the functional modules is merely used as an example, and in practical applications, the above function distribution may be completed by different functional modules according to needs, that is, the internal structure of the device may be divided into different functional modules to complete all or part of the above described functions.
In the several embodiments provided in the present application, it should be understood that the disclosed apparatus and method may be implemented in other ways. For example, the above-described device embodiments are merely illustrative, and for example, the division of the modules or units is only one logical functional division, and there may be other divisions when actually implemented, for example, a plurality of units or components may be combined or may be integrated into another device, or some features may be omitted, or not executed. In addition, the shown or discussed mutual coupling or direct coupling or communication connection may be an indirect coupling or communication connection through some interfaces, devices or units, and may be in an electrical, mechanical or other form.
The units described as separate parts may or may not be physically separate, and parts displayed as units may be one physical unit or a plurality of physical units, that is, may be located in one place, or may be distributed in a plurality of different places. Some or all of the units can be selected according to actual needs to achieve the purpose of the solution of the embodiment.
In addition, functional units in the embodiments of the present application may be integrated into one processing unit, or each unit may exist alone physically, or two or more units are integrated into one unit. The integrated unit can be realized in a form of hardware, and can also be realized in a form of a software functional unit.
The integrated unit, if implemented in the form of a software functional unit and sold or used as a stand-alone product, may be stored in a readable storage medium. Based on such understanding, the technical solutions of the embodiments of the present application may be embodied in the form of software products, such as: and (5) programming. The software product is stored in a program product, such as a computer readable storage medium, and includes several instructions for causing a device (which may be a single chip, a chip, or the like) or a processor (processor) to perform all or part of the steps of the methods described in the embodiments of the present application. And the aforementioned storage medium includes: various media capable of storing program codes, such as a U disk, a removable hard disk, a ROM, a RAM, a magnetic disk, or an optical disk.
For example, embodiments of the present application may also provide a computer-readable storage medium having stored thereon computer program instructions. The computer program instructions, when executed by the electronic device, cause the electronic device to implement the method as described in the preceding method embodiments.
The above description is only an embodiment of the present application, but the scope of the present application is not limited thereto, and any changes or substitutions within the technical scope of the present disclosure should be covered by the scope of the present application. Therefore, the protection scope of the present application shall be subject to the protection scope of the claims.

Claims (10)

1. A video recording method is applied to terminal equipment with a video recording function, and the method comprises the following steps:
displaying an initial video frame including a first subject;
in response to a selection operation on the first subject, determining that the first subject is a follow-up shooting object;
in the video recording process after the first main body is determined to be the follow-shot object, the proportion of the area occupied by the first main body in each frame of video recording picture to the video recording picture is adjusted to be a first proportion.
2. The method of claim 1, further comprising:
in the video recording process after the first main body is determined to be the follow-up shooting object, acquiring coordinate information of the first main body in an ith frame image, and adjusting a focusing focus when the (i + 1) th frame image is shot according to the coordinate information of the first main body in the ith frame image; wherein i is an integer greater than 0.
3. The method of claim 2, further comprising:
if the coordinate information of the first main body in the ith frame image is not acquired, adjusting a focusing focus when the (i + 1) th frame image is shot according to the coordinate information of the first main body in the (i-1) th frame image.
4. The method according to any one of claims 1-3, further comprising:
in the video recording process before the first main body is determined to be the follow-up shooting object, coordinate information of a central point of a j frame image is obtained, and a focusing focus when the j +1 frame image is shot is adjusted to the central point according to the coordinate information of the central point of the j frame image; wherein j is an integer greater than 0.
5. The method according to any one of claims 1-4, further comprising:
in the video recording process after the first main body is determined to be the follow-shot object, when the video recording picture corresponding to each frame of picture is detected that the first main body is located at the picture edge, prompt information is displayed in a first area of the video recording picture, and the prompt information is used for prompting a user to move the terminal equipment to a first direction.
6. The method of claim 5, wherein the prompt is a direction indicator; the method further comprises the following steps:
and displaying moving reference lines on the upper side and the lower side of the direction mark, wherein the moving reference lines are used for prompting a user to control the direction mark not to exceed the range of the moving reference lines when the user moves the terminal equipment.
7. The method of any of claims 1-6, wherein the initial video frame further comprises a second subject; the method further comprises the following steps:
in the video recording process after the first subject is determined to be the follow-up subject, responding to a first selection operation on the second subject, and switching the follow-up subject from the first subject to the second subject;
in the video recording process after the follow-shot object is switched from the first main body to the second main body, the proportion of the area occupied by the second main body in each video recording picture and the video recording picture is adjusted to be a first proportion.
8. The method of any of claims 1-7, wherein the initial video frame further comprises at least a second subject; the method further comprises the following steps:
in the video recording process after the first subject is determined to be the follow-up shooting object, responding to a second selection operation on at least one second subject, and determining at least one second subject to be the follow-up shooting object;
in the video recording process after at least one second main body is also determined as a follow-up shooting object, the proportion of the area occupied by the first main body and the at least one second main body in each video recording picture together to the video recording picture is adjusted to be a first proportion.
9. An electronic device, comprising: a processor, a memory for storing the processor-executable instructions;
the processor is configured to, when executing the instructions, cause the electronic device to implement the method of any of claims 1-8.
10. A computer readable storage medium having stored thereon computer program instructions; it is characterized in that the preparation method is characterized in that,
the computer program instructions, when executed by an electronic device, cause the electronic device to implement the method of any of claims 1-8.
CN202011142554.7A 2020-10-22 2020-10-22 Video recording method, electronic device and storage medium Active CN114390191B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202011142554.7A CN114390191B (en) 2020-10-22 2020-10-22 Video recording method, electronic device and storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202011142554.7A CN114390191B (en) 2020-10-22 2020-10-22 Video recording method, electronic device and storage medium

Publications (2)

Publication Number Publication Date
CN114390191A true CN114390191A (en) 2022-04-22
CN114390191B CN114390191B (en) 2023-11-10

Family

ID=81194899

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202011142554.7A Active CN114390191B (en) 2020-10-22 2020-10-22 Video recording method, electronic device and storage medium

Country Status (1)

Country Link
CN (1) CN114390191B (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN116112782A (en) * 2022-05-25 2023-05-12 荣耀终端有限公司 Video recording method and related device

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107015572A (en) * 2014-07-30 2017-08-04 深圳市大疆创新科技有限公司 Target tracking system and method
CN107135377A (en) * 2017-05-27 2017-09-05 深圳市景阳科技股份有限公司 Monitor automatic tracking method and device
CN107948523A (en) * 2017-12-19 2018-04-20 维沃移动通信有限公司 A kind of image pickup method and mobile terminal
CN108683849A (en) * 2018-05-15 2018-10-19 维沃移动通信有限公司 A kind of image acquiring method and terminal
CN109246351A (en) * 2018-07-20 2019-01-18 维沃移动通信有限公司 A kind of patterning process and terminal device

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107015572A (en) * 2014-07-30 2017-08-04 深圳市大疆创新科技有限公司 Target tracking system and method
CN107135377A (en) * 2017-05-27 2017-09-05 深圳市景阳科技股份有限公司 Monitor automatic tracking method and device
CN107948523A (en) * 2017-12-19 2018-04-20 维沃移动通信有限公司 A kind of image pickup method and mobile terminal
CN108683849A (en) * 2018-05-15 2018-10-19 维沃移动通信有限公司 A kind of image acquiring method and terminal
CN109246351A (en) * 2018-07-20 2019-01-18 维沃移动通信有限公司 A kind of patterning process and terminal device

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN116112782A (en) * 2022-05-25 2023-05-12 荣耀终端有限公司 Video recording method and related device
WO2023226721A1 (en) * 2022-05-25 2023-11-30 荣耀终端有限公司 Video recording method and related apparatus
CN116112782B (en) * 2022-05-25 2024-04-02 荣耀终端有限公司 Video recording method and related device

Also Published As

Publication number Publication date
CN114390191B (en) 2023-11-10

Similar Documents

Publication Publication Date Title
KR102381713B1 (en) Photographic method, photographic apparatus, and mobile terminal
US20200221014A1 (en) Image pickup device and method of tracking subject thereof
WO2022042776A1 (en) Photographing method and terminal
WO2022062318A1 (en) Photographing method and device
EP4135310A1 (en) Photographing method and device
CN112153272B (en) Image shooting method and electronic equipment
CN114339102B (en) Video recording method and equipment
CN114845059B (en) Shooting method and related equipment
CN113727018B (en) Shooting method and equipment
CN115484403B (en) Video recording method and related device
KR20220090158A (en) Electronic device for editing video using objects of interest and operating method thereof
CN114697530B (en) Photographing method and device for intelligent view finding recommendation
CN114390191B (en) Video recording method, electronic device and storage medium
CN112532854B (en) Image processing method and electronic equipment
CN113497880A (en) Method for shooting image and electronic equipment
EP4366289A1 (en) Photographing method and related apparatus
CN115278043B (en) Target tracking method and related device
CN117278839A (en) Shooting method, electronic equipment and storage medium
CN115914860A (en) Shooting method and electronic equipment
CN116405776A (en) Multi-lens video recording method and related equipment
CN114827445B (en) Image processing method and related device
CN115623319B (en) Shooting method and electronic equipment
US20240007740A1 (en) Photographing control method and device
CN116723382B (en) Shooting method and related equipment
CN116939123A (en) Photographing method and electronic equipment

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant