CN111654622B - Shooting focusing method and device, electronic equipment and storage medium - Google Patents

Shooting focusing method and device, electronic equipment and storage medium Download PDF

Info

Publication number
CN111654622B
CN111654622B CN202010465750.1A CN202010465750A CN111654622B CN 111654622 B CN111654622 B CN 111654622B CN 202010465750 A CN202010465750 A CN 202010465750A CN 111654622 B CN111654622 B CN 111654622B
Authority
CN
China
Prior art keywords
video
target
signal
expression
shooting
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202010465750.1A
Other languages
Chinese (zh)
Other versions
CN111654622A (en
Inventor
曾柏泉
陈露兰
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Vivo Mobile Communication Co Ltd
Original Assignee
Vivo Mobile Communication Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Vivo Mobile Communication Co Ltd filed Critical Vivo Mobile Communication Co Ltd
Priority to CN202010465750.1A priority Critical patent/CN111654622B/en
Publication of CN111654622A publication Critical patent/CN111654622A/en
Application granted granted Critical
Publication of CN111654622B publication Critical patent/CN111654622B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/60Control of cameras or camera modules
    • H04N23/67Focus control based on electronic image sensor signals
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/60Control of cameras or camera modules
    • H04N23/62Control of parameters via user interfaces
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/60Control of cameras or camera modules
    • H04N23/67Focus control based on electronic image sensor signals
    • H04N23/675Focus control based on electronic image sensor signals comprising setting of focusing regions

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Human Computer Interaction (AREA)
  • Studio Devices (AREA)

Abstract

The application discloses a shooting focusing method and device, electronic equipment and a storage medium, and belongs to the field of electronic equipment. The method comprises the following steps: under the condition that a shooting preview interface is displayed, acquiring a first video including a target shooting object, wherein the first video is formed by a plurality of frames of historical preview pictures displayed on the shooting preview interface within a first preset time period; amplifying expression actions of a target shooting object in the first video to obtain a second video; identifying an expression area of a target shooting object in a second video; focusing the expression area on the shooting preview interface. By utilizing the embodiment of the application, the accuracy of the focusing area can be ensured, and the user does not need to manually operate the focusing, so that the user can conveniently shoot.

Description

Shooting focusing method and device, electronic equipment and storage medium
Technical Field
The application belongs to the field of electronic equipment, and particularly relates to a shooting focusing method and device, electronic equipment and a storage medium.
Background
With the development of shooting technology, there are various ways to realize shooting focusing at present. In one mode, a user touches a certain area of the shooting preview interface with a finger to focus the area. The other mode is that the target shooting object on the shooting preview interface is identified by using an artificial intelligence technology, and the target shooting object is focused.
However, the method of focusing by finger touch can achieve accurate focusing, but requires manual operation by a user, which is very inconvenient for the user. Although the manner of focusing on the recognized target photographic subject does not require manual operation by the user, there are cases where the focusing area is not very accurate. For example, focusing is performed on a person who is confused in the shooting preview interface.
Disclosure of Invention
The embodiment of the application aims to provide a shooting focusing method, a shooting focusing device, electronic equipment and a storage medium, and can solve the problems that a user needs to manually operate focusing and the accuracy of a focusing area is not high.
In order to solve the technical problem, the present application is implemented as follows:
in a first aspect, an embodiment of the present application provides a shooting focusing method, including:
under the condition that a shooting preview interface is displayed, acquiring a first video including a target shooting object, wherein the first video is formed by a plurality of frames of historical preview pictures displayed on the shooting preview interface within a first preset time period;
amplifying the expression action of the target shooting object in the first video to obtain a second video;
identifying an expression area of the target photographic object in the second video;
focusing the expression area on the shooting preview interface.
In a second aspect, an embodiment of the present application provides a shooting focusing apparatus, including:
the video acquisition module is used for acquiring a first video comprising a target shooting object under the condition that a shooting preview interface is displayed, wherein the first video is formed by a plurality of frames of historical preview pictures displayed on the shooting preview interface within a first preset time period;
the amplification processing module is used for amplifying expression actions of the target shooting object in the first video to obtain a second video;
the expression area identification module is used for identifying the expression area of the target shooting object in the second video;
and the area focusing module is used for focusing the expression area on the shooting preview interface.
In a third aspect, an embodiment of the present application provides an electronic device, which includes a processor, a memory, and a program or instructions stored in the memory and executable on the processor, and when executed by the processor, the program or instructions implement the steps of the method according to the first aspect.
In a fourth aspect, embodiments of the present application provide a readable storage medium, on which a program or instructions are stored, which when executed by a processor implement the steps of the method according to the first aspect.
In a fifth aspect, an embodiment of the present application provides a chip, where the chip includes a processor and a communication interface, where the communication interface is coupled to the processor, and the processor is configured to execute a program or instructions to implement the method according to the first aspect.
In the embodiment of the present application, in the case where a shooting preview interface is displayed, a first video formed by a multi-frame history preview screen is acquired. And then amplifying the expression action of the target shooting object in the first video to obtain a second video. And then recognizing the expression area of the target shooting object in the second video, and focusing the expression area on the shooting preview interface. The expression action in the first video is amplified, so that the expression action which is not obvious in the first video is more obvious, and the expression area of the target shooting object can be more accurately identified. Then, focusing is carried out based on the expression area, and the accuracy of the focusing area can be ensured. In addition, the embodiment of the application can automatically focus without manual operation of a user, so that the user can conveniently shoot.
Drawings
FIG. 1 is a schematic flowchart of a photographing focusing method according to an embodiment of the present disclosure;
FIG. 2 is a schematic flowchart of a photographing focusing method according to another embodiment of the present disclosure;
FIG. 3 is a schematic diagram of a preview interface for shooting including a plurality of emotions according to an embodiment of the present disclosure;
FIG. 4 is a schematic diagram of a capture preview interface including a sensitivity adjustment control according to one embodiment provided herein;
FIG. 5 is a schematic structural diagram of a photographing focusing device according to an embodiment provided in the present application;
fig. 6 is a schematic diagram of a hardware structure of an electronic device implementing the embodiment of the present application.
Detailed Description
The technical solutions in the embodiments of the present application will be clearly and completely described below with reference to the drawings in the embodiments of the present application, and it is obvious that the described embodiments are some, but not all, embodiments of the present application. All other embodiments, which can be derived by a person skilled in the art from the embodiments given herein without making any creative effort, shall fall within the protection scope of the present application.
The terms first, second and the like in the description and in the claims of the present application are used for distinguishing between similar elements and not necessarily for describing a particular sequential or chronological order. It is to be understood that the data so used is interchangeable under appropriate circumstances such that the embodiments of the application are capable of operation in sequences other than those illustrated or described herein. In addition, "and/or" in the specification and claims means at least one of connected objects, a character "/" generally means that a preceding and succeeding related objects are in an "or" relationship.
The following describes the shooting focusing method provided in the embodiments of the present application in detail through specific embodiments and application scenarios thereof with reference to the accompanying drawings.
Fig. 1 is a schematic flowchart of a shooting focusing method according to an embodiment of the present disclosure. The shooting focusing method can be applied to a shooting focusing device, and as shown in fig. 1, the shooting focusing method includes:
and 102, under the condition that the shooting preview interface is displayed, acquiring a first video including a target shooting object, wherein the first video is formed by a plurality of frames of historical preview pictures displayed on the shooting preview interface within a first preset time period.
Note that, while the shooting preview interface is displayed, the preview screen displayed on the shooting preview interface is cached. Therefore, the multi-frame historical preview picture displayed on the shooting preview interface in the first preset time period can be obtained from the cached preview picture, and then the first video is formed.
For example, when a shooting preview interface is displayed, a multi-frame history preview screen displayed on the shooting preview interface within the last 5 seconds is acquired in the cached preview screen.
And 104, amplifying the expression action of the target shooting object in the first video to obtain a second video.
And 106, identifying the expression area of the target shooting object in the second video.
Optionally, identifying an expressive area of the target photographic subject in the second video may include: and extracting a plurality of regions in the second video, and inputting each extracted region into a pre-trained deep learning model to obtain whether each extracted region is an expression region.
Optionally, identifying an expressive area of the target photographic subject in the second video may include: extracting a plurality of regions in the second video; and performing similarity matching on each extracted area and a prestored facial expression image to obtain whether each extracted area is an expression area.
And step 108, focusing the expression area on the shooting preview interface.
For example, the expression area in the second video is a screenshot of a smile expression action on a certain frame image of the second video. And identifying an expression area matched with the screenshot on the shooting preview interface, and focusing the expression area matched with the screenshot on the shooting preview interface.
In the embodiment of the present application, in the case where a shooting preview interface is displayed, a first video formed by a multi-frame history preview screen is acquired. And then, amplifying the expression action of the target shooting object in the first video to obtain a second video. And then recognizing the expression area of the target shooting object in the second video, and focusing the expression area on the shooting preview interface. The expression action in the first video is amplified, so that the expression action which is not obvious in the first video becomes more obvious, and the expression area of the target shooting object can be more accurately identified. Then, focusing is carried out based on the expression area, and the accuracy of the focusing area can be ensured. In addition, through the embodiment of the application, the user can focus the user by making the expression action without manual operation of the user, so that the user can conveniently shoot.
In one or more embodiments of the present application, step 104 can include:
filtering the signal of the first video by using a target filter to obtain a first signal of the first video, which is related to the expression action;
amplifying the first signal according to the target amplification factor to obtain a target signal;
and generating a second video according to the target signal.
Optionally, before the target filter is used to filter the signal of the first video to obtain the first signal of the first video related to the expression motion, the initial filter used for amplification of the european style video is trained by using sample data including the expression motion to obtain the target filter used for amplification of the expression motion. The target filter may be used directly after the camera is started. Alternatively, the target filter may be trained once, and after the training is completed, the target filter may be used directly for filtering, and then the training may not be performed again.
Alternatively, the target magnification may be a preset magnification or a magnification set by the user.
The expression action can be amplified by using the European video amplification technology in the embodiment of the application. Specifically, the target filter is used to filter the signal of the first video, so as to obtain a first signal related to the expression motion, and a second signal unrelated to the expression motion is filtered out. And then, amplifying the first signal related to the expression action according to the target amplification factor to obtain a target signal after the expression action is amplified. And generating a second video according to the target signal amplified by the expression action. Compared with the first video, the expression action of the target shooting object in the second video is more obvious, so that the expression area of the target shooting object can be more accurately identified by using the second video.
In one or more embodiments of the present application, before filtering, by using a target filter, a signal of a first video to obtain a first signal of the first video related to an expressive motion, the shooting focusing method further includes:
converting a time domain signal of a first video into frequency domain signals of a plurality of frequency bands;
the method for filtering the signal of the first video by using the target filter to obtain the first signal of the first video related to the expression and the motion includes:
filtering the frequency domain signal by using a target filter to obtain a first signal related to the expression action in the frequency domain signal;
generating a second video from the target signal, comprising:
and converting the target signal and the second signal from a frequency domain to a time domain to obtain a second video.
Optionally, generating the second video according to the target signal may include: and generating a second video according to the target signal and a second signal irrelevant to the expression and the action.
Generating a second video according to the target signal and a second signal unrelated to the expression and the motion may specifically include: and performing frequency domain-to-time domain conversion on the target signal and the second signal which is filtered out during filtering and is irrelevant to the expression action to obtain a second video.
The following describes a photographing focusing method according to an embodiment of the present application with a specific example. The shooting focusing method is applied to a shooting focusing device. As shown in fig. 2, the photographing focusing method includes:
in the step 202, under the condition that the shooting preview interface is displayed, a first video including a target shooting object is acquired, wherein the first video is formed by a plurality of frames of history preview pictures displayed on the shooting preview interface within a first preset time period.
Step 204, converting the time domain signal of the first video into a frequency domain signal of a plurality of frequency bands.
And step 206, filtering the frequency domain signal by using a target filter to obtain a first signal related to the expression action and a second signal unrelated to the expression action in the frequency domain signal.
And 208, amplifying the first signal according to the target amplification factor to obtain a target signal.
Step 210, converting the target signal from a frequency domain to a time domain to obtain a second video.
In step 212, the expression area of the target photographic subject in the second video is identified.
And step 214, focusing the expression area on the shooting preview interface.
It should be noted that step 202 in the present embodiment is similar to step 102 in the embodiment of fig. 1, step 212 in the present embodiment is similar to step 106 in the embodiment of fig. 1, and step 214 in the present embodiment is similar to step 108 in the embodiment of fig. 1, and thus, the detailed description is not repeated here. The following mainly describes steps 204 to 210 in the embodiment of the present application.
In step 204 of the embodiment of the present application, a time domain signal of the first video may be converted into frequency domain signals of a plurality of frequency bands by using fast fourier transform.
It should be noted that the time domain describes a mathematical function or a physical signal versus time. For example, a time domain waveform of a signal may express the change of the signal over time. If discrete time is considered, the function or signal in the time domain, the value at each discrete time point is known. If continuous time is considered, the value of the function or signal at any time is known. When studying a time domain signal, an oscilloscope is often used to convert the signal into its time domain waveform. A time domain signal is waveform data in which a wave of a certain frequency has a time axis as the x-axis and an amplitude as the y-axis.
The video time domain signal is a digital signal obtained by encoding a video analog signal, and the digital signal is waveform data in the direction of the time axis. Wherein, the analog signal of the video is the real image signal obtained for each frame of image.
The frequency domain is a coordinate system used to describe the frequency characteristics of a signal. The argument of the frequency domain is the frequency, i.e. the horizontal axis is the frequency and the vertical axis is the amplitude of the frequency signal, i.e. the spectrogram in general. The spectrogram describes the frequency structure of a signal and the relationship of the frequency to the amplitude of the frequency signal. That is, the frequency signal is spectrum data at a certain time with the frequency as the x-axis and the amplitude as the y-axis.
In step 206 of the embodiment of the present application, the frequency domain signal is filtered by using a target filter, so as to obtain a first signal related to the expression motion in the frequency domain signal. The first signal may include some signals of micro expression actions. For example, the first signal includes at least one of a signal that the user's mouth is slightly raised, a signal that the user's eyebrows are raised, and a signal that the user is blinking.
In addition, by filtering the frequency domain signal, a second signal that is not related to the expressive action is also obtained from the frequency domain signal, and the second signal is a signal other than the signal related to the expressive action from the frequency domain signal.
In step 208 of the embodiment of the present application, the amplitude of the first signal is amplified, so that the first signal is amplified.
In step 210 of this embodiment of the present application, since the target signal after signal amplification is a frequency domain signal, in order to perform video restoration, the target signal is subjected to frequency domain to time domain conversion by using inverse fourier transform to obtain a second video, thereby implementing video restoration.
In the embodiment of the application, a time domain signal of a first video is converted into a frequency domain signal, then a first signal related to an expression action in the frequency domain signal is obtained through filtering, and the first signal is amplified. After the first signal is amplified, video restoration is performed based on the amplified target signal. Therefore, the expression action in the first video is amplified, so that the expression action of the user can be more accurately identified, and the expression action of the user can be more accurately focused.
In one or more embodiments of the present application, before filtering, by using a target filter, a signal of a first video to obtain a first signal of the first video related to an expressive action, the shooting focusing method further includes:
displaying at least one piece of identification information on a shooting preview interface, wherein each piece of identification information is used for indicating a corresponding expression action, and the expression action is obtained by amplifying a multi-frame history preview picture displayed on the shooting preview interface within a preset time period, wherein the preset time period is a time period before the time period of the first video;
receiving a first input of a user to target identification information in at least one piece of identification information, wherein the target identification information is used for indicating a target expression action;
the method for filtering the signal of the first video by using the target filter to obtain the first signal of the first video related to the expression and the motion includes:
and responding to the first input, and filtering the signals of the first video by using a target filter corresponding to the target expression action to obtain first signals of the first video, wherein the first signals are related to the target expression action.
The following describes an embodiment of the present application with reference to fig. 3.
As shown in fig. 3, a plurality of expressive motions are displayed on the shooting preview interface, and when a first input that the user selects a middle blinking motion is received, a signal of the first video is filtered by using a target filter corresponding to the blinking motion, so that a first signal related to the blinking motion is obtained. And then, amplifying the first signal, and performing video restoration to obtain a second video. Then the blinking motion is more pronounced in the second video than in the first video. Therefore, the fact that the user makes the blinking motion can be more accurately recognized according to the second video.
In the embodiment of the application, the user can select the expression action needing focusing according to the requirement of the user, so that the shooting focusing device can focus the expression action selected by the user on the shooting preview interface, and the focusing area can meet the shooting requirement of the user.
The following describes embodiments of the present application by way of a specific example.
After the camera is started, the preview picture displayed on the shooting preview interface is cached. When the preview screen of 5 seconds has been buffered, a one-segment video is generated from the preview screen of 5 seconds. And amplifying the expression action of the target shooting object of the video. The amplification processing manner here is similar to the above-described amplification processing manner for the first video, and is not repeated here. The filter used for enlarging the video generated from the 5-second preview screen may be a preset filter.
And then, the expression actions of the target shooting object in the video after the amplification processing are recognized, and 3 expression actions are recognized. And displaying the 3 expressive actions on a shooting preview interface for the user to select a target expressive action from the 3 expressive actions.
And after the user selects the target expression action, a preview screen displayed on the shooting preview interface within the last 5 seconds is acquired again in the cached history preview screen, and a first video is formed by the preview screen. And then, filtering the signal of the first video by using a target filter corresponding to the target expression motion. And then, signal amplification, video restoration, expression area identification and focusing are performed, and the steps are already described above, so that details are not repeated here.
In the embodiment of the application, a series of expression actions identified after video amplification processing are displayed on a shooting preview interface, so that a user can select different expression actions as focusing objects, and a focusing area meets shooting requirements of the user.
In one or more embodiments of the present application, acquiring a plurality of expressive motions of a target photographic subject includes: and recognizing a plurality of expression actions of the target shooting object on the shooting preview interface by utilizing face recognition.
In one or more embodiments of the present application, before the first signal is amplified according to a predetermined amplification factor to obtain the target signal, the shooting focusing method further includes:
receiving a second input of the user to the sensitivity adjusting control under the condition that the sensitivity adjusting control is displayed on the shooting preview interface;
in response to the second input, obtaining a target sensitivity value associated with the second input;
and determining the target magnification factor associated with the target sensitivity value according to the association relationship between the preset sensitivity value and the magnification factor.
For example, referring to fig. 4, a sensitivity adjustment control 302 is displayed on the shooting preview interface, and the user may slide on the sensitivity adjustment control 302 to adjust the target sensitivity value.
Wherein, the higher the target sensitivity value is, the larger the target magnification is, so that the more obvious the expression action in the second video is. Therefore, even if the user makes a tiny expressive motion, the expressive area with the expressive motion can be identified, and the expressive area on the shooting preview interface can be focused. Conversely, the lower the target sensitivity value is, the smaller the target magnification is, and the user needs to make an expression motion of a larger magnitude in order to identify an expression region having the expression motion.
In the embodiment of the application, the user can adjust the target sensitivity value according to the own needs, so that the target magnification factor is adjusted, and the shooting focusing requirements of the user can be met. Wherein, the user can adjust sensitivity to suitable numerical value, can prevent the condition that the mistake was focused.
In one or more embodiments of the present application, after the expression area of the target photographic object in the second video is identified and before focusing is performed on the expression area on the photographic preview interface, the photographic focusing method further includes:
displaying at least one expression area on a shooting preview interface under the condition that the expression area comprises at least one expression area;
receiving a third input of the user to a target expression area in the at least one expression area;
focusing an expression area on a shooting preview interface, comprising:
and responding to a third input, and focusing the target expression area on the shooting preview interface.
For example, when a person a is photographed, a nearby person B is confused with the shooting preview interface. In this case, the smiling expression region of the person a and the blinking expression region of the person B in the second video are recognized. Then, a smiling expression area and a blinking expression area are displayed on the photographing preview interface. And focusing the smiling expression area on the shooting preview interface under the condition that the user selects the smiling expression area. Therefore, focusing on the expression of the person A is achieved.
In the embodiment of the application, the at least one identified expression area is displayed on the shooting preview interface, and the user selects to focus the target expression area, so that the focusing result meets the requirements of the user.
In one or more embodiments of the present application, after focusing on an expressive area on a shooting preview interface, the shooting focusing method further includes:
and shooting to obtain a target shooting image under the condition that the expression action in the expression area is a preset expression action.
Optionally, after the expression area of the target photographic object in the second video is identified, the expression action in the expression area in the second video is acquired. And then, taking a picture or a video to obtain a target shooting image under the condition that the expression action in the expression area is a preset expression action. The target captured image may be a photograph or a frame image in a video.
The following description will be given taking a predetermined expression motion as a smiling motion as an example.
For example, an expression area of the user a in the second video is recognized, and the expression action in the expression area is recognized as a smiling action. Based on the above, focusing is performed on the area of the smiling action of the user A on the shooting preview interface, and shooting is automatically performed after focusing to obtain a picture or a video of the user A.
In the embodiment of the application, when the expression action in the expression area is the preset expression action, the shooting is automatically performed without manually clicking a shooting button by a user, so that the user is prevented from manually triggering a shooting instruction during the shooting, for example, the user is not required to click the shooting button. Therefore, the user can conveniently shoot.
It should be noted that, in the above embodiments, the execution subject is taken as an example of the shooting focusing apparatus to describe the shooting focusing method. However, the executing subject of the shooting focusing method provided by the embodiment of the present application is not limited to the shooting focusing device, and may also be a control module in the shooting focusing device for executing the loading of the shooting focusing method.
Correspondingly to the shooting focusing method provided by the embodiment of the application, the embodiment of the application provides a shooting focusing device. Fig. 5 is a schematic structural diagram of a photographing focusing device according to an embodiment provided in the present application. As shown in fig. 5, the photographing focusing device includes:
a video obtaining module 402, configured to, in a case where a shooting preview interface is displayed, obtain a first video including a target shooting object, where the first video is formed by multiple frames of history preview pictures displayed on the shooting preview interface within a first predetermined time period;
the amplification processing module 404 is configured to perform amplification processing on the expression action of the target shooting object in the first video to obtain a second video;
an expression area recognition module 406, configured to recognize an expression area of the target photographic object in the second video;
and the area focusing module 408 is configured to focus the expressive area on the shooting preview interface.
In the embodiment of the present application, in the case where a shooting preview interface is displayed, a first video formed by a multi-frame history preview screen is acquired. And then amplifying the expression action of the target shooting object in the first video to obtain a second video. And then recognizing the expression area of the target shooting object in the second video, and focusing the expression area on the shooting preview interface. The expression action in the first video is amplified, so that the expression action which is not obvious in the first video becomes more obvious, and the expression area of the target shooting object can be more accurately identified. Then, focusing is carried out based on the expression area, and the accuracy of the focusing area can be ensured. In addition, the embodiment of the application can automatically focus without manual operation of a user, so that the user can conveniently shoot.
In one or more embodiments of the present application, the amplification processing module 404 includes:
the signal filtering module is used for filtering the signal of the first video by using a target filter to obtain a first signal of the first video related to the expression action;
the signal amplification module is used for amplifying the first signal according to the target amplification factor to obtain a target signal;
and the video restoration module is used for generating a second video according to the target signal.
In the embodiment of the present application, a target filter is first used to filter a signal of a first video, so as to obtain a first signal related to an expression motion, and a second signal unrelated to the expression motion is filtered. And then, amplifying the first signal related to the expression action according to the target amplification factor to obtain a target signal after the expression action is amplified. And then, according to the target signal after the expression action is amplified and the second signal which is filtered out during filtering and is irrelevant to the expression action, carrying out video restoration to obtain a second video. Compared with the first video, the expression action of the target shooting object in the second video is more obvious, so that the expression area of the target shooting object can be more accurately identified by using the second video.
In one or more embodiments of the present application, the photographing focusing apparatus further includes:
the signal conversion module is used for converting the time domain signals of the first video into frequency domain signals of a plurality of frequency bands;
the signal filtering module is specifically used for filtering the frequency domain signal by using a target filter to obtain a first signal related to the expression action in the frequency domain signal;
and the video restoration module is specifically used for converting the target signal from a frequency domain to a time domain to obtain a second video.
In the embodiment of the application, a time domain signal of a first video is converted into a frequency domain signal, then a first signal related to an expression action in the frequency domain signal is obtained through filtering, and the first signal is amplified. After the first signal is amplified, video restoration is performed based on the amplified target signal. Therefore, the expression action in the first video is amplified, so that the expression action of the user can be more accurately identified, and the expression action of the user can be more accurately focused.
In one or more embodiments of the present application, the photographing focusing apparatus further includes:
the expression action display module is used for displaying at least one piece of identification information on the shooting preview interface, each piece of identification information is used for indicating a corresponding expression action, and the expression action is obtained by amplifying a multi-frame history preview picture displayed on the shooting preview interface within a preset time period, wherein the preset time period is a time period before the time period of the first video;
the first input receiving module is used for receiving first input of a user on target identification information in at least one piece of identification information, wherein the target identification information is used for indicating a target expression action;
the signal filtering module is specifically configured to filter, in response to the first input, a signal of the first video by using a target filter corresponding to the target expression action, so as to obtain a first signal of the first video related to the target expression action.
In the embodiment of the application, the user can select the expression action needing focusing according to the requirement of the user, so that the shooting focusing device can focus the expression action selected by the user on the shooting preview interface, and the focusing area meets the shooting requirement of the user.
In one or more embodiments of the present application, the photographing focusing apparatus further includes:
the second input receiving module is used for receiving second input of the sensitivity adjusting control by a user under the condition that the sensitivity adjusting control is displayed on the shooting preview interface;
a second input response module for, in response to a second input, obtaining a target sensitivity value associated with the second input;
and the amplification factor determining module is used for determining the target amplification factor related to the target sensitivity value according to the incidence relation between the preset sensitivity value and the amplification factor.
In the embodiment of the application, the user can adjust the sensitivity value of the target according to the requirement of the user, so that the target magnification factor is adjusted, and the shooting focusing requirement of the user can be met. The user can adjust the sensitivity to a proper value, and the situation of mis-focusing can be prevented.
In one or more embodiments of the present application, the photographing focusing apparatus further includes:
the expression area display module is used for displaying at least one expression area on the shooting preview interface under the condition that the expression area comprises at least one expression area;
the second input receiving module is used for receiving second input of a user to a target expression area in the at least one expression area;
the area focusing module 408 is specifically configured to focus the target expression area on the shooting preview interface in response to the second input.
In the embodiment of the application, the at least one identified expression area is displayed on the shooting preview interface, and the user selects to focus the target expression area, so that the focusing result meets the requirements of the user.
In one or more embodiments of the present application, the photographing focusing apparatus further includes:
and the shooting module is used for shooting under the condition that the expression action in the expression area is a preset expression action to obtain a target shooting image.
In the embodiment of the application, when the expression action in the expression area is the preset expression action, the shooting is automatically performed without manually clicking a shooting button by a user, so that the user is prevented from manually triggering a shooting instruction during shooting, for example, the user is not required to click the shooting button. Therefore, the user can conveniently shoot.
The shooting focusing device in the embodiment of the present application may be a device, or may be a component, an integrated circuit, or a chip in the device. The device can be mobile electronic equipment or non-mobile electronic equipment. By way of example, the mobile electronic device may be a mobile phone, a tablet computer, a notebook computer, a palm top computer, a vehicle-mounted electronic device, a wearable device, an ultra-mobile personal computer (UMPC), a netbook or a Personal Digital Assistant (PDA), and the like, and the non-mobile electronic device may be a server, a Network Attached Storage (NAS), a Personal Computer (PC), a Television (TV), a teller machine or a self-service machine, and the like, and the embodiments of the present application are not particularly limited.
The shooting focusing device in the embodiment of the present application may be a device having an operating system. The operating system may be an Android (Android) operating system, an ios operating system, or other possible operating systems, and embodiments of the present application are not limited specifically.
The shooting focusing device provided in the embodiment of the present application can implement each process implemented by the shooting focusing device in the method embodiment of fig. 1 or fig. 2, and is not described herein again to avoid repetition.
Optionally, an embodiment of the present application further provides an electronic device, which includes a processor, a memory, and a program or an instruction stored in the memory and executable on the processor, where the program or the instruction is executed by the processor to implement each process of the above shooting focusing method embodiment, and can achieve the same technical effect, and in order to avoid repetition, details are not repeated here.
It should be noted that the electronic devices in the embodiments of the present application include the mobile electronic devices and the non-mobile electronic devices described above.
Fig. 6 is a schematic diagram of a hardware structure of an electronic device implementing an embodiment of the present application.
The electronic device 500 includes, but is not limited to: a radio frequency unit 501, a network module 502, an audio output unit 503, an input unit 504, a sensor 505, a display unit 506, a user input unit 507, an interface unit 508, a memory 509, a processor 510, and the like.
The input unit 504 may include, among other things, a graphics processor 5041 and a microphone 5042. The display unit 506 includes a display panel 5061. The user input unit 507 may include a touch panel 5071 and other input devices 5072. The memory 509 may include application programs and an operating system.
Those skilled in the art will appreciate that the electronic device 500 may further include a power supply (e.g., a battery) for supplying power to various components, and the power supply may be logically connected to the processor 510 via a power management system, so as to implement functions of managing charging, discharging, and power consumption via the power management system. The electronic device structure shown in fig. 5 does not constitute a limitation of the electronic device, and the electronic device may include more or less components than those shown, or combine some components, or arrange different components, and thus, the description is omitted here.
The processor 510 is configured to, in a case where the display unit 506 displays a shooting preview interface, acquire a first video including a target shooting object, where the first video is formed by multiple frames of history preview pictures displayed on the shooting preview interface within a first predetermined time period;
the processor 510 is further configured to amplify the expression action of the target photographic object in the first video to obtain a second video;
a processor 510, further configured to identify an expressive region of the target photographic subject in the second video;
the processor 510 is further configured to focus on the expressive area on the shooting preview interface.
In the embodiment of the present application, in the case where a shooting preview interface is displayed, a first video formed by a multi-frame history preview screen is acquired. And then, amplifying the expression action of the target shooting object in the first video to obtain a second video. And then recognizing the expression area of the target shooting object in the second video, and focusing the expression area on the shooting preview interface. The expression action in the first video is amplified, so that the expression action which is not obvious in the first video is more obvious, and the expression area of the target shooting object can be more accurately identified. Then, focusing is carried out based on the expression area, and the accuracy of the focusing area can be ensured. In addition, the embodiment of the application can realize automatic focusing, and does not need manual focusing operation of a user, so that the user can conveniently shoot.
In one or more embodiments of the present application, the processor 510 is specifically configured to filter, by using a target filter, a signal of a first video to obtain a first signal of the first video, where the first signal is related to an expressive action; amplifying the first signal according to the target amplification factor to obtain a target signal; and generating a second video according to the target signal.
In the embodiment of the application, the target filter is used to filter the signal of the first video, so as to obtain the first signal related to the expression motion, and the second signal unrelated to the expression motion is filtered. And then, amplifying the first signal related to the expression action according to the target amplification factor to obtain a target signal after the expression action is amplified. And then, video restoration is carried out according to the target signal after the expression action is amplified and the second signal which is filtered during filtering and is irrelevant to the expression action, so that a second video is obtained. Compared with the first video, the expression action of the target shooting object in the second video is more obvious, so that the expression area of the target shooting object can be more accurately identified by using the second video.
In one or more embodiments of the present application, the processor 510 is further configured to convert a time domain signal of the first video into frequency domain signals of a plurality of frequency bands;
a processor 510, specifically configured to filter the frequency domain signal by using a target filter, to obtain a first signal related to an expression action in the frequency domain signal;
the processor 510 is specifically configured to perform frequency-domain to time-domain conversion on the target signal to obtain a second video.
In the embodiment of the application, a time domain signal of a first video is converted into a frequency domain signal, then a first signal related to an expression action in the frequency domain signal is obtained through filtering, and the first signal is amplified. After the first signal is amplified, video restoration is performed based on the amplified target signal. Therefore, the expression action in the first video is amplified, so that the expression action of the user can be more accurately recognized, and the expression action of the user can be more accurately focused.
In one or more embodiments of the present application, the processor 510 is further configured to obtain a plurality of expressive motions of the target photographic subject;
a display unit 506, configured to display at least one piece of identification information on the shooting preview interface, where each piece of identification information is used to indicate a corresponding expressive action, and the expressive action is obtained by performing amplification processing on a multi-frame history preview picture displayed on the shooting preview interface within a predetermined time period, where the predetermined time period is a time period before a time period of the first video;
a user input unit 507, configured to receive a first input of target identification information from the at least one identification information by a user, where the target identification information is used to indicate a target expression action;
the processor 510 is specifically configured to filter, in response to the first input, a signal of the first video by using a target filter corresponding to the target expression and action, so as to obtain a first signal of the first video related to the target expression and action.
In the embodiment of the application, the user can select the target identification information for indicating the target expression action according to the requirement of the user, so that the shooting focusing device can focus the target expression action on the shooting preview interface, and the focusing area meets the shooting requirement of the user.
In one or more embodiments of the present application, the user input unit 507 is further configured to receive a second input of the sensitivity adjustment control from the user when the sensitivity adjustment control is displayed on the shooting preview interface;
processor 510, further configured to, in response to a second input, obtain a target sensitivity value associated with the second input; and determining the target magnification factor associated with the target sensitivity value according to the association relation between the preset sensitivity value and the magnification factor.
In the embodiment of the application, the user can adjust the sensitivity value of the target according to the requirement of the user, so that the target magnification factor is adjusted, and the shooting focusing requirement of the user can be met. The user can adjust the sensitivity to a proper value, and the situation of mis-focusing can be prevented.
In one or more embodiments of the present application, the display unit 506 is configured to display at least one emoji area on the shooting preview interface if the emoji area includes at least one emoji area;
a user input unit 507 for receiving a third input of the user to a target expression area of the at least one expression area;
the processor 510 is specifically configured to focus on the target expression area on the shooting preview interface in response to a third input.
In the embodiment of the application, the at least one identified expression area is displayed on the shooting preview interface, and the user selects to focus the target expression area, so that the focusing result meets the requirements of the user.
In one or more embodiments of the application, the processor 510 is further configured to perform shooting to obtain a target shooting image when the expression motion in the expression area is a predetermined expression motion.
In the embodiment of the application, when the expression action in the expression area is the preset expression action, the shooting is automatically performed without manually clicking a shooting button by a user, so that the user is prevented from manually triggering a shooting instruction during shooting, for example, the user is not required to click the shooting button. Therefore, the user can conveniently shoot.
The embodiment of the present application further provides a readable storage medium, where a program or an instruction is stored on the readable storage medium, and when the program or the instruction is executed by a processor, the program or the instruction implements each process of the shooting focusing method embodiment, and can achieve the same technical effect, and in order to avoid repetition, details are not repeated here.
The processor is the processor in the electronic device in the above embodiment. Readable storage media include computer readable storage media such as Read-Only Memory (ROM), random Access Memory (RAM), magnetic or optical disk, and so on.
The embodiment of the present application further provides a chip, where the chip includes a processor and a communication interface, the communication interface is coupled to the processor, and the processor is configured to execute a program or an instruction to implement each process of the above shooting focusing method embodiment, and can achieve the same technical effect, and in order to avoid repetition, the description is omitted here.
It should be understood that the chips mentioned in the embodiments of the present application may also be referred to as system-on-chip, system-on-chip or system-on-chip, etc.
It should be noted that, in this document, the terms "comprises," "comprising," or any other variation thereof, are intended to cover a non-exclusive inclusion, such that a process, method, article, or apparatus that comprises a list of elements does not include only those elements but may include other elements not expressly listed or inherent to such process, method, article, or apparatus. Without further limitation, an element defined by the phrases "comprising a component of' 8230; \8230;" does not exclude the presence of another like element in a process, method, article, or apparatus that comprises the element. Further, it should be noted that the scope of the methods and apparatus of the embodiments of the present application is not limited to performing the functions in the order illustrated or discussed, but may include performing the functions in a substantially simultaneous manner or in a reverse order based on the functions involved, e.g., the methods described may be performed in an order different than that described, and various steps may be added, omitted, or combined. In addition, features described with reference to certain examples may be combined in other examples.
Through the description of the foregoing embodiments, it is clear to those skilled in the art that the method of the foregoing embodiments may be implemented by software plus a necessary general hardware platform, and certainly may also be implemented by hardware, but in many cases, the former is a better implementation. Based on such understanding, the technical solutions of the present application or portions thereof that contribute to the prior art may be embodied in the form of a software product, which is stored in a storage medium (such as ROM/RAM, magnetic disk, optical disk) and includes instructions for enabling a terminal (which may be a mobile phone, a computer, a server, an air conditioner, or a network device) to execute the method according to the embodiments of the present application.
While the present embodiments have been described with reference to the accompanying drawings, it is to be understood that the present embodiments are not limited to those precise embodiments, which are intended to be illustrative rather than restrictive, and that various changes and modifications may be effected therein by one skilled in the art without departing from the scope of the appended claims.

Claims (9)

1. A shooting focusing method is characterized by comprising the following steps:
under the condition that a shooting preview interface is displayed, acquiring a first video including a target shooting object, wherein the first video is formed by a plurality of frames of historical preview pictures displayed on the shooting preview interface within a first preset time period;
amplifying the expression action of the target shooting object in the first video to obtain a second video;
identifying an expression area of the target photographic object in the second video;
focusing the expression area on the shooting preview interface;
the method for amplifying the expression action of the target shooting object in the first video to obtain a second video comprises the following steps:
filtering the signal of the first video by using a target filter to obtain a first signal of the first video, wherein the first signal is related to expression actions, and the target filter is used for amplifying the expression actions;
amplifying the first signal according to a target amplification factor to obtain a target signal;
and generating the second video according to the target signal.
2. The method of claim 1, wherein before the filtering the signal of the first video with the target filter to obtain the first signal of the first video related to the expressive action, the method further comprises:
converting a time domain signal of the first video into a plurality of frequency domain signals of a plurality of frequency bands;
wherein the filtering, by using a target filter, the signal of the first video to obtain a first signal of the first video related to the expression action includes:
filtering the frequency domain signal by using the target filter to obtain the first signal related to the expression action in the frequency domain signal;
the generating the second video according to the target signal includes:
and converting the target signal from a frequency domain to a time domain to obtain the second video.
3. The method of claim 1, wherein before filtering the signal of the first video with the target filter to obtain the first signal of the first video related to the expressive motion, the method further comprises:
displaying at least one piece of identification information on the shooting preview interface, wherein each piece of identification information is used for indicating a corresponding expression action, and the expression action is obtained by amplifying a multi-frame history preview picture displayed on the shooting preview interface within a preset time period, wherein the preset time period is a time period before the time period of the first video;
receiving a first input of a user to target identification information in the at least one piece of identification information, wherein the target identification information is used for indicating a target expression action;
wherein the filtering, by using a target filter, the signal of the first video to obtain a first signal of the first video related to an expression action includes:
and responding to the first input, and filtering the signal of the first video by using the target filter corresponding to the target expression action to obtain a first signal of the first video, wherein the first signal is related to the target expression action.
4. The method of claim 1, wherein before the amplifying the first signal according to the predetermined amplification factor to obtain the target signal, the method further comprises:
receiving a second input of a user to the sensitivity adjusting control under the condition that the sensitivity adjusting control is displayed on the shooting preview interface;
in response to the second input, obtaining a target sensitivity value associated with the second input;
and determining the target magnification factor related to the target sensitivity value according to the correlation between the preset sensitivity value and the magnification factor.
5. The method of claim 1, wherein after identifying an expressive region of the target photographic subject in the second video and before focusing on the expressive region on the photographic preview interface, the method further comprises:
displaying at least one expression area on the shooting preview interface under the condition that the expression area comprises the at least one expression area;
receiving a third input of a user to a target expression area in the at least one expression area;
focusing the expression area on the shooting preview interface comprises the following steps:
responding to the third input, and focusing the target expression area on the shooting preview interface.
6. The method according to any one of claims 1 to 5, wherein after focusing the expressive area on the shooting preview interface, the method further comprises:
and shooting to obtain a target shooting image under the condition that the expression action in the expression area is a preset expression action.
7. A photographing focusing device, comprising:
the video acquisition module is used for acquiring a first video comprising a target shooting object under the condition of displaying a shooting preview interface, wherein the first video is formed by a plurality of frames of historical preview pictures displayed on the shooting preview interface within a first preset time period;
the amplification processing module is used for amplifying the expression action of the target shooting object in the first video to obtain a second video;
the expression area identification module is used for identifying the expression area of the target shooting object in the second video;
the area focusing module is used for focusing the expression area on the shooting preview interface;
wherein, enlarge the processing module and include:
the signal filtering module is used for filtering the signal of the first video by using a target filter to obtain a first signal of the first video related to the expression action, and the target filter is used for amplifying the expression action;
the signal amplification module is used for amplifying the first signal according to a target amplification factor to obtain a target signal;
and the video restoration module is used for generating the second video according to the target signal.
8. An electronic device comprising a processor, a memory and a program or instructions stored on the memory and executable on the processor, the program or instructions when executed by the processor implementing the steps of the shooting focusing method of any one of claims 1 to 6.
9. A readable storage medium, characterized in that it stores thereon a program or instructions which, when executed by a processor, implement the steps of the shooting focusing method according to any one of claims 1 to 6.
CN202010465750.1A 2020-05-28 2020-05-28 Shooting focusing method and device, electronic equipment and storage medium Active CN111654622B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202010465750.1A CN111654622B (en) 2020-05-28 2020-05-28 Shooting focusing method and device, electronic equipment and storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202010465750.1A CN111654622B (en) 2020-05-28 2020-05-28 Shooting focusing method and device, electronic equipment and storage medium

Publications (2)

Publication Number Publication Date
CN111654622A CN111654622A (en) 2020-09-11
CN111654622B true CN111654622B (en) 2022-10-14

Family

ID=72344764

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202010465750.1A Active CN111654622B (en) 2020-05-28 2020-05-28 Shooting focusing method and device, electronic equipment and storage medium

Country Status (1)

Country Link
CN (1) CN111654622B (en)

Families Citing this family (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112511748A (en) * 2020-11-30 2021-03-16 努比亚技术有限公司 Lens target intensified display method and device, mobile terminal and storage medium
CN112714253B (en) * 2020-12-28 2022-08-26 维沃移动通信有限公司 Video recording method and device, electronic equipment and readable storage medium
CN113810599A (en) * 2021-08-12 2021-12-17 惠州Tcl云创科技有限公司 Method for focusing on designated area by AI action recognition, mobile terminal and storage medium

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102244807A (en) * 2010-06-02 2011-11-16 微软公司 Microsoft Corporation
WO2019071440A1 (en) * 2017-10-10 2019-04-18 深圳传音通讯有限公司 Photographing focusing method and device

Family Cites Families (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7986458B2 (en) * 2007-12-26 2011-07-26 Sony Corporation Zoom lens and image pickup apparatus
CN102594990A (en) * 2012-02-10 2012-07-18 中兴通讯股份有限公司 Smart mobile phone base, mobile phone and implementation methods thereof
CN103780826A (en) * 2012-10-19 2014-05-07 爱国者数码科技有限公司 Digital image capturing device for reminding image shooting picture composition effect
KR102146858B1 (en) * 2014-06-03 2020-08-21 삼성전자주식회사 Photographing apparatus and method for making a video
WO2016103962A1 (en) * 2014-12-26 2016-06-30 富士フイルム株式会社 Focus control device, imaging device, focus control method, and focus control program
CN106888354A (en) * 2017-03-30 2017-06-23 维沃移动通信有限公司 A kind of singlehanded photographic method and mobile terminal
CN107680071B (en) * 2017-10-23 2020-08-07 深圳市云之梦科技有限公司 Method and system for fusion processing of human face and human body

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102244807A (en) * 2010-06-02 2011-11-16 微软公司 Microsoft Corporation
WO2019071440A1 (en) * 2017-10-10 2019-04-18 深圳传音通讯有限公司 Photographing focusing method and device

Also Published As

Publication number Publication date
CN111654622A (en) 2020-09-11

Similar Documents

Publication Publication Date Title
CN111654622B (en) Shooting focusing method and device, electronic equipment and storage medium
CN110517185B (en) Image processing method, device, electronic equipment and storage medium
CN107025419B (en) Fingerprint template inputting method and device
US11455491B2 (en) Method and device for training image recognition model, and storage medium
CN106228556B (en) image quality analysis method and device
CN104035558A (en) Terminal device control method and device
CN105205494B (en) Similar pictures recognition methods and device
EP3975046A1 (en) Method and apparatus for detecting occluded image and medium
CN110717399A (en) Face recognition method and electronic terminal equipment
US20180173394A1 (en) Method and apparatus for inputting expression information
CN111722775A (en) Image processing method, device, equipment and readable storage medium
CN108519811A (en) Screenshot method and Related product
CN113727021A (en) Shooting method and device and electronic equipment
CN112948704B (en) Model training method and device for information recommendation, electronic equipment and medium
CN113794831B (en) Video shooting method, device, electronic equipment and medium
CN113312967B (en) Detection method and device for detection
CN113852756A (en) Image acquisition method, device, equipment and storage medium
CN111416936B (en) Image processing method, image processing device, electronic equipment and storage medium
CN105224950A (en) The recognition methods of filter classification and device
CN115170424B (en) Heart ultrasonic image artifact removing method and device
CN116266418A (en) Motion recognition method, motion recognition device, electronic equipment and storage medium
US20230093165A1 (en) Information processing apparatus, information processing method, and program
CN117642817A (en) Method, device and storage medium for identifying audio data category
CN113873165A (en) Photographing method and device and electronic equipment
CN114565962A (en) Face image processing method and device, electronic equipment and storage medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant