CN112954209B - Photographing method and device, electronic equipment and medium - Google Patents

Photographing method and device, electronic equipment and medium Download PDF

Info

Publication number
CN112954209B
CN112954209B CN202110170800.8A CN202110170800A CN112954209B CN 112954209 B CN112954209 B CN 112954209B CN 202110170800 A CN202110170800 A CN 202110170800A CN 112954209 B CN112954209 B CN 112954209B
Authority
CN
China
Prior art keywords
preview
target
user
sub
area
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202110170800.8A
Other languages
Chinese (zh)
Other versions
CN112954209A (en
Inventor
陈露兰
林燕芝
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Vivo Mobile Communication Hangzhou Co Ltd
Original Assignee
Vivo Mobile Communication Hangzhou Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Vivo Mobile Communication Hangzhou Co Ltd filed Critical Vivo Mobile Communication Hangzhou Co Ltd
Priority to CN202110170800.8A priority Critical patent/CN112954209B/en
Publication of CN112954209A publication Critical patent/CN112954209A/en
Application granted granted Critical
Publication of CN112954209B publication Critical patent/CN112954209B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/60Control of cameras or camera modules
    • H04N23/61Control of cameras or camera modules based on recognised objects
    • H04N23/611Control of cameras or camera modules based on recognised objects where the recognised objects include parts of the human body
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/60Control of cameras or camera modules
    • H04N23/665Control of cameras or camera modules involving internal camera communication with the image sensor, e.g. synchronising or multiplexing SSIS control signals
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/90Arrangement of cameras or camera modules, e.g. multiple cameras in TV studios or sports stadiums
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N5/00Details of television systems
    • H04N5/222Studio circuitry; Studio devices; Studio equipment
    • H04N5/262Studio circuits, e.g. for mixing, switching-over, change of character of image, other special effects ; Cameras specially adapted for the electronic generation of special effects
    • H04N5/2624Studio circuits, e.g. for mixing, switching-over, change of character of image, other special effects ; Cameras specially adapted for the electronic generation of special effects for obtaining an image which is composed of whole input images, e.g. splitscreen

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Studio Devices (AREA)
  • User Interface Of Digital Computer (AREA)

Abstract

The application discloses a photographing method, a photographing device, electronic equipment and a photographing medium, and belongs to the technical field of camera shooting. The method comprises the following steps: the method comprises the steps that under the condition that a shooting preview interface is displayed, first motion information of eyeballs of a user is obtained, wherein the shooting preview interface comprises N preview sub-areas, each preview sub-area is associated with a camera, and N is a positive integer; determining a target preview subregion of the user's eye gaze based on the first motion information, the N preview subregions including the target preview subregion; and under the condition that the time length of the eyeballs of the user watching the target preview subarea is within a preset time length range, controlling a target camera associated with the target preview subarea to shoot a target image. According to the method and the device, the camera required by the user can be conveniently and quickly selected to photograph through the movement of the eyeballs of the user, so that the photographing mode is more flexible.

Description

Photographing method and device, electronic equipment and medium
Technical Field
The present application relates to the field of camera technologies, and in particular, to a photographing method and apparatus, an electronic device, and a medium.
Background
With the continuous development of mobile communication technology, intelligent electronic devices (such as mobile phones) are becoming popular. The functions of the intelligent electronic device are increasing, such as taking pictures, taking videos, playing audios and videos, browsing webpages, sending and receiving mails and the like. Among them, taking pictures using electronic devices is also popular among users.
With the continuous progress of the photographing technology of the electronic equipment, more and more cameras are used for photographing, the photographing modes of different cameras are generally fixed at present, the existing cameras are switched, a user can only switch the cameras by selecting the modes or opening a certain function, the switching mode is single, and the independent selection space of the user is small.
Disclosure of Invention
The embodiment of the application aims to provide a photographing method, a photographing device, electronic equipment and a photographing medium, and can solve the problem that a camera switching mode of existing electronic equipment is single.
In order to solve the technical problem, the present application is implemented as follows:
in a first aspect, an embodiment of the present application provides a photographing method, where the method includes:
acquiring first motion information of eyeballs of a user under the condition that a shooting preview interface is displayed, wherein the shooting preview interface comprises N preview sub-areas, each preview sub-area is associated with a camera, and N is a positive integer;
determining a target preview subregion of the user's eye gaze based on the first motion information, the N preview subregions including the target preview subregion;
and under the condition that the time length of the eyeballs of the user watching the target preview subarea is within a preset time length range, controlling a target camera associated with the target preview subarea to shoot a target image.
In a second aspect, an embodiment of the present application further provides a photographing apparatus, including:
the device comprises a first acquisition module, a second acquisition module and a third acquisition module, wherein the first acquisition module is used for acquiring first motion information of eyeballs of a user under the condition that a shooting preview interface is displayed, the shooting preview interface comprises N preview sub-areas, each preview sub-area is associated with a camera, and N is a positive integer;
a first processing module, configured to determine, based on the first motion information, a target preview sub-area at which the eyeballs of the user gaze, where the N preview sub-areas include the target preview sub-area;
and the photographing module is used for controlling a target camera associated with the target preview sub-area to photograph a target image under the condition that the time length of the eyeball of the user watching the target preview sub-area is within a preset time length range.
In a third aspect, an embodiment of the present application provides an electronic device, which includes a processor, a memory, and a program or instructions stored on the memory and executable on the processor, where the program or instructions, when executed by the processor, implement the steps of the photographing method according to the first aspect.
In a fourth aspect, the present application further provides a readable storage medium, on which a program or instructions are stored, and when the program or instructions are executed by a processor, the program or instructions implement the steps of the photographing method according to the first aspect.
In a fifth aspect, an embodiment of the present application provides a chip, where the chip includes a processor and a communication interface, where the communication interface is coupled to the processor, and the processor is configured to execute a program or instructions to implement the photographing method according to the first aspect.
In the embodiment of the application, first motion information of eyeballs of a user is acquired under the condition that a shooting preview interface is displayed, wherein the shooting preview interface comprises N preview sub-areas, each preview sub-area is associated with a camera, and N is a positive integer; determining a target preview subregion watched by the eyeballs of the user based on the first motion information, wherein the N preview subregions comprise the target preview subregion, and the N preview subregions comprise the target preview subregion; under the condition that the time length of the target preview subarea watched by the eyeballs of the user is within the preset time length range, the target camera associated with the target preview subarea is controlled to shoot a target image, so that the camera needed by the user can be conveniently and quickly selected to shoot through the movement of the eyeballs of the user, and the shooting mode is more flexible.
Drawings
Fig. 1 is a schematic flowchart of a photographing method provided in an embodiment of the present application;
FIG. 2 is a schematic view of an interface operation corresponding to the method of the embodiment of the present application;
FIG. 3 is a second schematic view of an interface operation corresponding to the method of the embodiment of the present application;
FIG. 4 is a third schematic view of an interface operation corresponding to the method of the embodiment of the present application;
FIG. 5 is a fourth schematic view illustrating interface operations corresponding to the method of the embodiment of the present application;
FIG. 6 is a fifth exemplary view illustrating operations of an interface corresponding to the method of the present embodiment;
fig. 7 is a schematic structural diagram of a photographing device according to an embodiment of the present application;
fig. 8 is a schematic structural diagram of an electronic device according to an embodiment of the present application;
fig. 9 is a schematic hardware structure diagram of an electronic device according to an embodiment of the present application.
Detailed Description
The technical solutions in the embodiments of the present application will be clearly and completely described below with reference to the drawings in the embodiments of the present application, and it is obvious that the described embodiments are some, but not all, embodiments of the present application. All other embodiments, which can be derived by a person skilled in the art from the embodiments given herein without making any creative effort, shall fall within the protection scope of the present application.
The terms first, second and the like in the description and in the claims of the present application are used for distinguishing between similar elements and not necessarily for describing a particular sequential or chronological order. It will be appreciated that the data so used may be interchanged under appropriate circumstances such that embodiments of the application may be practiced in sequences other than those illustrated or described herein, and that the terms "first," "second," and the like are generally used herein in a generic sense and do not limit the number of terms, e.g., the first term can be one or more than one. In addition, "and/or" in the specification and claims means at least one of connected objects, a character "/" generally means that a preceding and succeeding related objects are in an "or" relationship.
The photographing method provided by the embodiment of the present application will be described in detail by specific embodiments and application scenarios thereof with reference to the accompanying drawings.
Fig. 1 is a schematic flow chart of a photographing method according to an embodiment of the present disclosure. The implementation of the method is described in detail below with reference to this figure.
101, acquiring first motion information of eyeballs of a user under the condition that a shooting preview interface is displayed, wherein the shooting preview interface comprises N preview sub-areas, each preview sub-area is associated with a camera, and N is a positive integer;
in this step, the electronic device is triggered to enter a shooting preview interface when receiving a shooting request.
Here, the shooting preview interface may be divided into N preview sub-areas by a preset input of the user to the shooting preview interface. Here, the preset input is used to trigger the electronic device to enter the first photographing mode. It should be noted that the first photographing mode is a photographing mode in which a camera used for photographing is determined by eye movement of a user.
It should be noted that each preview sub-area is associated with one different camera corresponding to different preview sub-areas.
Here, each preview sub-area displays a preview image based on a picture captured by the corresponding camera. The preview images displayed in different preview sub-areas are different, so that a user can conveniently and visually check the imaging effects of different cameras of the electronic equipment.
Step 102, determining target preview sub-areas watched by eyeballs of the user based on the first motion information, wherein the N preview sub-areas comprise the target preview sub-areas;
103, under the condition that the time length that the eyeballs of the user watch the target preview subarea is within a preset time length range, controlling a target camera associated with the target preview subarea to shoot a target image.
In this step, in order to avoid that the gaze of the eyeball of the user on the target preview subregion is invalid gaze, under the condition that the time length of the eyeball of the user gazing on the target preview subregion is within the preset time length range, the camera associated with the target preview subregion is determined to be the target camera, that is, the target camera is the camera selected by the user and finally used for photographing; and controls the target camera to shoot the target image.
For example, the preset time duration range is 3 seconds to 5 seconds, and the time duration when the eyeballs of the user watch the target preview sub-area is less than 3 seconds or more than 5 seconds, the user is judged to watch the target preview sub-area inefficiently. When the time length that the eyeballs of the user watch the target preview sub-area is less than 3 seconds, the eyeballs of the user watch the target preview sub-area for a short time, a large possibility that the subsequent sight line leaves the area is provided, when the time length that the eyeballs of the user watch the target preview sub-area is more than 5 seconds, the user is considered to be in a dull state or meditation, and the camera which is finally used for shooting and is wanted by the user is not provided, and the above situations are considered to be invalid watching.
According to the photographing method, first motion information of eyeballs of a user is acquired under the condition that a photographing preview interface is displayed, wherein the photographing preview interface comprises N preview sub-areas, each preview sub-area is associated with one camera, and N is a positive integer; determining a target preview subregion watched by the eyeballs of the user based on the first motion information, wherein the N preview subregions comprise the target preview subregion, and the N preview subregions comprise the target preview subregion; under the condition that the time length of the target preview subregion is watched by eyeballs of the user, the target camera associated with the target preview subregion is controlled to shoot a target image, so that the camera required by the user can be conveniently and quickly selected to shoot through the movement of the eyeballs of the user, the shooting mode is more flexible, the camera is particularly suitable for scenes in which the hands of the user are inconvenient to operate, manual operation is not needed, and the required camera can be flexibly and quickly selected to shoot the target image by controlling the movement of the eyeballs to select one region.
Optionally, each of the N preview sub-areas includes an area identifier, and the area identifier has a first display mode and a second display mode;
under the condition that the area identification is in the first display mode, the area identification is used for indicating the eyeball of the user to watch the preview sub-area where the area identification is located;
and under the condition that the area identification is in the second display mode, the area identification is used for indicating that eyeballs of the user do not watch the preview sub-area where the area identification is located.
Here, the area identification of each preview sub-area is used to indicate whether the current user's eye looks at the preview sub-area.
In one example, through user operation, the electronic device opens a camera application and enters a shooting preview interface; displaying 6 preview sub-areas on a shooting preview interface through preset input of a user on the shooting preview interface, as shown in fig. 2; wherein the lower right corner of each preview sub-area displays an identification of an eye pattern (i.e. a region identification). When the eyeball of the user does not watch the shooting preview interface, the identification of the eye pattern of each preview subarea is a second display mode 1, such as being grayed; when the user's eyes are focused on one of the 6 preview sub-areas, the eye pattern of the preview sub-area is identified as the first display mode 2, such as presenting a preset color.
Based on this, after determining, in step 102, a target preview sub-area at which the user looks at based on the first motion information, an embodiment of the present application may further include:
and updating the area identification in the target preview subregion to the first display mode.
Continuing with the above example illustration, that is, when the user looks at the target preview sub-area, the area identifier of the target preview sub-area, such as the eye style identifier, is updated to present the preset color from the original grayed-out. Therefore, the user can intuitively know the preview subarea watched by the eyeballs in real time.
As an optional implementation manner, the shooting preview interface further includes a position identifier, where the position identifier is used to indicate first motion information of the eyeballs of the user, and the first motion information includes a real-time position watched by the eyeballs of the user;
based on this, the step 102 of determining, based on the first motion information, a target preview sub-area at which the user looks at the eyeball may include:
in a target time period, under the condition that the display position of the position mark is located in the display range of a first preview subregion, determining the first preview subregion as a target preview subregion watched by eyeballs of a user;
wherein the first preview sub-area is any one of the N preview sub-areas.
It should be noted that, when the user looks at the shooting preview interface, the position identifier moves on the shooting preview interface along with the movement of the eyeballs of the user, so that the staying position of the user's gaze on the screen, that is, the real-time position watched by the eyeballs of the user, can be visually displayed.
Based on this, through the position identification, the motion information of the eyeballs of the user, namely the real-time position watched by the eyeballs of the user, can be obtained, so that the target preview sub-area watched by the target eyeballs is determined, and the processing efficiency of determining the target preview sub-area watched by the target eyeballs can be improved.
Of course, the first motion information of the eyeball of the user can also be acquired through the front camera of the electronic equipment. Wherein the first motion information includes but is not limited to: the offset direction of the user's eyeball and the offset angle of the user's eyeball.
Here, in the case where the photographing preview interface displays N regions, the front camera is turned on, and first movement information of the user's eyes is acquired in real time through the front camera. And calculating to obtain a target preview subregion watched by eyeballs of the user according to the first motion information.
In one example, as shown in fig. 3, in the case that the shooting preview interface displays 6 preview sub-areas, if the user looks at the shooting preview interface, the shooting preview interface displays an identifier 3 (i.e. a position identifier) of a semi-transparent eye pattern at a position of an area where the user's eyes watch, where a dotted line in the figure represents the identifier of the semi-transparent eye pattern. For example, in fig. 3, the current region where the eyes of the user gaze is in the preview sub-region 4 in the middle of the left side, and the translucent eye-pattern mark 3 is displayed in the preview sub-region 4. If the display position of the semi-transparent eye-style identifier 3 is within the display range of the preview sub-area 4 in the middle of the left side in fig. 3 in the target time period, the preview sub-area 4 is determined as a target preview sub-area watched by the eyeballs of the user.
How to display the N preview sub-regions on the photo preview interface is implemented, as an optional implementation manner, the embodiment of the present application may further include:
receiving preset input of a user on a shooting preview interface;
here, the preset input may include, but is not limited to, at least one of a click input, a press input, a long press input, a pinch input, a drag input, a slide input, and a swipe input, that is, the preset input may be one of the above-described inputs, or may also be a combination input of two or more thereof.
And responding to the preset input, and displaying N preview sub-areas on the shooting preview interface.
Optionally, the camera associated with the preview sub-area is a rear camera. The number of the cameras in the realization mode is equal to the number of the preview subareas displayed in the shooting preview interface.
In one example, through user operation, the electronic device opens a camera application and enters a shooting preview interface; the shooting preview interface is slid downwards by two fingers of a user to enter a first shooting mode, namely all rear cameras of the electronic equipment are opened, and if the number of the rear cameras of the electronic equipment is 6, 6 preview sub-areas are displayed on the shooting preview interface, which is shown in fig. 2. Namely, the shooting preview interface is divided into 6 corresponding preview subareas according to the number of the rear cameras, each preview subarea is associated with one rear camera, the cameras associated with different preview subareas are different, and each preview subarea displays image data acquired by the associated camera. In this way, the user can preview the image data of 6 cameras at the same time.
As an optional implementation manner, in the case of displaying a shooting preview interface, an embodiment of the present application may further include:
receiving first input of a user to M preview sub-regions in the N preview sub-regions;
here, the first input is a preset input, and the first input may include, but is not limited to, at least one of an eye control input/click input, a press input, a long press input, a pinch input, a drag input, a slide input, and a swipe input, that is, the first input may be one of the above-mentioned inputs, or may also be a combination input of two or more thereof. Here, the eye control input is controlled by the movement of the user's eyeball.
In response to the first input, combining the M preview sub-areas into a second preview sub-area, performing image synthesis on preview images acquired by M cameras displayed in the M preview sub-areas to generate a target preview image, and displaying the target preview image in the second preview sub-area; wherein M is an integer greater than 1.
The combined area, namely the target preview image displayed in the second preview subregion is an image generated by synthesizing the preview images acquired by the M cameras displayed in the M preview subregions, so that not only can the combination of any two preview subregions be controlled conveniently, but also a user can directly view multiple-shot preview images after image synthesis processing, and the user has a larger selection space for the camera used in the final shooting, thereby realizing more flexible and convenient shooting and meeting the requirements of the user.
In one example, through user operation, the electronic device opens a camera application and enters a shooting preview interface; the shooting preview interface is slid downwards by two fingers of a user to enter a first shooting mode, namely all rear cameras of the electronic equipment are opened, and if the number of the rear cameras of the electronic equipment is 6, 6 preview sub-areas are displayed on the shooting preview interface, which is shown in fig. 2. Namely, the shooting preview interface is divided into 6 corresponding preview subareas according to the number of the rear cameras, each preview subarea is associated with one rear camera, the cameras associated with different preview subareas are different, and each preview subarea displays image data acquired by the associated camera.
Then, referring to the left diagram of fig. 4, a user's slide input (slide to the right in fig. 4) to the shooting preview interface is received, the trajectory of the slide input crossing the two preview sub-areas 5 and 6; in response to the slide input, the two preview sub-areas 5 and 6 are merged into one preview sub-area 7, see the right hand diagram of fig. 4. At this time, the number of preview sub-areas displayed on the shooting preview interface is updated from the original 6 to 5. At this time, the number of cameras associated with the combined preview sub-area 7 is two. Then, the preview images acquired by the 2 cameras displayed in the two preview sub-areas 5 and 6 are synthesized to generate a target preview image, and the target preview image is displayed in the combined preview sub-area 7.
Based on this, if the user wants to restore the original split state again, as an optional implementation manner, in response to the first input, the M preview sub-areas are merged into the second preview sub-area, and the preview images of the M preview sub-areas are image-synthesized to generate a target preview image, and after the target preview image is displayed in the second preview sub-area, the embodiment of the present application may further include:
receiving a second input of the user to the second preview subarea;
here, the second input is a preset input, and the second input may include, but is not limited to, at least one of an eye control input/click input, a press input, a long press input, a pinch input, a drag input, a slide input, and a swipe input, that is, the second input may be one of the above-mentioned inputs, or may also be a combination input of two or more thereof. Here, the eye control input is controlled by the movement of the user's eyeball.
Responding to the second input, splitting the second preview subregion into M preview subregions, and respectively displaying a preview image acquired by each camera in the M cameras in each preview subregion in the M preview subregions.
Here, each preview sub-area of the split M preview sub-areas is associated with one camera, that is, each preview sub-area displays a preview image acquired by one camera.
By a second input, such as a reverse input (left slide input) of the slide input when merging the regions in the previous example, specifically, the identification of the eye pattern of the merging region is clicked and released after sliding a distance to the left.
Optionally, under the condition that a preview image displayed by a third preview sub-area of the N preview sub-areas is obtained by synthesizing preview images acquired by at least two cameras, the third preview sub-area further includes a camera number identifier, and the camera number identifier is used for indicating the number of cameras associated with the third preview sub-area.
Here, when the number of the cameras associated with the preview sub-area exceeds two, the preview sub-area, here, referred to as a third preview sub-area, includes a camera number identifier for indicating the number of the cameras associated with the third preview sub-area, so that a user can intuitively know the number of the cameras associated with the preview sub-area in a shooting preview interface.
Specifically, referring to the right side of fig. 4, after the two preview sub-areas 5 and 6 are merged into one preview sub-area 7 in response to a sliding input of the user to the shooting preview interface, a camera number identifier 2 is displayed at the lower right corner of the merged preview sub-area 7, so as to indicate that the number of cameras associated with the preview sub-area is 2.
As an optional implementation manner, in a case that a duration of time that the user's eyeball focuses on the target preview sub-area exceeds a preset first duration, the method of the embodiment of the present application may further include:
under the condition of acquiring voice information input by a user, determining a camera obtained by recognizing the voice information as a target camera; alternatively, the first and second electrodes may be,
under the condition that selection input of a user on a fourth preview subarea in the N preview subareas is received, determining a camera related to the fourth preview subarea as a target camera;
and the preset first time length is greater than or equal to the maximum value of the preset time length range.
Here, when the duration of the target sub-preview area watched by the eyes of the user exceeds a preset first duration, that is, the gaze of the eyes of the user is invalid, the selected result of the camera finally used for photographing is selected as invalid through the movement information of the eyes of the user.
In this scenario, one of the cameras starts a voice recognition function by default to acquire voice information input by a user, and determines a camera obtained by recognizing the voice information as a target camera when the voice information input by the user is acquired. That is, the target camera, which is the camera ultimately used for taking a picture, can be selected through the voice command.
In an example, referring to the left diagram of fig. 5, if the time length of the user's eyeball gazing at the middle left preview sub-area 8 exceeds 5 seconds or longer, it is determined that the user's eyeball gazing is invalid gazing, the voice recognition function is started by default, and the user can directly select one camera through voice recognition, for example, the user says: the "wide-angle camera" or, if the camera associated with the preview sub-area is numbered, the user may also say that the number of the camera is, for example, "number 2", at this time, the target camera, that is, the camera finally used for taking a picture, is determined. Further, referring to the middle diagram of fig. 5, the identifier of the eye pattern of the preview sub-area 9 corresponding to the wide-angle camera is presented with a preset color, and the identifier 10 of the eye pattern of the left area 8 in the middle of the shooting preview interface is grayed out.
In addition, after the camera identified by the user voice information is determined as the target camera (i.e. the camera finally used for photographing is selected by the user voice instruction), a photographing instruction is not received within a preset time period, and updated motion information of the user eyeballs is obtained within the preset time period (i.e. the user eye sight line starts to change and does not look at a place any more) in the preset time period (corresponding to the right side diagram of fig. 5), the method returns to the step after the step 101 of continuing to execute, that is, the camera finally used for photographing is determined by the motion information of the user eyeballs.
And the target camera, namely the camera finally used for photographing is determined through the selection input of the user to the preview subarea.
As an optional implementation manner, before controlling the target camera associated with the target preview sub-area to shoot the target image, the method in this embodiment of the application may further include:
displaying a shooting parameter setting window under the condition that a third input of the user to the target preview sub-area is received, wherein the shooting parameter setting window comprises at least one shooting parameter;
optionally, in a case where a third input of the area identification of the target preview sub-area by the user is received, the shooting parameter setting window is displayed.
Receiving a fourth input of the shooting parameter setting window by the user;
here, in a case where the shooting parameter setting window includes an editing area in which the user can edit the shooting parameters, the fourth input is an editing input for the shooting parameters. That is, by the user manually editing the input, the photographing parameters desired by the user are determined.
In a case where the shooting parameter setting window includes a shooting parameter selection item selectable by the user, the fourth input is a selection input of the target shooting parameter selection item. That is, through the selection input of the user, the photographing parameters desired by the user are determined.
Determining a first photographing parameter in response to the fourth input;
the controlling the target camera associated with the target preview sub-area to shoot the target image comprises the following steps:
and controlling a target camera associated with the target preview subregion to shoot a target image according to the first shooting parameter.
In the implementation mode, the shooting parameters meeting the user requirements can be set through the shooting parameter setting window, and the target image can be shot by adopting the target camera according to the shooting parameters meeting the user requirements, so that the target image meeting the shooting effect desired by the user can be obtained.
In one example, through user operation, the electronic device opens a camera application and enters a shooting preview interface; entering a first photographing mode through the downslide input of two fingers of a user on a photographing preview interface, namely opening all rear cameras of the electronic equipment, and displaying 6 preview sub-areas on the photographing preview interface if the number of the rear cameras of the electronic equipment is 6, referring to fig. 6, wherein an eye-style identifier (namely an area identifier) is displayed at the lower right corner of each preview sub-area; when the eyeball of the user is focused on the right area 10 (target preview sub-area) in the middle of the 6 preview sub-areas, the identification of the eye style of the preview sub-area 10 is a first display mode 2, for example, a preset color is presented, and the time length for the eyeball of the user to watch the preview sub-area 10 is within a preset time length range, it is determined that the camera associated with the preview sub-area 10 is the target camera; then, clicking the mark of the eye pattern of the preview subregion 10, popping up a shooting parameter setting window 11, and selecting shooting time delay and shooting number as a first shooting parameter through user operation; or, the shooting parameter setting window, the subsequent shooting time delay selection and the number of the shot can be realized through eye control input; and finally, clicking a photographing button, and photographing a target image by using a target camera according to the first photographing parameter.
As an optional implementation manner, before controlling the target camera associated with the target preview sub-area to shoot the target image, the embodiment of the present application may further include:
acquiring second motion information of eyeballs of the user;
controlling the position mark to move according to the second motion information;
determining a second shooting parameter according to the moving distance of the position identifier;
that is to say, through the removal of eye control position sign to confirm the second shooting parameter, so, saved the step that the user edited the selection shooting parameter manually, for the user provides convenience when inconvenient use hand and interface are interacted.
Correspondingly, the controlling the target camera associated with the target preview sub-area to shoot the target image includes:
and controlling the target camera associated with the target preview subarea to shoot the target image according to the second shooting parameter.
According to the photographing method, first motion information of eyeballs of a user is acquired under the condition that a photographing preview interface is displayed, wherein the photographing preview interface comprises N preview sub-areas, each preview sub-area is associated with one camera, and N is a positive integer; determining a target preview subregion watched by the eyeballs of the user based on the first motion information, wherein the N preview subregions comprise the target preview subregion, and the N preview subregions comprise the target preview subregion; under the condition that the time length of the target preview subarea watched by the eyeballs of the user is within the preset time length range, the target camera associated with the target preview subarea is controlled to shoot a target image, so that the camera needed by the user can be conveniently and quickly selected to shoot through the movement of the eyeballs of the user, and the shooting mode is more flexible.
It should be noted that, in the photographing method provided in the embodiment of the present application, the execution main body may be a photographing device, or a control module used for executing the photographing method in the photographing device. The embodiment of the present application takes a photographing apparatus executing a photographing method as an example, and the photographing apparatus provided in the embodiment of the present application is described.
Fig. 7 is a schematic structural diagram of a photographing device according to an embodiment of the present disclosure. The photographing apparatus 700 may include:
a first obtaining module 701, configured to obtain first motion information of an eyeball of a user when a shooting preview interface is displayed, where the shooting preview interface includes N preview sub-areas, each preview sub-area is associated with one camera, and N is a positive integer;
a first processing module 702, configured to determine, based on the first motion information, target preview sub-areas at which the eyes of the user gaze, where the N preview sub-areas include the target preview sub-area;
the photographing module 703 is configured to control a target camera associated with the target preview sub-region to photograph a target image when a duration that the eyeballs of the user watch the target preview sub-region is within a preset duration range.
Optionally, each preview sub-region of the N preview sub-regions includes a region identifier, and the region identifier has a first display mode and a second display mode;
under the condition that the area identification is in the first display mode, the area identification is used for indicating the eyeball of the user to watch the preview sub-area where the area identification is located;
and under the condition that the area identification is in the second display mode, the area identification is used for indicating that eyeballs of the user do not watch the preview sub-area where the area identification is located.
Optionally, the apparatus 700 further comprises:
and the second processing module is used for updating the area identification in the target preview subarea into the first display mode.
Optionally, the shooting preview interface further includes a position identifier, where the position identifier is used to indicate first motion information of the user's eyes, and the first motion information includes a real-time position where the user's eyes gaze;
the first processing module 702 comprises:
the first processing unit is used for determining the first preview sub-area as a target preview sub-area watched by eyeballs of a user under the condition that the display position of the position mark is located in the display range of the first preview sub-area in a target time period;
wherein the first preview sub-area is any one of the N preview sub-areas.
Optionally, the apparatus 700 further comprises:
the first receiving module is used for receiving first input of a user to M preview sub-areas in the N preview sub-areas;
the third processing module is used for responding to the first input, combining the M preview sub-areas into a second preview sub-area, performing image synthesis on preview images acquired by the M cameras displayed in the M preview sub-areas, generating a target preview image, and displaying the target preview image in the second preview sub-area;
wherein M is an integer greater than 1.
Optionally, the apparatus 700 further comprises:
the second receiving module is used for receiving a second input of the user to the second preview subarea;
and the fourth processing module is configured to split the second preview sub-area into M preview sub-areas in response to the second input, and display a preview image acquired by each of the M cameras in each of the M preview sub-areas.
Optionally, under the condition that a preview image displayed by a third preview sub-area of the N preview sub-areas is obtained by synthesizing preview images acquired by at least two cameras, the third preview sub-area further includes a camera number identifier, and the camera number identifier is used for indicating the number of cameras associated with the third preview sub-area.
Optionally, the apparatus 700 further comprises:
the fifth processing module is used for determining a camera obtained by recognizing the voice information as a target camera under the condition of acquiring the voice information input by the user; alternatively, the first and second electrodes may be,
the sixth processing module is used for determining a camera associated with a fourth preview sub-area as a target camera under the condition that selection input of a user on the fourth preview sub-area in the N preview sub-areas is received;
and the preset first duration is greater than or equal to the maximum value of the preset duration range.
Optionally, the apparatus 700 further comprises:
the first display module is used for displaying a shooting parameter setting window under the condition that third input of a user to the target preview sub-area is received, wherein the shooting parameter setting window comprises at least one shooting parameter;
the third receiving module is used for receiving fourth input of the shooting parameter setting window by a user;
a seventh processing module, configured to determine a first shooting parameter in response to the fourth input;
the photographing module 703 includes:
and the first photographing unit is used for controlling the target camera associated with the target preview subarea to photograph the target image according to the first photographing parameter.
Optionally, the apparatus 700 further comprises:
the second acquisition module is used for acquiring second motion information of eyeballs of the user;
the eighth processing module controls the position identifier to move according to the second motion information;
the ninth processing module is used for determining a second shooting parameter according to the moving distance of the position identifier;
the photographing module 703 includes:
and the second photographing unit is used for controlling the target camera associated with the target preview sub-area to photograph the target image according to the second photographing parameter.
The photographing device in the embodiment of the present application may be a device, or may be a component, an integrated circuit, or a chip in a terminal. The device can be mobile electronic equipment or non-mobile electronic equipment. By way of example, the mobile electronic device may be a mobile phone, a tablet computer, a notebook computer, a palm top computer, a vehicle-mounted electronic device, a wearable device, an ultra-mobile personal computer (UMPC), a netbook or a Personal Digital Assistant (PDA), and the like, and the non-mobile electronic device may be a Network Attached Storage (NAS), a personal computer (personal computer, PC), a Television (TV), a teller machine, a self-service machine, and the like, and the embodiments of the present application are not limited in particular.
The photographing device in the embodiment of the present application may be a device having an operating system. The operating system may be an Android (Android) operating system, an ios operating system, or other possible operating systems, and embodiments of the present application are not limited specifically.
The photographing device provided in the embodiment of the present application can implement each process implemented by the method embodiments of fig. 1 to 6, and is not described here again to avoid repetition.
According to the photographing device, the first movement information of the eyeballs of the user is acquired through the first acquisition module under the condition that the photographing preview interface is displayed, wherein the photographing preview interface comprises N preview sub-areas, each preview sub-area is associated with one camera, and N is a positive integer; the first processing module determines target preview sub-areas watched by eyeballs of the user based on the first motion information, wherein the N preview sub-areas comprise the target preview sub-areas; the photographing module controls the target camera associated with the target preview subarea to photograph the target image under the condition that the time length of the target preview subarea watched by the eyeballs of the user is within the preset time length range, so that the camera required by the user can be conveniently and quickly selected to photograph through the movement of the eyeballs of the user, and the photographing mode is more flexible.
Optionally, as shown in fig. 8, an electronic device 800 is further provided in this embodiment of the present application, and includes a processor 801, a memory 802, and a program or an instruction stored in the memory 802 and executable on the processor 801, where the program or the instruction is executed by the processor 801 to implement each process of the foregoing photographing method embodiment, and can achieve the same technical effect, and in order to avoid repetition, details are not repeated here.
It should be noted that the electronic device in the embodiment of the present application includes the mobile electronic device and the non-mobile electronic device described above.
Fig. 9 is a schematic hardware structure diagram of an electronic device implementing various embodiments of the present application.
The electronic device 900 includes, but is not limited to: a radio frequency unit 901, a network module 902, an audio output unit 903, an input unit 904, a sensor 905, a display unit 906, a user input unit 907, an interface unit 908, a memory 909, a processor 910, and a power supply 911.
Those skilled in the art will appreciate that the electronic device 900 may further include a power source (e.g., a battery) for supplying power to various components, and the power source may be logically connected to the processor 910 through a power management system, so as to manage charging, discharging, and power consumption management functions through the power management system. The electronic device structure shown in fig. 9 does not constitute a limitation of the electronic device, and the electronic device may include more or less components than those shown, or combine some components, or arrange different components, and thus, the description is not repeated here.
The processor 910 is configured to acquire first motion information of an eyeball of a user under the condition that a shooting preview interface is displayed, where the shooting preview interface includes N preview sub-areas, each preview sub-area is associated with a camera, and N is a positive integer; determining a target preview subregion of the user's eye gaze based on the first motion information, the N preview subregions including the target preview subregion; and under the condition that the time length of the eyeballs of the user watching the target preview subarea is within a preset time length range, controlling a target camera associated with the target preview subarea to shoot a target image.
In the embodiment of the application, the camera required by the user can be conveniently and quickly selected to take a picture through the movement of the eyeballs of the user, so that the picture taking mode is more flexible.
Optionally, each of the N preview sub-areas includes an area identifier, and the area identifier has a first display mode and a second display mode;
under the condition that the area identification is in the first display mode, the area identification is used for indicating the eyeball of the user to watch the preview sub-area where the area identification is located;
and under the condition that the area identification is in the second display mode, the area identification is used for indicating that eyeballs of the user do not watch the preview sub-area where the area identification is located.
Optionally, the processor 910 is further configured to:
and updating the area identification in the target preview subregion to the first display mode.
Optionally, the shooting preview interface further includes a position identifier, where the position identifier is used to indicate first motion information of the user's eyes, and the first motion information includes a real-time position where the user's eyes gaze; processor 910, further configured to:
in a target time period, under the condition that the display position of the position mark is located in the display range of a first preview subregion, determining the first preview subregion as a target preview subregion watched by eyeballs of a user;
wherein the first preview sub-area is any one of the N preview sub-areas.
Optionally, the user input unit 907 is configured to receive a first input, by a user, to M preview sub-regions of the N preview sub-regions; processor 910, further configured to:
in response to the first input, combining the M preview sub-areas into a second preview sub-area, performing image synthesis on preview images acquired by the M cameras displayed in the M preview sub-areas, generating a target preview image, and displaying the target preview image on the second preview sub-area through a display unit 906; wherein M is an integer greater than 1.
Optionally, the user input unit 907 is further configured to receive a second input from the user to the second preview sub-area; in response to the second input, splitting the second preview sub-area into M preview sub-areas, and displaying a preview image acquired by each camera in the M cameras in each preview sub-area in the M preview sub-areas through the display unit 906.
Optionally, in a case that a preview image displayed by a third preview sub-area of the N preview sub-areas is obtained by synthesizing preview images acquired by at least two cameras, the third preview sub-area further includes a camera number identifier, and the camera number identifier is used to indicate the number of cameras associated with the third preview sub-area.
Optionally, the processor 910 is further configured to:
under the condition of acquiring voice information input by a user, determining a camera obtained by recognizing the voice information as a target camera; or, under the condition that a selection input of a user for a fourth preview sub-area in the N preview sub-areas is received, determining a camera associated with the fourth preview sub-area as a target camera; and the preset first time length is greater than or equal to the maximum value of the preset time length range.
Optionally, the display unit 906 is further configured to display a shooting parameter setting window in a case that a third input of the user to the target preview sub-area is received, where the shooting parameter setting window includes at least one shooting parameter; a user input unit 907 further configured to receive a fourth input of the shooting parameter setting window from the user; a processor 910, further configured to determine a first shooting parameter in response to the fourth input; and controlling a target camera associated with the target preview subregion to shoot a target image according to the first shooting parameter.
Optionally, the processor 910 is further configured to obtain second motion information of the user's eyeball; controlling the position mark to move according to the second motion information; determining a second shooting parameter according to the moving distance of the position identifier; and controlling the target camera associated with the target preview subarea to shoot the target image according to the second shooting parameter.
In the embodiment of the application, the camera required by the user can be conveniently and quickly selected to take pictures through the movement of the eyeballs of the user, so that the picture taking mode is more flexible.
It should be understood that, in the embodiment of the present application, the input Unit 904 may include a Graphics Processing Unit (GPU) 9041 and a microphone 9042, and the Graphics Processing Unit 9041 processes image data of a still picture or a video obtained by an image capturing device (such as a camera) in a video capturing mode or an image capturing mode. The display unit 906 may include a display panel 9061, and the display panel 9061 may be configured in the form of a liquid crystal display, an organic light emitting diode, or the like. The user input unit 907 includes a touch panel 9071 and other input devices 9072. A touch panel 9071 also referred to as a touch screen. The touch panel 9071 may include two parts, a touch detection device and a touch controller. Other input devices 9072 may include, but are not limited to, a physical keyboard, function keys (e.g., volume control keys, switch keys, etc.), a trackball, a mouse, and a joystick, which are not described in detail herein. Memory 909 can be used to store software programs as well as various data including, but not limited to, application programs and operating systems. The processor 910 may integrate an application processor, which primarily handles operating systems, user interfaces, applications, etc., and a modem processor, which primarily handles wireless communications. It will be appreciated that the modem processor described above may not be integrated into processor 910.
The embodiment of the present application further provides a readable storage medium, where a program or an instruction is stored on the readable storage medium, and when the program or the instruction is executed by a processor, the program or the instruction implements each process of the foregoing photographing method embodiment, and can achieve the same technical effect, and in order to avoid repetition, details are not repeated here.
The processor is the processor in the electronic device described in the above embodiment. The readable storage medium includes a computer readable storage medium, such as a Read-Only Memory (ROM), a Random Access Memory (RAM), a magnetic disk or an optical disk, and so on.
The embodiment of the present application further provides a chip, where the chip includes a processor and a communication interface, the communication interface is coupled to the processor, and the processor is configured to run a program or an instruction to implement each process of the foregoing photographing method embodiment, and can achieve the same technical effect, and the details are not repeated here to avoid repetition.
It should be understood that the chips mentioned in the embodiments of the present application may also be referred to as system-on-chip, system-on-chip or system-on-chip, etc.
It should be noted that, in this document, the terms "comprises," "comprising," or any other variation thereof, are intended to cover a non-exclusive inclusion, such that a process, method, article, or apparatus that comprises a list of elements does not include only those elements but may include other elements not expressly listed or inherent to such process, method, article, or apparatus. Without further limitation, an element defined by the phrase "comprising a … …" does not exclude the presence of another identical element in a process, method, article, or apparatus that comprises the element. Further, it should be noted that the scope of the methods and apparatus of the embodiments of the present application is not limited to performing the functions in the order illustrated or discussed, but may include performing the functions in a substantially simultaneous manner or in a reverse order based on the functions involved, e.g., the methods described may be performed in an order different than that described, and various steps may be added, omitted, or combined. In addition, features described with reference to certain examples may be combined in other examples.
Through the above description of the embodiments, those skilled in the art will clearly understand that the method of the above embodiments can be implemented by software plus a necessary general hardware platform, and certainly can also be implemented by hardware, but in many cases, the former is a better implementation manner. Based on such understanding, the technical solutions of the present application may be embodied in the form of a software product, which is stored in a storage medium (such as ROM/RAM, magnetic disk, optical disk) and includes instructions for enabling a terminal (such as a mobile phone, a computer, a server, an air conditioner, or a network device) to execute the method according to the embodiments of the present application.
While the present embodiments have been described with reference to the accompanying drawings, it is to be understood that the invention is not limited to the precise embodiments described above, which are meant to be illustrative and not restrictive, and that various changes may be made therein by those skilled in the art without departing from the spirit and scope of the invention as defined by the appended claims.

Claims (11)

1. A method of taking a picture, comprising:
the method comprises the steps that under the condition that a shooting preview interface is displayed, first motion information of eyeballs of a user is obtained, wherein the shooting preview interface comprises N preview sub-areas, each preview sub-area is associated with a camera, and N is a positive integer;
determining a target preview subregion of the user's eye gaze based on the first motion information, the N preview subregions including the target preview subregion;
under the condition that the time length of the eyeballs of the user watching the target preview subarea is within a preset time length range, controlling a target camera associated with the target preview subarea to shoot a target image;
in the case of displaying the photographing preview interface, the method further includes:
receiving first input of a user to M preview sub-areas in the N preview sub-areas;
in response to the first input, combining the M preview sub-areas into a second preview sub-area, performing image synthesis on preview images acquired by the M cameras displayed in the M preview sub-areas to generate a target preview image, and displaying the target preview image in the second preview sub-area;
wherein M is an integer greater than 1;
and under the condition that the preview image displayed by a third preview subregion in the N preview subregions is obtained by synthesizing preview images acquired by at least two cameras, the third preview subregion further comprises a camera number identifier, and the camera number identifier is used for indicating the number of cameras associated with the third preview subregion.
2. The method of claim 1, wherein each of the N preview sub-areas comprises an area identifier having a first display mode and a second display mode;
under the condition that the area identification is in the first display mode, the area identification is used for indicating the eyeball of the user to watch the preview sub-area where the area identification is located;
and under the condition that the area identification is in the second display mode, the area identification is used for indicating that eyeballs of the user do not watch the preview sub-area where the area identification is located.
3. The method of claim 2, wherein after determining a target preview subregion at which the user's eye is gazed based on the first motion information, the method further comprises:
and updating the area identification in the target preview subarea into the first display mode.
4. The method of claim 1, wherein the capture preview interface further comprises a location indicator indicating first motion information of the user's eye, the first motion information comprising a real-time location of the user's eye gaze;
the determining a target preview sub-area at which the user's eye gazes based on the first motion information includes:
in a target time period, under the condition that the display position of the position mark is located in the display range of a first preview subregion, the first preview subregion is determined to be a target preview subregion watched by eyeballs of a user;
wherein the first preview sub-area is any one of the N preview sub-areas.
5. The method of claim 1, wherein said combining the M preview sub-regions into a second preview sub-region and image combining the preview images of the M preview sub-regions to generate a target preview image in response to the first input, and further comprising, after the second preview sub-region displays the target preview image:
receiving a second input of the user to the second preview subarea;
responding to the second input, splitting the second preview subregion into M preview subregions, and respectively displaying a preview image acquired by each camera in the M cameras in each preview subregion in the M preview subregions.
6. The method of claim 1, wherein in a case that the user looks at the target preview subregion for a period of time longer than a preset first period of time, the method further comprises:
under the condition of acquiring voice information input by a user, determining a camera obtained by recognizing the voice information as a target camera; alternatively, the first and second electrodes may be,
under the condition that selection input of a user on a fourth preview subarea in the N preview subareas is received, determining a camera related to the fourth preview subarea as a target camera;
and the preset first duration is greater than or equal to the maximum value of the preset duration range.
7. The method of claim 1, wherein prior to controlling the target camera associated with the target preview subregion to capture a target image, the method further comprises:
displaying a shooting parameter setting window under the condition that a third input of the user to the target preview sub-area is received, wherein the shooting parameter setting window comprises at least one shooting parameter;
receiving a fourth input of the shooting parameter setting window by the user;
determining a first photographing parameter in response to the fourth input;
the controlling the target camera associated with the target preview sub-area to shoot the target image comprises:
and controlling a target camera associated with the target preview subregion to shoot a target image according to the first shooting parameter.
8. The method of claim 4, wherein prior to controlling the target camera associated with the target preview subregion to capture a target image, the method further comprises:
acquiring second motion information of eyeballs of the user;
controlling the position mark to move according to the second motion information;
determining a second shooting parameter according to the moving distance of the position identifier;
the controlling the target camera associated with the target preview sub-area to shoot the target image comprises the following steps:
and controlling the target camera associated with the target preview subarea to shoot the target image according to the second shooting parameter.
9. A photographing apparatus, comprising:
the device comprises a first acquisition module, a second acquisition module and a third acquisition module, wherein the first acquisition module is used for acquiring first motion information of eyeballs of a user under the condition that a shooting preview interface is displayed, the shooting preview interface comprises N preview sub-areas, each preview sub-area is associated with a camera, and N is a positive integer;
a first processing module, configured to determine, based on the first motion information, target preview sub-areas at which eyeballs of the user gaze, where the N preview sub-areas include the target preview sub-areas;
the photographing module is used for controlling a target camera associated with the target preview subarea to photograph a target image under the condition that the time length for the eyeballs of the user to watch the target preview subarea is within a preset time length range;
the device further comprises:
the first receiving module is used for receiving first input of a user to M preview sub-areas in the N preview sub-areas;
the third processing module is used for responding to the first input, combining the M preview sub-areas into a second preview sub-area, performing image synthesis on preview images acquired by the M cameras displayed in the M preview sub-areas, generating a target preview image, and displaying the target preview image in the second preview sub-area;
wherein M is an integer greater than 1;
and under the condition that the preview image displayed by a third preview subregion in the N preview subregions is obtained by synthesizing preview images acquired by at least two cameras, the third preview subregion further comprises a camera number identifier, and the camera number identifier is used for indicating the number of cameras associated with the third preview subregion.
10. An electronic device comprising a processor, a memory, and a program or instructions stored on the memory and executable on the processor, the program or instructions when executed by the processor implementing the steps of the photographing method according to any one of claims 1 to 8.
11. A readable storage medium, characterized in that the readable storage medium has stored thereon a program or instructions which, when executed by a processor, implement the steps of the photographing method according to any one of claims 1 to 8.
CN202110170800.8A 2021-02-08 2021-02-08 Photographing method and device, electronic equipment and medium Active CN112954209B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202110170800.8A CN112954209B (en) 2021-02-08 2021-02-08 Photographing method and device, electronic equipment and medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202110170800.8A CN112954209B (en) 2021-02-08 2021-02-08 Photographing method and device, electronic equipment and medium

Publications (2)

Publication Number Publication Date
CN112954209A CN112954209A (en) 2021-06-11
CN112954209B true CN112954209B (en) 2023-02-17

Family

ID=76243436

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202110170800.8A Active CN112954209B (en) 2021-02-08 2021-02-08 Photographing method and device, electronic equipment and medium

Country Status (1)

Country Link
CN (1) CN112954209B (en)

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113784040B (en) * 2021-08-05 2023-11-14 维沃移动通信(杭州)有限公司 Shooting method, shooting device, electronic equipment and storage medium
CN114257824B (en) * 2021-11-25 2024-03-19 广州方硅信息技术有限公司 Live broadcast display method and device, storage medium and computer equipment

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107995429A (en) * 2017-12-22 2018-05-04 维沃移动通信有限公司 A kind of image pickup method and mobile terminal
CN109803110A (en) * 2019-01-29 2019-05-24 维沃移动通信有限公司 A kind of image processing method, terminal device and server
CN110225252A (en) * 2019-06-11 2019-09-10 Oppo广东移动通信有限公司 Camera control method and Related product
CN110505411A (en) * 2019-09-03 2019-11-26 RealMe重庆移动通信有限公司 Image capturing method, device, storage medium and electronic equipment
WO2020022548A1 (en) * 2018-07-24 2020-01-30 엘지전자 주식회사 Mobile terminal and control method therefor
CN111614893A (en) * 2020-04-28 2020-09-01 宇龙计算机通信科技(深圳)有限公司 Automatic photographing method and device, storage medium and intelligent glasses
CN111970566A (en) * 2020-08-26 2020-11-20 北京达佳互联信息技术有限公司 Video playing method and device, electronic equipment and storage medium

Family Cites Families (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR20160128119A (en) * 2015-04-28 2016-11-07 엘지전자 주식회사 Mobile terminal and controlling metohd thereof
CN107395703A (en) * 2017-07-11 2017-11-24 中国医学科学院医学信息研究所 A kind of resource supplying method and apparatus
US10466779B1 (en) * 2017-07-24 2019-11-05 Facebook Technologies, Llc Event camera for eye tracking
CN107566725A (en) * 2017-09-15 2018-01-09 广东小天才科技有限公司 A kind of camera control method and device
CN108197620B (en) * 2018-01-23 2021-11-26 广东小天才科技有限公司 Photographing and question searching method and system based on eye positioning and handheld photographing equipment
CN108595492B (en) * 2018-03-15 2020-12-29 腾讯科技(深圳)有限公司 Content pushing method and device, storage medium and electronic device
CN110516553A (en) * 2019-07-31 2019-11-29 北京航空航天大学 The monitoring method and device of working condition
CN110457699B (en) * 2019-08-06 2023-07-04 腾讯科技(深圳)有限公司 Method and device for mining stop words, electronic equipment and storage medium
CN111225157B (en) * 2020-03-03 2022-01-14 Oppo广东移动通信有限公司 Focus tracking method and related equipment

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107995429A (en) * 2017-12-22 2018-05-04 维沃移动通信有限公司 A kind of image pickup method and mobile terminal
WO2020022548A1 (en) * 2018-07-24 2020-01-30 엘지전자 주식회사 Mobile terminal and control method therefor
CN109803110A (en) * 2019-01-29 2019-05-24 维沃移动通信有限公司 A kind of image processing method, terminal device and server
CN110225252A (en) * 2019-06-11 2019-09-10 Oppo广东移动通信有限公司 Camera control method and Related product
CN110505411A (en) * 2019-09-03 2019-11-26 RealMe重庆移动通信有限公司 Image capturing method, device, storage medium and electronic equipment
CN111614893A (en) * 2020-04-28 2020-09-01 宇龙计算机通信科技(深圳)有限公司 Automatic photographing method and device, storage medium and intelligent glasses
CN111970566A (en) * 2020-08-26 2020-11-20 北京达佳互联信息技术有限公司 Video playing method and device, electronic equipment and storage medium

Also Published As

Publication number Publication date
CN112954209A (en) 2021-06-11

Similar Documents

Publication Publication Date Title
CN112135049B (en) Image processing method and device and electronic equipment
CN112135046B (en) Video shooting method, video shooting device and electronic equipment
EP4246287A1 (en) Method and system for displaying virtual prop in real environment image, and storage medium
CN112954210B (en) Photographing method and device, electronic equipment and medium
CN112954199B (en) Video recording method and device
CN112954214B (en) Shooting method, shooting device, electronic equipment and storage medium
CN111669507A (en) Photographing method and device and electronic equipment
CN111970456B (en) Shooting control method, device, equipment and storage medium
CN112954209B (en) Photographing method and device, electronic equipment and medium
CN112565613B (en) Preview image display control method, device, electronic equipment and medium
CN112911147B (en) Display control method, display control device and electronic equipment
CN112887617B (en) Shooting method and device and electronic equipment
CN113794829B (en) Shooting method and device and electronic equipment
CN112333382B (en) Shooting method and device and electronic equipment
WO2023083089A1 (en) Photographing control display method and apparatus, and electronic device and medium
CN112637495B (en) Shooting method, shooting device, electronic equipment and readable storage medium
CN112887618B (en) Video shooting method and device
CN113596331B (en) Shooting method, shooting device, shooting equipment and storage medium
CN112492205B (en) Image preview method and device and electronic equipment
CN115499589A (en) Shooting method, shooting device, electronic equipment and medium
CN112333395B (en) Focusing control method and device and electronic equipment
CN114245017A (en) Shooting method and device and electronic equipment
CN113923392A (en) Video recording method, video recording device and electronic equipment
CN113891002A (en) Shooting method and device
CN113873168A (en) Shooting method, shooting device, electronic equipment and medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant