CN110650289B - Shooting depth of field control method, equipment and computer readable storage medium - Google Patents

Shooting depth of field control method, equipment and computer readable storage medium Download PDF

Info

Publication number
CN110650289B
CN110650289B CN201910924135.XA CN201910924135A CN110650289B CN 110650289 B CN110650289 B CN 110650289B CN 201910924135 A CN201910924135 A CN 201910924135A CN 110650289 B CN110650289 B CN 110650289B
Authority
CN
China
Prior art keywords
depth
field
shooting
image information
preview area
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201910924135.XA
Other languages
Chinese (zh)
Other versions
CN110650289A (en
Inventor
杨亮
邵雪纯
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Nubia Technology Co Ltd
Original Assignee
Nubia Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Nubia Technology Co Ltd filed Critical Nubia Technology Co Ltd
Priority to CN201910924135.XA priority Critical patent/CN110650289B/en
Publication of CN110650289A publication Critical patent/CN110650289A/en
Application granted granted Critical
Publication of CN110650289B publication Critical patent/CN110650289B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/60Control of cameras or camera modules
    • H04N23/61Control of cameras or camera modules based on recognised objects
    • H04N23/611Control of cameras or camera modules based on recognised objects where the recognised objects include parts of the human body
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/60Control of cameras or camera modules
    • H04N23/63Control of cameras or camera modules by using electronic viewfinders
    • H04N23/631Graphical user interfaces [GUI] specially adapted for controlling image capture or setting capture parameters
    • H04N23/632Graphical user interfaces [GUI] specially adapted for controlling image capture or setting capture parameters for displaying or modifying preview images prior to image capturing, e.g. variety of image resolutions or capturing parameters
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02DCLIMATE CHANGE MITIGATION TECHNOLOGIES IN INFORMATION AND COMMUNICATION TECHNOLOGIES [ICT], I.E. INFORMATION AND COMMUNICATION TECHNOLOGIES AIMING AT THE REDUCTION OF THEIR OWN ENERGY USE
    • Y02D30/00Reducing energy consumption in communication networks
    • Y02D30/70Reducing energy consumption in communication networks in wireless communication networks

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Human Computer Interaction (AREA)
  • User Interface Of Digital Computer (AREA)

Abstract

The application discloses a shooting depth of field control method, equipment and a computer readable storage medium, wherein the method comprises the following steps: acquiring a wearing state and a current shooting state of the wearing equipment, wherein the shooting state comprises a shooting component and a shooting azimuth; then, obtaining image information obtained by the shooting assembly under the shooting direction, and analyzing the image information to obtain depth information of the image information; then, extracting a depth-of-field object according to the depth-of-field information, dividing a depth-of-field preview area according to the wearing state, and placing the depth-of-field object in the depth-of-field preview area; and finally, monitoring rotation parameters of the wearable equipment, and selecting a target depth object in the depth preview area according to the rotation parameters. The rapid depth-of-field regulation and control scheme is realized, the user can perform rapid and accurate depth-of-field regulation and control without using extra hands, the operation efficiency is improved, and the user experience is enhanced.

Description

Shooting depth of field control method, equipment and computer readable storage medium
Technical Field
The present disclosure relates to the field of mobile communications, and in particular, to a method and apparatus for controlling a photographing depth of field, and a computer readable storage medium.
Background
In the prior art, with the rapid development of intelligent terminal equipment, wearable equipment different from conventional smart phones, such as wearable equipment like a smart watch or a smart bracelet, appears. Because wearing formula equipment compares in traditional smart mobile phone, its particularities such as software, hardware environment, operation mode and operation environment, if the scheme of controlling of traditional smart mobile phone is transferred to wearing formula equipment, can bring inconvenience, user experience bad for user's operation.
Disclosure of Invention
In order to solve the technical defects in the prior art, the invention provides a shooting depth control method, which comprises the following steps:
acquiring a wearing state and a current shooting state of the wearing equipment, wherein the shooting state comprises a shooting component and a shooting azimuth;
acquiring image information obtained by the shooting assembly under the shooting direction, and analyzing the image information to obtain depth information of the image information;
extracting a depth-of-field object according to the depth-of-field information, dividing a depth-of-field preview area according to the wearing state, and placing the depth-of-field object in the depth-of-field preview area;
and monitoring rotation parameters of the wearable equipment, and selecting a target depth object in the depth preview area according to the rotation parameters.
Optionally, the acquiring the wearing state and the current shooting state of the wearable device, where the shooting state includes a shooting component and a shooting azimuth, includes:
acquiring the wearing state, wherein the wearing state comprises a current wearing gesture and an operation gesture;
and acquiring the shooting state, wherein the shooting component is identified according to a shooting instruction, and the shooting direction is identified according to the wearing gesture.
Optionally, the acquiring the image information obtained by the shooting component under the shooting direction, analyzing the image information, and obtaining depth information of the image information includes:
monitoring the motion state of the wearable device;
and if the motion state is stable in a preset period, acquiring the image information under the shooting direction through the shooting assembly.
Optionally, the acquiring the image information obtained by the shooting component under the shooting direction, analyzing the image information to obtain depth information of field of the image information, and further includes:
analyzing the image information, and dividing a shooting object in the image information into depth levels of a preset level;
And extracting depth parameters corresponding to the depth of field levels and a shooting object, and generating the depth of field information by the shooting object and the corresponding depth of field parameters.
Optionally, the extracting the depth of field object according to the depth of field information, and meanwhile, dividing a depth of field preview area according to the wearing state, and placing the depth of field object in the depth of field preview area includes:
extracting a depth object in the depth information and a depth parameter corresponding to the depth object;
and obtaining the depth of field grade according to the depth of field parameter.
Optionally, the extracting the depth of field object according to the depth of field information, meanwhile, dividing a depth of field preview area according to the wearing state, and placing the depth of field object in the depth of field preview area, further includes:
determining a shooting preview area and a shooting control area in the wearing state, wherein a current shooting image is previewed in the shooting preview area, and a depth of field preview area corresponding to the depth of field level is divided in the shooting control area;
and respectively placing the depth-of-field objects in the corresponding depth-of-field preview areas according to the depth-of-field levels.
Optionally, the monitoring the rotation parameter of the wearable device, and selecting the target depth of field object in the depth of field preview area according to the rotation parameter includes:
Receiving a preview lock signal in the shooting preview area, wherein the preview lock signal and a preview lock release signal are generated according to a pinching operation in the shooting preview area;
and monitoring the rotation parameters of the wearable equipment when the shooting preview area is in a preview locking state.
Optionally, the monitoring the rotation parameter of the wearable device, selecting a target depth of field object in the depth of field preview area according to the rotation parameter, further includes:
dividing the rotation parameters into a first rotation parameter in a first direction and a second rotation parameter in a second direction;
selecting the corresponding depth of field grade in the depth of field preview area according to the first rotation parameter;
and selecting a depth object as the target depth object in each depth level according to the second rotation parameter.
The invention also provides shooting depth-of-field control equipment, which comprises:
a memory, a processor, and a computer program stored on the memory and executable on the processor;
the computer program implementing the steps of the method according to any of the preceding claims when executed by the processor.
The invention also proposes a computer readable storage medium having stored thereon a shooting depth of field control program which, when executed by a processor, implements the steps of the shooting depth of field control method as described in any one of the above.
The method has the beneficial effects that the wearing state and the current shooting state of the wearing equipment are obtained, wherein the shooting state comprises a shooting component and a shooting azimuth; then, obtaining image information obtained by the shooting assembly under the shooting direction, and analyzing the image information to obtain depth information of the image information; then, extracting a depth-of-field object according to the depth-of-field information, dividing a depth-of-field preview area according to the wearing state, and placing the depth-of-field object in the depth-of-field preview area; and finally, monitoring rotation parameters of the wearable equipment, and selecting a target depth object in the depth preview area according to the rotation parameters. The rapid depth-of-field regulation and control scheme is realized, the user can perform rapid and accurate depth-of-field regulation and control without using extra hands, the operation efficiency is improved, and the user experience is enhanced.
Drawings
The accompanying drawings, which are incorporated in and constitute a part of this specification, illustrate embodiments consistent with the invention and together with the description, serve to explain the principles of the invention.
In order to more clearly illustrate the embodiments of the invention or the technical solutions of the prior art, the drawings which are used in the description of the embodiments or the prior art will be briefly described, and it will be obvious to a person skilled in the art that other drawings can be obtained from these drawings without inventive effort.
Fig. 1 is a schematic hardware structure of an implementation manner of a wearable device according to an embodiment of the present invention;
fig. 2 is a hardware schematic of an implementation of a wearable device provided in an embodiment of the present application;
fig. 3 is a hardware schematic of an implementation of a wearable device provided in an embodiment of the present application;
fig. 4 is a hardware schematic of an implementation of a wearable device provided in an embodiment of the present application;
fig. 5 is a hardware schematic of an implementation of a wearable device provided in an embodiment of the present application;
FIG. 6 is a flowchart of a first embodiment of a photographing depth of field control method of the present invention;
FIG. 7 is a flowchart of a second embodiment of a photographing depth of field control method of the present invention;
FIG. 8 is a flowchart of a third embodiment of a shooting depth of field control method of the present invention;
fig. 9 is a flowchart of a fourth embodiment of a photographing depth of field control method of the present invention;
fig. 10 is a flowchart of a fifth embodiment of a photographing depth of field control method of the present invention;
FIG. 11 is a flowchart of a sixth embodiment of a shooting depth of field control method of the present invention;
fig. 12 is a flowchart of a seventh embodiment of a photographing depth of field control method of the present invention;
fig. 13 is a flowchart of an eighth embodiment of the photographing depth of field control method of the present invention.
Detailed Description
It should be understood that the specific embodiments described herein are for purposes of illustration only and are not intended to limit the scope of the invention.
In the following description, suffixes such as "module", "component", or "unit" for representing elements are used only for facilitating the description of the present invention, and have no specific meaning per se. Thus, "module," "component," or "unit" may be used in combination.
The wearable device provided by the embodiment of the invention comprises a mobile terminal such as an intelligent bracelet, an intelligent watch and an intelligent mobile phone. With the continuous development of screen technology, mobile terminals such as smart phones and the like can also be used as wearable devices due to the appearance of screen forms such as flexible screens, folding screens and the like. The wearable device provided in the embodiment of the invention can comprise: RF (Radio Frequency) unit, wiFi module, audio output unit, A/V (audio/video) input unit, sensor, display unit, user input unit, interface unit, memory, processor, and power supply.
In the following description, a wearable device will be taken as an example, please refer to fig. 1, which is a schematic hardware structure of a wearable device implementing various embodiments of the present invention, where the wearable device 100 may include: an RF (Radio Frequency) unit 101, a WiFi module 102, an audio output unit 103, an a/V (audio/video) input unit 104, a sensor 105, a display unit 106, a user input unit 107, an interface unit 108, a memory 109, a processor 110, and a power supply 111. Those skilled in the art will appreciate that the wearable device structure shown in fig. 1 does not constitute a limitation of the wearable device, and that the wearable device may include more or fewer components than shown, or certain components in combination, or a different arrangement of components.
The following describes the various components of the wearable device in detail with reference to fig. 1:
the radio frequency unit 101 may be used to send and receive information or send signals in a call process, specifically, the radio frequency unit 101 may send uplink information to the base station, or may send downlink information sent by the base station to the processor 110 of the wearable device to process the downlink information, where the downlink information sent by the base station to the radio frequency unit 101 may be generated according to the uplink information sent by the radio frequency unit 101, or may be actively pushed to the radio frequency unit 101 after detecting that the information of the wearable device is updated, for example, after detecting that the geographic position where the wearable device is located changes, the base station may send a notification of the change of the geographic position to the radio frequency unit 101 of the wearable device, after receiving the notification of the message, the radio frequency unit 101 may send the notification of the message to the processor 110 of the wearable device to process, and the processor 110 of the wearable device may control the notification of the message to be displayed on the display panel 1061 of the wearable device; typically, the radio frequency unit 101 includes, but is not limited to, an antenna, at least one amplifier, a transceiver, a coupler, a low noise amplifier, a duplexer, and the like. In addition, the radio frequency unit 101 may also communicate with a network and other devices through wireless communication, which may specifically include: through wireless communication with a server in a network system, for example, the wearable device can download file resources from the server through wireless communication, for example, an application program can be downloaded from the server, after the wearable device finishes downloading a certain application program, if the file resources corresponding to the application program in the server are updated, the server can push a message notification of the resource update to the wearable device through wireless communication so as to remind a user to update the application program. The wireless communication may use any communication standard or protocol, including but not limited to GSM (Global System of Mobile communication, global System for Mobile communications), GPRS (General Packet Radio Service ), CDMA2000 (Code Division Multiple Access, CDMA 2000), WCDMA (Wideband Code Division Multiple Access ), TD-SCDMA (Time Division-Synchronous Code Division Multiple Access, time Division synchronous code Division multiple Access), FDD-LTE (Frequency DivisionDuplexing-Long Term Evolution, frequency Division Duplex Long term evolution), and TDD-LTE (Time Division Duplexing-Long Term Evolution, time Division Duplex Long term evolution), etc.
In one embodiment, the wearable device 100 may access an existing communication network by inserting a SIM card.
In another embodiment, the wearable device 100 may access an existing communication network by setting an esim card (Embedded-SIM), and by adopting the esim card, the internal space of the wearable device may be saved and the thickness may be reduced.
It will be appreciated that although fig. 1 shows a radio frequency unit 101, it will be appreciated that the radio frequency unit 101 is not an essential component of a wearable device and may be omitted entirely as required within the scope of not changing the essence of the invention. The wearable device 100 may implement communication connection with other devices or communication networks through the wifi module 102 alone, which is not limited by the embodiment of the present invention.
WiFi belongs to a short-distance wireless transmission technology, and the wearable device can help a user to send and receive emails, browse webpages, access streaming media and the like through the WiFi module 102, so that wireless broadband Internet access is provided for the user. Although fig. 1 shows a WiFi module 102, it is understood that it does not belong to the necessary constitution of the wearable device, and can be omitted entirely as required within the scope of not changing the essence of the invention.
The audio output unit 103 may convert audio data received by the radio frequency unit 101 or the WiFi module 102 or stored in the memory 109 into an audio signal and output as sound when the wearable device 100 is in a call signal reception mode, a talk mode, a recording mode, a voice recognition mode, a broadcast reception mode, or the like. Also, the audio output unit 103 may also provide audio output (e.g., call signal reception sound, message reception sound, etc.) related to a specific function performed by the wearable device 100. The audio output unit 103 may include a speaker, a buzzer, and the like.
The a/V input unit 104 is used to receive an audio or video signal. The a/V input unit 104 may include a graphics processor (Graphics Processing Unit, GPU) 1041 and a microphone 1042, the graphics processor 1041 processing image data of still pictures or video obtained by an image capturing device (e.g., a camera) in a video capturing mode or an image capturing mode. The processed image frames may be displayed on the display unit 106. The image frames processed by the graphics processor 1041 may be stored in the memory 109 (or other storage medium) or transmitted via the radio frequency unit 101 or the WiFi module 102. The microphone 1042 can receive sound (audio data) via the microphone 1042 in a phone call mode, a recording mode, a voice recognition mode, and the like, and can process such sound into audio data. The processed audio (voice) data may be converted into a format output that can be transmitted to the mobile communication base station via the radio frequency unit 101 in the case of a telephone call mode. The microphone 1042 may implement various types of noise cancellation (or suppression) algorithms to cancel (or suppress) noise or interference generated in the course of receiving and transmitting the audio signal.
In one embodiment, the wearable device 100 includes one or more cameras, and by opening the cameras, capturing of images, photographing, video recording and other functions can be achieved, and the positions of the cameras can be set as required.
The wearable device 100 further comprises at least one sensor 105, such as a light sensor, a motion sensor, and other sensors. Specifically, the light sensor includes an ambient light sensor and a proximity sensor, wherein the ambient light sensor can adjust the brightness of the display panel 1061 according to the brightness of ambient light, and the proximity sensor can turn off the display panel 1061 and/or the backlight when the wearable device 100 moves to the ear. As one type of motion sensor, the accelerometer sensor can detect the acceleration in all directions (typically three axes), and can detect the gravity and direction when stationary, and can be used for applications for recognizing the gesture of a mobile phone (such as horizontal-vertical screen switching, related games, magnetometer gesture calibration), vibration recognition related functions (such as pedometer, knocking), and the like.
In one embodiment, the wearable device 100 further comprises a proximity sensor, by employing the proximity sensor, the wearable device is able to achieve non-contact manipulation, providing more modes of operation.
In one embodiment, the wearable device 100 further comprises a heart rate sensor, which when worn, enables detection of heart rate by being in close proximity to the user.
In one embodiment, the wearable device 100 may further include a fingerprint sensor, by reading a fingerprint, security verification or the like can be achieved.
The display unit 106 is used to display information input by a user or information provided to the user. The display unit 106 may include a display panel 1061, and the display panel 1061 may be configured in the form of a liquid crystal display (Liquid Crystal Display, LCD), an Organic Light-Emitting Diode (OLED), or the like.
In one embodiment, the display panel 1061 employs a flexible display screen, and the wearable device employing the flexible display screen is capable of bending when worn, thereby fitting more. Optionally, the flexible display screen may be an OLED screen body and a graphene screen body, and in other embodiments, the flexible display screen may also be other display materials, which is not limited to this embodiment.
In one embodiment, the display panel 1061 of the wearable device may take a rectangular shape for ease of wrapping when worn. In other embodiments, other approaches may be taken as well.
The user input unit 107 may be used to receive input numeric or character information and to generate key signal inputs related to user settings and function control of the wearable device. In particular, the user input unit 107 may include a touch panel 1071 and other input devices 1072. The touch panel 1071, also referred to as a touch screen, may collect touch operations thereon or thereabout by a user (e.g., operations of the user on the touch panel 1071 or thereabout by using any suitable object or accessory such as a finger, a stylus, etc.) and drive the corresponding connection device according to a predetermined program. The touch panel 1071 may include two parts of a touch detection device and a touch controller. The touch detection device detects the touch azimuth of a user, detects a signal brought by touch operation and transmits the signal to the touch controller; the touch controller receives touch information from the touch detection device, converts it into touch point coordinates, and sends the touch point coordinates to the processor 110, and can receive and execute commands sent from the processor 110. Further, the touch panel 1071 may be implemented in various types such as resistive, capacitive, infrared, and surface acoustic wave. The user input unit 107 may include other input devices 1072 in addition to the touch panel 1071. In particular, other input devices 1072 may include, but are not limited to, one or more of a physical keyboard, function keys (e.g., volume control keys, switch keys, etc.), a trackball, mouse, joystick, etc., as specifically not limited herein.
In one embodiment, the sides of the wearable device 100 may be provided with one or more buttons. The button can realize a plurality of modes such as short pressing, long pressing, rotation and the like, thereby realizing a plurality of operation effects. The number of the buttons can be multiple, and different buttons can be combined for use, so that multiple operation functions are realized.
Further, the touch panel 1071 may overlay the display panel 1061, and when the touch panel 1071 detects a touch operation thereon or thereabout, the touch panel 1071 is transferred to the processor 110 to determine the type of touch event, and then the processor 110 provides a corresponding visual output on the display panel 1061 according to the type of touch event. Although in fig. 1, the touch panel 1071 and the display panel 1061 are two independent components for implementing the input and output functions of the wearable device, in some embodiments, the touch panel 1071 may be integrated with the display panel 1061 to implement the input and output functions of the wearable device, which is not limited herein. For example, when a message notification of a certain application is received through the rf unit 101, the processor 110 may control the message notification to be displayed in a certain preset area of the display panel 1061, where the preset area corresponds to a certain area of the touch panel 1071, and may control the message notification displayed in the corresponding area on the display panel 1061 by performing a touch operation on the certain area of the touch panel 1071.
The interface unit 108 serves as an interface through which at least one external device can be connected with the wearable apparatus 100. For example, the external devices may include a wired or wireless headset port, an external power (or battery charger) port, a wired or wireless data port, a memory card port, a port for connecting a device having an identification module, an audio input/output (I/O) port, a video I/O port, an earphone port, and the like. The interface unit 108 may be used to receive input (e.g., data information, power, etc.) from an external device and transmit the received input to one or more elements within the wearable apparatus 100 or may be used to transmit data between the wearable apparatus 100 and the external device.
In one embodiment, the interface unit 108 of the wearable device 100 adopts a contact structure, and is connected with other corresponding devices through the contact, so as to realize functions of charging, connection and the like. The contact can also be waterproof.
Memory 109 may be used to store software programs as well as various data. The memory 109 may mainly include a storage program area that may store an operating system, application programs required for at least one function (such as a sound playing function, an image playing function, etc.), and a storage data area; the storage data area may store data (such as audio data, phonebook, etc.) created according to the use of the handset, etc. In addition, memory 109 may include high-speed random access memory, and may also include non-volatile memory, such as at least one magnetic disk storage device, flash memory device, or other volatile solid-state storage device.
The processor 110 is a control center of the wearable device, connects various parts of the entire wearable device with various interfaces and lines, performs various functions of the wearable device and processes data by running or executing software programs and/or modules stored in the memory 109, and invoking data stored in the memory 109, thereby performing overall monitoring of the wearable device. Processor 110 may include one or more processing units; preferably, the processor 110 may integrate an application processor that primarily handles operating systems, user interfaces, applications, etc., with a modem processor that primarily handles wireless communications. It will be appreciated that the modem processor described above may not be integrated into the processor 110.
The wearable device 100 may further include a power source 111 (such as a battery) for powering the various components, and preferably, the power source 111 may be logically connected to the processor 110 through a power management system, so as to perform functions of managing charging, discharging, and power consumption management through the power management system.
Although not shown in fig. 1, the wearable device 100 may further include a bluetooth module or the like, which is not described herein. The wearable device 100 can be connected with other terminal devices through bluetooth to realize communication and information interaction.
Fig. 2 to fig. 4 are schematic structural diagrams of a wearable device according to an embodiment of the present invention. The wearable device comprises a flexible screen. When the wearable device is unfolded, the flexible screen is in a strip shape; when the wearable device is in a wearing state, the flexible screen is bent to be annular. Fig. 2 and 3 show schematic structural diagrams of the wearable device screen when unfolded, and fig. 4 shows schematic structural diagrams of the wearable device screen when bent.
Based on the above embodiments, it can be seen that if the device is a wristwatch, a bracelet, or a wearable device, the screen of the device may not cover the watchband area of the device, or may cover the watchband area of the device. In this embodiment, the device may be a wristwatch, a bracelet, or a wearable device, and the device includes a screen and a connection portion. The screen may be a flexible screen and the connection may be a wristband. Alternatively, the screen of the device or the display area of the screen may be partially or fully overlaid on the wristband of the device. Fig. 5 is a schematic hardware diagram of an implementation manner of a wearable device according to an embodiment of the present application, where a screen of the device extends to two sides, and a part of the screen is covered on a watchband of the device. In other embodiments, the screen of the device may also be entirely covered on the watchband of the device, which is not limited to this embodiment.
Example 1
Fig. 6 is a flowchart of a first embodiment of a photographing depth of field control method of the present invention. A photographing depth of field control method, the method comprising:
s1, acquiring a wearing state and a current shooting state of a wearing device, wherein the shooting state comprises a shooting component and a shooting azimuth;
s2, acquiring image information obtained by the shooting assembly under the shooting direction, and analyzing the image information to obtain depth information of the image information;
s3, extracting a depth-of-field object according to the depth-of-field information, dividing a depth-of-field preview area according to the wearing state, and placing the depth-of-field object in the depth-of-field preview area;
and S4, monitoring rotation parameters of the wearable equipment, and selecting a target depth object in the depth preview area according to the rotation parameters.
In this embodiment, first, a wearing state and a current shooting state of a wearable device are obtained, where the shooting state includes a shooting component and a shooting azimuth; then, obtaining image information obtained by the shooting assembly under the shooting direction, and analyzing the image information to obtain depth information of the image information; then, extracting a depth-of-field object according to the depth-of-field information, dividing a depth-of-field preview area according to the wearing state, and placing the depth-of-field object in the depth-of-field preview area; and finally, monitoring rotation parameters of the wearable equipment, and selecting a target depth object in the depth preview area according to the rotation parameters.
Optionally, the wearing state and the current shooting state of the wearable device are obtained, wherein the shooting state comprises a shooting component and a shooting azimuth. In this embodiment, the wearable device includes a wearable device of a wrist, a wearable device of an arm, a wearable device of a finger, and the like, and the wearable device has one or more cameras, and when the photographing depth of field control scheme of this embodiment is implemented, the one or more cameras are started to achieve hardware or software requirements of depth of field recognition, specifically, when the depth of field photographing is started, a wearing state and a current photographing state of the wearable device are obtained, where the wearing state is used for determining a body part of the wearable device relative to a user, and a specific position of the body part, so that on one hand, judgment of a motion state is conveniently performed subsequently, and on the other hand, a photographing preview image is displayed in a suitable area, and a photographing control area is provided, so that user operation and viewing are facilitated. In this embodiment, the method further includes obtaining a shooting state, where the shooting state includes a shooting component and a shooting direction, and it can be understood that the shooting component is generally fixed on the wearable device, so that a current corresponding shooting direction can be determined according to a currently enabled shooting component and a current wearing state, so as to facilitate a subsequent execution of a determination of a motion state;
Optionally, the image information obtained by the shooting component under the shooting direction is obtained, the image information is analyzed to obtain the depth information of the image information, wherein in this embodiment, the depth information of the shooting object included in the image information is obtained by analyzing the image information, it can be understood that, according to different shooting scenes or objects, each time or each shooting direction of the obtained image information may have different shooting objects and depth information thereof, in order to facilitate subsequent extraction and selection operations, in this embodiment, the shooting objects and the depth information thereof are extracted in a grading manner, that is, the current image information is marked with marks each including the depth information;
optionally, a depth object is extracted according to the depth information, and meanwhile, a depth preview area is divided according to the wearing state, and the depth object is placed in the depth preview area. The display scheme of the current wearable device is determined according to the wearable state, and specifically, the display scheme is divided into a division of a display area and a division of a control area. For example, a part of a display screen of the wearable device is divided into a depth preview area of the present embodiment, and it is understood that the depth preview area is arranged in parallel with the display area, so as to improve the utilization rate of the display screen of the wearable device, and then the classified photographic objects are sequentially placed in the depth preview area, and it is understood that the depth preview area of the present embodiment is an area classified by depth information, and in each area, one or more depth objects with the same depth information mark are provided;
Optionally, the rotation parameters of the wearable device are monitored, a target depth of view object is selected in the depth of view preview area according to the rotation parameters, and as described above, when the current wearing state and the shooting state are determined, at this stage, the rotation parameters of the wearable device are monitored, and it can be understood that, because the wearable device is in the stereoscopic space in the wearing state, any movement mode can be regarded as generating a certain rotation angle with the original state, in this embodiment, the rotation angle is collected, and the corresponding selection operation is performed in the depth of view preview area by using the angle, for example, the correspondence between the angle and each level of the depth of view preview area is first determined, and then the depth of view preview area of the corresponding level is determined according to the monitored angle.
The method has the advantages that the wearing state and the current shooting state of the wearing equipment are obtained, wherein the shooting state comprises a shooting component and a shooting azimuth; then, obtaining image information obtained by the shooting assembly under the shooting direction, and analyzing the image information to obtain depth information of the image information; then, extracting a depth-of-field object according to the depth-of-field information, dividing a depth-of-field preview area according to the wearing state, and placing the depth-of-field object in the depth-of-field preview area; and finally, monitoring rotation parameters of the wearable equipment, and selecting a target depth object in the depth preview area according to the rotation parameters. The rapid depth-of-field regulation and control scheme is realized, the user can perform rapid and accurate depth-of-field regulation and control without using extra hands, the operation efficiency is improved, and the user experience is enhanced.
Example two
Fig. 7 is a flowchart of a second embodiment of a photographing depth of field control method according to the present invention, based on the foregoing embodiment, the acquiring a wearing state of a wearable device and a current photographing state, where the photographing state includes a photographing component and a photographing azimuth, and includes:
s11, acquiring the wearing state, wherein the wearing state comprises a current wearing gesture and an operation gesture;
s12, acquiring the shooting state, wherein the shooting assembly is identified according to shooting instructions, and the shooting direction is identified according to the wearing gesture.
In this embodiment, first, the wearing state is obtained, where the wearing state includes a current wearing gesture and an operation gesture; then, the shooting state is acquired, wherein the shooting component is identified according to shooting instructions, and the shooting direction is identified according to the wearing gesture.
Optionally, the wearing state is obtained through equipment edge pressure sensing identification of the wearing equipment and at least two built-in gravity sensors at different positions, wherein the wearing state comprises current wearing gestures and operation gestures;
optionally, the shooting state is obtained, the shooting component is identified according to a shooting instruction, and the shooting azimuth is identified according to the wearing gesture, wherein the shooting azimuth includes a shooting azimuth corresponding to the shooting component capable of obtaining depth information.
The method has the beneficial effects that the wearing state is obtained, wherein the wearing state comprises the current wearing gesture and the operation gesture; then, the shooting state is acquired, wherein the shooting component is identified according to shooting instructions, and the shooting direction is identified according to the wearing gesture. The rapid depth-of-field regulation and control scheme is realized, the user can perform rapid and accurate depth-of-field regulation and control without using extra hands, the operation efficiency is improved, and the user experience is enhanced.
Example III
Fig. 8 is a flowchart of a third embodiment of a shooting depth control method according to the present invention, based on the above embodiment, the obtaining image information obtained by the shooting assembly in the shooting direction, analyzing the image information, and obtaining depth information of the image information, includes:
s21, monitoring the motion state of the wearable equipment;
s22, if the motion state is stable in a preset period, acquiring the image information under the shooting direction through the shooting assembly.
In this embodiment, first, a motion state of the wearable device is monitored; and then, if the motion state is stable in a preset period, acquiring the image information under the shooting direction through the shooting assembly.
Optionally, presetting a motion range corresponding to a current shooting scene or shooting requirement, and simultaneously monitoring a motion state of the wearable equipment;
optionally, when the motion state is in the preset or corresponding motion range, detecting in real time whether the motion state is stable in a preset period, and if the motion state is stable in the preset period, acquiring the image information by the shooting assembly under the shooting direction.
The beneficial effects of the embodiment are that the movement state of the wearable equipment is monitored; and then, if the motion state is stable in a preset period, acquiring the image information under the shooting direction through the shooting assembly. The rapid depth-of-field regulation and control scheme is realized, the user can perform rapid and accurate depth-of-field regulation and control without using extra hands, the operation efficiency is improved, and the user experience is enhanced.
Example IV
Fig. 9 is a flowchart of a fourth embodiment of a shooting depth of field control method according to the present invention, based on the above embodiment, the obtaining image information obtained by the shooting assembly in the shooting direction, analyzing the image information to obtain depth of field information of the image information, and further includes:
S23, analyzing the image information, and dividing a shooting object in the image information into depth levels of a preset level;
s24, extracting depth parameters corresponding to the depth of field levels and shooting objects, and generating the depth of field information by the shooting objects and the corresponding depth of field parameters.
In this embodiment, first, the image information is analyzed, and the shooting object in the image information is divided into depth levels of a preset level; and then, extracting depth parameters corresponding to the depth levels and the shooting objects, and generating the depth information by the shooting objects and the corresponding depth parameters.
Optionally, analyzing the image information, dividing the shooting objects in the image information into depth levels of preset levels, and it can be understood that traversing and identifying according to shooting objects contained in the current image information to determine depth parameters corresponding to each shooting object, and then dividing the depth parameters into a plurality of depth levels by combining the current shooting requirements and correlations among the shooting objects;
optionally, the depth of field parameters corresponding to the depth of field level and the photographed object are extracted, and the depth of field information is generated by the photographed object and the corresponding depth of field parameters thereof, where it may be understood that in the above step, the depth of field parameters corresponding to the photographed objects are determined by traversing and identifying the photographed object included in the current image information, and in this embodiment, in order to facilitate the subsequent extraction of the depth of field object, in this step, the attribute of the depth of field object and the depth of field parameters of the depth of field object are associated and integrated, so as to generate the depth of field information corresponding to the image information, where the depth of field information includes the attribute of the photographed object (for example, the name or the identifier of each photographed object) and the corresponding depth of field parameter thereof (for example, the depth of field level to which the depth of field parameter belongs).
The method has the advantages that the shooting objects in the image information are divided into depth-of-field levels of preset levels by analyzing the image information; and then, extracting depth parameters corresponding to the depth levels and the shooting objects, and generating the depth information by the shooting objects and the corresponding depth parameters. The rapid depth-of-field regulation and control scheme is realized, the user can perform rapid and accurate depth-of-field regulation and control without using extra hands, the operation efficiency is improved, and the user experience is enhanced.
Example five
Fig. 10 is a flowchart of a fifth embodiment of a shooting depth control method according to the present invention, based on the above embodiment, the extracting a depth object according to the depth information, and dividing a depth preview area according to the wearing state, and placing the depth object in the depth preview area, where the method includes:
s31, extracting a depth object in the depth information and a depth parameter corresponding to the depth object;
s32, obtaining the depth of field grade according to the depth of field parameter.
In this embodiment, first, a depth object and a depth parameter corresponding to the depth object in the depth information are extracted; and then, obtaining the depth of field grade according to the depth of field parameter.
Optionally, extracting the depth field object and the corresponding depth field parameter in the depth field information, and it can be understood that in this embodiment, if the image is formed more singly, the corresponding depth field object may be extracted according to each different depth field parameter;
optionally, the depth field objects and the corresponding depth field parameters in the depth field information are extracted, and it can be understood that in this embodiment, if the image is more complex, one or more depth field objects under the corresponding levels may be extracted according to each different depth field level.
The embodiment has the beneficial effects that the depth field object and the corresponding depth field parameter in the depth field information are extracted; and then, obtaining the depth of field grade according to the depth of field parameter. The rapid depth-of-field regulation and control scheme is realized, the user can perform rapid and accurate depth-of-field regulation and control without using extra hands, the operation efficiency is improved, and the user experience is enhanced.
Example six
Fig. 11 is a flowchart of a sixth embodiment of a shooting depth control method according to the present invention, based on the above embodiment, the depth object is extracted according to the depth information, and meanwhile, according to the wearing state, a depth preview area is divided, and the depth object is placed in the depth preview area, and further includes:
S33, determining a shooting preview area and a shooting control area in the wearing state, wherein a current shooting image is previewed in the shooting preview area, and a depth of field preview area corresponding to the depth of field level is divided in the shooting control area;
s34, respectively placing the depth of field objects in the corresponding depth of field preview areas according to the depth of field levels.
In this embodiment, first, in the wearing state, a shooting preview area and a shooting control area are determined, wherein a current shooting image is previewed in the shooting preview area, and a depth of field preview area corresponding to the depth of field level is divided in the shooting control area; and then, respectively placing the depth-of-field objects in the corresponding depth-of-field preview areas according to the depth-of-field levels.
Optionally, in the wearing state, determining a shooting preview area and a shooting control area matched with the sight range and the operation range of the user, wherein the current shooting image is previewed in the shooting preview area, and a depth of field preview area corresponding to the depth of field level is divided in the shooting control area;
optionally, according to the depth-of-field level, the depth-of-field objects are respectively placed in the corresponding depth-of-field preview areas, where each depth-of-field preview area includes one or more depth-of-field objects with specifically same or similar depth-of-field parameters.
The method has the advantages that a shooting preview area and a shooting control area are determined in the wearing state, wherein a current shooting image is previewed in the shooting preview area, and a depth of field preview area corresponding to the depth of field level is divided in the shooting control area; and then, respectively placing the depth-of-field objects in the corresponding depth-of-field preview areas according to the depth-of-field levels. The rapid depth-of-field regulation and control scheme is realized, the user can perform rapid and accurate depth-of-field regulation and control without using extra hands, the operation efficiency is improved, and the user experience is enhanced.
Example seven
Fig. 12 is a flowchart of a seventh embodiment of a shooting depth of field control method according to the present invention, based on the foregoing embodiment, the monitoring a rotation parameter of the wearable device, and selecting a target depth of field object in the depth of field preview area according to the rotation parameter includes:
s41, receiving a preview locking signal in the shooting preview area, wherein the preview locking signal and a preview unlocking signal are generated according to the kneading operation in the shooting preview area;
and S42, monitoring rotation parameters of the wearable equipment when the shooting preview area is in a preview locking state.
In this embodiment, first, a preview lock signal is received in the shooting preview area, wherein the preview lock signal and a preview lock release signal are generated according to a pinching operation in the shooting preview area; and then, monitoring the rotation parameters of the wearable equipment when the shooting preview area is in a preview locking state.
Optionally, a preview lock signal is received in the shooting preview area, wherein the preview lock signal and the preview lock release signal are generated according to a pinching operation in the shooting preview area, and specifically, the preview lock signal and the preview lock release signal are generated when a pinching operation in a width direction of a display of the wearable device is received in the shooting preview area;
optionally, when the shooting preview area is in a preview locked state, monitoring a rotation parameter of the wearable device, and simultaneously displaying a current rotation state in the preview locked interface, where the rotation state includes a rotation direction and a rotation amplitude.
The embodiment has the beneficial effects that the preview locking signal and the preview locking releasing signal are generated according to the kneading operation in the shooting preview area by receiving the preview locking signal in the shooting preview area; and then, monitoring the rotation parameters of the wearable equipment when the shooting preview area is in a preview locking state. The rapid depth-of-field regulation and control scheme is realized, the user can perform rapid and accurate depth-of-field regulation and control without using extra hands, the operation efficiency is improved, and the user experience is enhanced.
Example eight
Fig. 13 is a flowchart of an eighth embodiment of a shooting depth of field control method according to the present invention, based on the foregoing embodiment, the monitoring a rotation parameter of the wearable device, selecting a target depth of field object in the depth of field preview area according to the rotation parameter, and further includes:
s43, dividing the rotation parameters into first rotation parameters in a first direction and second rotation parameters in a second direction;
s44, selecting the corresponding depth of field grades in the depth of field preview area according to the first rotation parameters;
s45, selecting a depth object as the target depth object in each depth level according to the second rotation parameters.
In this embodiment, first, the rotation parameters are divided into a first rotation parameter in a first direction and a second rotation parameter in a second direction; then, selecting corresponding depth levels in the depth preview area according to the first rotation parameters; and finally, selecting a depth object as the target depth object in each depth level according to the second rotation parameters.
Optionally, if the current preview locked image is more complex, the number of depth of field objects under the corresponding one or more depth of field levels is greater, so as to facilitate the user to perform finer adjustment and control operations, in this embodiment, selection of the depth of field objects is performed in two steps, first, each depth of field level corresponding to the depth of field preview area is selected according to the first rotation parameter, and then, the depth of field object is selected as the target depth of field object in each depth of field level according to the second rotation parameter;
Alternatively, in order to distinguish the above two types of rotation, in the present embodiment, rotation in two directions may be performed in the horizontal direction in consideration of the state in which the wearing device is horizontally placed, and thus the rotation parameters are classified into a first rotation parameter in the first direction and a second rotation parameter in the second direction.
The embodiment has the advantages that the rotation parameters are divided into a first rotation parameter in a first direction and a second rotation parameter in a second direction; then, selecting corresponding depth levels in the depth preview area according to the first rotation parameters; and finally, selecting a depth object as the target depth object in each depth level according to the second rotation parameters. The rapid depth-of-field regulation and control scheme is realized, the user can perform rapid and accurate depth-of-field regulation and control without using extra hands, the operation efficiency is improved, and the user experience is enhanced.
Example nine
Based on the above embodiment, the present invention further provides a shooting depth-of-field control apparatus, including:
a memory, a processor, and a computer program stored on the memory and executable on the processor;
The computer program implementing the steps of the method according to any of the preceding claims when executed by the processor.
Specifically, in this embodiment, first, a wearing state and a current shooting state of a wearable device are obtained, where the shooting state includes a shooting component and a shooting azimuth; then, obtaining image information obtained by the shooting assembly under the shooting direction, and analyzing the image information to obtain depth information of the image information; then, extracting a depth-of-field object according to the depth-of-field information, dividing a depth-of-field preview area according to the wearing state, and placing the depth-of-field object in the depth-of-field preview area; and finally, monitoring rotation parameters of the wearable equipment, and selecting a target depth object in the depth preview area according to the rotation parameters.
Optionally, the wearing state and the current shooting state of the wearable device are obtained, wherein the shooting state comprises a shooting component and a shooting azimuth. In this embodiment, the wearable device includes a wearable device of a wrist, a wearable device of an arm, a wearable device of a finger, and the like, and the wearable device has one or more cameras, and when the photographing depth of field control scheme of this embodiment is implemented, the one or more cameras are started to achieve hardware or software requirements of depth of field recognition, specifically, when the depth of field photographing is started, a wearing state and a current photographing state of the wearable device are obtained, where the wearing state is used for determining a body part of the wearable device relative to a user, and a specific position of the body part, so that on one hand, judgment of a motion state is conveniently performed subsequently, and on the other hand, a photographing preview image is displayed in a suitable area, and a photographing control area is provided, so that user operation and viewing are facilitated. In this embodiment, the method further includes obtaining a shooting state, where the shooting state includes a shooting component and a shooting direction, and it can be understood that the shooting component is generally fixed on the wearable device, so that a current corresponding shooting direction can be determined according to a currently enabled shooting component and a current wearing state, so as to facilitate a subsequent execution of a determination of a motion state;
Optionally, the image information obtained by the shooting component under the shooting direction is obtained, the image information is analyzed to obtain the depth information of the image information, wherein in this embodiment, the depth information of the shooting object included in the image information is obtained by analyzing the image information, it can be understood that, according to different shooting scenes or objects, each time or each shooting direction of the obtained image information may have different shooting objects and depth information thereof, in order to facilitate subsequent extraction and selection operations, in this embodiment, the shooting objects and the depth information thereof are extracted in a grading manner, that is, the current image information is marked with marks each including the depth information;
optionally, a depth object is extracted according to the depth information, and meanwhile, a depth preview area is divided according to the wearing state, and the depth object is placed in the depth preview area. The display scheme of the current wearable device is determined according to the wearable state, and specifically, the display scheme is divided into a division of a display area and a division of a control area. For example, a part of a display screen of the wearable device is divided into a depth preview area of the present embodiment, and it is understood that the depth preview area is arranged in parallel with the display area, so as to improve the utilization rate of the display screen of the wearable device, and then the classified photographic objects are sequentially placed in the depth preview area, and it is understood that the depth preview area of the present embodiment is an area classified by depth information, and in each area, one or more depth objects with the same depth information mark are provided;
Optionally, the rotation parameters of the wearable device are monitored, a target depth of view object is selected in the depth of view preview area according to the rotation parameters, and as described above, when the current wearing state and the shooting state are determined, at this stage, the rotation parameters of the wearable device are monitored, and it can be understood that, because the wearable device is in the stereoscopic space in the wearing state, any movement mode can be regarded as generating a certain rotation angle with the original state, in this embodiment, the rotation angle is collected, and the corresponding selection operation is performed in the depth of view preview area by using the angle, for example, the correspondence between the angle and each level of the depth of view preview area is first determined, and then the depth of view preview area of the corresponding level is determined according to the monitored angle.
The method has the advantages that the wearing state and the current shooting state of the wearing equipment are obtained, wherein the shooting state comprises a shooting component and a shooting azimuth; then, obtaining image information obtained by the shooting assembly under the shooting direction, and analyzing the image information to obtain depth information of the image information; then, extracting a depth-of-field object according to the depth-of-field information, dividing a depth-of-field preview area according to the wearing state, and placing the depth-of-field object in the depth-of-field preview area; and finally, monitoring rotation parameters of the wearable equipment, and selecting a target depth object in the depth preview area according to the rotation parameters. The rapid depth-of-field regulation and control scheme is realized, the user can perform rapid and accurate depth-of-field regulation and control without using extra hands, the operation efficiency is improved, and the user experience is enhanced.
Examples ten
Based on the above embodiments, the present invention also proposes a computer readable storage medium having a bitmap processing program stored thereon, which when executed by a processor implements the steps of the bitmap processing method according to any one of the above.
The bitmap processing method, the bitmap processing device and the computer readable storage medium are implemented by acquiring the wearing state and the current shooting state of the wearing device, wherein the shooting state comprises a shooting component and a shooting azimuth; then, obtaining image information obtained by the shooting assembly under the shooting direction, and analyzing the image information to obtain depth information of the image information; then, extracting a depth-of-field object according to the depth-of-field information, dividing a depth-of-field preview area according to the wearing state, and placing the depth-of-field object in the depth-of-field preview area; and finally, monitoring rotation parameters of the wearable equipment, and selecting a target depth object in the depth preview area according to the rotation parameters. The rapid depth-of-field regulation and control scheme is realized, the user can perform rapid and accurate depth-of-field regulation and control without using extra hands, the operation efficiency is improved, and the user experience is enhanced.
It should be noted that, in this document, the terms "comprises," "comprising," or any other variation thereof, are intended to cover a non-exclusive inclusion, such that a process, method, article, or apparatus that comprises a list of elements does not include only those elements but may include other elements not expressly listed or inherent to such process, method, article, or apparatus. Without further limitation, an element defined by the phrase "comprising one … …" does not exclude the presence of other like elements in a process, method, article, or apparatus that comprises the element.
The foregoing embodiment numbers of the present invention are merely for the purpose of description, and do not represent the advantages or disadvantages of the embodiments.
From the above description of the embodiments, it will be clear to those skilled in the art that the above-described embodiment method may be implemented by means of software plus a necessary general hardware platform, but of course may also be implemented by means of hardware, but in many cases the former is a preferred embodiment. Based on such understanding, the technical solution of the present invention may be embodied essentially or in a part contributing to the prior art in the form of a software product stored in a storage medium (e.g. ROM/RAM, magnetic disk, optical disk) comprising instructions for causing a terminal (which may be a mobile phone, a computer, a server, an air conditioner, or a network device, etc.) to perform the method according to the embodiments of the present invention.
The embodiments of the present invention have been described above with reference to the accompanying drawings, but the present invention is not limited to the above-described embodiments, which are merely illustrative and not restrictive, and many forms may be made by those having ordinary skill in the art without departing from the spirit of the present invention and the scope of the claims, which are to be protected by the present invention.

Claims (5)

1. A shooting depth of field control method, the method comprising:
acquiring a wearing state and a current shooting state of the wearing equipment, wherein the shooting state comprises a shooting component and a shooting azimuth;
acquiring image information obtained by the shooting assembly under the shooting direction, and analyzing the image information to obtain depth information of the image information;
extracting a depth-of-field object according to the depth-of-field information, dividing a depth-of-field preview area according to the wearing state, and placing the depth-of-field object in the depth-of-field preview area;
monitoring rotation parameters of the wearable equipment, and selecting a target depth of field object in the depth of field preview area according to the rotation parameters;
the obtaining the image information obtained by the shooting assembly under the shooting direction, analyzing the image information, and obtaining the depth of field information of the image information includes:
Analyzing the image information, and dividing a shooting object in the image information into depth levels of a preset level;
extracting depth parameters corresponding to the depth of field levels and shooting objects, and generating the depth of field information by the shooting objects and the corresponding depth of field parameters;
the method comprises the steps of associating and integrating the attribute of the depth of field object with the depth of field parameter of the depth of field object to generate the depth of field information corresponding to the image information, wherein the depth of field information comprises the attribute of the depth of field object, the attribute of the depth of field object comprises the name or the identification of each shooting object, and the depth of field parameter of the depth of field object comprises the depth of field grade to which the current depth of field parameter belongs;
and extracting a depth object according to the depth information, dividing a depth preview area according to the wearing state, and placing the depth object in the depth preview area, wherein the depth object comprises:
extracting a depth object in the depth information and a depth parameter corresponding to the depth object;
obtaining a depth of field grade according to the depth of field parameter;
extracting corresponding depth-of-field objects according to different depth-of-field parameters if the composition of the image information is single, and extracting one or more depth-of-field objects under corresponding levels according to different depth-of-field levels if the composition of the image information is not single;
And extracting a depth-of-field object according to the depth-of-field information, dividing a depth-of-field preview area according to the wearing state, and placing the depth-of-field object in the depth-of-field preview area, wherein the method further comprises:
determining a shooting preview area and a shooting control area in the wearing state, wherein a current shooting image is previewed in the shooting preview area, and a depth of field preview area corresponding to the depth of field level is divided in the shooting control area;
respectively placing the depth-of-field objects in the corresponding depth-of-field preview areas according to the depth-of-field levels;
wherein each depth of field preview region comprises one or more depth of field objects of the same or similar depth of field parameters;
the monitoring of the rotation parameter of the wearable device, selecting a target depth of field object in the depth of field preview area according to the rotation parameter, includes:
receiving a preview lock signal in the shooting preview area, wherein the preview lock signal and a preview lock release signal are generated according to a pinching operation in the shooting preview area;
when the shooting preview area is in a preview locking state, monitoring rotation parameters of the wearable equipment;
displaying a current rotation state in the preview locking interface, wherein the rotation state comprises a rotation direction and a rotation amplitude;
The monitoring of the rotation parameter of the wearable device, selecting a target depth of field object in the depth of field preview area according to the rotation parameter, and further comprises:
dividing the rotation parameters into a first rotation parameter in a first direction and a second rotation parameter in a second direction;
selecting the corresponding depth of field grade in the depth of field preview area according to the first rotation parameter;
selecting a depth object as the target depth object within each depth class according to the second rotation parameters
When the wearable equipment is in a transverse state, rotation in two directions is determined in the horizontal direction, and the rotation parameters are divided into the first rotation parameters corresponding to the first direction and the second rotation parameters corresponding to the second direction.
2. The photographing depth of view control method according to claim 1, wherein the acquiring the wearing state and the current photographing state of the wearable device, wherein the photographing state includes a photographing component and a photographing azimuth, includes:
acquiring the wearing state, wherein the wearing state comprises a current wearing gesture and an operation gesture;
and acquiring the shooting state, wherein the shooting component is identified according to a shooting instruction, and the shooting direction is identified according to the wearing gesture.
3. The method according to claim 2, wherein the acquiring the image information obtained by the shooting assembly in the shooting direction, analyzing the image information, and obtaining depth information of the image information, includes:
monitoring the motion state of the wearable device;
and if the motion state is stable in a preset period, acquiring the image information under the shooting direction through the shooting assembly.
4. A photographing depth of view control apparatus, the apparatus comprising:
a memory, a processor, and a computer program stored on the memory and executable on the processor;
the computer program implementing the steps of the method according to any one of claims 1 to 3 when executed by the processor.
5. A computer-readable storage medium, wherein a shooting depth of view control program is stored on the computer-readable storage medium, which when executed by a processor, implements the steps of the shooting depth of view control method according to any one of claims 1 to 3.
CN201910924135.XA 2019-09-27 2019-09-27 Shooting depth of field control method, equipment and computer readable storage medium Active CN110650289B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201910924135.XA CN110650289B (en) 2019-09-27 2019-09-27 Shooting depth of field control method, equipment and computer readable storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910924135.XA CN110650289B (en) 2019-09-27 2019-09-27 Shooting depth of field control method, equipment and computer readable storage medium

Publications (2)

Publication Number Publication Date
CN110650289A CN110650289A (en) 2020-01-03
CN110650289B true CN110650289B (en) 2023-06-30

Family

ID=68992848

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910924135.XA Active CN110650289B (en) 2019-09-27 2019-09-27 Shooting depth of field control method, equipment and computer readable storage medium

Country Status (1)

Country Link
CN (1) CN110650289B (en)

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113572961A (en) * 2021-07-23 2021-10-29 维沃移动通信(杭州)有限公司 Shooting processing method and electronic equipment

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105933532A (en) * 2016-06-06 2016-09-07 广东欧珀移动通信有限公司 Image processing method and device, and mobile terminal
CN110049248A (en) * 2019-04-29 2019-07-23 努比亚技术有限公司 A kind of reference object regulation method, equipment and computer readable storage medium
CN110062167A (en) * 2019-04-28 2019-07-26 努比亚技术有限公司 A kind of linkage filming control method, equipment and computer readable storage medium
CN110177209A (en) * 2019-05-31 2019-08-27 努比亚技术有限公司 A kind of video parameter regulation method, equipment and computer readable storage medium
CN110187769A (en) * 2019-05-31 2019-08-30 努比亚技术有限公司 A kind of preview image inspection method, equipment and computer readable storage medium

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR20160131720A (en) * 2015-05-08 2016-11-16 엘지전자 주식회사 Mobile terminal and method for controlling the same

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105933532A (en) * 2016-06-06 2016-09-07 广东欧珀移动通信有限公司 Image processing method and device, and mobile terminal
CN110062167A (en) * 2019-04-28 2019-07-26 努比亚技术有限公司 A kind of linkage filming control method, equipment and computer readable storage medium
CN110049248A (en) * 2019-04-29 2019-07-23 努比亚技术有限公司 A kind of reference object regulation method, equipment and computer readable storage medium
CN110177209A (en) * 2019-05-31 2019-08-27 努比亚技术有限公司 A kind of video parameter regulation method, equipment and computer readable storage medium
CN110187769A (en) * 2019-05-31 2019-08-30 努比亚技术有限公司 A kind of preview image inspection method, equipment and computer readable storage medium

Also Published As

Publication number Publication date
CN110650289A (en) 2020-01-03

Similar Documents

Publication Publication Date Title
CN110299100B (en) Display direction adjustment method, wearable device and computer readable storage medium
CN109982179B (en) Audio signal output method and device, wearable device and storage medium
CN110096195B (en) Sports icon display method, wearable device and computer readable storage medium
CN110099218B (en) Interactive control method and device in shooting process and computer readable storage medium
CN110225282B (en) Video recording control method, device and computer readable storage medium
CN110620875B (en) Screenshot control method, equipment and computer readable storage medium in video shooting process
CN110187769B (en) Preview image viewing method, equipment and computer readable storage medium
CN109947524B (en) Interface display method, wearable device and computer readable storage medium
CN110177209B (en) Video parameter regulation and control method, device and computer readable storage medium
CN110139270B (en) Wearable device pairing method, wearable device and computer readable storage medium
CN109933294B (en) Data processing method and device, wearable device and storage medium
CN110060062B (en) Information exchange method after wearable device is lost, wearable device and storage medium
CN110098845B (en) Message processing method, wearable device, mobile terminal and readable storage medium
CN109933400B (en) Display interface layout method, wearable device and computer readable storage medium
CN110933293A (en) Shooting method, terminal and computer readable storage medium
CN110072071B (en) Video recording interaction control method, equipment and computer readable storage medium
CN110177208B (en) Video recording association control method, equipment and computer readable storage medium
CN110198411B (en) Depth of field control method and device in video shooting process and computer readable storage medium
CN110113529B (en) Shooting parameter regulation and control method and device and computer readable storage medium
CN110069193B (en) Interface switching method of wearable device, wearable device and storage medium
CN110069200B (en) Wearable device input control method, wearable device and storage medium
CN110650289B (en) Shooting depth of field control method, equipment and computer readable storage medium
CN110620876B (en) Image preview interaction method, device and computer readable storage medium
CN110049253B (en) Focusing control method and device and computer readable storage medium
CN110191282B (en) Shooting parameter regulation and control method and device and computer readable storage medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
TA01 Transfer of patent application right
TA01 Transfer of patent application right

Effective date of registration: 20220507

Address after: 1801, building 2, Chongwen Park, Nanshan Zhiyuan, No. 3370, Liuxian Avenue, Fuguang community, Taoyuan Street, Nanshan District, Shenzhen, Guangdong 518000

Applicant after: NUBIA TECHNOLOGY Co.,Ltd.

Address before: 518000 6-8 / F, 10-11 / F, 6 / F, 6-10 / F, C zone, Han's innovation building, No. 9018, Beihuan Avenue, high tech Zone, Nanshan District, Shenzhen City, Guangdong Province

Applicant before: NUBIA TECHNOLOGY Co.,Ltd.

Applicant before: NANCHANG NUBIA TECHNOLOGY Co.,Ltd.

SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant