CN113840092B - Camera switching method and electronic equipment - Google Patents

Camera switching method and electronic equipment Download PDF

Info

Publication number
CN113840092B
CN113840092B CN202111413464.1A CN202111413464A CN113840092B CN 113840092 B CN113840092 B CN 113840092B CN 202111413464 A CN202111413464 A CN 202111413464A CN 113840092 B CN113840092 B CN 113840092B
Authority
CN
China
Prior art keywords
camera
sensor
mobile phone
mode
electronic equipment
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202111413464.1A
Other languages
Chinese (zh)
Other versions
CN113840092A (en
Inventor
翟巍
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Honor Device Co Ltd
Original Assignee
Honor Device Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Honor Device Co Ltd filed Critical Honor Device Co Ltd
Priority to CN202111413464.1A priority Critical patent/CN113840092B/en
Publication of CN113840092A publication Critical patent/CN113840092A/en
Application granted granted Critical
Publication of CN113840092B publication Critical patent/CN113840092B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/60Control of cameras or camera modules
    • H04N23/66Remote control of cameras or camera parts, e.g. by remote control devices
    • H04N23/663Remote control of cameras or camera parts, e.g. by remote control devices for controlling interchangeable camera parts based on electronic image sensor signals
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for program control, e.g. control units
    • G06F9/06Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
    • G06F9/44Arrangements for executing specific programs
    • G06F9/4401Bootstrapping
    • G06F9/4411Configuring for operating with peripheral devices; Loading of device drivers
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/60Control of cameras or camera modules
    • H04N23/63Control of cameras or camera modules by using electronic viewfinders
    • H04N23/631Graphical user interfaces [GUI] specially adapted for controlling image capture or setting capture parameters
    • H04N23/632Graphical user interfaces [GUI] specially adapted for controlling image capture or setting capture parameters for displaying or modifying preview images prior to image capturing, e.g. variety of image resolutions or capturing parameters
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/60Control of cameras or camera modules
    • H04N23/667Camera operation mode switching, e.g. between still and video, sport and normal or high- and low-resolution modes

Abstract

A camera switching method and electronic equipment relate to the technical field of terminals and achieve the purpose of automatic switching of front and rear cameras. The method is applied to the electronic equipment and comprises the following steps: after the camera application program of the electronic equipment is completely started, the electronic equipment inputs information collected by at least one sensor into a trained first decision model for camera mode recognition to obtain a target camera mode; the first decision model has a mapping relation between information acquired by at least one sensor and a corresponding camera mode; the target camera mode comprises a front camera mode or a rear camera mode; the electronic equipment displays a first preview interface; the first preview interface comprises image information collected by a camera of the camera application program in a target camera mode.

Description

Camera switching method and electronic equipment
Technical Field
The present application relates to the field of terminal technologies, and in particular, to a method for switching a camera and an electronic device.
Background
In order to meet the photographing requirements of users, more and more electronic devices have a photographing function. For example, the electronic device may be a mobile phone or a tablet computer. The existing electronic equipment is generally provided with a front camera and a rear camera, and a user can manually switch the front camera and the rear camera according to the photographing requirement of the user when using the electronic equipment to photograph so as to determine whether to use the front camera or the rear camera. For example, when a user uses the electronic device to perform self-shooting, the user needs to manually switch to the front camera; when a user uses the electronic equipment to shoot a scene or shoot another person, the user needs to manually switch to the rear camera.
In the related art, the camera shooting function of the electronic equipment needs a user to frequently and manually switch the camera, so that the camera shooting function is complicated to use, and the shooting physical examination of the user is poor.
Disclosure of Invention
The embodiment of the application provides a camera switching method and electronic equipment. Therefore, automatic switching of the front camera and the rear camera is realized.
In order to achieve the purpose, the technical scheme is as follows:
in a first aspect, an embodiment of the present application provides a method for switching a camera, which is applied to an electronic device, and includes:
after an Application (APP) of the electronic equipment is started, the electronic equipment inputs information collected by at least one sensor into a trained first decision model for camera mode recognition to obtain a target camera mode; the first decision model has (e.g., has learned) a mapping between information collected by the at least one sensor and the corresponding camera mode; the target camera mode comprises a front camera mode or a rear camera mode; the electronic equipment displays a first preview interface; the first preview interface comprises image information collected by a camera of the camera APP in the target camera mode.
Therefore, after the camera APP of the electronic equipment is started, the information collected by at least one sensor is input into the trained first decision model to identify the target camera mode, the front camera and the rear camera are automatically switched to the target camera mode under the control of the electronic equipment, and the automatic switching of the front camera and the rear camera is realized.
In a possible implementation manner, the target camera mode is a front-facing camera mode, and after the electronic device inputs information acquired by at least one sensor into a trained first decision model to perform camera mode recognition, and the target camera mode is obtained, the method includes automatically switching the camera from the front-facing camera mode to a rear-facing camera mode if the electronic device detects that image information acquired by the front-facing camera includes preset image information.
The preset image information comprises image information corresponding to the hand of the user. And when the electronic equipment detects that the image information acquired by the front camera comprises the image information corresponding to the hand of the user, determining that the user acquires the image by adopting the rear camera, and controlling the camera to be automatically switched from the front camera mode to the rear camera mode by the electronic equipment.
In another possible implementation manner, after the electronic device displays the first preview interface, the method includes:
displaying the first prompt message; the first prompt message is used for reminding a user that the camera is switched to a front camera mode or a rear camera mode. Therefore, the electronic equipment reminds the user that the camera is automatically switched by displaying the prompt message.
In another possible implementation manner, after the electronic device displays the first preview interface, the method includes:
in the process that the electronic equipment displays the first prompt message, if the electronic equipment receives a first control instruction of a user within a first preset time after the camera switching is completed, the electronic equipment responds to the first control instruction to control the camera switching.
The manual switching operation of the user includes a sliding operation of the user on the display screen, a touch operation of the user on a camera switching button, and the like.
Therefore, after the electronic equipment automatically switches the camera, if the target camera mode to which the electronic equipment switches the camera is not the camera mode required by the user to actually shoot the scene, the user can manually switch the camera to switch the camera mode to the mode required by the user to actually shoot the scene.
In another possible implementation manner, the camera switching method further includes:
the electronic equipment displays second prompt information after the camera APP is completely started and the automatic camera switching mode is started for the first time; the second prompt message is used for reminding a user whether to train the automatic camera switching mode.
From this, before the user uses this kind of APP of making a video recording to shoot or make a video recording for the first time, electronic equipment can remind the user to train camera automatic switch-over mode, is favorable to improving the degree of accuracy of electronic equipment automatic identification target camera mode.
In another possible implementation manner, before the electronic device inputs information acquired by at least one sensor into the trained first decision model for camera mode recognition to obtain a target camera mode, the method further includes:
the electronic equipment determines that the electronic equipment is in a stable state according to the information collected by the at least one sensor.
Therefore, after the electronic equipment identifies that the electronic equipment is in a stable state, the target camera mode is automatically determined according to the information acquired by the at least one sensor, and the problem that the identification accuracy of the target camera mode is low due to inaccurate information acquired by the sensor when the electronic equipment shakes is avoided.
In another possible implementation manner, after the electronic device displays the first preview interface, the method further includes:
and if a second control instruction of the user to the photographing button on the first preview interface is received, performing iterative training on the first decision model by using information acquired by at least one sensor.
Therefore, after the electronic equipment detects that the user operates the photographing button on the first preview interface, the target camera mode identified according to the information collected by the at least one sensor is determined to be correct, iterative training is performed on the first decision model by using the information collected by the at least one sensor, and accuracy of the first decision model for identifying the camera mode according to the information collected by the sensor is improved.
In another possible implementation manner, the first preview interface is a preview interface of the camera APP in a photographing process or a preview interface of the camera APP in a recording process.
In another possible implementation, the sensor includes at least one of an image sensor, a gyroscope sensor, a gravity sensor, a structured light sensor, a proximity light sensor, or an infrared sensor.
In a second aspect, an embodiment of the present application provides another method for switching a camera, which is applied to an electronic device, and includes:
when the electronic equipment is in a formal starting stage, if the electronic equipment is detected not to be connected with external equipment, acquiring information acquired by at least one sensor of the electronic equipment within a second preset time before the formal starting of the camera application; inputting information acquired by at least one sensor into a trained second decision model for camera mode recognition, and determining a target camera mode; the second decision model has a mapping relation between information acquired by at least one sensor and a corresponding camera mode; and the electronic equipment displays a second preview interface after the camera application program is completely started, wherein the second preview interface comprises image information acquired by a camera of the camera application program in a target camera mode.
Therefore, when the electronic equipment is not connected with the external equipment, the electronic equipment can determine a target camera mode according to information collected by at least one sensor in a second preset range between the starting of the shooting type APP, and therefore automatic switching of the front camera and the rear camera is completed in the non-sensing state of a user in the starting process of the shooting type APP. In addition, the trained second decision model is adopted to identify the camera mode of the electronic equipment, so that the accuracy of camera mode identification is improved.
In another possible implementation manner, the method further includes:
and if the electronic equipment is detected to be connected with the external equipment, determining the target camera mode according to the external equipment.
Therefore, the electronic equipment can determine the target camera mode according to the external equipment connected with the electronic equipment and related to shooting so as to complete automatic switching of the camera in the starting process of the camera APP.
In another possible implementation manner, the method for switching the camera further includes:
and if a third control instruction of the user to a photographing button on the second preview interface is received within a third preset time after the camera type APP is started, the electronic equipment performs iterative training on the second decision model by using information acquired by at least one sensor.
Therefore, after the electronic equipment detects that the user operates the photographing button on the second preview interface, the target camera mode identified according to the information collected by the at least one sensor is correct in the starting process of the photographing APP, iterative training is performed on the second decision model by adopting the information collected by the at least one sensor, and the accuracy of the second decision model for identifying the camera mode according to the information collected by the sensor is improved.
In another possible implementation manner, the method for switching the camera further includes:
the electronic equipment displays a third preview interface; the third preview interface comprises image information acquired by a camera of the camera APP in the first camera mode;
after the camera APP is restarted, the electronic equipment displays a fourth preview interface; the fourth preview interface comprises image information acquired by a camera of the camera APP in the second camera mode; the second camera mode is different from the first camera mode.
In another possible implementation manner, the second preview interface, the third preview interface, and the fourth preview interface are preview interfaces of the camera APP in a photographing process or preview interfaces in a recording process.
Electronic equipment all carries out camera automatic switch-over at the type APP that adopts making a video recording and shoots and record a video in-process, does not do the restriction to this in the embodiment of this application.
In another possible implementation, the sensor includes at least one of an image sensor, a gyroscope sensor, a gravity sensor, a structured light sensor, a proximity light sensor, or an infrared sensor.
In a third aspect, the present application provides an electronic device having a function of implementing the method of the first aspect. The function can be realized by hardware, and can also be realized by executing corresponding software by hardware. The hardware or software includes one or more modules corresponding to the functions described above. For example, the electronic device includes an intelligent decision module, where the intelligent decision module is used for inputting information acquired by at least one sensor into a trained first decision model for camera mode recognition after the electronic device is started up in a full APP manner for a camera application program, so as to obtain a target camera mode; the target camera mode comprises a front camera mode or a rear camera mode; the electronic equipment displays a first preview interface; the first preview interface comprises image information collected by a camera of the camera APP in the target camera mode.
In a fourth aspect, the present application provides an electronic device comprising: one or more processors; a memory; wherein the memory has stored therein one or more computer programs, the one or more computer programs comprising instructions, which when executed by the electronic device, cause the electronic device to perform the method of switching cameras as described in any of the first or second aspects above.
In a fifth aspect, the present application provides a computer-readable storage medium having instructions stored therein, which when run on an electronic device, cause the electronic device to perform the method for switching the camera according to any one of the first aspect or the second aspect.
In a sixth aspect, the present application provides a computer program product comprising computer instructions which, when run on an electronic device, cause the electronic device to perform the method for switching a camera according to any one of the first or second aspects.
It is to be understood that the electronic device according to the third and fourth aspects, the computer storage medium according to the fifth aspect, and the computer program product according to the sixth aspect are all configured to perform the corresponding method provided above, and therefore, the beneficial effects achieved by the electronic device according to the third and fourth aspects can refer to the beneficial effects in the corresponding method provided above, and are not described herein again.
Drawings
Fig. 1 is a first application scenario diagram of a switching method of a camera according to an embodiment of the present application;
fig. 2 is a first schematic structural diagram of an electronic device according to an embodiment of the present disclosure;
fig. 3 is a schematic structural diagram of an electronic device according to an embodiment of the present application;
fig. 4 is a schematic view of an application scenario of a method for switching a camera according to an embodiment of the present application;
fig. 5 is a schematic flowchart of a method for switching a camera according to an embodiment of the present disclosure;
fig. 6 is a schematic view of an application scenario three of a switching method of a camera provided in the embodiment of the present application;
fig. 7 is a schematic view of an application scenario of a method for switching a camera according to an embodiment of the present application;
FIG. 8 is a schematic diagram of a first decision model training process provided by an embodiment of the present application;
FIG. 9 is a flowchart illustrating another first decision model training process according to an embodiment of the present disclosure;
fig. 10 is a schematic flowchart of another switching method for a camera according to an embodiment of the present disclosure;
FIG. 11 is a diagram illustrating a second decision model training process according to an embodiment of the present disclosure;
FIG. 12 is a flowchart illustrating a process of training a second decision model in an online state according to an embodiment of the present disclosure;
fig. 13 is a third schematic structural diagram of an electronic device according to an embodiment of the present application.
Detailed Description
The technical solutions in the embodiments of the present application will be described below with reference to the drawings in the embodiments of the present application. In the description of the embodiments herein, "/" means "or" unless otherwise specified, for example, a/B may mean a or B; "and/or" herein is merely an association describing an associated object, and means that there may be three relationships, e.g., a and/or B, which may mean: a exists alone, A and B exist simultaneously, and B exists alone.
In the following, the terms "first", "second" are used for descriptive purposes only and are not to be understood as indicating or implying relative importance or implicitly indicating the number of technical features indicated. Thus, a feature defined as "first" or "second" may explicitly or implicitly include one or more of that feature. In the description of the embodiments of the present application, "a plurality" means two or more unless otherwise specified.
In the embodiments of the present application, words such as "exemplary" or "for example" are used to mean serving as an example, instance, or illustration. Any embodiment or design described herein as "exemplary" or "e.g.," is not necessarily to be construed as preferred or advantageous over other embodiments or designs. Rather, use of the word "exemplary" or "such as" is intended to present concepts related in a concrete fashion.
At present, when the electronic equipment switches the front camera and the rear camera during shooting, a user needs to manually switch the front camera and the rear camera. Fig. 1 is a schematic view of an application scenario of a camera switching method according to an embodiment of the present application. As shown in fig. 1, after the electronic device opens a camera Application (APP) in response to a user operation, a camera mode currently opened by the electronic device is not a camera mode required by the user. For example, when the user uses the electronic device to take a self-timer, but the electronic device turns on the camera APP, the camera mode turned on by the electronic device is the rear camera mode. In this case, the electronic apparatus may switch the camera mode from the rear camera mode to the front camera mode in response to a user clicking or touching a camera switching control (e.g., the camera switching control in fig. 1) or the like.
In addition, the electronic device may also have a memory function, for example, if the electronic device is started when shooting by using the camera APP last time, the electronic device still has a front camera mode when the electronic device defaults to start the camera again. However, such simple memory of the electronic device does not represent the real intention of the user to open the camera APP. If the camera mode is not the camera mode really needed by the user when the electronic equipment starts the camera APP again, the user still needs to manually switch the camera.
In the embodiment of the application, the starting process of the camera APP comprises a preparation starting phase, a formal starting phase and a complete starting phase. The preparation start phase refers to a phase in which the camera APP is prepared to start before the formal start phase. The formal starting stage refers to a stage that the camera APP starts to start but the preview interface of the mobile phone does not display image information acquired by the camera. The complete starting stage is a stage in which a preview interface of the mobile phone displays image information acquired by a camera of the camera APP.
In order to solve the problems that in the related art, in the use process of a camera APP of electronic equipment, a user manually switches a camera according to an actual shooting scene, so that the user operation is frequent, the shooting physical examination is poor and the like, the embodiment of the application provides a switching method of the camera, and the method enables the electronic equipment to control the automatic switching of the front camera and the rear camera when the camera function is used. For example, when a camera APP of the electronic device is in a formal start stage, after the electronic device identifies a target camera mode, the front camera and the rear camera are controlled to automatically switch to the target camera mode. Therefore, the electronic equipment completes automatic switching of the camera in a state of no perception of the user. Or, under the condition that the camera automatic switching mode is started by the camera APP of the electronic equipment, after the camera APP is completely started, the electronic equipment identifies the target camera mode, and controls the front camera and the rear camera to automatically switch to the target camera mode. From this, realized the in-process that camera APP shot, electronic equipment has improved user's shooting experience according to the purpose of camera around the shooting scene automatic switch-over of reality.
For example, the method for switching the camera provided in the embodiment of the present application may be applied to an electronic device with a camera function, such as a mobile phone, a tablet Computer, a Personal Computer (PC), a Personal Digital Assistant (PDA), a smart watch, a netbook, a wearable electronic device, an Augmented Reality (AR) device, a Virtual Reality (VR) device, and a vehicle-mounted device, and the embodiment of the present application does not limit the method.
Fig. 2 is a first schematic structural diagram of an electronic device according to an embodiment of the present application. As shown in fig. 2, the electronic device 200 may include a processor 210, an external memory interface 220, an internal memory 221, a Universal Serial Bus (USB) interface 230, a charging management Module 240, a power management Module 241, a battery 242, an antenna 1, an antenna 2, a mobile communication Module 250, a wireless communication Module 260, an audio Module 270, a speaker 270A, a receiver 270B, a microphone 270C, an earphone interface 270D, a sensor Module 280, a button 290, a motor 291, an indicator 292, a camera 293, a display screen 294, a Subscriber Identity Module (SIM) card interface 295, and the like. The sensor module 280 may include a pressure sensor 280A, a gyroscope sensor 280B, an air pressure sensor 280C, a magnetic sensor 280D, an acceleration sensor 280E, a distance sensor 280F, a proximity light sensor 280G, a fingerprint sensor 280H, a temperature sensor 280J, a touch sensor 280K, an ambient light sensor 280L, a bone conduction sensor 280M, and the like.
It is to be understood that the illustrated structure of the embodiment of the present application does not specifically limit the electronic device 200. In other embodiments of the present application, the electronic device 200 may include more or fewer components than shown, or combine certain components, or split certain components, or a different arrangement of components. The illustrated components may be implemented in hardware, software, or a combination of software and hardware.
Processor 210 may include one or more processing units, such as: the Processor 210 may include an Application Processor (AP), a modem Processor, a Graphics Processing Unit (GPU), an Image Signal Processor (ISP), a controller, a memory, a video codec, a Digital Signal Processor (DSP), a baseband Processor, and/or a Neural-Network Processing Unit (NPU), etc. The different processing units may be separate devices or may be integrated into one or more processors.
The controller may be, among other things, a neural center and a command center of the electronic device 200. The controller can generate an operation control signal according to the instruction operation code and the timing signal to complete the control of instruction fetching and instruction execution.
A memory may also be provided in processor 210 for storing instructions and data. In some embodiments, the memory in the processor 210 is a cache memory. The memory may hold instructions or data that have just been used or recycled by processor 210. If the processor 210 needs to reuse the instruction or data, it may be called directly from memory. Avoiding repeated accesses reduces the latency of the processor 210, thereby increasing the efficiency of the system.
In some embodiments, processor 210 may include one or more interfaces. The Interface may include an integrated Circuit (I2C) Interface, an Inter-integrated Circuit built-in audio source (I2S) Interface, a Pulse Code Modulation (PCM) Interface, a Universal Asynchronous Receiver/Transmitter (UART) Interface, a Mobile Industry Processor Interface (MIPI), a General-Purpose Input/Output (GPIO) Interface, a Subscriber Identity Module (SIM) Interface, and/or a Universal Serial Bus (USB) Interface, etc.
It should be understood that the interfacing relationship between the modules illustrated in the embodiments of the present application is only an illustration, and does not limit the structure of the electronic device 200. In other embodiments of the present application, the electronic device 200 may also adopt different interface connection manners or a combination of multiple interface connection manners in the above embodiments.
The charge management module 240 is configured to receive a charging input from a charger. The charger may be a wireless charger or a wired charger. In some wired charging embodiments, the charging management module 240 may receive charging input from a wired charger via the USB interface 230. In some wireless charging embodiments, the charging management module 240 may receive a wireless charging input through a wireless charging coil of the electronic device 200. The charging management module 240 may also supply power to the electronic device through the power management module 241 while charging the battery 242.
The power management module 241 is used to connect the battery 242, the charging management module 240 and the processor 210. The power management module 241 receives input from the battery 242 and/or the charging management module 240, and provides power to the processor 210, the internal memory 221, the external memory, the display 294, the camera 293, and the wireless communication module 260. The power management module 241 may also be used to monitor parameters such as battery capacity, battery cycle number, battery state of health (leakage, impedance), etc. In some other embodiments, the power management module 241 may also be disposed in the processor 210. In other embodiments, the power management module 241 and the charging management module 240 may be disposed in the same device.
The wireless communication function of the electronic device 200 may be implemented by the antenna 1, the antenna 2, the mobile communication module 250, the wireless communication module 260, the modem processor, the baseband processor, and the like.
The antennas 1 and 2 are used for transmitting and receiving electromagnetic wave signals. Each antenna in the electronic device 200 may be used to cover a single or multiple communication bands. Different antennas can also be multiplexed to improve the utilization of the antennas. For example: the antenna 1 may be multiplexed as a diversity antenna of a wireless local area network. In other embodiments, the antenna may be used in conjunction with a tuning switch.
The mobile communication module 250 may provide a solution including 2G/3G/4G/5G wireless communication applied on the electronic device 200. The mobile communication module 250 may include at least one filter, a switch, a power Amplifier, a Low Noise Amplifier (LNA), and the like. The mobile communication module 250 may receive the electromagnetic wave from the antenna 1, filter, amplify, etc. the received electromagnetic wave, and transmit the electromagnetic wave to the modem processor for demodulation. The mobile communication module 250 may also amplify the signal modulated by the modem processor, and convert the signal into electromagnetic wave through the antenna 1 to radiate the electromagnetic wave. In some embodiments, at least some of the functional modules of the mobile communication module 250 may be disposed in the processor 210. In some embodiments, at least some of the functional modules of the mobile communication module 250 may be disposed in the same device as at least some of the modules of the processor 210.
The Wireless Communication module 260 may provide solutions for Wireless Communication applied to the electronic device 200, including Wireless Local Area Networks (WLANs) (e.g., Wireless Fidelity (Wi-Fi) network), Bluetooth (BT), Global Navigation Satellite System (GNSS), Frequency Modulation (FM), Near Field Communication (NFC), Infrared (IR), and the like. The wireless communication module 260 may be one or more devices integrating at least one communication processing module. The wireless communication module 260 receives electromagnetic waves via the antenna 2, performs frequency modulation and filtering processing on electromagnetic wave signals, and transmits the processed signals to the processor 210. The wireless communication module 260 may also receive a signal to be transmitted from the processor 210, frequency-modulate and amplify the signal, and convert the signal into electromagnetic waves via the antenna 2 to radiate the electromagnetic waves.
In some embodiments, antenna 1 of electronic device 200 is coupled to mobile communication module 250 and antenna 2 is coupled to wireless communication module 260, such that electronic device 200 may communicate with networks and other devices via wireless communication techniques. The wireless communication technologies may include Global System for Mobile communications (GSM), General Packet Radio Service (GPRS), Code Division Multiple Access (Code Division Multiple Access, CDMA), Wideband Code Division Multiple Access (WCDMA), Time-Division Multiple Access (Time-Division Code Division Multiple Access, TD-CDMA), Long Term Evolution (Long Term Evolution, LTE), BT, GNSS, WLAN, NFC, FM, and/or IR technologies, among others. GNSS may include Global Positioning System (GPS), Global Navigation Satellite System (GLONASS), BeiDou Navigation Satellite System (BDS), Quasi-Zenith Satellite System (QZSS), and/or Satellite Based Augmentation System (SBAS).
The electronic device 200 implements display functions via the GPU, the display screen 294, and the application processor. The GPU is a microprocessor for image processing, and is connected to the display screen 294 and an application processor. The GPU is used to perform mathematical and geometric calculations for graphics rendering. Processor 210 may include one or more GPUs that execute program instructions to generate or alter display information.
The display screen 294 is used to display images, video, and the like. The display screen 294 includes a display panel. The Display panel may be a Liquid Crystal Display (LCD), an Organic Light-Emitting Diode (OLED), an Active Matrix Organic Light-Emitting Diode (Active-Matrix Organic Light-Emitting Diode, AMOLED), a flexible Light-Emitting Diode (FLED), a miniature, a Micro-oeled, a Quantum dot Light-Emitting Diode (QLED), or the like. In some embodiments, the electronic device 200 may include 1 or N display screens 294, N being a positive integer greater than 1.
The electronic device 200 may implement a shooting function through the ISP, the camera 293, the video codec, the GPU, the display screen 294, and the application processor.
The ISP is used to process the data fed back by the camera 293. For example, when a photo is taken, the shutter is opened, light is transmitted to the camera photosensitive element through the lens, the optical signal is converted into an electrical signal, and the camera photosensitive element transmits the electrical signal to the ISP for processing and converting into an image visible to naked eyes. The ISP can also carry out algorithm optimization on the noise, brightness and skin color of the image. The ISP can also optimize parameters such as exposure, color temperature and the like of a shooting scene. In some embodiments, the ISP may be provided in camera 293.
The camera 293 is used to capture still images or video. The camera 293 includes a front camera and a rear camera. The object generates an optical image through the lens and projects the optical image to the photosensitive element. The photosensitive element may be a Charge Coupled Device (CCD) or a Complementary Metal-Oxide-Semiconductor (CMOS) phototransistor. The light sensing element converts the optical signal into an electrical signal, which is then passed to the ISP where it is converted into a digital image signal. And the ISP outputs the digital image signal to the DSP for processing. The DSP converts the digital image signal into image signal in standard RGB, YUV and other formats. In some embodiments, electronic device 200 may include 1 or N cameras 293, N being a positive integer greater than 1.
The digital signal processor is used for processing digital signals, and can process digital image signals and other digital signals. For example, when the electronic device 200 selects a frequency bin, the digital signal processor is used to perform fourier transform or the like on the frequency bin energy.
Video codecs are used to compress or decompress digital video. The electronic device 200 may support one or more video codecs. In this way, the electronic device 200 may play or record video in a variety of encoding formats, such as: moving Picture Experts Group (MPEG) 1, MPEG2, MPEG3, MPEG4, and the like.
The NPU is a Neural-Network (NN) computing processor, which processes input information quickly by using a biological Neural Network structure, for example, by using a transfer mode between neurons of a human brain, and can also learn by itself continuously. The NPU can implement applications such as intelligent recognition of the electronic device 200, for example: image recognition, face recognition, speech recognition, text understanding, and the like.
The external memory interface 220 may be used to connect an external memory card, such as a Micro SD card, to extend the memory capability of the electronic device 200. The external memory card communicates with the processor 210 through the external memory interface 220 to implement a data storage function. For example, files such as music, video, etc. are saved in an external memory card.
Internal memory 221 may be used to store computer-executable program code, which includes instructions. The processor 210 executes various functional applications of the electronic device 200 and data processing by executing instructions stored in the internal memory 221. The internal memory 221 may include a program storage area and a data storage area. The storage program area may store an operating system, an application program (such as a sound playing function, an image playing function, etc.) required by at least one function, and the like. The storage data area may store data (e.g., audio data, a phone book, etc.) created during use of the electronic device 200, and the like. In addition, the internal memory 221 may include a high-speed random access memory, and may further include a nonvolatile memory, such as at least one magnetic disk Storage device, a Flash memory device, a Universal Flash Storage (UFS), and the like.
Electronic device 200 may implement audio functions via audio module 270, speaker 270A, receiver 270B, microphone 270C, headset interface 270D, and an application processor, among other things. Such as music playing, recording, etc.
The pressure sensor 280A is used to sense a pressure signal, which can be converted into an electrical signal. In some embodiments, the pressure sensor 280A may be disposed on the display screen 294. The pressure sensor 280A can be of a wide variety of types, such as a resistive pressure sensor, an inductive pressure sensor, a capacitive pressure sensor, and the like. The capacitive pressure sensor may be a sensor comprising at least two parallel plates having an electrically conductive material. When a force acts on the pressure sensor 280A, the capacitance between the electrodes changes. The electronic device 200 determines the intensity of the pressure from the change in capacitance. When a touch operation is applied to the display screen 294, the electronic apparatus 200 detects the intensity of the touch operation based on the pressure sensor 280A. The electronic apparatus 200 may also calculate the touched position from the detection signal of the pressure sensor 280A. In some embodiments, the touch operations that are applied to the same touch position but different touch operation intensities may correspond to different operation instructions. For example: and when the touch operation with the touch operation intensity smaller than the first pressure threshold value acts on the short message application icon, executing an instruction for viewing the short message. And when the touch operation with the touch operation intensity larger than or equal to the first pressure threshold value acts on the short message application icon, executing an instruction of newly building the short message.
The gyro sensor 280B may be used to determine the motion pose of the electronic device 200. In some embodiments, the angular velocity of the electronic device 200 about three axes (i.e., x, y, and z axes) may be determined by the gyroscope sensor 280B. The gyro sensor 280B may be used for photographing anti-shake. For example, when the shutter is pressed, the gyro sensor 280B detects a shake angle of the electronic device 200, calculates a distance to be compensated for by the lens module according to the shake angle, and allows the lens to counteract the shake of the electronic device 200 through a reverse movement, thereby achieving anti-shake. The gyro sensor 280B may also be used for navigation, somatosensory gaming scenes.
The air pressure sensor 280C is used to measure air pressure. In some embodiments, electronic device 200 calculates altitude, aiding in positioning and navigation, from barometric pressure values measured by barometric pressure sensor 280C.
The magnetic sensor 280D includes a hall sensor. The electronic device 200 may detect the opening and closing of the flip holster using the magnetic sensor 280D. In some embodiments, when the electronic device 200 is a flip, the electronic device 200 may detect the opening and closing of the flip according to the magnetic sensor 280D. And then according to the opening and closing state of the leather sheath or the opening and closing state of the flip cover, the automatic unlocking of the flip cover is set.
The acceleration sensor 280E may detect the magnitude of acceleration of the electronic device 200 in various directions (typically three axes). The magnitude and direction of gravity can be detected when the electronic device 200 is stationary. The method can also be used for recognizing the posture of the electronic equipment, and is applied to horizontal and vertical screen switching, pedometers and other applications.
A distance sensor 280F for measuring distance. The electronic device 200 may measure the distance by infrared or laser. In some embodiments, taking a picture of a scene, the electronic device 200 may utilize the distance sensor 280F to range for fast focus.
The proximity light sensor 280G may include, for example, a Light Emitting Diode (LED) and a light detector, such as a photodiode. The light emitting diode may be an infrared light emitting diode. The electronic apparatus 200 emits infrared light to the outside through the light emitting diode. The electronic device 200 detects infrared reflected light from nearby objects using a photodiode. When sufficient reflected light is detected, it can be determined that there is an object near the electronic device 200. When insufficient reflected light is detected, the electronic device 200 may determine that there are no objects near the electronic device 200. The electronic device 200 can utilize the proximity sensor 280G to detect that the user holds the electronic device 200 close to the ear for talking, so as to automatically turn off the screen to save power. The proximity light sensor 280G may also be used in a holster mode, a pocket mode automatically unlocks and locks the screen.
The ambient light sensor 280L is used to sense the ambient light level. The electronic device 200 may adaptively adjust the brightness of the display screen 294 based on the perceived ambient light level. The ambient light sensor 280L may also be used to automatically adjust the white balance when taking a picture. The ambient light sensor 280L may also cooperate with the proximity light sensor 280G to detect whether the electronic device 200 is in a pocket to prevent inadvertent contact.
The temperature sensor 280J is used to detect temperature. In some embodiments, the electronic device 200 implements a temperature processing strategy using the temperature detected by the temperature sensor 280J. For example, when the temperature reported by the temperature sensor 280J exceeds the threshold, the electronic device 200 performs a reduction in performance of a processor located near the temperature sensor 280J, so as to reduce power consumption and implement thermal protection. In other embodiments, the electronic device 200 heats the battery 242 when the temperature is below another threshold to avoid the low temperature causing the electronic device 200 to shut down abnormally. In other embodiments, when the temperature is below a further threshold, the electronic device 200 performs a boost on the output voltage of the battery 242 to avoid an abnormal shutdown due to low temperature.
The touch sensor 280K is also referred to as a "touch panel". The touch sensor 280K may be disposed on the display screen 294, and the touch sensor 280K and the display screen 294 form a touch screen, which is also called a "touch screen". The touch sensor 280K is used to detect a touch operation applied thereto or nearby. The touch sensor can communicate the detected touch operation to the application processor to determine the touch event type. Visual output related to touch operations may be provided through the display screen 294. In other embodiments, the touch sensor 280K can be disposed on a surface of the electronic device 200 at a different location than the display screen 294.
The software system of the electronic device may employ a layered architecture, an event-driven architecture, a micro-core architecture, a micro-service architecture, or a cloud architecture. The embodiment of the invention takes an Android system with a layered architecture as an example, and exemplarily illustrates a software structure of an electronic device.
Fig. 3 is a schematic structural diagram of an electronic device according to an embodiment of the present application.
It will be appreciated that the hierarchical architecture divides the software into several layers, each layer having a clear role and division of labor. The layers communicate with each other through a software interface. In some embodiments, the Android system may include an application layer (referred to as an application layer for short) and an application framework layer (referred to as a framework layer for short).
The application layer may include a series of application packages.
As shown in fig. 3, the application package includes system applications. The system application refers to an application that is set in the electronic device before the electronic device is shipped from the factory. Exemplary system applications may include programs for settings, electronic device stewards, cameras, galleries, calendars, music, short messages, and conversations.
The Application framework layer provides an Application Programming Interface (API) and a Programming framework for an Application program of the Application layer. The application framework layer includes a number of predefined functions.
As shown in FIG. 3, the application framework layers may include a window manager, content provider, view system, phone manager, resource manager, notification manager, and the like.
The window manager is used for managing window programs. The window manager can obtain the size of the display screen, judge whether a status bar exists, lock the screen, intercept the screen and the like.
The content provider is used to store and retrieve data and make it accessible to applications. The data may include video, images, audio, calls made and received, browsing history and bookmarks, phone books, etc.
The view system includes visual controls such as controls to display text, controls to display pictures, and the like. The view system may be used to build applications. The display interface may be composed of one or more views. For example, the display interface including the short message notification icon may include a view for displaying text and a view for displaying pictures.
The phone manager is used to provide communication functions of the electronic device 200. Such as management of call status (including on, off, etc.).
The resource manager provides various resources for the application, such as localized strings, icons, pictures, layout files, video files, and the like.
The notification manager enables the application to display notification information in the status bar, can be used to convey notification-type messages, can disappear automatically after a short dwell, and does not require user interaction. Such as a notification manager used to inform download completion, message alerts, etc. The notification manager may also be a notification that appears in the form of a chart or scroll bar text at the top status bar of the system, such as a notification of a background running application, or a notification that appears on the screen in the form of a dialog window. For example, text information is prompted in the status bar, a prompt tone is given, the mobile phone vibrates, and an indicator light flickers.
The Android Runtime environment (Android Runtime) comprises a core library and a virtual machine. The Android Runtime is responsible for scheduling and managing an Android system.
The core library comprises two parts: one part is a function which needs to be called by java language, and the other part is a core library of android.
The application layer and the application framework layer run in a virtual machine. And executing java files of the application program layer and the application program framework layer into a binary file by the virtual machine. The virtual machine is used for performing the functions of object life cycle management, stack management, thread management, safety and exception management, garbage collection and the like.
The system library may include a plurality of functional modules. For example: surface Manager (Surface Manager), Media Libraries (Media Libraries), three-dimensional graphics processing Libraries (e.g., OpenGL ES), two-dimensional graphics engines (e.g., SGL), and the like.
The surface manager is used to manage the display subsystem and provide a fusion of two-dimensional and three-dimensional layers for multiple applications.
The media library supports a variety of commonly used audio, video format playback and recording, and still image files, among others. The media library may support a variety of audio-video encoding formats, such as MPEG4, h.264, MP3, AAC, AMR, JPG, PNG, and the like.
The three-dimensional graphic processing library is used for realizing three-dimensional graphic drawing, image rendering, synthesis, layer processing and the like.
The two-dimensional graphics engine is a two-dimensional drawing engine.
The kernel layer is a layer between hardware and software. The kernel layer at least comprises a display driver, a camera driver, an audio driver, a sensor driver, an information acquisition module, an intelligent decision module, an intelligent analysis module and a model training module.
The information acquisition module is used for acquiring image information acquired by a front camera and a rear camera of the mobile phone, and information acquired by sensors (such as an infrared sensor, a gyroscope sensor, a gravity sensor and the like).
The intelligent decision module is used for determining a target camera mode (such as a front camera or a rear camera) according to at least one of the image information sent by the information acquisition module or the information acquired by the sensor.
The intelligent analysis module is used for determining whether the target camera mode determined by the intelligent decision module is correct or not according to whether the mobile phone detects that the user manually switches the camera or whether the mobile phone detects the shooting operation of the user.
The model training module is used for training the first decision-making model and/or the second decision-making model according to at least one of image information acquired by the front camera and the rear camera of the mobile phone or information acquired by the sensor. The first decision model is a model obtained by training according to information acquired by various sensors in the electronic equipment, image information acquired by a front camera and a rear camera and corresponding camera modes, and is suitable for automatically switching the front camera and the rear camera under the condition that a user senses after a camera APP of the electronic equipment is completely started. The second decision model is a model obtained by training according to information acquired by various sensors in the electronic equipment and corresponding camera modes, and is suitable for automatically switching front and rear cameras in a formal starting stage of a camera APP of the electronic equipment under the condition that a user does not perceive.
The technical solutions involved in the following embodiments can be implemented in the electronic device 200 having the above-described hardware structure and software architecture. The electronic device 200 is taken as a mobile phone as an example, and the present embodiment is exemplarily described below.
In a possible scenario of the embodiment of the present application, the camera APP of the mobile phone has a camera automatic switching mode. After the camera APP is completely started, under the condition that the mobile phone starts the automatic camera switching mode, the intelligent decision module can determine an actual shooting scene (for example, the shooting scene is self-portrait of a user, a shooting landscape, a video and the like) according to information collected by various sensors in the mobile phone and image information collected by the front camera and the rear camera. Furthermore, after the intelligent decision module determines a target camera mode according to an actual shooting scene, the camera driver can drive the front camera and the rear camera to automatically switch according to the target camera mode determined by the intelligent decision module. Therefore, in the process that the user acquires images by adopting the camera APP, automatic switching of the front camera and the rear camera under the condition that the user has perception is realized.
In some embodiments, after the camera APP is completely started, the mobile phone may start the camera automatic switching mode in response to an operation of the user. Exemplarily, fig. 4 is a schematic view of an application scenario of a camera switching method according to an embodiment of the present application. As shown in fig. 4, after the camera APP of the mobile phone is completely started, the mobile phone responds to the click or touch operation of the user on the setting icon of the camera APP, and the display screen of the mobile phone displays the setting interface of the camera APP. The mobile phone can respond to the operation of a user to turn on or turn off the automatic camera switching mode on the setting interface.
It should be explained that, after the camera APP starts the camera automatic switching mode, the mobile phone has a memory function, and after the mobile phone starts the camera APP again, the camera automatic switching mode is opened by default without manual operation of a user.
In the following, the automatic camera switching process will be explained in detail by the mobile phone in conjunction with fig. 5 when the mobile phone starts the automatic camera switching mode. Fig. 5 is a schematic flowchart of a switching method of a camera provided in an embodiment of the present application, and as shown in fig. 5, an automatic switching process of the camera may include the following steps:
step 501, after the starting process of the camera APP of the mobile phone is finished, the mobile phone detects the state of the mobile phone in real time according to the information acquired by the sensor.
Step 502, the mobile phone determines whether it is in a stable state.
In the embodiment of the application, when a user uses the camera APP to shoot, after the mobile phone responds to the operation of the user and opens the camera APP, the intelligent decision module of the mobile phone can detect the state of the mobile phone in real time according to information collected by each sensor in the mobile phone. The state of the mobile phone can be, for example, a stable state or a shaking state of the mobile phone. For example, the mobile phone can detect the state of the mobile phone in real time according to the information acquired by each low-power consumption sensor in the mobile phone. Because the low-power consumption sensors have low power consumption, the low-power consumption sensors in the mobile phone can acquire data in real time after the mobile phone is started. For example, when the mobile phone is in a screen-off state, each low-power consumption sensor in the mobile phone is also in a working state. In addition, in order to save the electric quantity of the mobile phone, a high-power consumption sensor in the mobile phone can acquire data in a bright screen state of the mobile phone, and the acquired data can also be used for judging whether the mobile phone is in a stable state or not.
It can be understood that after the mobile phone is turned on, the low power consumption sensors (for example, the gravity sensor, the gyroscope sensor, and the like) in the mobile phone are always in a normally open state (that is, the low power consumption sensors are always in a working state) to collect data in real time, and after the information acquisition module acquires the data collected by the low power consumption sensors in real time, the data collected by the low power consumption sensors are sent to the intelligent decision module. The intelligent decision module can determine the state of the mobile phone according to the acquired data acquired by the low-power consumption sensors. For example, the intelligent decision module may acquire data acquired by sensors such as a gravity sensor and/or a gyroscope sensor in real time. The intelligent decision module can determine that the state of the mobile phone is a stable state or a shaking state according to the acquired data acquired by the gravity sensor and/or the gyroscope sensor in real time.
In a possible case, after the intelligent decision module acquires the sensor data acquired by the gravity sensor and/or the gyroscope sensor in real time, the intelligent decision module determines that the difference between the sensor data acquired by the gravity sensor and/or the gyroscope sensor in real time is small or even remains unchanged within a fourth preset time period (for example, 2 seconds, 3 seconds, or the like). In this case, the intelligent decision module may determine that the handset is in a steady state. Namely, after the user opens the camera APP, the mobile phone is kept in a stable state to prepare for shooting.
Illustratively, the gyroscope sensor captures the angular velocity of the handset in real time about three axes (i.e., the x, y and z axes) while the camera APP of the handset is in a full start-up phase. After the information acquisition module acquires the angular velocity acquired by the gyroscope sensor in real time, the angular velocity acquired by the gyroscope sensor in real time is sent to the intelligent decision module. After the intelligent decision module acquires data acquired by the gyroscope sensor in real time, the intelligent decision module determines that the difference of the angular speed of the gyroscope sensor in the fourth preset time period is small or even keeps unchanged. In this case, the intelligent decision module may determine that the handset is in a steady state.
In another possible situation, after the intelligent decision module acquires the sensor data acquired by the gravity sensor and/or the gyroscope sensor in real time, the intelligent decision module determines that the difference of the sensor data acquired by the gravity sensor and/or the gyroscope sensor in real time is larger within a fourth preset time period. In this case, the intelligent decision module may determine that the mobile phone is in a shaking state. That is, after the user opens the camera APP, the user may adjust the photographing posture to cause the mobile phone to be in a shaking state.
Illustratively, the gyroscope sensor captures the angular velocity of the handset in real time about three axes (i.e., the x, y and z axes) while the camera APP of the handset is in a full start-up phase. After the information acquisition module acquires the angular velocity acquired by the gyroscope sensor in real time, the angular velocity acquired by the gyroscope sensor in real time is sent to the intelligent decision module. After the intelligent decision module acquires data acquired by the gyroscope sensor in real time, the intelligent decision module determines that the angular speed of the gyroscope sensor has larger difference within a fourth preset time. In this case, the intelligent decision module may determine that the mobile phone is in a shaking state.
It should be explained that the above-mentioned intelligent decision module determines whether the mobile phone is in a stable state according to the data of the sensor collected by the gyroscope sensor in real time only as an exemplary description. In the embodiment of the application, the intelligent decision module can also determine whether the mobile phone is in a stable state according to the sensor data acquired by the gravity sensor in real time. In order to improve the accuracy of determining the state of the mobile phone, the intelligent decision module can also determine whether the mobile phone is in a stable state according to sensor data acquired by the gravity sensor and the gyroscope sensor simultaneously. Of course, the mobile phone may also determine whether the mobile phone is in a stable state by using data acquired by other sensors in real time, which is not limited herein.
If the mobile phone detects that the mobile phone is in a stable state in step 502, step 503 is executed, otherwise, step 501 is executed.
Step 503, the intelligent decision module of the mobile phone performs camera mode recognition according to the sensor information and the image information collected by the front and rear cameras.
In the embodiment of the application, the intelligent decision module determines that the mobile phone is in a stable state, that is, the user is ready to use the mobile phone for shooting. In this case, the intelligent decision module may determine the target camera mode according to at least one of information such as sensor information of the mobile phone, image information acquired by front and rear cameras, and the like.
In some embodiments, the intelligent decision module may input at least one of the acquired sensor information of the mobile phone, image information acquired by front and rear cameras, and the like into the first decision model, and the intelligent decision module determines the target camera mode according to an output of the first decision model.
The first decision model is obtained by training the model training module by adopting at least one of information collected by each sensor of the mobile phone, image information collected by front and back cameras and the like, and a target camera mode can be accurately identified.
In the embodiment of the application, when the camera APP of the mobile phone is in the complete start stage, the information acquisition module may acquire information such as image information obtained by scanning of the front camera and the rear camera when the mobile phone is shot by the front camera or the rear camera, a distance between the mobile phone and a face, and a horizontal and vertical screen state of the mobile phone. The model training module trains the first decision model according to at least one of information acquired by the information acquisition module from a sensor of the mobile phone and information such as image information acquired by front and rear cameras in a shooting scene of the front camera or the rear camera.
When a camera APP of the mobile phone is in a complete starting stage, the front camera and the rear camera scan corresponding shooting areas in real time. Meanwhile, various sensors (such as an infrared sensor, a gyroscope sensor, a gravity sensor and the like) in the mobile phone also collect information. The information acquisition module can acquire image information acquired by the front camera and the rear camera and sensor information acquired by each sensor. The information acquisition module sends the acquired image information and the sensor information to the model training module, and the model training module can train the first decision-making model according to at least one of the image information and the sensor information acquired by the front camera and the rear camera.
It can be understood that when the camera APP of the mobile phone is in the complete start stage, and the user uses the front camera or the rear camera to shoot, the face information collected by the front camera is different, the distance between the mobile phone and the face of the user is also different, and the horizontal and vertical screen states of the mobile phone are also different. The model training module can train the first decision model according to face information scanned by the front camera, the distance between the mobile phone and the face, the horizontal and vertical screen states of the mobile phone and the camera mode. Therefore, after the model training module trains the first decision-making model, the first decision-making model can accurately identify the target camera mode according to at least one of the input image information acquired by the front camera and the rear camera or the sensor information acquired by the sensor. The specific training process of the first decision model is referred to the following detailed explanation process, which is not described here.
When a camera APP of a mobile phone is in a full-boot stage, a front-facing camera (e.g., a structured light camera) may scan face information of a user to obtain face information (e.g., full face information or partial face information). The information acquisition module sends the face information obtained by scanning of the front camera to the intelligent decision module, and the intelligent decision module identifies the face information of the user in the face information. Such as the user's facial expression, whether the eyes are looking at the screen, whether there are scissors gestures and whether the user's complete face is included, etc. Meanwhile, the intelligent decision-making module determines the distance between the mobile phone and the face according to face information scanned by the front camera or information collected by the infrared sensor and the structured light camera. The intelligent decision-making module can also determine the horizontal and vertical screen states of the mobile phone according to information acquired by sensors such as a gyroscope sensor and/or a gravity sensor. For example, when the mobile phone is in a rear camera mode, the mobile phone may be in a landscape state; when the mobile phone is in the front camera mode, the mobile phone may be in a vertical screen state. And then, the intelligent decision module identifies a target camera mode according to at least one of the face information scanned by the front camera, the distance between the mobile phone and the face and the horizontal and vertical screen states of the mobile phone.
The first decision model learns the mapping relation between the image information scanned by the front-mounted camera and the rear-mounted camera, the information collected by the sensor and the camera mode. Therefore, when the camera APP of the mobile phone is in the complete start stage and the mobile phone is in a stable state, the intelligent decision module inputs information acquired by each sensor of the mobile phone and image information scanned by the front and rear cameras into the first decision model, and then the first decision model can identify the target camera mode.
In a possible scenario, if the image information scanned by the front camera of the mobile phone may include calm user expression and no gesture, the human eye may stare at the screen or the human eye may switch between a distant scene and an eye screen many times. The distance between the mobile phone and the face can be preliminarily measured by the mobile phone through information collected by the infrared sensor or the structured light camera. Meanwhile, because the distance between the mobile phone and the face is short, the face information of the user in the image information scanned by the front camera may not be complete, for example, the image information scanned by the front camera may include partial face information, such as only nose and mouth or only eyes and nose. The intelligent decision-making module can also determine the transverse and vertical screen states of the mobile phone according to sensors such as a gyroscope sensor, and the like, such as a transverse screen commonly used when a user adopts a rear camera. The intelligent decision module can identify that the user wants to adopt the rear camera mode at present according to at least one of image information scanned by a front camera and a rear camera of the mobile phone, the distance between the mobile phone and the face and the horizontal and vertical screen states of the mobile phone. The intelligent decision module inputs at least one of image information scanned by a front camera and a rear camera of the mobile phone and information acquired by a low-power consumption sensor of the mobile phone into a first decision model, and the first decision model can identify that a target camera mode is a rear camera.
In another possible scenario, if the image information scanned by the front camera of the mobile phone may include rich user expressions and gesture actions, the eyes of a person may stare at the screen. The distance between the mobile phone and the face can be preliminarily measured by the mobile phone through information collected by the infrared sensor or the structured light camera. Meanwhile, as the distance between the mobile phone and the face is long, the image information scanned by the front camera comprises the complete face information of the user. The intelligent decision-making module can determine the horizontal and vertical screen states of the mobile phone according to sensors such as a gyroscope sensor, and the like, and the user frequently uses the vertical screen when taking a self-timer by adopting a front camera. The intelligent decision module can identify that the user wants to adopt the front camera mode at present by combining the image information scanned by the front camera and the rear camera of the mobile phone, the distance between the mobile phone and the face and the horizontal and vertical screen states of the mobile phone. The intelligent decision module inputs at least one of image information scanned by a front camera and a rear camera of the mobile phone and information acquired by a low-power consumption sensor of the mobile phone into a first decision model, and the first decision model can identify that a target camera mode is the front camera.
In another possible scenario, when a camera APP of a mobile phone is in a complete start stage, after the information acquisition module acquires image information scanned by the rear camera, the image information scanned by the rear camera is sent to the first decision model. The first decision model can determine a camera mode according to image information scanned by the rear camera. For example, assuming that the first decision model determines that the image information scanned by the rear camera includes preset information, such as a two-dimensional code or a barcode, the first decision model may identify the target camera mode as the rear camera.
In step 504, the intelligent decision module determines whether an exclusive condition exists.
The exclusive condition refers to the condition that the front camera of the mobile phone is blocked. For example, an intelligent decision module of the mobile phone determines that a human hand is blocked within a preset range of a front camera according to data acquired by a sensor (e.g., a structured light sensor or an infrared sensor) or image information scanned by the front camera. In this case, the intelligent decision module determines that an exclusive condition exists.
If the intelligent decision module determines in step 504 that an exclusive condition exists, then step 505 is performed, otherwise step 506 is performed.
And 505, the intelligent decision module corrects the camera mode output by the first decision model under the exclusive condition to determine the corrected camera mode.
In the embodiment of the application, if the intelligent decision module of the mobile phone determines that the hand is blocked in the preset range of the front camera according to the data collected by the sensor or the image information scanned by the front camera. At this time, the intelligent decision module can determine that the target camera mode of the mobile phone is a rear camera. In this case, when the camera mode output by the first decision model is a front camera, the intelligent decision module may modify the camera mode output by the first decision model (i.e., the front camera) to determine a modified camera mode (i.e., a rear camera).
In some embodiments, the intelligent decision module determines that the camera mode output by the first decision model is not the camera mode desired by the user under exclusive conditions. For example, the intelligent decision module identifies that the camera mode to be switched by the mobile phone is a front-facing camera according to at least one of image information scanned by front and back cameras and information acquired by a low-power consumption sensor of the mobile phone. However, the intelligent decision module determines that the human hand is blocked in the preset range of the front camera according to the image information scanned by the front camera, and the intelligent decision module modifies the camera mode from the front camera to the rear camera.
Exemplarily, fig. 6 is a third application scenario schematic diagram of a camera switching method provided in the embodiment of the present application. As shown in fig. 6, at the complete start stage of the camera APP of the mobile phone, it is assumed that the intelligent decision module inputs information acquired by the sensor of the mobile phone in fig. 6 and images acquired by front and rear cameras into the first decision model, and the intelligent decision module determines that the camera mode is the front camera according to the output of the first decision model. I.e. the camera mode in the first image in fig. 6. As shown in the second diagram of fig. 6, it is assumed that the intelligent decision module determines that the front camera has a hand occlusion according to the image information collected by the front camera. Then, as shown in the third diagram in fig. 6, the intelligent decision module corrects the camera mode, and corrects the camera mode from the front camera to the rear camera. The camera drives and controls the front camera to be closed and controls the rear camera to be started. Or under the condition that the front camera and the rear camera are both started, the mobile phone switches the image information acquired by the camera APP from the information acquired by the front camera to the information acquired by the rear camera.
And step 506, according to the camera mode, the camera drives and controls the front camera and the rear camera to complete automatic switching.
In the embodiment of the application, after the intelligent decision module determines the camera mode, the camera driver drives the corresponding camera to be started according to the camera mode determined by the intelligent decision module. At this moment, a first preview interface is displayed in a display screen of the mobile phone, wherein the first preview interface comprises image information acquired by a camera of the camera APP in a camera mode. For example, if the starting process of the camera APP is finished, the mobile phone starts a rear camera, and the camera mode determined by the intelligent decision module is a front camera. A first preview interface displayed in a display screen of the mobile phone comprises image information collected by a front camera of a camera APP. Under the condition, the camera drive can control the rear camera to be closed and the front camera to be opened, so that the automatic switching of the front camera and the rear camera is realized.
Step 507, the mobile phone prompts the user that the camera is automatically switched.
In the embodiment of the application, after the front camera and the rear camera are automatically switched under the drive control of the cameras, the mobile phone can prompt a user that the front camera and the rear camera are automatically switched.
For example, after the front and rear cameras are automatically switched under the drive control of the cameras, the mobile phone may prompt the user that the front and rear cameras have been automatically switched by displaying the first prompt information. For example, in the first diagram in fig. 7, assuming that the camera of the mobile phone is switched from the front camera to the rear camera, the display screen of the mobile phone may display a prompt message that "the camera has been switched to the rear camera" to remind the user that the front camera and the rear camera have completed automatic switching.
It should be explained that the mobile phone prompts the user by displaying the prompt information only as an exemplary expression, and the mobile phone may further control the photographing button or the front-facing camera to flash for a preset time, or control the buzzer to emit a buzzing sound (for example, a buzzing sound of 2 seconds), or control the display screen to alternately turn on and off, and the like, which is not limited herein.
Therefore, when the camera APP starts the automatic camera switching mode, after the camera APP is started, the intelligent decision module identifies the target camera mode, and the camera drive controls the front camera and the rear camera to be automatically switched according to the target camera mode. From this, realized the in-process that camera APP shot, the purpose of camera around the cell-phone is according to the shooting scene automatic switch of reality, improved user's shooting experience.
The training process of the first decision model is explained in detail below. Fig. 8 is a schematic diagram of a training process of a first decision model according to an embodiment of the present disclosure. As shown in fig. 8, the training process of the first decision model may include a training process in an off-line state, a parameter adjustment process for the first decision model, and a training process in an on-line state. The training process in the offline state may refer to a training process performed on the first decision model before the mobile phone leaves the factory. The parameter adjustment process for the first decision model may refer to a process of adjusting parameters of the first decision model after the mobile phone leaves the factory and the mobile phone is in an idle state. The training process in the online state may refer to a training process performed on the first decision model after the mobile phone leaves a factory and during a process in which the user uses the mobile phone. In one or more of the three processes, the model training module may train the first decision model or adjust parameters of the first decision model using information of the sensor, image information of front and rear cameras, and corresponding camera modes.
For example, as shown in fig. 8, when the mobile phone is in an offline state before shipping, the mobile phone obtains information of each sensor and image information of front and rear cameras, which are preset manually. Including but not limited to image sensors, gyroscope sensors, proximity light sensors, and structured light sensors, among others. And the model training module of the mobile phone trains the general decision model by adopting preset information of each sensor, image information of front and rear cameras and corresponding camera modes to obtain a trained first decision model.
When the mobile phone is in an idle state, the mobile phone can acquire information of each sensor and image information of front and back cameras when a mobile phone user uses the camera APP to acquire images, and a camera mode adopted by the user when shooting. The model training module of the mobile phone can train the first decision model by adopting the information of the sensor, the image information of the front camera and the back camera and the corresponding camera modes, which are collected by the mobile phone, when the mobile phone is in an idle state. Therefore, the accuracy of the first decision model for identifying the target camera mode is improved by adjusting the parameters of the first decision model by the mobile phone.
The camera APP of the mobile phone is in a complete starting stage, a user adopts the camera APP to shoot, and the model training module can perform online training on the first decision model according to information collected by a sensor in the mobile phone, image information of front and back cameras and a corresponding camera mode. The training process of the first decision model in the online state is explained in detail below with reference to fig. 9. Fig. 9 is a flowchart illustrating another training process of a first decision model according to an embodiment of the present disclosure. As shown in fig. 9, the training process may include the following steps:
step 901, the mobile phone prompts the user that the front camera and the rear camera have been automatically switched.
In the embodiment of the application, after the front camera and the rear camera are automatically switched under the drive control of the cameras, the mobile phone can prompt a user that the front camera and the rear camera are automatically switched. The specific prompting manner is similar to that in step 507, and is not described herein again.
Step 902, the mobile phone determines whether a first control instruction of the user is received within a first preset time period.
The first control instruction is used for controlling switching of the front camera and the rear camera.
In the embodiment of the application, in the process that the mobile phone prompts the user that the automatic switching of the front and rear cameras is completed, the mobile phone determines whether a first control instruction of the user is received within a first preset time (for example, 1 second or 2 seconds) after the automatic switching of the cameras. For example, the first control instruction may be a manual operation by a user (e.g., a quick screen sliding operation or a click or touch camera switching button operation, etc.). And the intelligent decision module determines whether the determined camera mode is correct or not according to whether the mobile phone receives a first control instruction of the user within a first preset time length. That is to say, after the mobile phone prompts the user that the front and rear cameras are automatically switched, if the current camera mode of the camera APP cannot meet the user requirement, the mobile phone continues to switch the front and rear cameras again in response to the manual operation of the user. If the current camera mode of the camera APP is the camera mode required by the user, the mobile phone does not need to switch the front camera and the rear camera, and the user can shoot in the current camera mode.
In a possible situation, in the process that the mobile phone prompts the user to complete automatic switching of the front camera and the rear camera, the mobile phone receives a first control instruction of the user within a first preset time length. For example, as shown in fig. 7, it is assumed that, in the process that the mobile phone in the first picture in fig. 7 displays the prompt message "the camera has switched to the rear camera", the user quickly slides up the screen, as shown in the second picture in fig. 7. If the mobile phone detects a control instruction of the user for fast sliding the screen, the intelligent analysis module determines that the camera mode identified at this time is wrong. Under the condition, the mobile phone reminds the user that the camera is switched from the rear camera to the front camera in a mode of displaying prompt information. As shown in the third picture in fig. 7, a prompt message "the camera has been switched to the rear camera" is displayed on the display screen of the mobile phone to prompt the user that the camera has been switched from the rear camera to the front camera.
In another possible situation, in the process that the mobile phone prompts the user to complete automatic switching of the front camera and the rear camera, the mobile phone does not receive the first control instruction of the user within the first preset time. For example, assuming that the mobile phone does not detect any operation of the display screen by the user in the process of flashing the photographing button of the mobile phone, the intelligent analysis module determines that the recognized camera mode is correct at this time.
It should be explained that the first preset time duration may be less than or equal to the preset time duration for the mobile phone to prompt the user. For example, the mobile phone controls the photographing button to flash for 3 seconds to prompt the user that the front camera and the rear camera are automatically switched, and the mobile phone can detect whether the user performs manual operation within 2 seconds or 3 seconds of flashing of the photographing button.
If the mobile phone receives the first control instruction of the user within the first preset time duration in step 902, then step 903 is executed, otherwise step 904 is executed.
Step 903, the intelligent analysis module determines that the decision is wrong, and the detection information does not participate in iterative training of model training.
In the embodiment of the application, when the intelligent analysis module determines that the decision of the camera mode determined by the intelligent decision module is wrong, the image information scanned by the camera before and after the time and the information acquired by the low-power consumption sensor do not participate in the iterative training process of the first decision model.
In step 904, the mobile phone determines whether a second control command of the user to the photo button is received.
The second control instruction is used for controlling the photographing button to photograph.
In a possible case, after the automatic switching of the front camera and the rear camera of the camera APP is completed, the mobile phone receives a second control instruction of the photographing button from the user, that is, the user operates the photographing button (for example, clicks or touches the photographing button), and the mobile phone controls the camera to acquire an image in response to the second control instruction.
Under another possible condition, after the automatic switching of the front camera and the rear camera of the camera APP is completed, the mobile phone does not receive a second control instruction of the photographing button from the user, that is, the user does not operate the photographing button. In this case, the detection information does not participate in the iterative training of the model training.
If the mobile phone receives a second control instruction of the user to the photo button in step 904, step 905 is executed, otherwise step 903 is executed.
Step 905, the intelligent analysis module determines that the decision is correct, and the detection information participates in iterative training of model training.
In the embodiment of the application, the intelligent analysis module determines that the decision of the intelligent decision module is correct, and the image information scanned by the front camera and the back camera and the information acquired by the low-power consumption sensor, which are acquired by the information acquisition module, can participate in the iterative training process of the model training of the first decision model. Therefore, the information of the low-power consumption sensor is used for carrying out iterative training on the first decision model, and the model accuracy of the first decision model is improved.
Illustratively, after the starting process of the camera APP is finished, the intelligent decision module identifies a camera mode required by a user according to image information acquired by front and back cameras and/or information acquired by a low-power consumption sensor, and the mobile phone detects that the user takes a picture in the camera mode. At this moment, the intelligent analysis module determines that the intelligent decision module identifies that the camera mode required by the user is correct according to the image information acquired by the front and rear cameras and/or the information acquired by the low-power consumption sensor acquired after the starting process of the camera APP is finished. The intelligent analysis module determines image information acquired by front and back cameras and/or information acquired by a low-power consumption sensor after the camera APP is started, and can participate in an iterative training process of the first decision model. The model training module inputs image information acquired by front and back cameras and/or information acquired by a low-power consumption sensor acquired after the starting process of the camera APP is finished into a first decision model. And then, the model training module adjusts parameters of the first decision model according to the difference between the camera mode output by the first decision model and the camera mode required by the user at this time, so that the camera mode output by the first decision model is the camera mode required by the user.
In a possible scenario of the embodiment of the application, when the mobile phone first starts the automatic camera switching mode, the mobile phone may prompt the user to train the first decision model in the automatic camera switching mode, so that the first decision model learns the correspondence between the information acquired by each sensor, the image information acquired by the front and rear cameras, and the camera mode.
Returning to fig. 4, as shown in fig. 4, after the mobile phone starts the camera auto-switching mode at the setting interface of the camera APP in response to the operation of the user, the mobile phone may display a second prompt message on the display interface of the display screen, for example, display a "prompt: whether to train for automatic camera switching. The mobile phone responds to the operation of the user on the 'yes' control, and the camera APP prompts the user to acquire images under the front camera and/or the rear camera. After the automatic camera switching mode training is completed, the mobile phone can display prompt information on the display screen to remind a user that the training is completed. For example, in fig. 4, the mobile phone displays a prompt message "prompt: the training is complete ". In the embodiment of the application, the mobile phone can prompt the user to shoot a plurality of images under the condition that the front camera and/or the rear camera are started. In the process that the camera APP acquires images by using the front camera or the rear camera, the information acquisition module of the mobile phone acquires information acquired by each sensor, for example, information acquired by sensors such as a structured light sensor, an infrared sensor, a gravity sensor or a gyroscope sensor. The model training module trains the first decision model according to the information acquired by the sensors and the image information acquired by the front camera and the rear camera acquired by the information acquisition module.
In some embodiments, after the mobile phone is turned on, when the mobile phone detects that the camera APP is in a formal start stage, the mobile phone may determine to turn on the front camera or the rear camera according to information collected by a connected external device (e.g., a selfie stick, a cradle head, or the like) and/or a sensor related to shooting. The sensor may be a low power consumption sensor, such as an image sensor, a structured light sensor, a proximity light sensor, or a gyroscope sensor. Therefore, the mobile phone realizes automatic switching of the front camera and the rear camera in the user non-perception state at the formal starting stage of the camera APP, and shooting experience of the user is improved.
The above-mentioned automatic camera switching method will be described in detail with reference to fig. 10. Exemplarily, fig. 10 is a schematic flowchart of another camera switching method provided in the embodiment of the present application. As shown in fig. 10, the automatic camera switching method may include the steps of:
step 1001, after the mobile phone is started, the mobile phone detects that the camera APP is in a formal starting stage.
In step 1002, the mobile phone detects whether the mobile phone is connected to an external device related to shooting.
When the mobile phone detects that the camera APP is in a formal starting stage, the mobile phone can detect whether the mobile phone is connected with external equipment related to shooting, and then the mobile phone can determine to start the front camera or the rear camera according to the connected external equipment.
If the cellular phone detects that the cellular phone is connected to the external device related to photographing in step S1002, step S1003 is executed, otherwise step S1004 is executed.
And 1003, switching the front camera and the rear camera by the mobile phone according to the connection condition of the external equipment.
In a possible scenario, the mobile phone detects that the camera APP is in a formal start stage, and the mobile phone is connected to a self-timer device (such as a self-timer stick). In this case, the user may use the camera APP to take a self-timer, and the camera driver may control the front camera to start. For example, when the camera APP is in the formal start stage, if the rear camera of the camera APP is started, the mobile phone detects that the mobile phone is connected with the self-timer, the camera drives and controls the rear camera to be closed, and controls the front camera to be opened, so that the automatic switching of the front camera and the rear camera is completed before the camera APP is completely started and in a user-unaware state.
In another possible scenario, the mobile phone detects that the camera APP is in a formal start stage, and the mobile phone is connected with a support device (e.g., a cradle head or a tripod). In this case, the user may use the camera APP to photograph a landscape, a person, or the like, and the camera drive may control the rear camera to start. For example, when camera APP is in the formal start stage, if the front camera of camera APP starts, the mobile phone detects that the mobile phone is connected with the cradle head, the front camera is controlled to be closed by the camera drive, and the rear camera is controlled to be opened, so that the automatic switching of the front camera and the rear camera is completed before the camera APP starts and in a user non-sensing state.
And 1004, the information acquisition module acquires state information within a second preset time before the camera APP is formally started.
The state information may include a current state of the mobile phone (for example, the mobile phone is in a landscape state, a portrait state, or the like), a state between the mobile phone and the user (for example, the mobile phone is closer to the user, or the like), or a state of the user (for example, the user has a rich expression and has a gesture, or the like).
In some embodiments, after the mobile phone is powered on, each low-power sensor disposed in the mobile phone is in a normally open state (i.e., the low-power sensor is always in a working state) to collect data. For example, after the mobile phone is turned on, an image sensor in the mobile phone collects image information, a gyroscope sensor collects angular velocity information, a proximity light sensor collects distance information between a nearby object and the mobile phone, or a structured light sensor collects information. The mobile phone can determine state information within a second preset time before the camera APP is formally started according to the information acquired by each sensor. If the mobile phone detects that the mobile phone is not connected with the external device related to shooting, the information acquisition module quickly acquires the state information within a second preset time (for example, the second preset time may be 2 seconds, 3 seconds, 5 seconds, or the like) before the camera APP is formally started.
It can be understood that after the mobile phone is started, each low-power consumption sensor arranged in the mobile phone is always in a working state to collect data in real time, and the collected data is stored in the memory. Therefore, when the camera APP is in the formal start stage, the information acquisition module can quickly acquire the information acquired by each low-power-consumption sensor within the second preset time before the formal start of the camera APP from the memory, so that the mobile phone determines the corresponding state information according to the information acquired by each low-power-consumption sensor within the second preset time before the formal start of the camera APP.
As an example, when the camera APP is in the formal start stage, the information obtaining module may quickly obtain, from the memory, image information collected by an image sensor 3 seconds before the formal start of the camera APP, angular velocity information collected by a gyroscope sensor, distance information between a nearby object collected by a proximity light sensor and a mobile phone, or information collected by a structured light sensor, and the like.
Step 1005, the intelligent decision module identifies a target camera mode according to information acquired by the sensor, and completes automatic camera switching before the camera APP is completely started.
In the embodiment of the application, when the camera APP is in the formal start stage, after the information acquisition module acquires the information acquired by each sensor within the second preset time before the formal start of the camera APP, the information acquisition module sends the information acquired by each sensor to the intelligent decision module. And after receiving the information acquired by each sensor, the intelligent decision module inputs the information acquired by the sensors into the second decision model, and determines a target camera mode according to the output of the second decision model. It can be understood that the target camera mode is a camera mode required by a user to shoot when the camera APP is started. Furthermore, the camera driver can complete automatic switching of the camera according to the target camera mode identified by the intelligent decision module when the camera APP is in the formal starting stage. And then, the display screen of the mobile phone displays a second preview interface after the camera APP is completely started. And the second preview interface comprises image information acquired by a camera of the camera APP in the target camera mode.
It should be explained that the second decision model is obtained by training information acquired by each sensor, and can accurately identify the target camera mode according to the information of the sensor input by the intelligent decision module.
In the embodiment of the application, the camera APP is in a formal starting stage, the intelligent decision module inputs information acquired by at least one sensor into a trained second decision model after receiving the information acquired by at least one sensor, and the intelligent decision module determines a target camera mode according to the output of the second decision model.
In a possible scenario, it is assumed that the information of the sensor acquired by the intelligent decision module includes information such as distance information between the mobile phone and the human body acquired by a time of flight (TOF) camera, angle information acquired by a gyroscope sensor when the mobile phone is shot, and image information acquired by an image sensor. If the mobile phone determines that the distance between the mobile phone and the human body is long according to the distance information between the mobile phone and the human body acquired by the TOF, the shooting state of the mobile phone is determined by the angle information acquired by the gyroscope sensor during the shooting of the mobile phone, the image information acquired by the image sensor comprises a face image and the like, the intelligent decision-making module inputs the information of the sensor into a second decision-making model, and the second decision-making model determines that the target camera mode is a front camera. Under the condition, the camera drive can control the rear camera to be closed and the front camera to be opened before the camera APP is completely started. From this, after the target camera mode was discerned to intelligence decision-making module, for example, the user probably used this camera APP to carry out the auto heterodyne, and the cell-phone is automatic to be switched over the camera to the target camera mode, before having realized camera APP and starting, accomplished the automatic switch-over of camera under the user does not have the perception state.
In another possible scenario, it is assumed that the information of the sensor acquired by the intelligent decision module includes distance information between the mobile phone and the human body acquired by the TOF, angle information acquired by the gyroscope sensor when the mobile phone is used for shooting, image information acquired by the image sensor, and the like. If the mobile phone determines that the distance between the mobile phone and the human body is short according to the distance information between the mobile phone and the human body acquired by the TOF, the shooting of the mobile phone is in a horizontal screen state according to the angle information acquired by the gyroscope sensor during the shooting of the mobile phone, the image information acquired by the image sensor does not include a face image and the like, the intelligent decision-making module inputs the information of the sensor into a second decision-making model, and the second decision-making model determines that the target camera mode is a rear camera. Under the condition, the camera drive can be in the formal start-up stage at camera APP, and the control front camera is closed, and the control rear camera is opened. From this, after the intelligent decision-making module discerned target camera mode, for example, the user probably used this camera APP to shoot scenery or object etc. the cell-phone switches the camera to target camera mode automatically, before having realized camera APP and having started completely, accomplishes the automatic switch-over of camera under the user does not have the perception state. The training process of the second decision model is explained in detail below with reference to fig. 11. Fig. 11 is a schematic diagram illustrating a training process of a second decision model according to an embodiment of the present application. As shown in fig. 11, the training process of the second decision model may include a training process in an offline state, a parameter adjustment process for the second decision model, and a training process in an online state. The training process in the offline state may refer to a training process performed on the second decision model before the mobile phone leaves the factory. The parameter adjustment process for the second decision model may refer to a process of adjusting parameters of the second decision model after the mobile phone leaves the factory and the mobile phone is in an idle state. The training process in the online state may refer to a training process performed on the second decision model after the mobile phone leaves the factory and during the process of using the mobile phone by the user. In one or more of the three processes, the model training module may train the second decision model or adjust parameters of the second decision model by using information of the sensor, image information of front and rear cameras, and corresponding camera modes.
As shown in fig. 11, the mobile phone is in an off-line state before shipping, and the mobile phone can acquire information of each sensor preset manually. Including but not limited to image sensors, gyroscope sensors, proximity light sensors, and structured light sensors, among others. And the model training module of the mobile phone trains the general decision model by adopting the preset information of each sensor and the corresponding camera mode to obtain a trained second decision model. The camera mode comprises a front camera mode and a rear camera mode.
When the mobile phone is in an idle state (for example, a late-night sleep state or a charging state and the like), the information acquisition module can acquire the information acquired by each sensor stored in the memory and the corresponding camera mode used by the user. The information acquisition module sends the information acquired by each sensor and the corresponding camera mode to the model training module, and the model training module adjusts the model parameters of the second decision model by adopting the information of each sensor and the corresponding camera mode. Thus, the accuracy of the second decision model in identifying the target camera mode is improved.
After the camera APP completes automatic camera switching in the formal starting stage, the intelligent analysis module can determine whether the camera mode determined by the second decision model according to the information collected by the sensors is correct according to whether the camera is switched by a user within a third preset time after the automatic camera switching is completed so as to determine whether iterative training is performed on the second decision model in an online state by adopting the information collected by the sensors. The training process of the second decision model in the online state is explained in detail below with reference to fig. 12. Fig. 12 is a flowchart illustrating a training process of a second decision model in an online state according to an embodiment of the present application. As shown in fig. 12, the training process may include the following steps:
in step 1201, after the camera APP is completely started, the mobile phone determines whether a third control instruction of the user is received within a third preset time period. And the third control instruction is used for controlling the camera to switch.
In the embodiment of the application, the camera driver drives the corresponding camera to be opened according to the target camera mode identified by the intelligent decision module, and the camera is opened before the camera APP is completely started. The mobile phone determines whether a third control instruction of the user is received within a third preset time length (wherein the third preset time length may be 0.5 second, 1 second or 2 seconds) after the camera APP is started. Therefore, the mobile phone determines whether the target camera mode determined by the intelligent decision module is correct or not according to whether the camera is switched by the user within the third preset time length of the starting of the camera APP. It can be understood that, after the camera APP is completely started, if the mobile phone receives a third control instruction of the user to the camera within a third preset time period, for example, if the user manually switches the camera, it indicates that the target camera mode determined by the intelligent decision module may not be the camera actually used by the user for shooting. If the mobile phone does not receive a third control instruction of the user to the camera within a third preset time length, it indicates that the target camera mode determined by the intelligent decision module may be the camera actually used by the user for shooting.
If the mobile phone receives a third control instruction of the user in a third preset time length after the camera APP is completely started in step 1201, step 1202 is executed, otherwise step 1203 is executed.
Step 1202, the intelligent analysis module determines that the decision is wrong at this time, and information acquired by each sensor does not participate in the iterative training process of model training.
Under a possible scene, after the camera APP is completely started, the mobile phone detects that the camera APP is started and the user manually switches the camera mode within a third preset time. In this case, the intelligent analysis module determines that the decision of the intelligent decision module is wrong, and the information of each sensor acquired by the information acquisition module within a second preset time before the camera APP is started does not participate in the iterative training process of model training of the second decision model. Therefore, the problem that the accuracy of the model is reduced due to the fact that the second decision model is trained by using the information of the sensors is solved.
Illustratively, in the process of starting the camera APP, the intelligent decision module identifies that the target camera mode is the front camera according to information of the sensor acquired within a second preset time before the camera APP is started. However, after the camera APP is completely started, the mobile phone detects that the user switches the camera from the front camera to the rear camera immediately or manually within a preset time. Under the condition, the intelligent analysis module determines that the intelligent decision module recognizes that the target camera mode is a decision error of the front camera according to the information of the sensor acquired within the second preset time before the camera APP is started, and determines that the information of the low-power-consumption sensor acquired within the second preset time before the camera APP is started does not participate in the iterative training process of the second decision module.
In step 1203, the intelligent analysis module determines that the decision is correct, and information acquired by each sensor participates in an iterative training process of model training.
In step 1204, the model training module performs iterative training on the second decision model by using the information of each sensor within a second preset time period before the camera APP is formally started.
In another possible scene, after the camera APP is started, the mobile phone does not detect that the camera APP is completely started, and then the camera is manually switched by a user within a third preset time. Under the condition, the intelligent analysis module determines that the decision of the intelligent decision module is correct, and the information of each sensor acquired by the information acquisition module in a second preset time before the camera APP is started can participate in an iterative training process of model training of the second decision module. Therefore, the information of the sensor is used for carrying out iterative training on the second decision model, and the model accuracy of the second decision model is improved.
For example, as shown in fig. 11, assuming that the camera APP is in a formal start stage, the intelligent decision module identifies that the target camera mode is a front-facing camera according to information of the sensor acquired by a second preset time length before the start of the camera APP. However, within a third preset time (for example, 3 seconds) after the camera APP is completely started, the mobile phone does not detect that the user manually switches the mode of the camera from the front camera to the rear camera. Under the condition, the intelligent analysis module determines that the intelligent decision module obtains the information of the sensor according to the second preset time length before the starting of the camera APP, and the identified target camera mode is the correct decision of the front camera. The intelligent analysis module determines that the information of the sensor acquired at a second preset time before the starting of the camera APP can participate in the iterative training process of the second decision model.
It is to be understood that the training process of the second decision model by the model training module is a lifetime learning process. Namely, when the mobile phone is in an idle state, the model training module trains the second decision model by using the information of the low-power consumption sensor stored in the memory and the operation behavior of the user on the camera. After the camera APP is completely started, when the intelligent analysis module determines that the information of the sensor acquired in the second preset time before the camera APP is started can participate in the iterative training process of the second decision-making model, the model training module carries out iterative training on the second decision-making model by adopting the information of each sensor in the second preset time before the camera APP is formally started. Therefore, the model training module continuously trains and updates the second decision model in the whole life cycle of the mobile phone, and the accuracy of the second decision model for identifying the camera mode according to the input information of the sensor is improved.
Since the second decision model has learned the correspondence between the information of the sensor of the handset and the camera mode. In a scenario where a user changes a handset, the second decision model and the decision logic in the handset may be migrated to a new handset. The new mobile phone can continue iterative optimization on the second decision-making model according to the acquired information of the sensor, so that when the camera APP is in a formal starting stage, the intelligent decision-making module can more accurately identify a target camera mode.
Under a possible scene, under the condition that the mobile phone does not start the automatic camera switching mode, in the starting process of the camera APP, the mobile phone can determine a target camera mode according to information collected by connected external equipment and/or a sensor related to shooting, and when the camera APP is in a formal starting stage, the automatic switching of front and rear cameras in a user non-perception state is realized. Under the condition, after the camera APP is completely started, when a user switches the cameras according to an actual shooting scene, the mobile phone responds to the operation of the user to switch the front camera and the rear camera.
Under another possible scene, under the condition that the mobile phone starts the automatic camera switching mode, after the camera APP is completely started, the intelligent decision module identifies the camera mode required by the actual shooting scene, and the front camera and the rear camera are automatically switched under the drive control of the cameras.
In another possible scenario, when the mobile phone starts the automatic camera switching mode, when the camera APP is in the formal start stage, the mobile phone may determine the target camera mode according to information collected by the connected external device and/or the sensor related to shooting, and when the camera APP is in the formal start stage, switch the camera to the target camera mode. After the camera APP is completely started, after the intelligent decision module identifies a camera mode required by an actual shooting scene, the cameras are automatically switched before and after the camera drive control.
In summary, in the embodiment of the present application, before the camera APP is completely started, after the intelligent decision module identifies the target camera mode, the camera driver controls the front and rear cameras to automatically switch according to the target camera mode in a user-unaware state. From this, when camera APP is in formal start-up stage, the cell-phone has accomplished the automatic switch-over of camera under user's noninductive state, has avoided the in-process that the user adopted camera APP to shoot, needs the problem that the camera was frequently switched over to user's manual, has improved user's the experience of shooing. Under the condition that camera automatic switch mode is opened at camera APP, after camera APP starts completely, after the intelligent decision-making module discerns the target camera mode, camera drive is according to camera automatic switch around the control of target camera mode. From this, realized the in-process that camera APP shot, the purpose of camera around the cell-phone is according to the shooting scene automatic switch of reality, improved user's shooting experience.
Compared with the existing mobile phone shooting process, the camera APP requires the user to manually switch the front camera and the rear camera according to the actual shooting scene, so that the problems of frequent user operation, poor shooting experience and the like are caused. After the target camera mode is identified by the intelligent decision module, the camera driving can control the automatic switching of the front camera and the rear camera, the whole camera switching process does not need manual operation of a user, great convenience is brought to the shooting process of the user, and therefore the shooting experience of the user is improved.
It should be explained that the method for automatically switching the camera is not limited to the scene of photographing by the camera APP in the mobile phone, and may also be applied to the scene of photographing by the camera APP in the mobile phone, which is not limited herein.
In the embodiment of the application, after the camera APP is completely started, the intelligent decision module identifies that the target camera mode is the first camera mode. After the camera APP is restarted, the intelligent decision module identifies that the target camera mode is the second camera mode. Wherein the second camera mode is different from the first camera mode. For example, the first camera mode is a front camera mode, and the second camera mode is a rear camera mode. The method can be understood that after the camera APP is completely started, a third preview interface is displayed in a display screen of the mobile phone, wherein the third preview interface includes image information acquired by a camera of the camera APP in the first camera mode. And after the camera APP is restarted, a fourth preview interface is displayed in a display screen of the mobile phone, wherein the fourth preview interface comprises image information acquired by a camera in a second camera mode of the camera APP. Therefore, when the camera APP in the mobile phone is in the starting stage, the camera is not started according to the memory function of the camera APP, the purpose that the mobile phone automatically switches the front camera and the back camera according to the actual shooting scene is achieved, and the shooting experience of a user is improved.
In a possible situation, in the process that the camera APP adopts the front camera to shoot, the rear camera is in a closed state, and the intelligent decision module identifies that the target camera mode is the rear camera. Under the condition, the camera drives and controls the rear camera to be opened and controls the front camera to be closed. That is to say, at the in-process that camera APP was made a video recording, the cell-phone can carry out automatic switch-over to the front and back camera according to the scene of actually taking. Because the camera APP adopts the leading camera to carry out the in-process of making a video recording, the trailing camera is in the closed condition and does not gather image information, then the camera APP carries out automatic switch-over to the front and back camera at the in-process of making a video recording and probably influences the effect of making a video recording.
Under another possible condition, in the process that the camera APP adopts the front camera to shoot, the rear camera is in a starting state, and the intelligent decision module identifies that the target camera mode is the rear camera. In this case, the mobile phone acquires and maintains the image acquired by the rear camera. Because the camera APP adopts the leading camera to carry out the in-process of making a video recording, the trailing camera is in the on-state and continuously gathers image information, then the camera APP can not influence the effect of making a video recording to the in-process that the front and back camera carries out automatic switch-over in the process of making a video recording.
It should be explained that the switch states of the front and rear cameras in the shooting process of the camera APP are only used as an exemplary description, and are not limited herein.
As shown in fig. 13, fig. 13 is a schematic structural diagram three of an electronic device provided in the embodiment of the present application, where the electronic device may be the mobile phone. The electronic device may specifically include: a touch screen 1301, the touch screen 1301 comprising a touch sensor 1306 and a display 1307; one or more processors 1302; a memory 1303; one or more application programs (not shown); and one or more computer programs 1304, which may be connected via one or more communication buses 1305. Wherein the one or more computer programs 1304 are stored in the memory 1303 and configured to be executed by the one or more processors 1302, the one or more computer programs 1304 include instructions that can be used for executing the relevant steps in the above embodiments.
It is to be understood that the electronic devices and the like described above include hardware structures and/or software modules for performing the respective functions in order to realize the functions described above. Those of skill in the art will readily appreciate that the various illustrative elements and algorithm steps described in connection with the embodiments disclosed herein may be implemented as hardware or combinations of hardware and computer software. Whether a function is performed as hardware or computer software drives hardware depends upon the particular application and design constraints imposed on the solution. Skilled artisans may implement the described functionality in varying ways for each particular application, but such implementation decisions should not be interpreted as causing a departure from the scope of the present embodiments.
In the embodiment of the present application, the electronic device and the like may be divided into functional modules according to the method example, for example, each functional module may be divided according to each function, or two or more functions may be integrated into one processing module. The integrated module can be realized in a hardware mode, and can also be realized in a software functional module mode. It should be noted that, the division of the modules in the embodiment of the present invention is schematic, and is only a logic function division, and there may be another division manner in actual implementation.
In a case where the functional modules are divided according to the respective functions, a possible composition diagram of the electronic device related to the above embodiment may include: display unit, transmission unit and processing unit etc. It should be noted that all relevant contents of each step related to the above method embodiment may be referred to the functional description of the corresponding functional module, and are not described herein again.
Embodiments of the present application also provide an electronic device including one or more processors and one or more memories. The one or more memories are coupled to the one or more processors, the one or more memories are configured to store computer program code, the computer program code comprising computer instructions, which, when executed by the one or more processors, cause the electronic device to perform the above-described related method steps to implement the switching method of the camera in the above-described embodiments.
Embodiments of the present application further provide a computer-readable storage medium, where computer instructions are stored in the computer-readable storage medium, and when the computer instructions are run on an electronic device, the electronic device is caused to execute the above related method steps to implement the camera switching method in the above embodiments.
Embodiments of the present application further provide a computer program product, where the computer program product includes computer instructions, and when the computer instructions are run on an electronic device, the electronic device is caused to execute the above related method steps to implement the camera switching method in the above embodiments.
In addition, embodiments of the present application also provide an apparatus, which may be specifically a chip, a component or a module, and may include a processor and a memory connected to each other; the memory is used for storing computer execution instructions, and when the device runs, the processor can execute the computer execution instructions stored in the memory, so that the device can execute the camera switching method executed by the electronic equipment in the above method embodiments.
In addition, the electronic device, the computer-readable storage medium, the computer program product, or the apparatus provided in this embodiment are all configured to execute the corresponding method provided above, so that the beneficial effects achieved by the electronic device, the computer-readable storage medium, the computer program product, or the apparatus can refer to the beneficial effects in the corresponding method provided above, and are not described herein again.
Through the above description of the embodiments, it is clear to those skilled in the art that, for convenience and simplicity of description, the foregoing division of the functional modules is merely used as an example, and in practical applications, the above function distribution may be completed by different functional modules according to needs, that is, the internal structure of the device may be divided into different functional modules to complete all or part of the above described functions. For the specific working processes of the system, the apparatus and the unit described above, reference may be made to the corresponding processes in the foregoing method embodiments, and details are not described here again.
Each functional unit in the embodiments of the present application may be integrated into one processing unit, or each unit may exist alone physically, or two or more units are integrated into one unit. The integrated unit can be realized in a form of hardware, and can also be realized in a form of a software functional unit.
The integrated unit, if implemented in the form of a software functional unit and sold or used as a stand-alone product, may be stored in a computer readable storage medium. Based on such understanding, the technical solutions of the embodiments of the present application may be essentially implemented or make a contribution to the prior art, or all or part of the technical solutions may be implemented in the form of a software product stored in a storage medium and including several instructions for causing a computer device (which may be a personal computer, a server, or a network device) or a processor to execute all or part of the steps of the methods described in the embodiments of the present application. And the aforementioned storage medium includes: flash memory, removable hard drive, read only memory, random access memory, magnetic or optical disk, and the like.
The above description is only an embodiment of the present application, but the scope of the present application is not limited thereto, and any changes or substitutions within the technical scope of the present disclosure should be covered by the scope of the present application. Therefore, the protection scope of the present application shall be subject to the protection scope of the claims.

Claims (15)

1. A camera switching method is applied to electronic equipment, and is characterized by comprising the following steps:
when the electronic equipment is in a formal starting stage, if the electronic equipment is detected not to be connected with external equipment, acquiring information acquired by at least one sensor of the electronic equipment within a preset time length before the formal starting of the camera application;
inputting the information acquired by the at least one sensor into a trained decision model for camera mode recognition, and determining a first target camera mode, wherein the decision model has a mapping relation between the information acquired by the at least one sensor and the corresponding camera mode;
the electronic equipment displays a preview interface after the camera application program is completely started, wherein the preview interface comprises image information acquired by a camera of the camera application program in the first target camera mode;
after the camera application program is completely started, the electronic equipment inputs information acquired by at least one sensor of the electronic equipment after the camera application program is completely started into another trained decision model for camera mode recognition to obtain a second target camera mode, wherein the other decision model has a mapping relation between the information acquired by the at least one sensor and the corresponding camera mode, and the first target camera mode and the second target camera mode comprise a front camera mode or a rear camera mode; and
and the electronic equipment displays another preview interface, wherein the another preview interface comprises image information acquired by the camera of the camera application program in the second target camera mode.
2. The method of claim 1,
wherein the second target camera mode is the front-facing camera mode;
after the electronic device completely starts the camera application program, inputting information acquired by at least one sensor of the electronic device into another trained decision model for camera mode recognition to obtain a second target camera mode, the method further includes:
and if the electronic equipment detects that the image information acquired by the front camera comprises preset image information, automatically switching the camera from the front camera mode to the rear camera mode.
3. The method of claim 1 or 2, wherein after the electronic device displays another preview interface, the method further comprises:
displaying first prompt information, wherein the first prompt information is used for reminding a user that the camera is switched to the front camera mode or the rear camera mode.
4. The method of claim 3, wherein after the electronic device displays another preview interface, the method further comprises:
in the process that the electronic equipment displays the first prompt message, if the electronic equipment receives a first control instruction of a user within a first preset time after the camera switching is completed, the electronic equipment responds to the first control instruction to control the camera switching.
5. The method according to claim 1 or 2, characterized in that the method further comprises:
and the electronic equipment displays second prompt information after the camera application program is completely started and the automatic camera switching mode is started for the first time, wherein the second prompt information is used for reminding a user whether to train the automatic camera switching mode.
6. The method according to claim 1 or 2, wherein before the information collected by the at least one sensor of the electronic device enters the trained another decision model for camera pattern recognition to obtain a second target camera pattern after the electronic device completely starts the camera application program, the method further comprises:
and the electronic equipment determines that the electronic equipment is in a stable state according to the information acquired by the at least one sensor.
7. The method of claim 1, wherein after the electronic device displays another preview interface, the method further comprises:
and if a second control instruction of the user to the photographing button on the other preview interface is received, performing iterative training on the other decision model by using the information acquired by the at least one sensor.
8. The method according to claim 1 or 2, wherein the another preview interface is a preview interface of the camera application in a photographing process or a preview interface of the camera application in a recording process.
9. The method of claim 1 or 2, wherein the sensor comprises at least one of an image sensor, a gyroscope sensor, a gravity sensor, a structured light sensor, a proximity light sensor, or an infrared sensor.
10. The method of claim 1, further comprising:
and if the electronic equipment is detected to be connected with the external equipment, determining the target camera mode according to the external equipment.
11. The method of claim 1, further comprising:
and if a third control instruction of the user to a photographing button on the preview interface is received within a third preset time after the camera application program is started, the electronic equipment performs iterative training on the decision model by using the information acquired by the at least one sensor.
12. The method according to claim 10 or 11, characterized in that the method further comprises:
the electronic equipment displays a third preview interface, wherein the third preview interface comprises image information acquired by a camera of the camera application program in a first camera mode;
the electronic equipment displays a fourth preview interface after the camera application program is restarted; the fourth preview interface comprises image information acquired by a camera of the camera application program in a second camera mode; the second camera mode is different from the first camera mode.
13. The method of claim 12, wherein the one preview interface, the third preview interface, and the fourth preview interface are preview interfaces of the camera application in a photographing process or preview interfaces of the camera application in a recording process.
14. An electronic device, comprising:
a processor;
a memory;
wherein the memory has stored therein a computer program comprising instructions that, when executed, cause the processor to perform the method of switching a camera head according to any one of claims 1-13.
15. A computer-readable storage medium having instructions stored therein, which when run on an electronic device, cause the electronic device to perform the method of switching the camera of any one of claims 1-13.
CN202111413464.1A 2021-11-25 2021-11-25 Camera switching method and electronic equipment Active CN113840092B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202111413464.1A CN113840092B (en) 2021-11-25 2021-11-25 Camera switching method and electronic equipment

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202111413464.1A CN113840092B (en) 2021-11-25 2021-11-25 Camera switching method and electronic equipment

Publications (2)

Publication Number Publication Date
CN113840092A CN113840092A (en) 2021-12-24
CN113840092B true CN113840092B (en) 2022-04-15

Family

ID=78971418

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202111413464.1A Active CN113840092B (en) 2021-11-25 2021-11-25 Camera switching method and electronic equipment

Country Status (1)

Country Link
CN (1) CN113840092B (en)

Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103051795A (en) * 2012-12-18 2013-04-17 广东欧珀移动通信有限公司 Method for switching front camera and rear camera of mobile terminal and mobile terminal
CN104168418A (en) * 2014-06-30 2014-11-26 宇龙计算机通信科技(深圳)有限公司 Image acquisition device, image acquisition method and terminal
CN104536559A (en) * 2014-11-25 2015-04-22 深圳市金立通信设备有限公司 Terminal control method
CN106303260A (en) * 2016-10-18 2017-01-04 北京小米移动软件有限公司 Photographic head changing method, device and terminal unit
CN107105160A (en) * 2017-04-25 2017-08-29 维沃移动通信有限公司 A kind of method, system and mobile terminal for starting camera
CN108632536A (en) * 2018-07-31 2018-10-09 Oppo广东移动通信有限公司 A kind of camera control method and device, terminal, storage medium
CN109743552A (en) * 2019-01-17 2019-05-10 宇龙计算机通信科技(深圳)有限公司 A kind of object monitor method, apparatus, server and storage medium
CN109819168A (en) * 2019-01-31 2019-05-28 维沃移动通信有限公司 A kind of the starting method and mobile terminal of camera
WO2021015337A1 (en) * 2019-07-24 2021-01-28 엘지전자 주식회사 Method for switching between front camera and rear camera in mobile terminal and same mobile terminal

Family Cites Families (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9160912B2 (en) * 2012-06-08 2015-10-13 Apple Inc. System and method for automatic image capture control in digital imaging
CN104980644B (en) * 2014-04-14 2018-12-14 华为技术有限公司 A kind of image pickup method and device
IT201900012813A1 (en) * 2019-07-24 2021-01-24 Ambarella Int Lp SWITCHABLE DISPLAY DURING PARKING MANEUVERS
CN111654629B (en) * 2020-06-11 2022-06-24 展讯通信(上海)有限公司 Camera switching method and device, electronic equipment and readable storage medium

Patent Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103051795A (en) * 2012-12-18 2013-04-17 广东欧珀移动通信有限公司 Method for switching front camera and rear camera of mobile terminal and mobile terminal
CN104168418A (en) * 2014-06-30 2014-11-26 宇龙计算机通信科技(深圳)有限公司 Image acquisition device, image acquisition method and terminal
CN104536559A (en) * 2014-11-25 2015-04-22 深圳市金立通信设备有限公司 Terminal control method
CN106303260A (en) * 2016-10-18 2017-01-04 北京小米移动软件有限公司 Photographic head changing method, device and terminal unit
CN107105160A (en) * 2017-04-25 2017-08-29 维沃移动通信有限公司 A kind of method, system and mobile terminal for starting camera
CN108632536A (en) * 2018-07-31 2018-10-09 Oppo广东移动通信有限公司 A kind of camera control method and device, terminal, storage medium
CN109743552A (en) * 2019-01-17 2019-05-10 宇龙计算机通信科技(深圳)有限公司 A kind of object monitor method, apparatus, server and storage medium
CN109819168A (en) * 2019-01-31 2019-05-28 维沃移动通信有限公司 A kind of the starting method and mobile terminal of camera
WO2021015337A1 (en) * 2019-07-24 2021-01-28 엘지전자 주식회사 Method for switching between front camera and rear camera in mobile terminal and same mobile terminal

Also Published As

Publication number Publication date
CN113840092A (en) 2021-12-24

Similar Documents

Publication Publication Date Title
WO2021093793A1 (en) Capturing method and electronic device
CN113132620B (en) Image shooting method and related device
CN114816210B (en) Full screen display method and device of mobile terminal
CN111327814A (en) Image processing method and electronic equipment
CN113475057B (en) Video frame rate control method and related device
CN110874168A (en) Display method and electronic equipment
CN112527094A (en) Human body posture detection method and electronic equipment
CN113660408B (en) Anti-shake method and device for video shooting
CN114125130B (en) Method for controlling communication service state, terminal device and readable storage medium
CN114090102B (en) Method, device, electronic equipment and medium for starting application program
CN115589051B (en) Charging method and terminal equipment
CN113572956A (en) Focusing method and related equipment
CN114257671B (en) Image display method and electronic equipment
WO2023273323A9 (en) Focusing method and electronic device
CN115967851A (en) Quick photographing method, electronic device and computer readable storage medium
CN114422686A (en) Parameter adjusting method and related device
CN112866557A (en) Composition recommendation method and electronic device
CN113840092B (en) Camera switching method and electronic equipment
WO2022017270A1 (en) Appearance analysis method, and electronic device
CN114812381A (en) Electronic equipment positioning method and electronic equipment
CN116055872B (en) Image acquisition method, electronic device, and computer-readable storage medium
CN114245011B (en) Image processing method, user interface and electronic equipment
WO2022206783A1 (en) Photography method and apparatus, and electronic device and readable storage medium
WO2022028324A1 (en) Method and apparatus for starting application, and electronic device and medium
CN113849113A (en) Touch operation processing method, electronic device and computer-readable storage medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant