CN113542580B - Method and device for removing light spots of glasses and electronic equipment - Google Patents

Method and device for removing light spots of glasses and electronic equipment Download PDF

Info

Publication number
CN113542580B
CN113542580B CN202010323054.7A CN202010323054A CN113542580B CN 113542580 B CN113542580 B CN 113542580B CN 202010323054 A CN202010323054 A CN 202010323054A CN 113542580 B CN113542580 B CN 113542580B
Authority
CN
China
Prior art keywords
image
glasses
preview image
shooting
marking
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202010323054.7A
Other languages
Chinese (zh)
Other versions
CN113542580A (en
Inventor
季伟
陈国乔
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Huawei Technologies Co Ltd
Original Assignee
Huawei Technologies Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Huawei Technologies Co Ltd filed Critical Huawei Technologies Co Ltd
Priority to CN202010323054.7A priority Critical patent/CN113542580B/en
Publication of CN113542580A publication Critical patent/CN113542580A/en
Application granted granted Critical
Publication of CN113542580B publication Critical patent/CN113542580B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/80Camera processing pipelines; Components thereof
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • General Health & Medical Sciences (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Artificial Intelligence (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • Computational Linguistics (AREA)
  • Signal Processing (AREA)
  • Evolutionary Computation (AREA)
  • Health & Medical Sciences (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Multimedia (AREA)
  • Studio Devices (AREA)
  • Controls And Circuits For Display Device (AREA)

Abstract

The embodiment of the application provides a method and a device for removing light spots of glasses and electronic equipment, wherein the method comprises the following steps: acquiring a preview image of a shooting object before receiving a shutter instruction; marking a light spot area in the preview image according to the target characteristic information in the preview image; receiving the shutter instruction, and acquiring an image to be processed, wherein the image to be processed is an image for marking the light spot area; and removing the light spot area in the image to be processed to obtain a target image. According to the embodiment of the application, the distribution of the glasses light spots in the predicted images can be well predicted, so that the post-processing efficiency is improved, the processing cost is reduced, and the photo quality of a user wearing glasses when shooting through screen light supplement is improved.

Description

Method and device for removing light spots of glasses and electronic equipment
Technical Field
The application belongs to the technical field of information processing, and particularly relates to a method and a device for removing a light spot of a pair of glasses and electronic equipment.
Background
With the development and popularization of electronic devices, the camera function of the electronic devices is also more and more powerful. The portability of the mobile phone makes the mobile phone become the primary choice for people to take pictures in daily life so as to record the life of people at any time and any place.
At present, in the process of photographing a person image at night or in the case of weak light, the light is supplemented by instantly making the screen photograph completely white or previewing the high brightness around the interface screen, and after the environment is illuminated, the photographing effect under the dark light is improved. However, when a user wearing glasses shoots through screen light supplement, especially in the self-shooting process, colored light spots appear on the lenses, so that eye details are lost, and the shooting effect is seriously influenced; due to the complexity of the light spots reflected by the glasses, the distribution of the light spots cannot be well predicted, so that the post-processing efficiency is low, the cost is high, and the quality of the shot pictures is reduced.
Disclosure of Invention
The application discloses a method and a device for removing glasses light spots and electronic equipment, the distribution of the light spots in images can be well predicted, the image processing efficiency in the later stage is improved, and the picture quality of a user wearing glasses when shooting is carried out through screen light supplement.
In a first aspect, an embodiment of the present application provides a method for removing a light spot of glasses, where the method includes, before a shutter instruction is received, acquiring a preview image of a photographic subject; marking a light spot area in the preview image according to the target characteristic information in the preview image; receiving the shutter instruction, and acquiring an image to be processed, wherein the image to be processed is an image for marking the light spot area; and removing the light spot area in the image to be processed to obtain a target image.
According to the embodiment of the application, in the photographing process, the electronic equipment marks the glasses light spot area in the preview image, removes the glasses light spot area in the image to be processed after acquiring the image to be processed for marking the glasses light spot, and obtains the target image with the detail characteristics of the photographing object of the corresponding light spot area reserved; the distribution of light spots in the images and the marking of the light spots are predicted, and the light spot areas in the images are removed, so that the later image processing efficiency is improved; meanwhile, the quality of pictures shot by the user wearing the glasses through screen light supplement is improved.
With reference to the first aspect, in some embodiments, the marking a spot region in the preview image according to target feature information in the preview image includes:
determining the axial distance between a shooting device and a shooting object according to target characteristic information in the preview image, wherein the target characteristic information comprises a glasses area in the preview image; determining projection information of a screen of the shooting device on glasses of the shooting object according to the axial distance and the screen size information of the shooting device; and determining a light spot area of the glasses area in the preview image according to the projection information and the focal length of the shooting device.
In the method for identifying the light spot area of the glasses, the electronic equipment determines the projection information of the screen of the shooting device on the glasses of the shooting object by processing the preview image before receiving the shutter command in the shooting process, further determines the light spot information in the preview image according to the projection information, can well predict and determine the distribution and the size of the light spot in the image, provides more accurate and reliable reference for image processing in the later shooting period, improves the efficiency of image processing in the later shooting period, reduces the cost of image processing, and further improves the shooting effect and the picture quality of a user wearing the glasses when shooting through screen light supplement.
With reference to the first aspect, in some embodiments, before the determining, according to the target feature information in the preview image, an axial distance between the camera and the shooting object, the method includes:
extracting facial feature points of a human face region in the preview image; and marking the target feature information in the preview image according to the facial feature points.
Illustratively, by extracting facial feature points of a face region in a preview image, the region information of the face contour and the region information of the facial features in the preview image can be located. Therefore, the required target characteristic information, such as the information of the interpupillary distance in the preview image, the eyeglass area determined according to the eye area in the five sense organs and the like, can be marked according to the area information in the preview image in which the face contour and the five sense organs of the face are respectively positioned; therefore, the prediction and the positioning of the region where the light spot possibly appears in the preview image are realized, and a reliable data basis and an operation basis are provided for subsequent processing.
With reference to the first aspect, in some embodiments, the facial feature points comprise eye feature points;
the marking the target feature information in the preview image according to the facial feature points comprises:
and marking a first pupil distance in the preview image according to the coordinates of the eye feature points.
It can be understood that the eye region may be determined according to the eye feature point, so that the first pupil distance in the preview image, that is, the pupil distance in the image corresponding to the actual pupil distance, may be determined according to the coordinate information of the eye region.
With reference to the first aspect, in some embodiments, the determining an axial distance between the camera and the shooting object according to the target feature information in the preview image includes:
and calculating the axial distance between the shooting device and the glasses of the shooting object according to the first pupil distance, the preset second pupil distance and the focal length of the shooting device in the preview image.
Through the above embodiment, the electronic device stores the preset second pupil distance, which may be a conventional pupil distance of a human, for example, a typical pupil distance of a human eye is 63 mm; the electronic equipment can automatically focus by detecting the face of a person, and after focusing, the focal length can be determined; according to the camera imaging principle, the first pupil distance, the second pupil distance and the focal length, the axial distance between the shooting device and glasses in the shooting object or the distance in the x-axis direction can be calculated, and then the position information of the shooting object can be determined.
With reference to the first aspect, in some embodiments, the determining projection information of the screen of the shooting device on the glasses of the shooting object according to the axial distance and the screen size information of the shooting device includes:
determining first angle information of the shot object according to target characteristic information in the preview image; acquiring second angle information of the shooting device through a direction sensor of the shooting device; and determining the projection position and the projection size of the screen of the shooting device on the glasses of the shooting object according to the first angle information, the second angle information, the axial distance and the screen size of the shooting device.
Through the embodiment, the electronic equipment can determine the first angle information of the shooting object in the three-dimensional space coordinate system according to the image of the marked glasses area; determining second angle information of the shooting device in a three-dimensional space coordinate system according to the direction sensor, and calculating projection information of a screen of the shooting device on glasses of a shooting object according to the first angle information, the second angle information, the axial distance and the screen size of the shooting device, so as to preliminarily determine a region capable of generating light reflection on the shooting object. The problem of complexity of determining the light reflection area in the photographing process is solved, the light reflection area of the glasses of the photographed object can be predicted according to the angle information and the distance information before the shutter command is received, a reliable basis is provided for subsequent image processing, and the image processing efficiency is improved.
With reference to the first aspect, in some embodiments, the projection information includes a projection location and a projection size;
the determining a light spot area of the glasses area in the preview image according to the projection information and the focal length of the shooting device includes:
calculating the position of the light spot area in the glasses area in the preview image according to the projection position of the screen of the shooting device on the glasses of the shooting object and the focal length of the shooting device; and calculating the size of the glasses area of the light spot area in the preview image according to the projection size of the screen of the shooting device on the glasses of the shooting object and the focal length of the shooting device.
With reference to the first aspect, in some embodiments, the facial feature points comprise eye feature points;
the marking the target feature information in the preview image according to the facial feature points comprises:
marking the eye region in the preview image according to the coordinates of the eye feature points; inputting the image for marking the eye region into an image recognition model to obtain an image for marking the glasses region output by the image recognition model; the image recognition model is obtained through training of a first image training set and a second image training set, the first image training set is an image training set with glasses, and the second image training set is an image training set without glasses.
Understandably, the image recognition model can be an antagonistic neural network model, iterative training is carried out on the antagonistic neural network model through the first image training set and the second image training set, the image recognition model which can more accurately recognize the glasses area in the preview image is obtained, and therefore the image for marking the glasses area is obtained.
With reference to the first aspect, in some embodiments, after the inputting the image for marking the eye region into an image recognition model and obtaining the image for marking the eyeglass region output by the image recognition model, the method includes:
inputting the image for marking the glasses area into a neural network model to obtain an image for marking a light spot area output by the neural network model;
and the neural network model is a model for marking a light spot area in the preview image and is obtained by training according to a light spot image training set.
Illustratively, when shooting multiple persons through screen light supplement, the electronic equipment performs face detection to extract facial feature points of the multiple persons, and by means of the embodiment, whether the persons wear glasses or not and a region where the persons wear the glasses is marked are judged, so that rapid marking of target feature information of the multiple persons in the preview image is achieved.
With reference to the first aspect, in some embodiments, the removing the spot region in the image to be processed to obtain a target image includes:
and adjusting the brightness value and the gray value of the light spot area in the image to be processed according to preset control parameters to obtain the target image.
Through the embodiment, after the electronic equipment receives the shutter instruction, the electronic equipment collects the image to be processed of the marked light spot area, and retains the detail information of the light spot area by adjusting the brightness value and the gray value of the pixels of the light spot area, so as to obtain the target image of the light spot area.
In a second aspect, an embodiment of the present application provides an apparatus for removing an eye glasses spot, including:
an acquisition unit configured to acquire a preview image of a photographic subject before receiving a shutter instruction;
the calculation unit is used for marking a light spot area in the preview image according to the target characteristic information in the preview image;
the acquisition unit is used for receiving the shutter instruction and acquiring an image to be processed, wherein the image to be processed is an image for marking the light spot area;
and the processing unit is used for removing the light spot area in the image to be processed to obtain a target image.
In a third aspect, an embodiment of the present application provides an electronic device, including: one or more processors, a memory, a display screen and a camera; the memory, the display screen, the camera, and the one or more processors, the memory for storing computer program code, the computer program code comprising computer instructions; the computer instructions, when executed by the one or more processors, cause the electronic device to perform the method as provided by the first aspect and any one of the possible implementations of the first aspect.
In a fourth aspect, an embodiment of the present application provides a computer-readable storage medium, where a computer program is stored, where the computer program includes instructions that, when executed on an electronic device, cause the electronic device to perform the method as provided in the first aspect and any one of the possible implementation manners of the first aspect.
In a fifth aspect, an embodiment of the present application provides a computer program product including instructions, which, when run on an electronic device, cause the electronic device to perform the method as provided in the first aspect and any one of the possible implementation manners of the first aspect.
It is to be understood that the apparatus for removing eye-glasses spots according to the second aspect, the electronic device according to the third aspect, the computer-readable storage medium according to the fourth aspect, or the computer program product containing instructions according to the fifth aspect, all provided above, are configured to perform the method according to any one of the first aspect. Therefore, the beneficial effects achieved by the method can refer to the beneficial effects in the corresponding method, and are not described herein again.
Drawings
The drawings used in the embodiments of the present application are described below.
Fig. 1 is a schematic structural diagram of an electronic device 100 provided in an embodiment of the present application;
fig. 2 is a block diagram of a software structure of the electronic device 100 according to the embodiment of the present application;
fig. 3 (a) is a schematic diagram of a photographing scene provided in an embodiment of the present application;
fig. 3 (B) is a schematic diagram of another photographing scene provided in the embodiment of the present application;
FIG. 4 is a schematic diagram of the principle of camera imaging provided by an embodiment of the present application;
FIG. 5 is a schematic diagram for determining an axial distance between a camera and a photographic subject according to an embodiment of the present disclosure;
FIG. 6 is a schematic diagram of a projection principle provided by an embodiment of the present application;
fig. 7 is a schematic diagram of determining a light reflection region on glasses of a photographic subject according to an embodiment of the present disclosure;
fig. 8 is a schematic diagram of a to-be-processed image of a mark light spot area provided by an embodiment of the present application;
fig. 9 is a schematic view of a light spot mark of a scene photographed by multiple persons according to an embodiment of the present application;
FIG. 10 is a schematic diagram of a target image output after processing a spot area according to an embodiment of the present disclosure;
fig. 11 is a schematic flowchart of a method for identifying a light spot region of glasses in a photographing process according to an embodiment of the present application;
fig. 12 is a schematic structural diagram of an apparatus for identifying a spot area of glasses in a photographing process according to an embodiment of the present application.
Detailed Description
The embodiments of the present application are described below with reference to the drawings. The terminology used in the description of the embodiments herein is for the purpose of describing particular embodiments herein only and is not intended to be limiting of the application.
First, an electronic device according to an embodiment of the present application will be described. Referring to fig. 1, fig. 1 is a schematic structural diagram of an electronic device 100 according to an embodiment of the present disclosure.
The electronic device 100 may include a processor 110, an external memory interface 120, an internal memory 121, a Universal Serial Bus (USB) interface 130, a charging management module 140, a power management module 141, a battery 142, an antenna 1, an antenna 2, a mobile communication module 150, a wireless communication module 160, an audio module 170, a speaker 170A, a receiver 170B, a microphone 170C, an earphone interface 170D, a sensor module 180, a button 190, a motor 191, an indicator 192, a camera 193, a display screen 194, a Subscriber Identity Module (SIM) card interface 195, and the like. The sensor module 180 may include a pressure sensor 180A, a gyroscope sensor 180B, an air pressure sensor 180C, a magnetic sensor 180D, an acceleration sensor 180E, a distance sensor 180F, a proximity light sensor 180G, a fingerprint sensor 180H, a temperature sensor 180J, a touch sensor 180K, an ambient light sensor 180L, a bone conduction sensor 180M, and the like.
It is to be understood that the illustrated structure of the embodiment of the present invention does not specifically limit the electronic device 100. In other embodiments of the present application, the electronic device 100 may include more or fewer components than shown, or combine certain components, or split certain components, or arrange different components. The illustrated components may be implemented in hardware, software, or a combination of software and hardware.
Processor 110 may include one or more processing units, such as: the processor 110 may include an Application Processor (AP), a modem processor, a Graphics Processing Unit (GPU), an Image Signal Processor (ISP), a controller, a memory, a video codec, a Digital Signal Processor (DSP), a baseband processor, and/or a neural-Network Processing Unit (NPU), etc. Wherein, the different processing units may be independent devices or may be integrated in one or more processors.
The controller may be, among other things, a neural center and a command center of the electronic device 100. The controller can generate an operation control signal according to the instruction operation code and the timing signal to complete the control of instruction fetching and instruction execution.
A memory may also be provided in the processor 110 for storing instructions and data. In some embodiments, the memory in the processor 110 is a cache memory. The memory may hold instructions or data that have just been used or recycled by the processor 110. If the processor 110 needs to reuse the instruction or data, it can be called directly from the memory. Avoiding repeated accesses reduces the latency of the processor 110, thereby increasing the efficiency of the system.
In some embodiments, processor 110 may include one or more interfaces. The interface may include an integrated circuit (I2C) interface, an integrated circuit built-in audio (I2S) interface, a Pulse Code Modulation (PCM) interface, a universal asynchronous receiver/transmitter (UART) interface, a Mobile Industry Processor Interface (MIPI), a general-purpose input/output (GPIO) interface, a Subscriber Identity Module (SIM) interface, and/or a Universal Serial Bus (USB) interface, etc.
The I2C interface is a bidirectional synchronous serial bus comprising a serial data line (SDA) and a Serial Clock Line (SCL). In some embodiments, processor 110 may include multiple sets of I2C buses. The processor 110 may be coupled to the touch sensor 180K, a charger, a flash, a camera 193, etc. through different I2C bus interfaces, respectively. For example: the processor 110 may be coupled to the touch sensor 180K through an I2C interface, so that the processor 110 and the touch sensor 180K communicate through an I2C bus interface to implement a touch function of the electronic device 100.
The I2S interface may be used for audio communication. In some embodiments, processor 110 may include multiple sets of I2S buses. The processor 110 may be coupled to the audio module 170 through an I2S bus to enable communication between the processor 110 and the audio module 170. In some embodiments, the audio module 170 may transmit an audio signal to the wireless communication module 160 through the I2S interface, so as to implement a function of answering a call through a bluetooth headset.
The PCM interface may also be used for audio communication, sampling, quantizing and encoding analog signals. In some embodiments, the audio module 170 and the wireless communication module 160 may be coupled by a PCM bus interface. In some embodiments, the audio module 170 may also transmit the audio signal to the wireless communication module 160 through the PCM interface, so as to implement the function of answering a call through the bluetooth headset. Both the I2S interface and the PCM interface may be used for audio communication.
The UART interface is a universal serial data bus used for asynchronous communications. The bus may be a bidirectional communication bus. It converts the data to be transmitted between serial communication and parallel communication. In some embodiments, a UART interface is generally used to connect the processor 110 with the wireless communication module 160. For example: the processor 110 communicates with a bluetooth module in the wireless communication module 160 through a UART interface to implement a bluetooth function. In some embodiments, the audio module 170 may transmit the audio signal to the wireless communication module 160 through a UART interface, so as to realize the function of playing music through a bluetooth headset.
MIPI interfaces may be used to connect processor 110 with peripheral devices such as display screen 194, camera 193, and the like. The MIPI interface includes a Camera Serial Interface (CSI), a Display Serial Interface (DSI), and the like. In some embodiments, processor 110 and camera 193 communicate through a CSI interface to implement the capture functionality of electronic device 100. Processor 110 and display screen 194 communicate via a DSI interface to implement display functions of electronic device 100.
The GPIO interface may be configured by software. The GPIO interface may be configured as a control signal and may also be configured as a data signal. In some embodiments, a GPIO interface may be used to connect the processor 110 with the camera 193, the display 194, the wireless communication module 160, the audio module 170, the sensor module 180, and the like. The GPIO interface may also be configured as an I2C interface, I2S interface, UART interface, MIPI interface, and the like.
The USB interface 130 is an interface conforming to the USB standard specification, and may specifically be a Mini USB interface, a Micro USB interface, a USB Type C interface, or the like. The USB interface 130 may be used to connect a charger to charge the electronic device 100, and may also be used to transmit data between the electronic device 100 and a peripheral device. And the method can also be used for connecting a headset and playing audio through the headset. The interface may also be used to connect other electronic devices, such as AR devices and the like.
It should be understood that the connection relationship between the modules according to the embodiment of the present invention is only illustrative and is not limited to the structure of the electronic device 100. In other embodiments of the present application, the electronic device 100 may also adopt different interface connection manners or a combination of multiple interface connection manners in the above embodiments.
The charging management module 140 is configured to receive charging input from a charger. The charger can be a wireless charger or a wired charger. In some wired charging embodiments, the charging management module 140 may receive charging input from a wired charger via the USB interface 130. In some wireless charging embodiments, the charging management module 140 may receive a wireless charging input through a wireless charging coil of the electronic device 100. The charging management module 140 may also supply power to the electronic device through the power management module 141 while charging the battery 142.
The power management module 141 is used to connect the battery 142, the charging management module 140 and the processor 110. The power management module 141 receives input from the battery 142 and/or the charge management module 140 and provides power to the processor 110, the internal memory 121, the external memory, the display 194, the camera 193, the wireless communication module 160, and the like. The power management module 141 may also be used to monitor parameters such as battery capacity, battery cycle count, battery state of health (leakage, impedance), etc. In some other embodiments, the power management module 141 may also be disposed in the processor 110. In other embodiments, the power management module 141 and the charging management module 140 may also be disposed in the same device.
The wireless communication function of the electronic device 100 may be implemented by the antenna 1, the antenna 2, the mobile communication module 150, the wireless communication module 160, a modem processor, a baseband processor, and the like.
The antennas 1 and 2 are used for transmitting and receiving electromagnetic wave signals. Each antenna in the electronic device 100 may be used to cover a single or multiple communication bands. Different antennas can also be multiplexed to improve the utilization of the antennas. For example: the antenna 1 may be multiplexed as a diversity antenna of a wireless local area network. In other embodiments, the antenna may be used in conjunction with a tuning switch.
The mobile communication module 150 may provide a solution including 2G/3G/4G/5G wireless communication applied to the electronic device 100. The mobile communication module 150 may include at least one filter, a switch, a power amplifier, a Low Noise Amplifier (LNA), and the like. The mobile communication module 150 may receive the electromagnetic wave from the antenna 1, filter, amplify, etc. the received electromagnetic wave, and transmit the electromagnetic wave to the modem processor for demodulation. The mobile communication module 150 may also amplify the signal modulated by the modem processor, and convert the signal into electromagnetic wave through the antenna 1 to radiate the electromagnetic wave. In some embodiments, at least some of the functional modules of the mobile communication module 150 may be disposed in the processor 110. In some embodiments, at least some of the functional modules of the mobile communication module 150 may be provided in the same device as at least some of the modules of the processor 110.
The modem processor may include a modulator and a demodulator. The modulator is used for modulating a low-frequency baseband signal to be transmitted into a medium-high frequency signal. The demodulator is used for demodulating the received electromagnetic wave signal into a low-frequency baseband signal. The demodulator then passes the demodulated low frequency baseband signal to a baseband processor for processing. The low frequency baseband signal is processed by the baseband processor and then passed to the application processor. The application processor outputs a sound signal through an audio device (not limited to the speaker 170A, the receiver 170B, etc.) or displays an image or video through the display screen 194. In some embodiments, the modem processor may be a stand-alone device. In other embodiments, the modem processor may be provided in the same device as the mobile communication module 150 or other functional modules, independent of the processor 110.
The wireless communication module 160 may provide solutions for wireless communication applied to the electronic device 100, including Wireless Local Area Networks (WLANs) (e.g., wireless fidelity (Wi-Fi) networks), bluetooth (BT), global Navigation Satellite System (GNSS), frequency Modulation (FM), near Field Communication (NFC), infrared (IR), and the like. The wireless communication module 160 may be one or more devices integrating at least one communication processing module. The wireless communication module 160 receives electromagnetic waves via the antenna 2, performs frequency modulation and filtering processing on electromagnetic wave signals, and transmits the processed signals to the processor 110. The wireless communication module 160 may also receive a signal to be transmitted from the processor 110, perform frequency modulation and amplification on the signal, and convert the signal into electromagnetic waves through the antenna 2 to radiate the electromagnetic waves.
In some embodiments, antenna 1 of electronic device 100 is coupled to mobile communication module 150 and antenna 2 is coupled to wireless communication module 160 so that electronic device 100 can communicate with networks and other devices through wireless communication techniques. The wireless communication technology may include global system for mobile communications (GSM), general Packet Radio Service (GPRS), code division multiple access (code division multiple access, CDMA), wideband Code Division Multiple Access (WCDMA), time-division code division multiple access (time-division code division multiple access, TD-SCDMA), long Term Evolution (LTE), BT, GNSS, WLAN, NFC, FM, and/or IR technologies, etc. The GNSS may include a Global Positioning System (GPS), a global navigation satellite system (GLONASS), a beidou satellite navigation system (BDS), a quasi-zenith satellite system (QZSS), and/or a Satellite Based Augmentation System (SBAS).
The electronic device 100 implements display functions via the GPU, the display screen 194, and the application processor. The GPU is a microprocessor for image processing, and is connected to the display screen 194 and an application processor. The GPU is used to perform mathematical and geometric calculations for graphics rendering. The processor 110 may include one or more GPUs that execute program instructions to generate or alter display information.
The display screen 194 is used to display images, video, and the like. The display screen 194 includes a display panel. The display panel may be a Liquid Crystal Display (LCD), an organic light-emitting diode (OLED), an active-matrix organic light-emitting diode (active-matrix organic light-emitting diode, AMOLED), a flexible light-emitting diode (FLED), a miniature, a Micro-oeld, a quantum dot light-emitting diode (QLED), or the like. In some embodiments, the electronic device 100 may include 1 or N display screens 194, N being a positive integer greater than 1.
The electronic device 100 may implement a shooting function through the ISP, the camera 193, the video codec, the GPU, the display 194, the application processor, and the like.
The ISP is used to process the data fed back by the camera 193. For example, when a user takes a picture, the shutter is opened, light is transmitted to the camera photosensitive element through the lens, an optical signal is converted into an electric signal, and the camera photosensitive element transmits the electric signal to the ISP for processing and converting into an image visible to the naked eye. The ISP can also carry out algorithm optimization on the noise, brightness and skin color of the image. The ISP can also optimize parameters such as exposure, color temperature and the like of a shooting scene. In some embodiments, the ISP may be provided in camera 193.
The camera 193 is used to capture still images or video. The object generates an optical image through the lens and projects the optical image to the photosensitive element. The photosensitive element may be a Charge Coupled Device (CCD) or a complementary metal-oxide-semiconductor (CMOS) phototransistor. The light sensing element converts the optical signal into an electrical signal, which is then passed to the ISP where it is converted into a digital image signal. And the ISP outputs the digital image signal to the DSP for processing. The DSP converts the digital image signal into an image signal in a standard RGB, YUV and other formats. In some embodiments, electronic device 100 may include 1 or N cameras 193, N being a positive integer greater than 1.
The digital signal processor is used for processing digital signals, and can process digital image signals and other digital signals. For example, when the electronic device 100 selects a frequency bin, the digital signal processor is used to perform fourier transform or the like on the frequency bin energy.
Video codecs are used to compress or decompress digital video. The electronic device 100 may support one or more video codecs. In this way, the electronic device 100 may play or record video in a variety of encoding formats, such as: moving Picture Experts Group (MPEG) 1, MPEG2, MPEG3, MPEG4, and the like.
The NPU is a neural-network (NN) computing processor that processes input information quickly by using a biological neural network structure, for example, by using a transfer mode between neurons of a human brain, and can also learn by itself continuously. Applications such as intelligent recognition of the electronic device 100 can be realized through the NPU, for example: image recognition, face recognition, speech recognition, text understanding, and the like.
The external memory interface 120 may be used to connect an external memory card, such as a Micro SD card, to extend the memory capability of the electronic device 100. The external memory card communicates with the processor 110 through the external memory interface 120 to implement a data storage function. For example, files such as music, video, etc. are saved in an external memory card.
The internal memory 121 may be used to store computer-executable program code, which includes instructions. The processor 110 executes various functional applications of the electronic device 100 and data processing by executing instructions stored in the internal memory 121. The internal memory 121 may include a program storage area and a data storage area. The storage program area may store an operating system, an application program (such as a sound playing function, an image playing function, etc.) required by at least one function, and the like. The storage data area may store data (such as audio data, phone book, etc.) created during use of the electronic device 100, and the like. In addition, the internal memory 121 may include a high-speed random access memory, and may further include a nonvolatile memory, such as at least one magnetic disk storage device, a flash memory device, a universal flash memory (UFS), and the like.
The electronic device 100 may implement audio functions via the audio module 170, the speaker 170A, the receiver 170B, the microphone 170C, the headphone interface 170D, and the application processor. Such as music playing, recording, etc.
The audio module 170 is used to convert digital audio information into analog audio signals for output, and also used to convert analog audio inputs into digital audio signals. The audio module 170 may also be used to encode and decode audio signals. In some embodiments, the audio module 170 may be disposed in the processor 110, or some functional modules of the audio module 170 may be disposed in the processor 110.
The speaker 170A, also called a "horn", is used to convert the audio electrical signal into an acoustic signal. The electronic apparatus 100 can listen to music through the speaker 170A or listen to a handsfree call.
The receiver 170B, also called "earpiece", is used to convert the electrical audio signal into an acoustic signal. When the electronic apparatus 100 receives a call or voice information, it can receive voice by placing the receiver 170B close to the ear of the person.
The microphone 170C, also referred to as a "microphone," is used to convert sound signals into electrical signals. When making a call or transmitting voice information, the user can input a voice signal to the microphone 170C by speaking near the microphone 170C through the mouth. The electronic device 100 may be provided with at least one microphone 170C. In other embodiments, the electronic device 100 may be provided with two microphones 170C to achieve a noise reduction function in addition to collecting sound signals. In other embodiments, the electronic device 100 may further include three, four or more microphones 170C to collect sound signals, reduce noise, identify sound sources, and perform directional recording.
The earphone interface 170D is used to connect a wired earphone. The headset interface 170D may be the USB interface 130, or may be a 3.5mm open mobile electronic device platform (OMTP) standard interface, a cellular telecommunications industry association (cellular telecommunications industry association of the USA, CTIA) standard interface.
The pressure sensor 180A is used for sensing a pressure signal, and can convert the pressure signal into an electrical signal. In some embodiments, the pressure sensor 180A may be disposed on the display screen 194. The pressure sensor 180A can be of a variety of types, such as a resistive pressure sensor, an inductive pressure sensor, a capacitive pressure sensor, and the like. The capacitive pressure sensor may be a sensor comprising at least two parallel plates having an electrically conductive material. When a force acts on the pressure sensor 180A, the capacitance between the electrodes changes. The electronic device 100 determines the strength of the pressure from the change in capacitance. When a touch operation is applied to the display screen 194, the electronic device 100 detects the intensity of the touch operation according to the pressure sensor 180A. The electronic apparatus 100 may also calculate the touched position from the detection signal of the pressure sensor 180A. In some embodiments, the touch operations that are applied to the same touch position but different touch operation intensities may correspond to different operation instructions. For example: and when the touch operation with the touch operation intensity smaller than the first pressure threshold value acts on the short message application icon, executing an instruction for viewing the short message. And when the touch operation with the touch operation intensity larger than or equal to the first pressure threshold value acts on the short message application icon, executing an instruction of newly building the short message.
The gyro sensor 180B may be used to determine the motion attitude of the electronic device 100. In some embodiments, the angular velocity of electronic device 100 about three axes (i.e., x, y, and z axes) may be determined by gyroscope sensor 180B. The gyro sensor 180B may be used for photographing anti-shake. For example, when the shutter is pressed, the gyro sensor 180B detects a shake angle of the electronic device 100, calculates a distance to be compensated for by the lens module according to the shake angle, and allows the lens to counteract the shake of the electronic device 100 through a reverse movement, thereby achieving anti-shake. The gyroscope sensor 180B may also be used for navigation, somatosensory gaming scenes.
The air pressure sensor 180C is used to measure air pressure. In some embodiments, electronic device 100 calculates altitude from barometric pressure values measured by barometric pressure sensor 180C to assist in positioning and navigation.
The magnetic sensor 180D includes a hall sensor. The electronic device 100 may detect the opening and closing of the flip holster using the magnetic sensor 180D. In some embodiments, when the electronic device 100 is a flip phone, the electronic device 100 may detect the opening and closing of the flip according to the magnetic sensor 180D. And then according to the opening and closing state of the leather sheath or the opening and closing state of the flip cover, the automatic unlocking of the flip cover is set.
The acceleration sensor 180E may detect the magnitude of acceleration of the electronic device 100 in various directions (typically three axes). The magnitude and direction of gravity can be detected when the electronic device 100 is stationary. The method can also be used for identifying the posture of the electronic equipment, and is applied to horizontal and vertical screen switching, pedometers and the like.
A distance sensor 180F for measuring a distance. The electronic device 100 may measure the distance by infrared or laser. In some embodiments, taking a picture of a scene, the electronic device 100 may utilize the distance sensor 180F to range to achieve fast focus.
The proximity light sensor 180G may include, for example, a Light Emitting Diode (LED) and a light detector, such as a photodiode. The light emitting diode may be an infrared light emitting diode. The electronic device 100 emits infrared light to the outside through the light emitting diode. The electronic device 100 detects infrared reflected light from nearby objects using a photodiode. When sufficient reflected light is detected, it can be determined that there is an object near the electronic device 100. When insufficient reflected light is detected, the electronic device 100 may determine that there are no objects near the electronic device 100. The electronic device 100 can utilize the proximity light sensor 180G to detect that the user holds the electronic device 100 close to the ear for talking, so as to automatically turn off the screen to achieve the purpose of saving power. The proximity light sensor 180G can also be used in a holster mode, a pocket mode automatically unlocks and locks the screen.
The ambient light sensor 180L is used to sense ambient light brightness. Electronic device 100 may adaptively adjust the brightness of display screen 194 based on the perceived ambient light level. The ambient light sensor 180L may also be used to automatically adjust the white balance when taking a picture. The ambient light sensor 180L may also cooperate with the proximity light sensor 180G to detect whether the electronic device 100 is in a pocket to prevent accidental touches.
The fingerprint sensor 180H is used to collect a fingerprint. The electronic device 100 can utilize the collected fingerprint characteristics to unlock the fingerprint, access the application lock, photograph the fingerprint, answer an incoming call with the fingerprint, and so on.
The temperature sensor 180J is used to detect temperature. In some embodiments, electronic device 100 implements a temperature processing strategy using the temperature detected by temperature sensor 180J. For example, when the temperature reported by the temperature sensor 180J exceeds a threshold, the electronic device 100 performs a reduction in performance of a processor located near the temperature sensor 180J, so as to reduce power consumption and implement thermal protection. In other embodiments, the electronic device 100 heats the battery 142 when the temperature is below another threshold to avoid the low temperature causing the electronic device 100 to shut down abnormally. In other embodiments, when the temperature is lower than a further threshold, the electronic device 100 performs boosting on the output voltage of the battery 142 to avoid abnormal shutdown due to low temperature.
The touch sensor 180K is also referred to as a "touch panel". The touch sensor 180K may be disposed on the display screen 194, and the touch sensor 180K and the display screen 194 form a touch screen, which is also called a "touch screen". The touch sensor 180K is used to detect a touch operation acting thereon or nearby. The touch sensor may communicate the detected touch operation to the application processor to determine the touch event type. Visual output associated with the touch operation may be provided through the display screen 194. In other embodiments, the touch sensor 180K may be disposed on a surface of the electronic device 100, different from the position of the display screen 194.
The bone conduction sensor 180M can acquire a vibration signal. In some embodiments, the bone conduction sensor 180M may acquire a vibration signal of the human voice vibrating a bone mass. The bone conduction sensor 180M may also contact the human body pulse to receive the blood pressure pulsation signal. In some embodiments, the bone conduction sensor 180M may also be disposed in a headset, integrated into a bone conduction headset. The audio module 170 may analyze a voice signal based on the vibration signal of the bone mass vibrated by the sound part acquired by the bone conduction sensor 180M, so as to implement a voice function. The application processor can analyze heart rate information based on the blood pressure beating signal acquired by the bone conduction sensor 180M, so as to realize the heart rate detection function.
The keys 190 include a power-on key, a volume key, and the like. The keys 190 may be mechanical keys. Or may be touch keys. The electronic apparatus 100 may receive a key input, and generate a key signal input related to user setting and function control of the electronic apparatus 100.
The motor 191 may generate a vibration cue. The motor 191 may be used for incoming call vibration cues, as well as for touch vibration feedback. For example, touch operations applied to different applications (e.g., photographing, audio playing, etc.) may correspond to different vibration feedback effects. The motor 191 may also respond to different vibration feedback effects for touch operations applied to different areas of the display screen 194. Different application scenes (such as time reminding, receiving information, alarm clock, game and the like) can also correspond to different vibration feedback effects. The touch vibration feedback effect may also support customization.
Indicator 192 may be an indicator light that may be used to indicate a state of charge, a change in charge, or a message, missed call, notification, etc.
The SIM card interface 195 is used to connect a SIM card. The SIM card can be brought into and out of contact with the electronic apparatus 100 by being inserted into the SIM card interface 195 or being pulled out of the SIM card interface 195. The electronic device 100 may support 1 or N SIM card interfaces, N being a positive integer greater than 1. The SIM card interface 195 may support a Nano SIM card, a Micro SIM card, a SIM card, etc. The same SIM card interface 195 can be inserted with multiple cards at the same time. The types of the plurality of cards can be the same or different. The SIM card interface 195 may also be compatible with different types of SIM cards. The SIM card interface 195 is also compatible with external memory cards. The electronic device 100 interacts with the network through the SIM card to implement functions such as communication and data communication. In some embodiments, the electronic device 100 employs esims, namely: an embedded SIM card. The eSIM card can be embedded in the electronic device 100 and cannot be separated from the electronic device 100.
The software system of the electronic device 100 may employ a hierarchical architecture, an event-driven architecture, a micro-core architecture, a micro-service architecture, or a cloud architecture. The embodiment of the present invention uses an Android system with a layered architecture as an example to exemplarily illustrate a software structure of the electronic device 100.
Fig. 2 is a block diagram of a software structure of the electronic device 100 according to the embodiment of the present application.
The layered architecture divides the software into several layers, each layer having a clear role and division of labor. The layers communicate with each other through a software interface. In some embodiments, the Android system is divided into four layers, an application layer, an application framework layer, an Android runtime (Android runtime) and system library, and a kernel layer from top to bottom.
The application layer may include a series of application packages.
As shown in fig. 2, the application package may include applications such as camera, gallery, calendar, phone call, map, navigation, WLAN, bluetooth, music, video, short message, etc.
The application framework layer provides an Application Programming Interface (API) and a programming framework for the application program of the application layer. The application framework layer includes a number of predefined functions.
As shown in FIG. 2, the application framework layers may include a window manager, content provider, view system, phone manager, resource manager, notification manager, and the like.
The window manager is used for managing window programs. The window manager can obtain the size of the display screen, judge whether a status bar exists, lock the screen, intercept the screen and the like.
The content provider is used to store and retrieve data and make it accessible to applications. The data may include video, images, audio, calls made and received, browsing history and bookmarks, phone books, etc.
The view system includes visual controls such as controls to display text, controls to display pictures, and the like. The view system may be used to build applications. The display interface may be composed of one or more views. For example, the display interface including the short message notification icon may include a view for displaying text and a view for displaying pictures.
The phone manager is used to provide communication functions of the electronic device 100. Such as management of call status (including on, off, etc.).
The resource manager provides various resources for the application, such as localized strings, icons, pictures, layout files, video files, and the like.
The notification manager enables the application to display notification information in the status bar, can be used to convey notification-type messages, can disappear automatically after a short dwell, and does not require user interaction. Such as a notification manager used to inform download completion, message alerts, etc. The notification manager may also be a notification that appears in the form of a chart or scroll bar text at the top status bar of the system, such as a notification of a background running application, or a notification that appears on the screen in the form of a dialog window. For example, prompting text information in the status bar, sounding a prompt tone, vibrating the electronic device, flashing an indicator light, etc.
The Android Runtime comprises a core library and a virtual machine. The Android runtime is responsible for scheduling and managing an Android system.
The core library comprises two parts: one part is a function which needs to be called by java language, and the other part is a core library of android.
The application layer and the application framework layer run in a virtual machine. The virtual machine executes java files of the application layer and the application framework layer as binary files. The virtual machine is used for performing the functions of object life cycle management, stack management, thread management, safety and exception management, garbage collection and the like.
The system library may include a plurality of functional modules. For example: surface managers (surface managers), media Libraries (Media Libraries), three-dimensional graphics processing Libraries (e.g., openGL ES), 2D graphics engines (e.g., SGL), and the like.
The surface manager is used to manage the display subsystem and provide a fusion of the 2D and 3D layers for multiple applications.
The media library supports a variety of commonly used audio, video format playback and recording, and still image files, among others. The media library may support a variety of audio-video encoding formats such as MPEG4, h.264, MP3, AAC, AMR, JPG, PNG, etc.
The three-dimensional graphic processing library is used for realizing three-dimensional graphic drawing, image rendering, synthesis, layer processing and the like.
The 2D graphics engine is a drawing engine for 2D drawing.
The kernel layer is a layer between hardware and software. The inner core layer at least comprises a display driver, a camera driver, an audio driver and a sensor driver.
The following describes exemplary workflow of the software and hardware of the electronic device 100 in connection with capturing a photo scene.
When the touch sensor 180K receives a touch operation, a corresponding hardware interrupt is issued to the kernel layer. The kernel layer processes the touch operation into an original input event (including touch coordinates, timestamp of the touch operation, and the like). The raw input events are stored at the kernel layer. And the application program framework layer acquires the original input event from the kernel layer and identifies the control corresponding to the input event. Taking the touch operation as a touch click operation, and taking a control corresponding to the click operation as a control of a camera application icon as an example, the camera application calls an interface of an application framework layer, starts the camera application, further starts a camera drive by calling a kernel layer, and captures a still image or a video through the camera 193.
At present, with the progress of science and technology, the camera function of electronic equipment is more and more powerful, the number of pixels of the shooting device is also more and more high, and the image is more and more clear. Taking a mobile phone terminal as an example, the shooting function is more and more comprehensive, and the mobile phone shooting becomes the first choice for people to shoot daily due to the characteristic of portability. When a portrait is shot at night or in an environment with weak light, shooting is usually carried out through screen light supplement, and the shooting effect under dark light is improved; when a user wearing glasses shoots a portrait through screen light supplement, colored light spots can appear on the lenses due to light reflection of the glasses in images collected by the electronic equipment, a local overexposure state is presented, eye details are lost, and the shooting quality of the photos is affected.
The method for removing the glasses light spots is suitable for shooting scenes supplemented with light through screens of electronic devices. Referring to fig. 3 (a), a schematic diagram of a photographing scene provided in the embodiment of the present application is shown; when the shooting environment light is weak, the electronic equipment carries out light supplementing shooting by improving the brightness of the frame area of the screen on a preview interface. Fig. 3 (B) is a schematic diagram of another photographing scene provided in the embodiment of the present application; when shooting ambient light is weak, when the electronic equipment receives a shutter instruction, the face of a user is illuminated by the instant full white (or Automatic Exposure, AE) of the screen, and light supplementing shooting is performed by improving the brightness of the full screen area. In addition, when the screen is in a high-brightness state for portrait shooting, the glasses may reflect light, and the method for removing the light spots of the glasses provided by the embodiment of the application is also suitable for portrait self-shooting scenes when the screen is high in brightness.
According to the method for removing the light spots of the glasses, when the portrait is shot through screen light supplement, the preview image of the shot object can be obtained on the preview interface, the light spot areas in the preview image are marked, and after the sensor of the shooting device collects the image with the light spots, the marked light spot areas are removed, so that the target image with the light spots removed and the eye details reserved is obtained.
The method for marking the spot area in the preview image comprises the following steps: firstly, inputting a preview image into a trained light spot marking algorithm model to obtain an image which is output by the light spot marking algorithm model and marks a light spot area of the glasses, wherein the light spot marking algorithm model can be a neural network model and is obtained by training an image training set with light spots and an image training set without the light spots; secondly, determining a projection area of a screen of the shooting device on the glasses in a projection calculation mode, and further determining a light spot area on the glasses in the preview image according to the projection area.
Through above-mentioned mark mode, the distribution of facula in the prediction image that can be more accurate through the mark to the facula region, for later stage image processing provides reliable reference basis, improves later stage image processing's efficiency, and then improves the photo quality of the user of wearing glasses when shooing through the screen light filling.
In addition, the mode of removing the light spot area of the mark comprises the following steps: adjusting the brightness value and the gray value of the pixel points in the light spot area, and realizing smooth transition between the brightness value and the gray value of the pixel points in the light spot area and adjacent areas; the method for removing the light spot region further comprises the following steps: directly deleting the information of the marked pixel points of the light spot region, and supplementing the eye details of the corresponding light spot region through a trained detail backfill algorithm; thereby obtaining a target image which removes light spots and retains eye details.
A second way of marking the spot area in the preview image will be described in detail with reference to the drawings.
Referring to fig. 4, which is a schematic diagram of the principle of camera imaging provided in the embodiment of the present application, as shown in fig. 4, a photographing device acquires an image P ' of a photographing object P in a physical imaging plane, the photographing device is provided with a camera coordinate system O-x-y-z, the physical imaging plane is provided with an image coordinate system O ' -x ' -y ' -z ', and a pixel plane located in the physical imaging plane is provided with a pixel coordinate system O-u-v. By the imaging principle, when the focal length EFL of the photographing device, the object height of the photographing object, and the imaged image height are determined, the axial distance (object distance) between the photographing device and the photographing object, that is, the distance in the x-axis direction can be calculated.
Wherein, the image coordinate system O '-x' -y '-z' of the physical imaging plane and the pixel coordinate system O-u-v of the pixel plane can be converted through a preset coordinate relation.
In some embodiments of the application, before receiving a shutter instruction, the electronic device acquires a preview image of a shooting object, performs face detection on the preview image, and extracts facial feature points of a face region in the preview image. The facial feature points comprise feature points of the five sense organ regions and feature points of the face contour. Marking target feature information in the preview image according to the extracted facial feature points in the preview image; wherein the target characteristic information includes a first pupil distance.
Fig. 5 is a schematic diagram for determining an axial distance between a shooting device and a shooting object according to an embodiment of the present disclosure. The facial feature points include eye feature points, and eye feature information in the preview image is obtained according to the camera imaging principle in fig. 4. Extracting eye feature points in the preview image, acquiring coordinates of the eye feature points, and determining a first pupil distance in the preview image according to pixel coordinates or image coordinates of the eye feature points, as shown in fig. 5, a in a physical imaging plane; the first pupil distance can be represented by a pixel point, and can also be directly represented by a length unit.
In some embodiments of the present application, the electronic device stores a preset second pupil distance, which is a conventional human eye pupil distance, typically about 63 mm, as shown in fig. 5 as second pupil distance a. According to the imaging principle, the axial distance between the glasses of the shooting object and the shooting device, namely the distance in the x-axis direction can be calculated, and the calculation formula is as follows:
axial distance D = a/a OO' (1);
wherein, A is the second pupil distance, a is the first pupil distance in the preview image, and OO' is the focal length of the camera; when the electronic equipment previews, automatic focusing can be achieved when the face image is detected, and therefore the focal distance is determined.
In some embodiments, the projection area of the screen of the camera on the glasses of the photographic subject is determined by means of a three-dimensional geometric projection calculation.
Fig. 6 is a schematic diagram of a projection principle provided in the embodiment of the present application. The projection principle shown in fig. 6 is only exemplary and not limited to the projection manner shown in the figure. As shown in fig. 6, the projection mode includes oblique projection and projection based on the center point. As shown in fig. 6, the graph ABC in oblique projection is projected onto a projection surface through the projection direction of the projection line to obtain a projection graph ABC; the projection direction may also be a forward projection. Based on the projection of the central point, the center of the screen of the camera device is taken as the central point, the glasses area is taken as a projection surface, and the camera device can form a projection area on the projection surface according to the projection direction of the projection line. In determining the angle information of the photographing device, the angle information of the photographing object, and the axial distance between the photographing device and the photographing object, the coordinates of the projection area may be calculated in a three-dimensional coordinate system to determine the projection position and the projection size of the screen of the photographing device in the glasses area.
In some embodiments, a projection area of the photographing device on the glasses of the photographing object may be determined according to angle information of the photographing device, angle information of the photographing object, and a distance between the photographing device and the photographing object.
Fig. 7 is a schematic diagram illustrating a light reflection area determined on glasses of a photographic subject according to an embodiment of the present disclosure. And determining first angle information of the shooting object through the target characteristic information in the preview image. Determining the turning angle of the shooting object by the size ratio of the left eye to the right eye in the preview image; by the ratio of the length and width of the face contour in the preview image, the angle of the subject to be raised or lowered can be determined, thereby determining the first angle information of the subject in the three-dimensional coordinate system, as shown in fig. 7, the actual angle information (α 1, β 1, γ 1) of the subject determined from the target feature information in the preview image.
In addition, the actual angle information of the shot object can be determined through an angle algorithm model, and the angle algorithm model can be a neural network model obtained through multi-angle image training set training; the preview image is input to the angle algorithm model, and the angle algorithm model can output three-dimensional angle information of the shot object according to the characteristic information of the preview image.
In some embodiments, the second angle information in the three-dimensional coordinate system of the current state of the photographing apparatus, as shown in fig. 7, may be acquired through an orientation sensor or a gyroscope of the photographing apparatus, the acquired angle information (α 2, β 2, γ 2) of the photographing apparatus.
The value of the photographing device acquired by the gyroscope corresponding to a certain time of 1.585e +18 in the three-dimensional coordinate system as shown in table 1:
Time gyr_x gyr_y gyr_z
1.585E+18 0.091752 0.175842 -0.055501
TABLE 1
And determining the angle information of the shooting device at the current time in the three-dimensional coordinate system through the integral conversion relation between the gyroscope measurement value and the angle.
The first angle information of the shot object and the second angle information of the shooting device belong to angle information in the same reference coordinate system, and the angle information can be unified by changing the origin of the coordinate system; by combining the axial distance D between the shooting device and the shooting object and the screen size of the shooting device (the screen size of the shooting device is known), projection information of the screen of the shooting device on the glasses of the shooting object can be determined through projection calculation. As shown in fig. 7, the light reflection area on the glasses is a projection area of the screen of the photographing device on the glasses. The projection information includes a projection position and a projection size of the projection area.
In some embodiments, according to the imaging principle of the camera, the position of the light spot area in the preview image of the glasses area is calculated according to the projection position of the screen of the shooting device on the glasses of the shooting object and the focal length of the shooting device; calculating the size of a glasses area of a light spot area in a preview image according to the projection size of a screen of a shooting device on glasses of a shooting object and the focal length of the shooting device; thereby determining and marking the distribution information of the light spots in the preview image (as shown in fig. 8 for an example of an image corresponding to the marked light spot area).
Through the embodiment, the electronic equipment can determine the first angle information of the shooting object in the three-dimensional space coordinate system according to the target characteristic information in the preview image; determining second angle information of the shooting device in a three-dimensional space coordinate system according to the direction sensor; and calculating projection information of the screen of the shooting device on glasses of the shooting object according to the first angle information, the second angle information, the axial distance and the screen size of the shooting device, so as to preliminarily determine a region capable of generating light reflection on the shooting object. The problem of complexity of determining the light reflecting area in the photographing process is solved, the light reflecting area of the glasses of the photographing object can be predicted according to the angle information and the distance information before the shutter command is received, reliable data basis is provided for subsequent image processing, and the image processing efficiency is improved.
Through the embodiment of the application, the electronic equipment is in the shooting process, before receiving the shutter instruction, through the processing to the preview image, confirm the projection information of the screen of the shooting device on the glasses of the shooting object, further confirm the facula information in the preview image according to the projection information, the distribution and the size of the facula in the image can be well predicted and confirmed, more accurate and reliable reference basis is provided for the image processing in the later stage of shooting, the efficiency of the image processing in the later stage is improved, the cost of the image processing is reduced, the shooting effect and the picture quality of the user wearing the glasses when shooting through screen light supplement are further improved.
Fig. 8 is a schematic diagram of an image to be processed of a mark spot region according to an embodiment of the present application.
In some embodiments, the facial feature points comprise eye feature points; marking an eye region in the preview image according to the coordinates of the eye feature points; inputting the image of the marked eye region into the image recognition model, and obtaining an image of the marked glasses region output by the image recognition model (such as an image example corresponding to the marked glasses region shown in fig. 8); the image recognition model is obtained through training of a first image training set and a second image training set, the first image training set is an image training set with glasses, and the second image training set is an image training set without glasses.
Understandably, the image recognition model may be a countermeasure neural network model, and iterative training is performed on the countermeasure neural network model through the first image training set and the second image training set to obtain an image recognition model that can more accurately recognize the glasses area in the preview image, so as to obtain an image of the marked glasses area (such as the image example corresponding to the marked glasses area shown in fig. 8).
In some embodiments, the image for marking the glasses area is input to the neural network model, and an image for marking the light spot area output by the neural network model (for example, the image corresponding to the marking light spot area shown in fig. 8) is obtained; the neural network model is a model for marking a light spot area in the preview image and is obtained by training according to the light spot image training set.
Referring to fig. 9, which is a schematic view of a light spot mark of a multi-user shooting scene provided in the embodiment of the present application, when a multi-user shooting is performed through screen light supplement, an electronic device performs face detection, extracts facial feature points of the multi-user, and obtains target feature information of multiple objects in a preview image; with the above embodiment, it is determined whether or not a person wears glasses and the glasses area is marked. Determining eye feature points of a plurality of persons by extracting facial feature points, determining corresponding eye regions according to the eye feature points of the plurality of persons, and predicting and marking the glasses region according to the eye regions; or inputting the images for marking the plurality of eye regions into the image recognition model, and obtaining the image for marking at least one glasses region output by the image recognition model (such as the image example corresponding to the marked glasses region shown in fig. 9). The method comprises the steps that projection areas of a shooting device on glasses of a plurality of shooting objects are obtained, and light spot areas corresponding to the projection areas in a preview image are determined and marked according to the camera imaging principle; or inputting the image for marking the plurality of glasses areas into the neural network model, and obtaining the image for marking the plurality of light spot areas output by the neural network model (as shown in fig. 9, the image example for marking the light spot areas corresponds to). Therefore, the rapid marking of a plurality of glasses areas and a plurality of light spot areas in the preview image is realized.
Through the embodiment of the application, the electronic equipment is in the shooting process, before receiving the shutter instruction, through the processing to the preview image, confirm the projection information of the screen of the shooting device on the glasses of the shooting object, further confirm the facula information in the preview image according to the projection information, the distribution and the size of the facula in the image can be well predicted and confirmed, more accurate and reliable reference basis is provided for the image processing in the later stage of shooting, the efficiency of the image processing in the later stage is improved, the cost of the image processing is reduced, the shooting effect and the picture quality of the user wearing the glasses when shooting through screen light supplement are further improved.
Fig. 10 is a schematic diagram of a target image output after processing a spot area according to an embodiment of the present application. As shown in the figure, the image to be processed is obtained after the marking process, the image to be processed comprises the marked light spot area, the light spot area is removed through the post image processing, and the target image without the light spot area and with the image details reserved is obtained.
In some embodiments, removing the light spot region in the image to be processed to obtain a target image includes: and adjusting the brightness value and the gray value of the light spot area in the image to be processed according to preset control parameters to obtain the target image.
The control parameters comprise gray scale transition parameters of the light spot region and pixel points in an adjacent threshold range and the like, so that smooth transition between the light spot region and a peripheral adjacent image region is ensured, and sudden change of a picture is avoided.
In addition, the camera device can acquire a High-Dynamic Range (HDR) image through a sensor, can provide more Dynamic ranges and image details, and synthesizes a final HDR image according to Dynamic Range images of different exposure times and a low Dynamic Range image with optimal details corresponding to each exposure time; and more detailed information of the image can be reserved conveniently when the processing is carried out by adjusting the brightness value and the gray value of the light spot area in the later period.
In some embodiments, the neural network model is trained through a pre-manufactured image training set with glasses and an image training set without glasses, and the region of the glasses is determined through machine learning; training a neural network model through an image training set with light spots and an image training set without the light spots, and acquiring a light spot area on the glasses through machine learning; by utilizing a multi-frame exposure strategy, when the last frame is collected, an image without screen light supplement is obtained, a picture without light reflection is directly obtained, high-frequency component comparison is carried out on the picture without light reflection and the existing light reflection picture, and a light reflection region is marked; and (3) utilizing single frame noise reduction to brighten the area of the glasses in the non-reflection photo, and simultaneously utilizing the image to remove a light spot area on the glasses and backfill eye details.
Fig. 11 is a schematic flowchart of a method for removing light spots of glasses according to an embodiment of the present disclosure. As shown in fig. 11, the method comprises the steps of:
step S1101 of acquiring a preview image of a photographic subject before receiving a shutter instruction;
step S1102, marking a light spot area in the preview image according to the target characteristic information in the preview image;
step S1103, receiving the shutter instruction, and acquiring an image to be processed, wherein the image to be processed is an image for marking the light spot area;
and S1104, removing the spot area in the image to be processed to obtain a target image.
According to the embodiment of the application, in the photographing process, the electronic equipment marks the glasses light spot area in the preview image, and after the to-be-processed image for marking the glasses light spot is acquired, the glasses light spot area in the to-be-processed image is removed, so that the target image with the detail characteristics of the corresponding light spot area photographing object reserved is obtained; the light spot distribution and the light spot marking in the image are predicted, and the light spot area in the image is removed, so that the later image processing efficiency is improved; meanwhile, the quality of pictures shot by the user wearing the glasses through screen light supplement is improved.
It should be understood that, the sequence numbers of the steps in the foregoing embodiments do not imply an execution sequence, and the execution sequence of each process should be determined by its function and inherent logic, and should not constitute any limitation to the implementation process of the embodiments of the present application.
Corresponding to the embodiments of the method for removing a spectacle optical spot and the application scenario described in the foregoing embodiments, fig. 12 shows a block diagram of a device for removing a spectacle optical spot provided in another embodiment of the present application, and for convenience of description, only the portions related to the embodiments of the present application are shown.
Referring to fig. 12, the apparatus includes an acquisition unit 1201, a calculation unit 1202, an acquisition unit 1203, and a processing unit 1204. Wherein, each unit function is as follows:
an acquisition unit configured to acquire a preview image of a photographic subject before receiving a shutter instruction;
the calculating unit is used for marking a light spot area in the preview image according to the target characteristic information in the preview image;
the acquisition unit is used for receiving the shutter instruction and acquiring an image to be processed, wherein the image to be processed is an image for marking the light spot area;
and the processing unit is used for removing the light spot area in the image to be processed to obtain a target image.
Through the embodiment of the application, the electronic equipment is in the shooting process, before receiving the shutter instruction, through the processing to the preview image, confirm the projection information of the screen of the shooting device on the glasses of the shooting object, further confirm the facula information in the preview image according to the projection information, the distribution and the size of the facula in the image can be well predicted and confirmed, more accurate and reliable reference basis is provided for the post-processing of shooting, the efficiency of the post-processing of the image is improved, the cost of the image processing is reduced, the shooting effect and the picture quality of the user wearing the glasses when shooting through screen light supplement are further improved.
It should be clear to those skilled in the art that, for convenience and simplicity of description, the foregoing division of the functional units and modules is only used for illustration, and in practical applications, the above function distribution may be performed by different functional units and modules as needed, that is, the internal structure of the apparatus may be divided into different functional units or modules to perform all or part of the above described functions. Each functional unit and module in the embodiments may be integrated in one processing unit, or each unit may exist alone physically, or two or more units are integrated in one unit, and the integrated unit may be implemented in a form of hardware, or in a form of software functional unit. In addition, specific names of the functional units and modules are only used for distinguishing one functional unit from another, and are not used for limiting the protection scope of the present application. The specific working processes of the units and modules in the system may refer to the corresponding processes in the foregoing method embodiments, and are not described herein again.
The embodiment of the application also provides electronic equipment, which comprises one or more processors, a memory, a display screen and a shooting device; the memory, the display screen, the camera, and the one or more processors, the memory for storing computer program code, the computer program code comprising computer instructions; the computer instructions, when executed by the one or more processors, cause the electronic device to perform any method of removing eye glass spots. The electronic equipment can be terminal equipment such as a mobile phone, a tablet, a smart watch and the like.
Embodiments of the present application further provide a computer-readable storage medium, which stores a computer program, where the computer program includes instructions that, when executed on an electronic device, cause a computer or a processor to perform one or more steps of any one of the above methods.
The embodiment of the application also provides a computer program product containing instructions. When run on an electronic device or on a processor, the computer program product causes the computer or processor to perform one or more steps of any of the methods described above.
In the above embodiments, the implementation may be wholly or partially realized by software, hardware, firmware, or any combination thereof. When implemented in software, it may be implemented in whole or in part in the form of a computer program product. The computer program product includes one or more computer instructions. The procedures or functions described in accordance with the embodiments of the application are all or partially generated when the computer program instructions are loaded and executed on a computer. The computer may be a general purpose computer, a special purpose computer, a network of computers, or other programmable device. The computer instructions may be stored on or transmitted over a computer-readable storage medium. The computer instructions may be transmitted from one website site, computer, server, or data center to another website site, computer, server, or data center via wired (e.g., coaxial cable, fiber optics, digital subscriber line) or wireless (e.g., infrared, wireless, microwave, etc.). The computer-readable storage medium can be any available medium that can be accessed by a computer or a data storage device, such as a server, a data center, etc., that incorporates one or more of the available media. The usable medium may be a magnetic medium (e.g., floppy disk, hard disk, magnetic tape), an optical medium (e.g., DVD), or a semiconductor medium (e.g., solid State Disk (SSD)), among others.
Those skilled in the art can understand that all or part of the processes in the methods of the embodiments described above can be implemented by a computer program, which can be stored in a computer readable storage medium and can include the processes of the method embodiments described above when executed. And the aforementioned storage medium includes: various media capable of storing program codes, such as ROM or RAM, magnetic or optical disks, etc.
The integrated unit may be stored in a computer-readable storage medium if it is implemented in the form of a software functional unit and sold or used as a separate product. Based on such understanding, all or part of the processes in the methods of the embodiments described above can be implemented by a computer program, which can be stored in a computer-readable storage medium and can implement the steps of the embodiments of the methods described above when the computer program is executed by a processor. Wherein the computer program comprises computer program code, which may be in the form of source code, object code, an executable file or some intermediate form, etc. The computer-readable storage medium may include at least: any entity or device capable of carrying computer program code to a photographing apparatus/terminal apparatus, a recording medium, computer Memory, read-Only Memory (ROM), random Access Memory (RAM), electrical carrier wave signal, telecommunication signal, and software distribution medium. Such as a usb-drive, a removable hard drive, a magnetic or optical disk, etc. In certain jurisdictions, computer-readable media may not be an electrical carrier signal or a telecommunications signal in accordance with legislative and patent practice.
The Processor may be a Central Processing Unit (CPU), other general purpose Processor, a Digital Signal Processor (DSP), an Application Specific Integrated Circuit (ASIC), a Field-Programmable Gate Array (FPGA) or other Programmable logic device, discrete Gate or transistor logic, discrete hardware components, etc. A general purpose processor may be a microprocessor or the processor may be any conventional processor or the like.
The memory may include forms of volatile memory in a computer readable medium, random Access Memory (RAM) and/or non-volatile memory, such as Read Only Memory (ROM) or flash memory (flash RAM). The memory is an example of a computer-readable medium.
In the above embodiments, the descriptions of the respective embodiments have respective emphasis, and reference may be made to the related descriptions of other embodiments for parts that are not described or illustrated in a certain embodiment.
Those of ordinary skill in the art will appreciate that the various illustrative elements and algorithm steps described in connection with the embodiments disclosed herein may be implemented as electronic hardware or combinations of computer software and electronic hardware. Whether such functionality is implemented as hardware or software depends upon the particular application and design constraints imposed on the implementation. Skilled artisans may implement the described functionality in varying ways for each particular application, but such implementation decisions should not be interpreted as causing a departure from the scope of the present application.
In the embodiments provided in the present application, it should be understood that the disclosed apparatus/device and method may be implemented in other ways. For example, the above-described apparatus/device embodiments are merely illustrative, and for example, the division of the modules or units is only one type of logical function division, and other division manners may exist in actual implementation, for example, a plurality of units or components may be combined or integrated into another system, or some features may be omitted, or not executed. In addition, the shown or discussed mutual coupling or direct coupling or communication connection may be through some interfaces, indirect coupling or communication connection of devices or units, and may be in an electrical, mechanical or other form.
It will be understood that the terms "comprises" and/or "comprising," when used in this specification and the appended claims, specify the presence of stated features, integers, steps, operations, elements, and/or components, but do not preclude the presence or addition of one or more other features, integers, steps, operations, elements, components, and/or groups thereof.
It should also be understood that the term "and/or" as used in this specification and the appended claims refers to and includes any and all possible combinations of one or more of the associated listed items.
As used in this specification and the appended claims, the term "if" may be interpreted contextually as "when", "upon" or "in response to a determination" or "in response to a detection". Similarly, the phrase "if it is determined" or "if a [ described condition or event ] is detected" may be interpreted contextually to mean "upon determining" or "in response to determining" or "upon detecting [ described condition or event ]" or "in response to detecting [ described condition or event ]".
Furthermore, in the description of the present application and the appended claims, the terms "first," "second," "third," and the like are used for distinguishing between descriptions and not necessarily for describing a relative importance or importance.
Reference throughout this specification to "one embodiment" or "some embodiments," or the like, means that a particular feature, structure, or characteristic described in connection with the embodiment is included in one or more embodiments of the present application. Thus, appearances of the phrases "in one embodiment," "in some embodiments," "in other embodiments," or the like, in various places throughout this specification are not necessarily all referring to the same embodiment, but rather "one or more but not all embodiments" unless specifically stated otherwise. The terms "comprising," "including," "having," and variations thereof mean "including, but not limited to," unless expressly specified otherwise.
Finally, it should be noted that: the above embodiments are only used to illustrate the technical solutions of the present application, and not to limit the same; although the present application has been described in detail with reference to the foregoing embodiments, those of ordinary skill in the art will understand that: the technical solutions described in the foregoing embodiments may still be modified, or some or all of the technical features may be equivalently replaced; and these modifications or substitutions do not depart from the scope of the technical solutions of the embodiments of the present application.

Claims (12)

1. A method for removing flare from an eye, comprising:
acquiring a preview image of a shooting object before receiving a shutter instruction;
marking a light spot area in the preview image according to the target characteristic information in the preview image;
receiving the shutter instruction, and acquiring an image to be processed, wherein the image to be processed is an image for marking the light spot area;
removing the light spot area in the image to be processed to obtain a target image;
the marking the spot area in the preview image according to the target feature information in the preview image includes:
determining the axial distance between a shooting device and a shooting object according to target characteristic information in the preview image, wherein the target characteristic information comprises a glasses area in the preview image;
determining projection information of a screen of the shooting device on glasses of the shooting object according to the axial distance and the screen size information of the shooting device;
and determining a light spot area of the glasses area in the preview image according to the projection information and the focal length of the shooting device.
2. The method according to claim 1, before determining the axial distance between the shooting device and the shooting object according to the target characteristic information in the preview image, comprising:
extracting facial feature points of a human face area in the preview image;
and marking the target feature information in the preview image according to the facial feature points.
3. The method of claim 2, wherein the facial feature points comprise eye feature points;
the marking the target feature information in the preview image according to the facial feature points comprises:
and marking a first pupil distance in the preview image according to the coordinates of the eye feature points.
4. The method according to claim 3, wherein determining the axial distance between the shooting device and the shooting object according to the target feature information in the preview image comprises:
and calculating the axial distance between the shooting device and the glasses of the shooting object according to the first pupil distance, the preset second pupil distance and the focal length of the shooting device in the preview image.
5. The method of claim 1, wherein the determining projection information of the screen of the photographing apparatus on the glasses of the photographing object according to the axial distance and the screen size information of the photographing apparatus comprises:
determining first angle information of the shot object according to target characteristic information in the preview image;
acquiring second angle information of the shooting device through a direction sensor of the shooting device;
and determining the projection position and the projection size of the screen of the shooting device on the glasses of the shooting object according to the first angle information, the second angle information, the axial distance and the screen size of the shooting device.
6. The method of claim 1, wherein the projection information comprises a projection location and a projection size;
the determining a light spot area of the glasses area in the preview image according to the projection information and the focal length of the shooting device includes:
calculating the position of the light spot area in the glasses area in the preview image according to the projection position of the screen of the shooting device on the glasses of the shooting object and the focal length of the shooting device;
and calculating the size of the glasses area of the light spot area in the preview image according to the projection size of the screen of the shooting device on the glasses of the shooting object and the focal length of the shooting device.
7. The method of claim 2, wherein the facial feature points comprise eye feature points;
the marking the target feature information in the preview image according to the facial feature points comprises:
marking the eye region in the preview image according to the coordinates of the eye feature points;
inputting the image for marking the eye region into an image recognition model to obtain an image for marking the glasses region output by the image recognition model;
the image recognition model is obtained through training of a first image training set and a second image training set, the first image training set is an image training set with glasses, and the second image training set is an image training set without glasses.
8. The method according to claim 7, wherein after the inputting the image for marking the eye region to an image recognition model and obtaining the image for marking the glasses region output by the image recognition model, the method comprises:
inputting the image for marking the glasses area into a neural network model to obtain an image for marking a light spot area output by the neural network model;
and the neural network model is a model for marking a light spot area in the preview image and is obtained by training according to a light spot image training set.
9. The method according to claim 1, wherein the removing the spot region in the image to be processed to obtain a target image comprises:
and adjusting the brightness value and the gray value of the light spot area in the image to be processed according to preset control parameters to obtain the target image.
10. An apparatus for removing flare from a pair of glasses, comprising:
an acquisition unit configured to acquire a preview image of a photographic subject before receiving a shutter instruction;
the calculating unit is used for marking a light spot area in the preview image according to the target characteristic information in the preview image, wherein the light spot area is in the glasses area in the preview image;
the acquisition unit is used for receiving the shutter instruction and acquiring an image to be processed, wherein the image to be processed is an image for marking the light spot area;
the processing unit is used for removing the light spot area in the image to be processed to obtain a target image;
the computing unit is specifically configured to:
determining the axial distance between a shooting device and a shooting object according to target characteristic information in the preview image, wherein the target characteristic information comprises a glasses area in the preview image;
determining projection information of a screen of the shooting device on glasses of the shooting object according to the axial distance and the screen size information of the shooting device;
and determining a light spot area of the glasses area in the preview image according to the projection information and the focal length of the shooting device.
11. An electronic device, comprising one or more processors, a memory, a display screen, and a camera;
the memory, the display screen, the camera, and the one or more processors are coupled, the memory for storing computer program code, the computer program code comprising computer instructions;
the computer instructions, when executed by the one or more processors, cause the electronic device to perform the method of removing eye glass spots of any one of claims 1-9.
12. A computer-readable storage medium, characterized in that the computer-readable storage medium stores a computer program comprising instructions which, when run on an electronic device, cause the electronic device to carry out the method of removing spectacle spots according to any one of claims 1 to 9.
CN202010323054.7A 2020-04-22 2020-04-22 Method and device for removing light spots of glasses and electronic equipment Active CN113542580B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202010323054.7A CN113542580B (en) 2020-04-22 2020-04-22 Method and device for removing light spots of glasses and electronic equipment

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202010323054.7A CN113542580B (en) 2020-04-22 2020-04-22 Method and device for removing light spots of glasses and electronic equipment

Publications (2)

Publication Number Publication Date
CN113542580A CN113542580A (en) 2021-10-22
CN113542580B true CN113542580B (en) 2022-10-28

Family

ID=78124099

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202010323054.7A Active CN113542580B (en) 2020-04-22 2020-04-22 Method and device for removing light spots of glasses and electronic equipment

Country Status (1)

Country Link
CN (1) CN113542580B (en)

Families Citing this family (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115580690B (en) * 2022-01-24 2023-10-20 荣耀终端有限公司 Image processing method and electronic equipment
CN115546041B (en) * 2022-02-28 2023-10-20 荣耀终端有限公司 Training method of light supplementing model, image processing method and related equipment thereof
CN114418897B (en) * 2022-03-10 2022-07-19 深圳市一心视觉科技有限公司 Eye spot image restoration method and device, terminal equipment and storage medium
CN117078558A (en) * 2023-09-22 2023-11-17 荣耀终端有限公司 Image processing method and electronic equipment

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103793719A (en) * 2014-01-26 2014-05-14 深圳大学 Monocular distance-measuring method and system based on human eye positioning
CN108093170A (en) * 2017-11-30 2018-05-29 广东欧珀移动通信有限公司 User's photographic method, device and equipment
CN108564540A (en) * 2018-03-05 2018-09-21 广东欧珀移动通信有限公司 Remove image processing method, device and the terminal device that eyeglass is reflective in image
CN110223237A (en) * 2019-04-23 2019-09-10 维沃移动通信有限公司 Adjust the method and terminal device of image parameter

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
FR3048788B1 (en) * 2016-03-11 2019-04-05 Valeo Vision SCREEN IMAGE PROJECTOR WITH LIGHT SOURCE WITH ELECTROLUMINESCENT STICKS

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103793719A (en) * 2014-01-26 2014-05-14 深圳大学 Monocular distance-measuring method and system based on human eye positioning
CN108093170A (en) * 2017-11-30 2018-05-29 广东欧珀移动通信有限公司 User's photographic method, device and equipment
CN108564540A (en) * 2018-03-05 2018-09-21 广东欧珀移动通信有限公司 Remove image processing method, device and the terminal device that eyeglass is reflective in image
CN110223237A (en) * 2019-04-23 2019-09-10 维沃移动通信有限公司 Adjust the method and terminal device of image parameter

Also Published As

Publication number Publication date
CN113542580A (en) 2021-10-22

Similar Documents

Publication Publication Date Title
CN110506416B (en) Method for switching camera by terminal and terminal
CN113132620B (en) Image shooting method and related device
CN112130742B (en) Full screen display method and device of mobile terminal
CN112887583B (en) Shooting method and electronic equipment
CN113542580B (en) Method and device for removing light spots of glasses and electronic equipment
WO2020029306A1 (en) Image capture method and electronic device
CN113810601B (en) Terminal image processing method and device and terminal equipment
CN113573390B (en) Antenna power adjusting method, terminal device and storage medium
CN114650363B (en) Image display method and electronic equipment
CN113810603B (en) Point light source image detection method and electronic equipment
CN111510626B (en) Image synthesis method and related device
CN114095666B (en) Photographing method, electronic device, and computer-readable storage medium
CN111741284A (en) Image processing apparatus and method
CN111580671A (en) Video image processing method and related device
CN110138999B (en) Certificate scanning method and device for mobile terminal
CN112150499A (en) Image processing method and related device
CN110286975B (en) Display method of foreground elements and electronic equipment
CN114283195B (en) Method for generating dynamic image, electronic device and readable storage medium
CN114079725B (en) Video anti-shake method, terminal device, and computer-readable storage medium
CN113542574A (en) Shooting preview method under zooming, terminal, storage medium and electronic equipment
CN116389884B (en) Thumbnail display method and terminal equipment
CN115686182B (en) Processing method of augmented reality video and electronic equipment
CN116708751B (en) Method and device for determining photographing duration and electronic equipment
CN116017138B (en) Light measuring control display method, computer equipment and storage medium
CN115150543B (en) Shooting method, shooting device, electronic equipment and readable storage medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant