WO2021143269A1 - 一种长焦场景下的拍摄方法及移动终端 - Google Patents
一种长焦场景下的拍摄方法及移动终端 Download PDFInfo
- Publication number
- WO2021143269A1 WO2021143269A1 PCT/CN2020/124545 CN2020124545W WO2021143269A1 WO 2021143269 A1 WO2021143269 A1 WO 2021143269A1 CN 2020124545 W CN2020124545 W CN 2020124545W WO 2021143269 A1 WO2021143269 A1 WO 2021143269A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- image
- mobile terminal
- camera
- target
- image processing
- Prior art date
Links
- 238000000034 method Methods 0.000 title claims abstract description 85
- 230000000694 effects Effects 0.000 claims abstract description 23
- 238000012545 processing Methods 0.000 claims description 88
- 230000006870 function Effects 0.000 claims description 49
- 230000008569 process Effects 0.000 claims description 20
- 238000004422 calculation algorithm Methods 0.000 claims description 18
- 238000001914 filtration Methods 0.000 claims description 13
- 238000012549 training Methods 0.000 claims description 10
- 238000004590 computer program Methods 0.000 claims description 9
- 238000005516 engineering process Methods 0.000 claims description 9
- 230000004044 response Effects 0.000 claims description 9
- 238000000605 extraction Methods 0.000 claims description 6
- 238000003062 neural network model Methods 0.000 claims description 4
- 238000004891 communication Methods 0.000 abstract description 35
- 230000000007 visual effect Effects 0.000 abstract description 10
- 238000007726 management method Methods 0.000 description 21
- 230000005236 sound signal Effects 0.000 description 13
- 241001137251 Corvidae Species 0.000 description 11
- 238000010295 mobile communication Methods 0.000 description 11
- 238000010586 diagram Methods 0.000 description 9
- 241000894007 species Species 0.000 description 8
- 230000008859 change Effects 0.000 description 7
- 229920001621 AMOLED Polymers 0.000 description 4
- 239000003550 marker Substances 0.000 description 4
- 238000013528 artificial neural network Methods 0.000 description 3
- 238000013527 convolutional neural network Methods 0.000 description 3
- 238000013461 design Methods 0.000 description 3
- 230000003287 optical effect Effects 0.000 description 3
- 238000012546 transfer Methods 0.000 description 3
- 241000271566 Aves Species 0.000 description 2
- 241001465754 Metazoa Species 0.000 description 2
- 230000006399 behavior Effects 0.000 description 2
- 230000001413 cellular effect Effects 0.000 description 2
- 239000003086 colorant Substances 0.000 description 2
- 238000013500 data storage Methods 0.000 description 2
- 238000001514 detection method Methods 0.000 description 2
- 230000007613 environmental effect Effects 0.000 description 2
- 230000002093 peripheral effect Effects 0.000 description 2
- 238000009877 rendering Methods 0.000 description 2
- 230000008439 repair process Effects 0.000 description 2
- 241000272517 Anseriformes Species 0.000 description 1
- 241000272201 Columbiformes Species 0.000 description 1
- 241000196324 Embryophyta Species 0.000 description 1
- 241000282326 Felis catus Species 0.000 description 1
- 241000282412 Homo Species 0.000 description 1
- 241000287127 Passeridae Species 0.000 description 1
- 241000009328 Perro Species 0.000 description 1
- 241000555745 Sciuridae Species 0.000 description 1
- 230000001133 acceleration Effects 0.000 description 1
- 230000009471 action Effects 0.000 description 1
- 239000008186 active pharmaceutical agent Substances 0.000 description 1
- 230000003416 augmentation Effects 0.000 description 1
- 230000003190 augmentative effect Effects 0.000 description 1
- 230000003796 beauty Effects 0.000 description 1
- 230000002457 bidirectional effect Effects 0.000 description 1
- 238000013529 biological neural network Methods 0.000 description 1
- 230000015572 biosynthetic process Effects 0.000 description 1
- 210000000988 bone and bone Anatomy 0.000 description 1
- 210000004556 brain Anatomy 0.000 description 1
- 238000004364 calculation method Methods 0.000 description 1
- 230000019771 cognition Effects 0.000 description 1
- 230000000295 complement effect Effects 0.000 description 1
- 238000013135 deep learning Methods 0.000 description 1
- 230000003862 health status Effects 0.000 description 1
- 238000010191 image analysis Methods 0.000 description 1
- 238000003709 image segmentation Methods 0.000 description 1
- 230000003993 interaction Effects 0.000 description 1
- 239000004973 liquid crystal related substance Substances 0.000 description 1
- 230000007774 longterm Effects 0.000 description 1
- 210000002569 neuron Anatomy 0.000 description 1
- 238000003672 processing method Methods 0.000 description 1
- 239000002096 quantum dot Substances 0.000 description 1
- 230000005855 radiation Effects 0.000 description 1
- 230000009467 reduction Effects 0.000 description 1
- 239000004065 semiconductor Substances 0.000 description 1
- 238000000926 separation method Methods 0.000 description 1
- 238000006467 substitution reaction Methods 0.000 description 1
- 230000001360 synchronised effect Effects 0.000 description 1
- 238000003786 synthesis reaction Methods 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F1/00—Details not covered by groups G06F3/00 - G06F13/00 and G06F21/00
- G06F1/16—Constructional details or arrangements
- G06F1/1613—Constructional details or arrangements for portable computers
- G06F1/1633—Constructional details or arrangements of portable computers not specific to the type of enclosures covered by groups G06F1/1615 - G06F1/1626
- G06F1/1684—Constructional details or arrangements related to integrated I/O peripherals not covered by groups G06F1/1635 - G06F1/1675
- G06F1/1686—Constructional details or arrangements related to integrated I/O peripherals not covered by groups G06F1/1635 - G06F1/1675 the I/O peripheral being an integrated camera
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/60—Control of cameras or camera modules
- H04N23/67—Focus control based on electronic image sensor signals
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0484—Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range
- G06F3/04845—Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range for image manipulation, e.g. dragging, rotation, expansion or change of colour
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0484—Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range
- G06F3/04847—Interaction techniques to control parameter settings, e.g. interaction with sliders or dials
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/60—Control of cameras or camera modules
- H04N23/61—Control of cameras or camera modules based on recognised objects
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/60—Control of cameras or camera modules
- H04N23/62—Control of parameters via user interfaces
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/60—Control of cameras or camera modules
- H04N23/63—Control of cameras or camera modules by using electronic viewfinders
- H04N23/631—Graphical user interfaces [GUI] specially adapted for controlling image capture or setting capture parameters
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/60—Control of cameras or camera modules
- H04N23/63—Control of cameras or camera modules by using electronic viewfinders
- H04N23/631—Graphical user interfaces [GUI] specially adapted for controlling image capture or setting capture parameters
- H04N23/632—Graphical user interfaces [GUI] specially adapted for controlling image capture or setting capture parameters for displaying or modifying preview images prior to image capturing, e.g. variety of image resolutions or capturing parameters
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/60—Control of cameras or camera modules
- H04N23/63—Control of cameras or camera modules by using electronic viewfinders
- H04N23/633—Control of cameras or camera modules by using electronic viewfinders for displaying additional information relating to control or operation of the camera
- H04N23/635—Region indicators; Field of view indicators
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/60—Control of cameras or camera modules
- H04N23/64—Computer-aided capture of images, e.g. transfer from script file into camera, check of taken image quality, advice or proposal for image composition or decision on when to take image
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/60—Control of cameras or camera modules
- H04N23/69—Control of means for changing angle of the field of view, e.g. optical zoom objectives or electronic zooming
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/80—Camera processing pipelines; Components thereof
Definitions
- This application relates to the technical field of mobile terminals, and in particular to a shooting method and a mobile terminal in a telephoto scene.
- mobile phones can integrate multiple lenses, such as wide-angle cameras, medium-focus cameras, and telephoto cameras, to meet different shooting scenarios of users.
- the telephoto camera can be used to shoot objects far away from the user (ie, telephoto shooting scenes).
- the sharpness of the subject in the image obtained in the telephoto shooting scene is not high, and the subject is also It is not prominent, and the visual effect is not good.
- the shooting method and mobile terminal in a telephoto scene provided by the present application can improve the definition of the shooting object in the image, blur the background, highlight the shooting object, and improve the overall visual effect of the image.
- a shooting method in a telephoto scene is provided, which is applied to a mobile terminal including a camera.
- the method includes: the mobile terminal starts the camera, the mobile terminal displays a viewfinder frame, and the viewfinder frame is used to display a first preview image.
- the zoom magnification of the camera corresponding to a preview picture is the first magnification; the first operation to increase the zoom magnification of the camera input by the user is received; in response to the first operation, the viewfinder frame displays the second preview picture, and the second preview picture corresponds to the camera’s zoom magnification.
- the zoom magnification is the second magnification, and the second magnification is greater than the first magnification; if the second magnification is greater than or equal to the preset magnification, before the mobile terminal displays the second preview screen, the mobile terminal performs the first image processing on the image collected by the camera, Generate a second preview image; wherein, the first image processing includes: identifying the target subject and the background of the target subject according to the image collected by the camera; improving the image clarity of the area where the target subject is located, and the area where the background of the target subject is located Add a blur effect to the image.
- the embodiment of the present application provides a shooting method in a telephoto shooting scene, which can intelligently identify the target shooting object, enhance the details of the target shooting object, and improve the clarity of the part of the image of the shooting object in the collected original image. It can also automatically recognize the front and back scenes of the image according to the identified subject, blur the background outside the subject, highlight the target subject, and enhance the artistic sense of the image in the telephoto shooting scene, and enhance the user's visual experience.
- the method further includes: the mobile terminal displays a marker frame on the second preview screen for marking the area where the target photographed object is located.
- the user can further determine whether the target photographed object automatically recognized by the mobile terminal is the object to be photographed, and whether the automatically recognized target photographed object is complete, etc., according to the marked frame.
- the method further includes: the mobile terminal displays prompt information in the second preview screen for recommending the first image processing corresponding to the target shooting object.
- the mobile terminal performs the first image processing on the image captured by the camera, including: the mobile terminal receives a second operation input by the user to select the first image processing; in response to the second operation, the mobile terminal performs the first image processing on the camera The first image processing is performed on the image.
- adjusting the position of the target photographing object in the second preview image is specifically: adjusting the second preview image so that the target photographing object is located in the central area of the second preview image.
- the background of the shooting object in the image can be cropped or filled based on the identified position of the shooting object in the image, so that the position of the shooting object is located at the center of the image.
- the position of the photographed object may also be located in another position of the image, for example, used for a certain distance to the left or right of the center position, which is not limited in the embodiment of the application.
- the method further includes: receiving a third operation input by the user, the third operation is used to instruct the mobile terminal to turn off the first image processing function corresponding to the target shooting object ; In response to the third operation, the mobile terminal determines not to perform the first image processing on the image collected by the camera.
- increasing the image clarity of the area where the target photographed object is located is specifically: identifying the category of the target photographing object, and segmenting the first image of the target photographing object from the image collected by the camera.
- An image; the first image is input into the neural network model corresponding to the category of the target shooting object, and the second image of the target shooting object is output, wherein the definition of the second image is greater than the definition of the first image.
- the neural network model corresponding to the category of the target shooting object is obtained by training based on the images of multiple shooting objects in the category of the target shooting object.
- the mobile phone can pre-train an AI model that enhances the details of the subject based on the category of the subject.
- the mobile phone can also directly obtain the trained AI model from the server.
- training an AI model that enhances the details of the subject based on the category of the subject it is necessary to input a large number of training samples of the subject, including training samples of each category. For example, training samples for magpies in birds, including images of different species of magpies, images of different sizes or colors of the same species of magpies, images of different poses of the same species of magpies, etc., images of the same species of magpies in different environments Wait.
- the detail enhancement of the photographed object may include the use of an AI model for intelligent pixel filling of the blurred area in the photographed object, etc., to enhance the clarity of the image. It also includes: for the missing parts of the subject, the AI model is used for intelligent pixel filling, etc., to repair the missing parts of the subject, etc. It may also include: improving the overall sharpness of the image of the part of the subject.
- the first image processing to identify the background of the target subject is specifically: using a background extraction algorithm to identify the background of the target subject from the image collected by the camera, where the background extraction algorithm includes: frame Any one or several of the time difference method, background difference method, and environmental algorithm.
- the first image processing further includes: when the camera collects an image containing the target shooting object, automatically focusing on the target shooting object.
- focusing and exposure are generally performed based on a fixed position, or based on a position selected by the user (for example, a touch position of a click operation).
- focusing and exposure are performed based on the position of the target photographed object automatically recognized by the mobile phone, so as to improve the overall clarity of the target photographed object.
- the size of the focus frame is fixed, and is usually set by default in the mobile phone.
- the size of the focus frame will be automatically adjusted according to the size of the target shooting object recognized by the mobile terminal.
- the focus frame may be the same as the marker frame.
- the closest search distance during auto-focusing of the mobile terminal may be set, for example, 10 meters. That is to say, when the mobile terminal is auto-focusing, the search distance range is 10 meters away from the mobile terminal to infinity. In other words, the mobile terminal does not need to search in the range of 10 meters to the macro distance from the mobile terminal. As a result, the time for the mobile terminal to automatically focus can be reduced, and the efficiency of image processing can be improved.
- the first image processing further includes: controlling the camera to continuously collect N images containing the target subject, and using super-resolution technology to synthesize the N images containing the target subject into a third image;
- the third image enhances the image clarity of the area where the target photographed object is located, and adds a blur effect to the image of the area where the target photographed object is located.
- the target subject in the image is further highlighted, and the overall artistic sense of the image is enhanced.
- the first image processing is to add a blur effect to the image of the area where the background of the target object is located, specifically: using a blur algorithm to process the image of the area where the background of the target object is located, and the blur algorithm includes Gaussian filtering. Any one or several of, circular filtering, guided filtering, and domain filtering.
- a mobile terminal including: a processor, a memory, and a touch screen, the memory, the touch screen are coupled to the processor, and the memory is used to store computer program code, the computer program code including a computer Instruction, when the processor reads the computer instruction from the memory, so that the mobile terminal executes the method described in the foregoing aspects and any one of the possible implementation manners.
- a device in a third aspect, is provided, the device is included in a mobile terminal, and the device has the function of realizing the behavior of the mobile terminal in any of the foregoing aspects and possible implementation manners.
- This function can be realized by hardware, or by hardware executing corresponding software.
- the hardware or software includes at least one module or unit corresponding to the above-mentioned functions. For example, a receiving module or unit, a display module or unit, and a processing module or unit, etc.
- a chip system including a processor, and when the processor executes an instruction, the processor executes the method described in the foregoing aspects and any one of the possible implementation manners.
- a computer-readable storage medium including computer instructions, which, when the computer instructions run on a mobile terminal, cause the mobile terminal to execute the method described in the above aspects and any one of its possible implementation manners.
- a computer program product which when the computer program product runs on a computer, causes the computer to execute the method described in the foregoing aspects and any one of the possible implementation manners.
- FIG. 1 is a first structural diagram of a mobile terminal provided by an embodiment of this application.
- FIG. 2 is a second structural diagram of a mobile terminal provided by an embodiment of this application.
- FIG. 3 is a schematic diagram of user interfaces of some mobile terminals provided by embodiments of the application.
- FIG. 4 is a schematic diagram of user interfaces of other mobile terminals provided by embodiments of the application.
- FIG. 5 is a schematic diagram of user interfaces of still other mobile terminals provided by the embodiments of this application.
- FIG. 6 is a schematic diagram of user interfaces of still other mobile terminals provided by embodiments of this application.
- FIG. 7 is a schematic structural diagram of a chip system provided by an embodiment of the application.
- words such as “exemplary” or “for example” are used as examples, illustrations, or illustrations. Any embodiment or design solution described as “exemplary” or “for example” in the embodiments of the present application should not be construed as being more preferable or advantageous than other embodiments or design solutions. To be precise, words such as “exemplary” or “for example” are used to present related concepts in a specific manner.
- first and second are only used for descriptive purposes, and cannot be understood as indicating or implying relative importance or implicitly indicating the number of indicated technical features. Thus, the features defined with “first” and “second” may explicitly or implicitly include one or more of these features. In the description of the embodiments of the present application, unless otherwise specified, “plurality” means two or more.
- the shooting method provided in the embodiments of the present application can be applied to a mobile terminal equipped with a camera.
- the mobile terminal can be, for example, a mobile phone, a tablet computer, a personal computer (PC), a personal digital assistant (PDA), Smart watches, netbooks, wearable electronic devices, augmented reality (AR) devices, virtual reality (VR) devices, in-vehicle devices, smart cars, smart speakers, robots, etc.
- AR augmented reality
- VR virtual reality
- FIG. 1 shows a schematic diagram of the structure of the mobile terminal 100.
- the mobile terminal 100 may include a processor 110, an external memory interface 120, an internal memory 121, a universal serial bus (USB) interface 130, a charging management module 140, a power management module 141, a battery 142, an antenna 1, and an antenna 2.
- Mobile communication module 150 wireless communication module 160, audio module 170, speaker 170A, receiver 170B, microphone 170C, earphone jack 170D, sensor module 180, buttons 190, motor 191, indicator 192, camera 193, display screen 194, and Subscriber identification module (subscriber identification module, SIM) card interface 195, etc.
- SIM Subscriber identification module
- the sensor module 180 may include a pressure sensor 180A, a gyroscope sensor 180B, an air pressure sensor 180C, a magnetic sensor 180D, an acceleration sensor 180E, a distance sensor 180F, a proximity light sensor 180G, a fingerprint sensor 180H, a temperature sensor 180J, a touch sensor 180K, and ambient light Sensor 180L, bone conduction sensor 180M, etc.
- the structure illustrated in the embodiment of the present invention does not constitute a specific limitation on the mobile terminal 100.
- the mobile terminal 100 may include more or fewer components than shown, or combine certain components, or split certain components, or arrange different components.
- the illustrated components can be implemented in hardware, software, or a combination of software and hardware.
- the processor 110 may include one or more processing units.
- the processor 110 may include an application processor (AP), a modem processor, a graphics processing unit (GPU), and an image signal processor. (image signal processor, ISP), controller, video codec, digital signal processor (digital signal processor, DSP), baseband processor, and/or neural-network processing unit (NPU), etc.
- AP application processor
- modem processor modem processor
- GPU graphics processing unit
- image signal processor image signal processor
- ISP image signal processor
- controller video codec
- digital signal processor digital signal processor
- DSP digital signal processor
- NPU neural-network processing unit
- the different processing units may be independent devices or integrated in one or more processors.
- the controller can generate operation control signals according to the instruction operation code and timing signals to complete the control of fetching instructions and executing instructions.
- a memory may also be provided in the processor 110 to store instructions and data.
- the memory in the processor 110 is a cache memory.
- the memory can store instructions or data that have just been used or recycled by the processor 110. If the processor 110 needs to use the instruction or data again, it can be directly called from the memory. Repeated accesses are avoided, the waiting time of the processor 110 is reduced, and the efficiency of the system is improved.
- the processor 110 may include one or more interfaces.
- the interface can include an integrated circuit (inter-integrated circuit, I2C) interface, an integrated circuit built-in audio (inter-integrated circuit sound, I2S) interface, a pulse code modulation (pulse code modulation, PCM) interface, and a universal asynchronous transmitter (universal asynchronous) interface.
- I2C integrated circuit
- I2S integrated circuit built-in audio
- PCM pulse code modulation
- UART universal asynchronous transmitter
- MIPI mobile industry processor interface
- GPIO general-purpose input/output
- SIM subscriber identity module
- USB Universal Serial Bus
- the I2C interface is a bidirectional synchronous serial bus, including a serial data line (SDA) and a serial clock line (SCL).
- the processor 110 may include multiple sets of I2C buses.
- the processor 110 may be coupled to the touch sensor 180K, charger, flash, camera 193, etc., respectively through different I2C bus interfaces.
- the processor 110 may couple the touch sensor 180K through an I2C interface, so that the processor 110 and the touch sensor 180K communicate through an I2C bus interface to implement the touch function of the mobile terminal 100.
- the I2S interface can be used for audio communication.
- the processor 110 may include multiple sets of I2S buses.
- the processor 110 may be coupled with the audio module 170 through an I2S bus to implement communication between the processor 110 and the audio module 170.
- the audio module 170 may transmit audio signals to the wireless communication module 160 through an I2S interface, so as to realize the function of answering calls through a Bluetooth headset.
- the PCM interface can also be used for audio communication to sample, quantize and encode analog signals.
- the audio module 170 and the wireless communication module 160 may be coupled through a PCM bus interface.
- the audio module 170 may also transmit audio signals to the wireless communication module 160 through the PCM interface, so as to realize the function of answering calls through the Bluetooth headset. Both the I2S interface and the PCM interface can be used for audio communication.
- the UART interface is a universal serial data bus used for asynchronous communication.
- the bus can be a two-way communication bus. It converts the data to be transmitted between serial communication and parallel communication.
- the UART interface is generally used to connect the processor 110 and the wireless communication module 160.
- the processor 110 communicates with the Bluetooth module in the wireless communication module 160 through the UART interface to realize the Bluetooth function.
- the audio module 170 may transmit audio signals to the wireless communication module 160 through a UART interface, so as to realize the function of playing music through a Bluetooth headset.
- the MIPI interface can be used to connect the processor 110 with the display screen 194, the camera 193 and other peripheral devices.
- the MIPI interface includes a camera serial interface (camera serial interface, CSI), a display serial interface (display serial interface, DSI), and so on.
- the processor 110 and the camera 193 communicate through a CSI interface to implement the shooting function of the mobile terminal 100.
- the processor 110 and the display screen 194 communicate through a DSI interface to realize the display function of the mobile terminal 100.
- the GPIO interface can be configured through software.
- the GPIO interface can be configured as a control signal or as a data signal.
- the GPIO interface can be used to connect the processor 110 with the camera 193, the display screen 194, the wireless communication module 160, the audio module 170, the sensor module 180, and so on.
- the GPIO interface can also be configured as an I2C interface, I2S interface, UART interface, MIPI interface, etc.
- the USB interface 130 is an interface that complies with the USB standard specification, and specifically may be a Mini USB interface, a Micro USB interface, a USB Type C interface, and so on.
- the USB interface 130 can be used to connect a charger to charge the mobile terminal 100, and can also be used to transfer data between the mobile terminal 100 and peripheral devices. It can also be used to connect earphones and play audio through earphones. This interface can also be used to connect to other mobile terminals, such as AR devices.
- the interface connection relationship between the modules illustrated in the embodiment of the present invention is merely a schematic description, and does not constitute a structural limitation of the mobile terminal 100.
- the mobile terminal 100 may also adopt different interface connection modes or a combination of multiple interface connection modes in the foregoing embodiments.
- the charging management module 140 is used to receive charging input from the charger.
- the charger can be a wireless charger or a wired charger.
- the charging management module 140 may receive the charging input of the wired charger through the USB interface 130.
- the charging management module 140 may receive the wireless charging input through the wireless charging coil of the mobile terminal 100. While the charging management module 140 charges the battery 142, it can also supply power to the mobile terminal through the power management module 141.
- the power management module 141 is used to connect the battery 142, the charging management module 140 and the processor 110.
- the power management module 141 receives input from the battery 142 and/or the charging management module 140, and supplies power to the processor 110, the internal memory 121, the display screen 194, the camera 193, and the wireless communication module 160.
- the power management module 141 can also be used to monitor parameters such as battery capacity, battery cycle times, and battery health status (leakage, impedance).
- the power management module 141 may also be provided in the processor 110.
- the power management module 141 and the charging management module 140 may also be provided in the same device.
- the wireless communication function of the mobile terminal 100 may be implemented by the antenna 1, the antenna 2, the mobile communication module 150, the wireless communication module 160, the modem processor, and the baseband processor.
- the antenna 1 and the antenna 2 are used to transmit and receive electromagnetic wave signals.
- Each antenna in the mobile terminal 100 can be used to cover a single or multiple communication frequency bands. Different antennas can also be reused to improve antenna utilization.
- Antenna 1 can be multiplexed as a diversity antenna of a wireless local area network.
- the antenna can be used in combination with a tuning switch.
- the mobile communication module 150 may provide a wireless communication solution including 2G/3G/4G/5G and the like applied to the mobile terminal 100.
- the mobile communication module 150 may include at least one filter, a switch, a power amplifier, a low noise amplifier (LNA), and the like.
- the mobile communication module 150 can receive electromagnetic waves by the antenna 1, and perform processing such as filtering, amplifying and transmitting the received electromagnetic waves to the modem processor for demodulation.
- the mobile communication module 150 can also amplify the signal modulated by the modem processor, and convert it into electromagnetic waves for radiation via the antenna 1.
- at least part of the functional modules of the mobile communication module 150 may be provided in the processor 110.
- at least part of the functional modules of the mobile communication module 150 and at least part of the modules of the processor 110 may be provided in the same device.
- the modem processor may include a modulator and a demodulator.
- the modulator is used to modulate the low frequency baseband signal to be sent into a medium and high frequency signal.
- the demodulator is used to demodulate the received electromagnetic wave signal into a low-frequency baseband signal.
- the demodulator then transmits the demodulated low-frequency baseband signal to the baseband processor for processing.
- the application processor outputs a sound signal through an audio device (not limited to the speaker 170A, the receiver 170B, etc.), or displays an image or video through the display screen 194.
- the modem processor may be an independent device.
- the modem processor may be independent of the processor 110 and be provided in the same device as the mobile communication module 150 or other functional modules.
- the wireless communication module 160 can provide applications on the mobile terminal 100 including wireless local area networks (WLAN) (such as wireless fidelity (Wi-Fi) networks), bluetooth (BT), and global navigation satellites. System (global navigation satellite system, GNSS), frequency modulation (FM), near field communication (NFC), infrared technology (infrared, IR) and other wireless communication solutions.
- the wireless communication module 160 may be one or more devices integrating at least one communication processing module.
- the wireless communication module 160 receives electromagnetic waves via the antenna 2, frequency modulates and filters the electromagnetic wave signals, and sends the processed signals to the processor 110.
- the wireless communication module 160 may also receive a signal to be sent from the processor 110, perform frequency modulation, amplify, and convert it into electromagnetic waves to radiate through the antenna 2.
- the antenna 1 of the mobile terminal 100 is coupled with the mobile communication module 150, and the antenna 2 is coupled with the wireless communication module 160, so that the mobile terminal 100 can communicate with the network and other devices through wireless communication technology.
- the wireless communication technology may include global system for mobile communications (GSM), general packet radio service (GPRS), code division multiple access (CDMA), broadband Code division multiple access (wideband code division multiple access, WCDMA), time-division code division multiple access (TD-SCDMA), long term evolution (LTE), BT, GNSS, WLAN, NFC , FM, and/or IR technology, etc.
- the GNSS may include global positioning system (GPS), global navigation satellite system (GLONASS), Beidou navigation satellite system (BDS), quasi-zenith satellite system (quasi -zenith satellite system, QZSS) and/or satellite-based augmentation systems (SBAS).
- GPS global positioning system
- GLONASS global navigation satellite system
- BDS Beidou navigation satellite system
- QZSS quasi-zenith satellite system
- SBAS satellite-based augmentation systems
- the mobile terminal 100 implements a display function through a GPU, a display screen 194, and an application processor.
- the GPU is an image processing microprocessor, which is connected to the display screen 194 and the application processor.
- the GPU is used to perform mathematical and geometric calculations and is used for graphics rendering.
- the processor 110 may include one or more GPUs that execute program instructions to generate or change display information.
- the display screen 194 is used to display images, videos, and the like.
- the display screen 194 includes a display panel.
- the display panel can use liquid crystal display (LCD), organic light-emitting diode (OLED), active matrix organic light-emitting diode or active-matrix organic light-emitting diode (active-matrix organic light-emitting diode).
- LCD liquid crystal display
- OLED organic light-emitting diode
- active-matrix organic light-emitting diode active-matrix organic light-emitting diode
- AMOLED flexible light-emitting diode (FLED), Miniled, MicroLed, Micro-oLed, quantum dot light-emitting diode (QLED), etc.
- the mobile terminal 100 may include one or N display screens 194, and N is a positive integer greater than one.
- the mobile terminal 100 can implement a shooting function through an ISP, a camera 193, a video codec, a GPU, a display screen 194, and an application processor.
- the ISP is used to process the data fed back from the camera 193. For example, when taking a picture, the shutter is opened, the light is transmitted to the photosensitive element of the camera through the lens, the light signal is converted into an electrical signal, and the photosensitive element of the camera transmits the electrical signal to the ISP for processing and is converted into an image visible to the naked eye.
- ISP can also optimize the image noise, brightness, and skin color. ISP can also optimize the exposure, color temperature and other parameters of the shooting scene.
- the ISP may be provided in the camera 193.
- the camera 193 is used to capture still images or videos.
- the object generates an optical image through the lens and is projected to the photosensitive element.
- the photosensitive element may be a charge coupled device (CCD) or a complementary metal-oxide-semiconductor (CMOS) phototransistor.
- CMOS complementary metal-oxide-semiconductor
- the photosensitive element converts the optical signal into an electrical signal, and then transfers the electrical signal to the ISP to convert it into a digital image signal.
- ISP outputs digital image signals to DSP for processing.
- DSP converts digital image signals into standard RGB, YUV and other formats of image signals.
- the mobile terminal 100 may include 1 or N cameras 193, and N is a positive integer greater than 1.
- the above-mentioned one or N cameras 193 may include at least one telephoto camera, which may be used to capture a photographed object far away from the mobile terminal 100.
- the processor 110 (for example, it can be one or more of ISP, CPU, DSP, and NPU) can perform detail enhancement and background blurring processing on the image collected by the telephoto camera, and improve the telephoto shooting scene of the mobile terminal. The image quality enhances the user's visual experience.
- Digital signal processors are used to process digital signals. In addition to digital image signals, they can also process other digital signals. For example, when the mobile terminal 100 selects a frequency point, the digital signal processor is used to perform Fourier transform on the energy of the frequency point.
- Video codecs are used to compress or decompress digital video.
- the mobile terminal 100 may support one or more video codecs. In this way, the mobile terminal 100 can play or record videos in multiple encoding formats, such as: moving picture experts group (MPEG) 1, MPEG2, MPEG3, MPEG4, and so on.
- MPEG moving picture experts group
- MPEG2 MPEG2, MPEG3, MPEG4, and so on.
- NPU is a neural-network (NN) computing processor.
- NN neural-network
- applications such as intelligent cognition of the mobile terminal 100 can be realized, such as image recognition, face recognition, voice recognition, text understanding, and so on.
- the external memory interface 120 may be used to connect an external memory card, such as a Micro SD card, so as to expand the storage capacity of the mobile terminal 100.
- the external memory card communicates with the processor 110 through the external memory interface 120 to realize the data storage function. For example, save music, video and other files in an external memory card.
- the internal memory 121 may be used to store computer executable program code, where the executable program code includes instructions.
- the internal memory 121 may include a storage program area and a storage data area.
- the storage program area can store an operating system, an application program (such as a sound playback function, an image playback function, etc.) required by at least one function, and the like.
- the data storage area can store data (such as audio data, phone book, etc.) created during the use of the mobile terminal 100.
- the internal memory 121 may include a high-speed random access memory, and may also include a non-volatile memory, such as at least one magnetic disk storage device, a flash memory device, a universal flash storage (UFS), and the like.
- the processor 110 executes various functional applications and data processing of the mobile terminal 100 by running instructions stored in the internal memory 121 and/or instructions stored in a memory provided in the processor.
- the mobile terminal 100 can implement audio functions through the audio module 170, the speaker 170A, the receiver 170B, the microphone 170C, the earphone interface 170D, and the application processor. For example, music playback, recording, etc.
- the audio module 170 is used to convert digital audio information into an analog audio signal for output, and is also used to convert an analog audio input into a digital audio signal.
- the audio module 170 can also be used to encode and decode audio signals.
- the audio module 170 may be provided in the processor 110, or part of the functional modules of the audio module 170 may be provided in the processor 110.
- the speaker 170A also called “speaker” is used to convert audio electrical signals into sound signals.
- the mobile terminal 100 can listen to music through the speaker 170A, or listen to a hands-free call.
- the receiver 170B also called “earpiece” is used to convert audio electrical signals into sound signals.
- the mobile terminal 100 answers a call or voice message, it can receive the voice by bringing the receiver 170B close to the human ear.
- the microphone 170C also called “microphone”, “microphone”, is used to convert sound signals into electrical signals.
- the user can make a sound by approaching the microphone 170C through the human mouth, and input the sound signal into the microphone 170C.
- the mobile terminal 100 may be provided with at least one microphone 170C.
- the mobile terminal 100 may be provided with two microphones 170C, which can implement noise reduction functions in addition to collecting sound signals.
- the mobile terminal 100 may also be provided with three, four or more microphones 170C to collect sound signals, reduce noise, identify sound sources, and realize directional recording functions.
- the earphone interface 170D is used to connect wired earphones.
- the earphone interface 170D may be a USB interface 130, or a 3.5mm open mobile terminal platform (OMTP) standard interface, or a cellular telecommunications industry association (cellular telecommunications industry association of the USA, CTIA) standard interface.
- OMTP open mobile terminal platform
- CTIA cellular telecommunications industry association
- the button 190 includes a power-on button, a volume button, and so on.
- the button 190 may be a mechanical button. It can also be a touch button.
- the mobile terminal 100 may receive key input, and generate key signal input related to user settings and function control of the mobile terminal 100.
- the motor 191 can generate vibration prompts.
- the motor 191 can be used for incoming call vibration notification, and can also be used for touch vibration feedback.
- touch operations applied to different applications can correspond to different vibration feedback effects.
- Acting on touch operations in different areas of the display screen 194, the motor 191 can also correspond to different vibration feedback effects.
- Different application scenarios for example: time reminding, receiving information, alarm clock, games, etc.
- the touch vibration feedback effect can also support customization.
- the indicator 192 may be an indicator light, which may be used to indicate the charging status, power change, or to indicate messages, missed calls, notifications, and so on.
- the SIM card interface 195 is used to connect to the SIM card.
- the SIM card can be inserted into the SIM card interface 195 or pulled out from the SIM card interface 195 to achieve contact and separation with the mobile terminal 100.
- the mobile terminal 100 may support 1 or N SIM card interfaces, and N is a positive integer greater than 1.
- the SIM card interface 195 can support Nano SIM cards, Micro SIM cards, SIM cards, etc.
- the same SIM card interface 195 can insert multiple cards at the same time. The types of the multiple cards can be the same or different.
- the SIM card interface 195 can also be compatible with different types of SIM cards.
- the SIM card interface 195 may also be compatible with external memory cards.
- the mobile terminal 100 interacts with the network through the SIM card to implement functions such as call and data communication.
- the mobile terminal 100 adopts an eSIM, that is, an embedded SIM card.
- the eSIM card can be embedded in the mobile terminal 100 and cannot be separated from the mobile terminal 100.
- the software system of the mobile terminal 100 may adopt a layered architecture, an event-driven architecture, a micro-core architecture, a micro-service architecture, or a cloud architecture.
- an Android system with a layered architecture is taken as an example to illustrate the software structure of the mobile terminal 100 by way of example.
- FIG. 2 is a block diagram of the software structure of the mobile terminal 100 according to an embodiment of the present invention.
- the layered architecture divides the software into several layers, and each layer has a clear role and division of labor. Communication between layers through software interface.
- the Android system is divided into four layers, from top to bottom, the application layer, the application framework layer, the Android runtime and system library, and the kernel layer.
- the application layer may include a series of application packages, including applications preset before the mobile terminal leaves the factory, or applications installed by the user through, for example, an application market or other methods after the mobile terminal leaves the factory.
- applications include, but are not limited to, applications such as camera, gallery, calendar, call, map, navigation, WLAN, Bluetooth, music, video, SMS, browser, WeChat, Taobao, Alipay, etc. (only some of them are shown in the figure) .
- the application framework layer provides an application programming interface (application programming interface, API) and a programming framework for applications in the application layer.
- the application framework layer includes some predefined functions.
- the application framework layer can include a window manager, a content provider, a view system, a phone manager, a resource manager, and a notification manager.
- the window manager is used to manage window programs.
- the window manager can obtain the size of the display screen, determine whether there is a status bar, lock the screen, take a screenshot, etc.
- the content provider is used to store and retrieve data and make these data accessible to applications.
- the data may include videos, images, audios, phone calls made and received, browsing history and bookmarks, phone book, etc.
- the view system includes visual controls, such as controls that display text, controls that display pictures, and so on.
- the view system can be used to build applications.
- the display interface can be composed of one or more views.
- a display interface that includes a short message notification icon may include a view that displays text and a view that displays pictures.
- the phone manager is used to provide the communication function of the mobile terminal 100. For example, the management of the call status (including connecting, hanging up, etc.).
- the resource manager provides various resources for the application, such as localized strings, icons, pictures, layout files, video files, and so on.
- the notification manager enables the application to display notification information in the status bar, which can be used to convey notification-type messages, and it can automatically disappear after a short stay without user interaction. For example, the notification manager is used to notify download completion, message reminders, and so on.
- the notification manager can also be a notification that appears in the status bar at the top of the system in the form of a chart or a scroll bar text, such as a notification of an application running in the background, or a notification that appears on the screen in the form of a dialog window. For example, text messages are prompted in the status bar, a prompt sound is issued, the mobile terminal vibrates, and the indicator light flashes.
- Android Runtime includes core libraries and virtual machines. Android runtime is responsible for the scheduling and management of the Android system.
- the core library consists of two parts: one part is the function functions that the java language needs to call, and the other part is the core library of Android.
- the application layer and application framework layer run in a virtual machine.
- the virtual machine executes the java files of the application layer and the application framework layer as binary files.
- the virtual machine is used to perform functions such as object life cycle management, stack management, thread management, security and exception management, and garbage collection.
- the system library can include multiple functional modules. For example: surface manager (surface manager), media library (Media Libraries), three-dimensional graphics processing library (for example: OpenGL ES), 2D graphics engine (for example: SGL), etc.
- the surface manager is used to manage the display subsystem and provides a combination of 2D and 3D layers for multiple applications.
- the media library supports playback and recording of a variety of commonly used audio and video formats, as well as still image files.
- the media library can support multiple audio and video encoding formats, such as: MPEG4, H.264, MP3, AAC, AMR, JPG, PNG, etc.
- the 3D graphics processing library is used to implement 3D graphics drawing, image rendering, synthesis, and layer processing.
- the 2D graphics engine is a drawing engine for 2D drawing.
- the system library may also include a first module, which can be used to perform detail enhancement and background blurring processing on the image collected by the telephoto camera in the mobile terminal 100 to improve the telephoto shooting scene of the mobile terminal. Under the image quality, enhance the user’s visual experience.
- the kernel layer is the layer between hardware and software.
- the kernel layer contains at least display driver, camera driver, audio driver, and sensor driver.
- the corresponding hardware interrupt is sent to the kernel layer.
- the kernel layer processes touch operations into original input events (including touch coordinates, time stamps of touch operations, etc.).
- the original input events are stored in the kernel layer.
- the application framework layer obtains the original input event from the kernel layer and identifies the control corresponding to the input event. Taking the touch operation as a touch click operation, and the control corresponding to the click operation is the control of the camera application icon as an example, the camera application calls the interface of the application framework layer to start the camera application, and then starts the camera driver by calling the kernel layer.
- the camera 193 captures still images or videos.
- the user can instruct the mobile phone to start the camera application by touching a specific control on the mobile phone screen, pressing a specific physical key or key combination, inputting voice, or gestures in the air.
- the mobile phone starts the camera and displays the shooting interface.
- the user can click the "camera” application icon 301 on the mobile phone desktop to instruct the mobile phone to start the camera application, and the mobile phone displays the shooting interface shown in Figure 3 (2).
- the user when the mobile phone is in the locked screen state, the user can also instruct the mobile phone to open the camera application through a gesture of sliding right on the mobile phone screen, and the mobile phone can also display the shooting interface as shown in Figure 3 (2).
- the user can click the shortcut icon of the "camera” application on the lock screen interface to instruct the mobile phone to open the camera application, and the mobile phone can also display the shooting interface as shown in Figure 3 (2).
- the user can also click the corresponding control to make the mobile phone start the camera application to take pictures.
- the user can also select the control of the camera function to instruct the mobile phone to open the camera application to take pictures and shoot videos.
- the camera's shooting interface generally includes a viewfinder 302, shooting controls, and other functional controls ("large aperture”, “portrait”, “photograph”, “video”, etc.).
- the viewfinder frame 302 can be used to preview the image (or picture) collected by the camera, and the user can decide the timing to instruct the mobile phone to perform the shooting operation based on the image (or frame) in the viewfinder frame 302.
- the user instructing the mobile phone to perform a shooting operation may be, for example, an operation of the user clicking a shooting control, or an operation of the user pressing a volume key.
- a zoom magnification indicator 303 may also be included in the shooting interface.
- the default zoom magnification of the mobile phone is the basic magnification, which is "1 ⁇ ".
- the medium-focus camera is generally set as the main camera.
- the focal length of the main camera is set as the reference focal length, and the zoom magnification is "1 ⁇ ".
- the image captured by the main camera can be digitally zoomed (or called digital zoom), that is, the "1 ⁇ " image captured by the main camera can be captured by the ISP or other processor in the mobile phone.
- the area of each pixel is enlarged, and the viewing range of the image is reduced accordingly, so that the processed image is equivalent to the image taken by the main camera with other zoom magnifications (for example, "2 ⁇ "). That is to say, the image taken with the main camera can correspond to a range of zoom magnification, for example: “1 ⁇ " to "5 ⁇ ".
- the multiple of the focal length of the telephoto camera and the focal length of the main camera can be used as the zoom magnification of the telephoto camera.
- the focal length of the telephoto camera can be 5 times the focal length of the main camera, that is, the zoom magnification of the telephoto camera is "5 ⁇ ".
- the image captured by the telephoto camera can also be digitally zoomed. That is, the image taken with the telephoto camera can correspond to the zoom magnification of another interval, for example: "5 ⁇ " to "50 ⁇ ".
- the multiple of the focal length of the short-focus (wide-angle) camera and the focal length of the main camera can be used as the zoom magnification of the short-focus (wide-angle) camera.
- the focal length of the short-focus camera may be 0.5 times the focal length of the main camera, that is, the zoom magnification of the long-focus camera is "0.5 ⁇ ".
- the telephoto shooting scene in the embodiments of the present application can be understood as a shooting scene where the zoom magnification of the mobile phone is greater than the preset magnification (for example: "5 ⁇ ", "10 ⁇ ", "20 ⁇ ”).
- the mobile phone can use an image taken by a medium-focus camera or an image taken by a telephoto camera, which is not specifically limited in the embodiment of the present application.
- the embodiment of the present application provides a shooting method in a telephoto shooting scene, which can process the original image collected by the mobile phone camera. Specifically, it intelligently recognizes the subject and enhances the details of the subject to improve the clarity of the part of the subject in the original image. It can also automatically recognize the front and back scenes of the image according to the identified subject, blur the background outside the subject, enhance the artistic sense of the image in the telephoto shooting scene, and enhance the user's visual experience.
- the definition refers to the texture details of each part of the image and the clarity of the boundary.
- image 1 is the original image collected by a mobile phone camera. If the method provided in the embodiment of the present application is used, the image 1 is processed to obtain the image 2. The sharpness of the subject in image 2 is higher than the sharpness of the subject in image 1. If the image 1 and image 2 are enlarged by the same magnification, and the enlarged image 1 and image 2 are compared, the texture details of the subject in the enlarged image 2 are richer, and the boundaries of the various parts of the subject are also clearer.
- the phone switches to the telephoto shooting scene
- the user can manually adjust the zoom magnification used when shooting with the mobile phone.
- the user can adjust the zoom magnification used by the mobile phone by operating the zoom magnification indicator 303 in the shooting interface.
- the user can click the zoom magnification indicator 303 one or more times to change the zoom magnification used by the mobile phone to "5 ⁇ ", that is, the mobile phone displays as shown in Figure 4 (1 )
- the shooting interface In the shooting interface shown in Figure 4 (1), the viewing range of the image previewed in the viewing frame 302 is obviously smaller than the viewing range in the viewing frame 302 in Figure 3 (2), but the viewing area is previewed in the viewing frame 302
- the size of the photographed object for example: bird
- the shooting interface shown in (1) in FIG. 4 may continue to display the zoom magnification indicator 303, and the current zoom magnification is displayed as "5 ⁇ " at this time, so that the user can know the current zoom magnification.
- the user can reduce the zoom ratio used by the mobile phone by pinching two fingers (or three fingers) in the shooting interface, or slide two fingers (or three fingers) outwards
- the gesture in the opposite direction of pinching
- the user can also change the zoom ratio used by the mobile phone by dragging the zoom ruler 304 in the shooting interface.
- the user can also change the zoom ratio of the mobile phone by switching the currently used camera in the shooting interface or the shooting setting interface. For example, if the user chooses to switch to a telephoto camera, the phone automatically increases the zoom ratio.
- the user can also change the zoom ratio of the mobile phone by selecting options for telephoto shooting scenes or remote shooting scenes in the shooting interface or shooting setting interface.
- the mobile phone can also automatically recognize the specific scene of the image captured by the camera, and automatically adjust the zoom magnification according to the recognized specific scene. For example, if the mobile phone recognizes that the image captured by the camera is a scene with a larger field of view, such as the sea, mountains, woods, etc., the zoom factor can be automatically reduced. For another example, if the mobile phone recognizes that the image captured by the camera is a distant object, such as a bird in the distance, an athlete on a sports field, etc., the zoom factor can be automatically increased, which is not limited in this application.
- the mobile phone When the zoom magnification of the mobile phone is adjusted to be greater than or equal to the preset magnification, the mobile phone enters the telephoto shooting scene.
- the mobile phone displays the shooting interface as shown in (2) of FIG. 4.
- the view frame 302 includes a marking frame 401 for marking the photographic subject recognized by the mobile phone.
- the subject is a bird.
- the photographic subject shown in the figure is in a moving state.
- the method provided in the embodiment of the present application is not limited to whether the shooting object is in a moving state or a stationary state.
- the mobile phone may also display a prompt box 402 for recommending the image processing mode corresponding to the subject category to the user.
- the identified photographing object is a bird
- the user is asked whether to turn on the image processing mode corresponding to the bird (or called bird mode, bird effect).
- the mobile phone performs corresponding image processing on the collected images, and displays the shooting interface as shown in (3) in FIG. 4.
- the mobile phone after the mobile phone recognizes the category and location of the subject, it can also automatically use the image processing mode corresponding to the category of the subject for processing. That is, when the zoom magnification of the mobile phone is adjusted to be greater than or equal to the preset magnification, the mobile phone directly displays the shooting interface as shown in (3) of FIG. 4.
- the image displayed in the viewfinder frame 302 is an image after corresponding image processing. It can be seen that the identified subject is clearer and clearer.
- the mobile phone may also display a label 403 corresponding to the identified photographic subject.
- the mobile phone can also blur the background of the identified shooting object, so that the processed image has an artistic sense and enhances the visual effect of the image.
- the mobile phone after the mobile phone selects the image processing mode corresponding to the subject category by default, or the user manually turns on the image processing mode corresponding to the subject category, the user can also manually turn off the image processing mode corresponding to the subject category. , Or when the identified object category does not correspond to the corresponding image processing mode, the mobile phone automatically turns off the image processing mode corresponding to the object category.
- the ISP or other processor of the mobile phone performs image analysis on the image collected by the camera, recognizes the subject that the user intends to photograph from the image, and further recognizes the category and location of the subject.
- the category of the photographed object can be arbitrarily classified based on the objects that the user usually photographs, and it can be a one-level classification or a multi-level classification.
- the categories of the photographed objects may include: humans, animals, plants, and so on.
- the animal category can also include the second level of classification, such as: dog, cat, squirrel, bird, etc.
- the bird category may also include a third level classification, such as: sparrows, magpies, seagulls, seabirds, pigeons, wild geese and so on.
- the object with the largest area can be defaulted as the subject, or the object located in the central area of the image is the subject, or the object selected by the user (for example, the object selected by the click operation) is the subject .
- multiple objects may also be selected as the shooting objects by default or selected, which is not limited in the embodiment of the present application.
- the mobile phone's ISP or other processors can use deep learning-based target detection technologies, such as: Region-Convolutional Neural Networks (R-CNN), Fast R-CNN -CNN), Faster Regional Convolutional Neural Network (Faster R-CNN), etc., to identify the category and location of the subject.
- the category and location of the subject can be marked in the viewfinder of the mobile phone, so that the user can confirm whether the marked subject is the intended subject and whether the mobile phone accurately recognizes the category and location of the subject. If you confirm that the type and location of the object automatically recognized by the mobile phone are correct, you can use the corresponding image processing mode recommended by the mobile phone. Otherwise, the user can also choose not to use the image processing mode recommended by the mobile phone, or change the subject, etc., to avoid wrong image processing.
- the image of the subject can be processed. For example, based on the image segmentation technology, a part of the image corresponding to the subject can be segmented from the collected image, and then the segmented subject can be enhanced in detail.
- the mobile phone can pre-train an AI model that enhances the details of the subject based on the category of the subject.
- the mobile phone can also directly obtain the trained AI model from the server.
- training an AI model that enhances the details of the subject based on the category of the subject it is necessary to input a large number of training samples of the subject, including training samples of each category. For example, training samples for magpies in birds, including images of different species of magpies, images of different sizes or colors of the same species of magpies, images of different poses of the same species of magpies, etc., images of the same species of magpies in different environments Wait.
- the detail enhancement of the photographed object may include the use of an AI model for intelligent pixel filling of the blurred area in the photographed object, etc., to enhance the clarity of the image. It also includes: for the missing parts of the subject, the AI model is used for intelligent pixel filling, etc., to repair the missing parts of the subject, etc. It may also include: improving the overall sharpness of the image of the part of the subject.
- the mobile phone can also perform automatic focus (AF) and automatic exposure (AE) based on the identified subject to highlight the subject and improve The quality of the image.
- AF automatic focus
- AE automatic exposure
- focusing and exposure are generally performed based on a fixed position, or based on a position selected by the user (for example, a touch position of a click operation).
- focusing and exposure are performed based on the position of the photographed object automatically recognized by the mobile phone, so as to improve the overall clarity of the photographed object.
- the size of the focus frame is fixed, and is usually set by default in the mobile phone.
- the size of the focus frame will be automatically adjusted according to the size of the subject recognized by the mobile phone.
- the focus frame may be the same as the marking frame 401 shown in (2) in FIG. 4.
- the closest search distance during auto-focusing of the mobile phone can be set, for example, 10 meters.
- the search distance range is 10 meters away from the mobile phone to infinity.
- the mobile phone does not need to search in the range of 10 meters to the macro distance from the mobile phone.
- the mobile phone may also collect multiple (for example: 6) images containing the subject based on parameters determined during autofocus and automatic exposure. For multiple images containing the subject, super resolution technology is used to synthesize a high-definition image. Then use the synthesized high-definition image to enhance the details of the subject. As a result, the overall clarity of the image is further improved.
- a background extraction algorithm can also be used to identify the background of the subject based on the identified subject, and blur the background.
- the background extraction algorithm includes, but is not limited to, the inter-frame difference method, the background difference method, the environmental algorithm (ViBe algorithm, ViBe+ algorithm), etc., which are not limited in the embodiment of the present application.
- a blur algorithm can be used to process the image of the area where the background is located to obtain a background image with a blur effect.
- the fuzzy algorithm includes any one or several of Gaussian filtering, circular filtering, guided filtering, and domain filtering.
- the mobile phone can also adjust the position of the identified photographic object.
- the background of the shooting object in the image can be cropped or filled based on the identified position of the shooting object in the image, so that the position of the shooting object is located at the center of the image.
- the position of the photographed object may also be located in another position of the image, for example, used for a certain distance to the left or right of the center position, which is not limited in the embodiment of the application.
- the photographic subject in the image displayed in the viewfinder frame 302 is located in the central area of the viewfinder frame. In this way, you can highlight the subject and further enhance the visual effect of the image.
- the user when previewing, may also choose to switch between the image processing effect corresponding to the category of the photographed object and no processing (ie no effect).
- the mobile phone may also display an option 502. If the user selects the bird effect option in the option 502, the image displayed in the view frame 302 is an image processed according to the bird effect mode. If the user has no effect in the option 502, as shown in (2) in FIG. 5, the image displayed in the viewfinder frame 302 is an image that has not been processed in accordance with the bird effect mode.
- the sharpness of the image of the marked frame 501 in the view frame 302 is higher than the sharpness of the image in the marked frame 501 in Figure 5 (2) Big.
- the definition of the image (ie, the background) outside the marking frame 501 in the view frame 302 is lower than that of the marking frame 501 in FIG. 5 (2).
- the definition of the external image is small.
- the position of the marker frame 501 in the viewfinder frame 302 in the interface shown in (1) in FIG. 5 is also different from the position of the marker frame 501 in the interface shown in (2) in FIG. 5. It can be seen that the user can switch the effect before and after the image processing through the option 502, so that the user can compare and choose a preferred method for shooting, thereby improving the user experience.
- the user may further choose whether to adjust the position of the photographed object.
- an option control 503 may also be displayed.
- the mobile phone retains the original position of the subject. If the user selects the option control 503, the mobile phone displays the interface as shown in (4) in FIG. 5, and the marking frame 501 is located in the central area of the viewfinder frame 302. That is, the mobile phone adjusts the position of the subject to the center area of the image.
- the user can also manually turn on the image processing function based on the category of the photographed object provided in the embodiments of the present application.
- the mobile phone in response to detecting that the user operates the setting control 601, the mobile phone displays the shooting setting interface 602 as shown in (2) in FIG.
- a control 603 is displayed in the shooting setting interface 602 for enabling the image processing function based on the category of the shooting object. That is to say, when the user turns on this function, the mobile phone will automatically use the method provided in the embodiment of the present application to process the collected image when it is in a telephoto shooting scene.
- the user can also manually turn off the image processing function based on the subject category in the telephoto scene through the control 603.
- some commonly used category controls 605 may be displayed.
- the user can turn on the image processing function of the corresponding category through the category control 605. That is to say, when the user turns on the image processing function corresponding to a certain category, the mobile phone will automatically adopt the corresponding image processing method to process the collected images when it is in the telephoto shooting scene. For example, if the user chooses to enable the bird effect function, the mobile phone will process the collected images according to the bird's image processing mode.
- the setting interface 606 as shown in (4) in FIG. 6 displays multiple categories of options. The user can select the corresponding category according to the actual scene, so that the mobile phone processes the collected images according to the corresponding image processing mode.
- the mobile phone can no longer automatically recognize the category of the subject, and directly use the corresponding image processing mode, which can accelerate the image processing effect. Or, after the mobile phone recognizes the wrong category of the subject, the user can also manually select the image processing mode corresponding to the correct category.
- the user can decide the timing of shooting based on the preview in the viewfinder. After detecting that the user performs the shooting operation, the mobile phone performs corresponding image processing based on the image collected at the current moment to obtain the processed image.
- the method of image processing can refer to the image processing during preview, which will not be repeated here.
- the phone can save images after image processing based on the category of the subject in the album.
- the mobile phone can also save images that have not undergone image processing based on the subject category at the same time in the album.
- the method provided in this application can also be used to process the collected images accordingly.
- the camera collects multiple images, and the mobile phone can perform image processing based on the subject of each image to obtain multiple processed images.
- both slow motion and video are composed of one frame of images. Each frame of the image can be processed based on the subject, and then the processed images are formed into a new slow motion. Action and new video.
- the method provided in this application can also be used to perform image processing on the image based on the shooting object.
- the embodiment of the present application also provides a chip system.
- the chip system includes at least one processor 1101 and at least one interface circuit 1102.
- the processor 1101 and the interface circuit 1102 may be interconnected by wires.
- the interface circuit 1102 may be used to receive signals from other devices (such as the memory of the mobile terminal 100).
- the interface circuit 1102 may be used to send signals to other devices (such as the processor 1101).
- the interface circuit 1102 can read an instruction stored in the memory, and send the instruction to the processor 1101.
- the mobile terminal can be made to execute various steps executed by the mobile terminal 100 (for example, a mobile phone) in the above-mentioned embodiment.
- the chip system may also include other discrete devices, which are not specifically limited in the embodiment of the present application.
- An embodiment of the present application also provides a device included in a mobile terminal, and the device has a function of realizing the behavior of the mobile terminal in any of the methods in the foregoing embodiments.
- This function can be realized by hardware, or by hardware executing corresponding software.
- the hardware or software includes at least one module or unit corresponding to the above-mentioned functions. For example, a detection module or unit, and a determination module or unit, etc.
- the embodiments of the present application also provide a computer-readable storage medium, including computer instructions, which when the computer instructions are executed on the mobile terminal, cause the mobile terminal to execute any method as in the above-mentioned embodiments.
- the embodiments of the present application also provide a computer program product, which when the computer program product runs on a computer, causes the computer to execute any method as in the above-mentioned embodiments.
- the above-mentioned terminal and the like include hardware structures and/or software modules corresponding to each function.
- the embodiments of the present application can be implemented in the form of hardware or a combination of hardware and computer software. Whether a certain function is executed by hardware or computer software-driven hardware depends on the specific application and design constraint conditions of the technical solution. Professionals and technicians can use different methods for each specific application to implement the described functions, but such implementation should not be considered as going beyond the scope of the embodiments of the present invention.
- the embodiments of the present application may divide the above-mentioned terminal and the like into functional modules according to the above method examples.
- each functional module may be divided corresponding to each function, or two or more functions may be integrated into one processing module.
- the above-mentioned integrated modules can be implemented in the form of hardware or software functional modules. It should be noted that the division of modules in the embodiment of the present invention is illustrative, and is only a logical function division, and there may be other division methods in actual implementation.
- the functional units in the various embodiments of the embodiments of the present application may be integrated into one processing unit, or each unit may exist alone physically, or two or more units may be integrated into one unit.
- the above-mentioned integrated unit can be implemented in the form of hardware or software functional unit.
- the integrated unit is implemented in the form of a software functional unit and sold or used as an independent product, it can be stored in a computer readable storage medium.
- the technical solutions of the embodiments of the present application are essentially or the part that contributes to the prior art, or all or part of the technical solutions can be embodied in the form of software products, and the computer software products are stored in a storage
- the medium includes several instructions for causing a computer device (which may be a personal computer, a server, or a network device, etc.) or a processor to execute all or part of the steps of the methods described in the various embodiments of the present application.
- the aforementioned storage media include: flash memory, mobile hard disk, read-only memory, random access memory, magnetic disk or optical disk and other media that can store program codes.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Human Computer Interaction (AREA)
- Theoretical Computer Science (AREA)
- General Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Computer Hardware Design (AREA)
- Telephone Function (AREA)
Abstract
一种长焦场景下的拍摄方法及移动终端,涉及通信技术领域,可以提升图像中拍摄对象的清晰度,虚化背景,突出拍摄对象,提升图像整体的视觉效果,该方法具体包括:在长焦拍摄场景下,移动终端在显示预览或拍摄图像时,根据相机采集的图像识别出目标拍摄对象,以及目标拍摄对象的背景;提升目标拍摄对象所在区域的图像清晰度,为目标拍摄对象的背景所在区域的图像添加虚化效果。
Description
本申请要求于2020年1月14日提交国家知识产权局、申请号为202010038444.X、申请名称为“一种长焦场景下的拍摄方法及移动终端”的中国专利申请的优先权,其全部内容通过引用结合在本申请中。
本申请涉及移动终端技术领域,尤其涉及一种长焦场景下的拍摄方法及移动终端。
目前,手机上可以集成多个镜头,例如广角摄像头、中焦摄像头和长焦摄像头等,以满足用户的不同拍摄场景。其中,长焦摄像头可用于拍摄距离用户较远的拍摄对象(即长焦拍摄场景)。然而,在实际情况中,由于拍摄对象距离用户较远,且手机上集成的长焦摄像头的规格也有限,导致长焦拍摄场景下得到的图像中拍摄对象的清晰度不高,且拍摄对象也不突出,视觉效果不佳。
发明内容
本申请提供的一种长焦场景下的拍摄方法及移动终端,可以提升图像中拍摄对象的清晰度,虚化背景,突出拍摄对象,提升图像整体的视觉效果。
为了实现上述目的,本申请实施例提供了以下技术方案:
第一方面、提供一种长焦场景下的拍摄方法,应用于包含相机的移动终端中,该方法包括:移动终端启动相机,移动终端显示取景框,取景框用于显示第一预览画面,第一预览画面对应的相机的变焦倍率为第一倍率;接收用户输入的增大相机变焦倍率的第一操作;响应于第一操作,取景框显示第二预览画面,第二预览画面对应的相机的变焦倍率为第二倍率,第二倍率大于第一倍率;若第二倍率大于或等于预设倍率,则在移动终端显示第二预览画面之前,移动终端对相机采集的图像进行第一图像处理,生成第二预览画面;其中,第一图像处理包括:根据相机采集的图像识别出目标拍摄对象,以及目标拍摄对象的背景;提升目标拍摄对象所在区域的图像清晰度,为目标拍摄对象背景所在区域的图像添加虚化效果。
在长焦拍摄场景中,由于拍摄对象距离手机的位置较远,手机摄像头采集到的原始图像清晰度不高。因此,本申请实施例提供了一种长焦拍摄场景下拍摄方法,可以智能识别出目标拍摄对象,并对目标拍摄对象进行细节增强,提升采集的原图像中拍摄对象部分图像的清晰度。还能根据识别出的拍摄对象自动识别图像的前后景,虚化拍摄对象外的背景,突出目标拍摄对象,并提升长焦拍摄场景下的图像的艺术感,增强用户的视觉体验。
一种可能的实现方式中,该方法还包括:移动终端在第二预览画面上显示有标记框,用于标记目标拍摄对象所在的区域。
由此,用户可以根据标记框,进一步确定移动终端自动识别的目标拍摄对象是否为待拍摄对象,以及自动识别的目标拍摄对象是否完整等。
一种可能的实现方式中,该方法还包括:移动终端在第二预览画面中显示提示信 息,用于推荐目标拍摄对象对应的第一图像处理。
一种可能的实现方式中,移动终端对相机采集的图像进行第一图像处理,包括:移动终端接收用户输入的选择第一图像处理的第二操作;响应于第二操作,移动终端对相机采集的图像进行第一图像处理。
一种可能的实现方式中,调整目标拍摄对象在第二预览画面中的位置,具体为:调整第二预览画面,使得目标拍摄对象位于第二预览画面的中心区域。
在具体实现时,可以基于识别出的拍摄对象在图像中的位置,对图像中拍摄对象的背景进行裁剪或填充,使得拍摄对象的位置位于图像的中心位置上。或者,拍摄对象的位置也可以位于图像的其他位置,例如用于中心位置偏左或偏右一定距离的位置,本申请实施例对此不做限定。
一种可能的实现方式中,在移动终端显示第二预览画面之后,该方法还包括:接收用户输入的第三操作,第三操作用于指示移动终端关闭目标拍摄对象对应的第一图像处理功能;响应于第三操作,移动终端确定不对摄像头采集的图像进行第一图像处理。
由此,提供了一种关闭第一图像处理的方法,满足用户不同的使用需求。
一种可能的实现方式中,第一图像处理中增大目标拍摄对象所在区域的图像清晰度,具体为:识别出目标拍摄对象的类别,从摄像头采集到的图像中分割出目标拍摄对象的第一图像;将第一图像输入到目标拍摄对象的类别对应的神经网络模型中,输出目标拍摄对象的第二图像,其中,第二图像的清晰度大于第一图像的清晰度。
一种可能的实现方式中,目标拍摄对象的类别对应的神经网络模型,为根据目标拍摄对象的类别下的多个拍摄对象的图像训练得到。
可以理解,手机可以预先训练出基于拍摄对象的类别,对拍摄对象进行细节增强的AI模型。当然,手机也可以直接从服务器处获取已训练好的AI模型。需要说明的是,在训练基于拍摄对象的类别,对拍摄对象进行细节增强的AI模型时,需要输入大量拍摄对象的训练样本,包括各个类别的训练样本。例如,针对鸟类中的喜鹊的训练样本,包括不同品种的喜鹊的图像,同一品种喜鹊的不同体积或颜色的图像,同一品种喜鹊的不同姿态的图像等,同一品种喜鹊在不同环境中的图像等。其中,对拍摄对象进行细节增强,可以包括对拍摄对象中模糊的区域,采用AI模型进行智能的像素填充等,以提升图像的清晰度。还包括:对拍摄对象中缺失的部位,采用AI模型进行智能的像素填充等,以修复拍摄对象的缺失部分等。还可以包括:提升拍摄对象部分的图像的整体清晰度等。
一种可能的实现方式中,第一图像处理中识别出目标拍摄对象的背景,具体为:采用背景提取算法从相机采集的图像中识别出目标拍摄对象的背景,其中,背景提取算法包括:帧间差分法、背景差分法、环境算法中的任一项或任几项。
一种可能的实现方式中,第一图像处理,还包括:在相机采集包含目标拍摄对象的图像时,自动基于目标拍摄对象进行对焦。
可以注意的是,现有技术中一般基于固定位置进行对焦和曝光的,或者基于用户选择的位置(例如点击操作的触碰位置)进行对焦和曝光。然而,本申请实施例中是基于手机自动识别出的目标拍摄对象的位置进行对焦和曝光,以提升目标拍摄对象整 体的清晰度。
还可以注意到,现有技术中,对焦框的大小是固定的,通常为手机默认设置的。然而,在本申请实施例基于目标拍摄对象的自动对焦过程中,对焦框的大小会随着移动终端识别出的目标拍摄对象的大小而自动调整。一些示例中,对焦框可以与标记框相同。
可选的,由于本申请实施例提供的方法是运用于长焦拍摄场景中,故可以设置移动终端自动对焦时的最近搜索距离,例如10米。也就是说,移动终端在自动对焦时,搜索的距离范围为距离移动终端10米外至无穷远。换言之,移动终端不用搜索在距离移动终端10米至微距的范围。由此,可以减少移动终端自动对焦的时间,提升图像处理的效率。
一种可能的实现方式中,第一图像处理,还包括:控制相机连续采集N张包含目标拍摄对象的图像,采用超清晰度技术将N张包含目标拍摄对象的图像合成为第三图像;基于第三图像提升目标拍摄对象所在区域的图像清晰度,为目标拍摄对象背景所在区域的图像添加虚化效果。由此,进一步突出图像中的目标拍摄对象,提升图像整体的艺术感。
一种可能的实现方式中,第一图像处理中为目标拍摄对象背景所在区域的图像添加虚化效果,具体为:采用模糊算法对目标拍摄对象背景所在区域的图像进行处理,模糊算法包括高斯滤波、圆形滤波、导向滤波、域滤波中任一项或任几项。
第二方面、提供一种移动终端,包括:处理器、存储器和触摸屏,所述存储器、所述触摸屏与所述处理器耦合,所述存储器用于存储计算机程序代码,所述计算机程序代码包括计算机指令,当所述处理器从所述存储器中读取所述计算机指令,以使得所述移动终端执行上述方面中及其中任一种可能的实现方式中所述的方法。
第三方面、提供一种装置,该装置包含在移动终端中,该装置具有实现上述方面及可能的实现方式中任一方法中移动终端行为的功能。该功能可以通过硬件实现,也可以通过硬件执行相应的软件实现。硬件或软件包括至少一个与上述功能相对应的模块或单元。例如,接收模块或单元、显示模块或单元、以及处理模块或单元等。
第四方面、提供一种芯片系统,包括处理器,当处理器执行指令时,处理器执行如上述方面中及其中任一种可能的实现方式中所述的方法。
第五方面、提供一种计算机可读存储介质,包括计算机指令,当计算机指令在移动终端上运行时,使得移动终端执行如上述方面及其中任一种可能的实现方式中所述的方法。
第六方面、提供一种计算机程序产品,当计算机程序产品在计算机上运行时,使得计算机执行如上述方面中及其中任一种可能的实现方式中所述的方法。
图1为本申请实施例提供的一种移动终端的结构示意图一;
图2为本申请实施例提供的一种移动终端的结构示意图二;
图3为本申请实施例提供的一些移动终端的用户界面示意图;
图4为本申请实施例提供的另一些移动终端的用户界面示意图;
图5为本申请实施例提供的又一些移动终端的用户界面示意图;
图6为本申请实施例提供的又一些移动终端的用户界面示意图;
图7为本申请实施例提供的一种芯片系统的结构示意图。
在本申请实施例中,“示例性的”或者“例如”等词用于表示作例子、例证或说明。本申请实施例中被描述为“示例性的”或者“例如”的任何实施例或设计方案不应被解释为比其它实施例或设计方案更优选或更具优势。确切而言,使用“示例性的”或者“例如”等词旨在以具体方式呈现相关概念。
以下,术语“第一”、“第二”仅用于描述目的,而不能理解为指示或暗示相对重要性或者隐含指明所指示的技术特征的数量。由此,限定有“第一”、“第二”的特征可以明示或者隐含地包括一个或者更多个该特征。在本申请实施例的描述中,除非另有说明,“多个”的含义是两个或两个以上。
本申请实施例提供的拍摄方法可应用于安装有摄像头的移动终端中,该移动终端例如可以为手机、平板电脑、个人计算机(personal computer,PC)、个人数字助理(personal digital assistant,PDA)、智能手表、上网本、可穿戴电子设备、增强现实技术(augmented reality,AR)设备、虚拟现实(virtual reality,VR)设备、车载设备、智能汽车、智能音响、机器人等,本申请对该移动终端的具体形式不做特殊限制。
图1示出了移动终端100的结构示意图。移动终端100可以包括处理器110,外部存储器接口120,内部存储器121,通用串行总线(universal serial bus,USB)接口130,充电管理模块140,电源管理模块141,电池142,天线1,天线2,移动通信模块150,无线通信模块160,音频模块170,扬声器170A,受话器170B,麦克风170C,耳机接口170D,传感器模块180,按键190,马达191,指示器192,摄像头193,显示屏194,以及用户标识模块(subscriber identification module,SIM)卡接口195等。其中传感器模块180可以包括压力传感器180A,陀螺仪传感器180B,气压传感器180C,磁传感器180D,加速度传感器180E,距离传感器180F,接近光传感器180G,指纹传感器180H,温度传感器180J,触摸传感器180K,环境光传感器180L,骨传导传感器180M等。
可以理解的是,本发明实施例示意的结构并不构成对移动终端100的具体限定。在本申请另一些实施例中,移动终端100可以包括比图示更多或更少的部件,或者组合某些部件,或者拆分某些部件,或者不同的部件布置。图示的部件可以以硬件,软件或软件和硬件的组合实现。
处理器110可以包括一个或多个处理单元,例如:处理器110可以包括应用处理器(application processor,AP),调制解调处理器,图形处理器(graphics processing unit,GPU),图像信号处理器(image signal processor,ISP),控制器,视频编解码器,数字信号处理器(digital signal processor,DSP),基带处理器,和/或神经网络处理器(neural-network processing unit,NPU)等。其中,不同的处理单元可以是独立的器件,也可以集成在一个或多个处理器中。
控制器可以根据指令操作码和时序信号,产生操作控制信号,完成取指令和执行指令的控制。
处理器110中还可以设置存储器,用于存储指令和数据。在一些实施例中,处理 器110中的存储器为高速缓冲存储器。该存储器可以保存处理器110刚用过或循环使用的指令或数据。如果处理器110需要再次使用该指令或数据,可从所述存储器中直接调用。避免了重复存取,减少了处理器110的等待时间,因而提高了系统的效率。
在一些实施例中,处理器110可以包括一个或多个接口。接口可以包括集成电路(inter-integrated circuit,I2C)接口,集成电路内置音频(inter-integrated circuit sound,I2S)接口,脉冲编码调制(pulse code modulation,PCM)接口,通用异步收发传输器(universal asynchronous receiver/transmitter,UART)接口,移动产业处理器接口(mobile industry processor interface,MIPI),通用输入输出(general-purpose input/output,GPIO)接口,用户标识模块(subscriber identity module,SIM)接口,和/或通用串行总线(universal serial bus,USB)接口等。
I2C接口是一种双向同步串行总线,包括一根串行数据线(serial data line,SDA)和一根串行时钟线(derail clock line,SCL)。在一些实施例中,处理器110可以包含多组I2C总线。处理器110可以通过不同的I2C总线接口分别耦合触摸传感器180K,充电器,闪光灯,摄像头193等。例如:处理器110可以通过I2C接口耦合触摸传感器180K,使处理器110与触摸传感器180K通过I2C总线接口通信,实现移动终端100的触摸功能。
I2S接口可以用于音频通信。在一些实施例中,处理器110可以包含多组I2S总线。处理器110可以通过I2S总线与音频模块170耦合,实现处理器110与音频模块170之间的通信。在一些实施例中,音频模块170可以通过I2S接口向无线通信模块160传递音频信号,实现通过蓝牙耳机接听电话的功能。
PCM接口也可以用于音频通信,将模拟信号抽样,量化和编码。在一些实施例中,音频模块170与无线通信模块160可以通过PCM总线接口耦合。在一些实施例中,音频模块170也可以通过PCM接口向无线通信模块160传递音频信号,实现通过蓝牙耳机接听电话的功能。所述I2S接口和所述PCM接口都可以用于音频通信。
UART接口是一种通用串行数据总线,用于异步通信。该总线可以为双向通信总线。它将要传输的数据在串行通信与并行通信之间转换。在一些实施例中,UART接口通常被用于连接处理器110与无线通信模块160。例如:处理器110通过UART接口与无线通信模块160中的蓝牙模块通信,实现蓝牙功能。在一些实施例中,音频模块170可以通过UART接口向无线通信模块160传递音频信号,实现通过蓝牙耳机播放音乐的功能。
MIPI接口可以被用于连接处理器110与显示屏194,摄像头193等外围器件。MIPI接口包括摄像头串行接口(camera serial interface,CSI),显示屏串行接口(display serial interface,DSI)等。在一些实施例中,处理器110和摄像头193通过CSI接口通信,实现移动终端100的拍摄功能。处理器110和显示屏194通过DSI接口通信,实现移动终端100的显示功能。
GPIO接口可以通过软件配置。GPIO接口可以被配置为控制信号,也可被配置为数据信号。在一些实施例中,GPIO接口可以用于连接处理器110与摄像头193,显示屏194,无线通信模块160,音频模块170,传感器模块180等。GPIO接口还可以被配置为I2C接口,I2S接口,UART接口,MIPI接口等。
USB接口130是符合USB标准规范的接口,具体可以是Mini USB接口,Micro USB接口,USB Type C接口等。USB接口130可以用于连接充电器为移动终端100充电,也可以用于移动终端100与外围设备之间传输数据。也可以用于连接耳机,通过耳机播放音频。该接口还可以用于连接其他移动终端,例如AR设备等。
可以理解的是,本发明实施例示意的各模块间的接口连接关系,只是示意性说明,并不构成对移动终端100的结构限定。在本申请另一些实施例中,移动终端100也可以采用上述实施例中不同的接口连接方式,或多种接口连接方式的组合。
充电管理模块140用于从充电器接收充电输入。其中,充电器可以是无线充电器,也可以是有线充电器。在一些有线充电的实施例中,充电管理模块140可以通过USB接口130接收有线充电器的充电输入。在一些无线充电的实施例中,充电管理模块140可以通过移动终端100的无线充电线圈接收无线充电输入。充电管理模块140为电池142充电的同时,还可以通过电源管理模块141为移动终端供电。
电源管理模块141用于连接电池142,充电管理模块140与处理器110。电源管理模块141接收电池142和/或充电管理模块140的输入,为处理器110,内部存储器121,显示屏194,摄像头193,和无线通信模块160等供电。电源管理模块141还可以用于监测电池容量,电池循环次数,电池健康状态(漏电,阻抗)等参数。在其他一些实施例中,电源管理模块141也可以设置于处理器110中。在另一些实施例中,电源管理模块141和充电管理模块140也可以设置于同一个器件中。
移动终端100的无线通信功能可以通过天线1,天线2,移动通信模块150,无线通信模块160,调制解调处理器以及基带处理器等实现。
天线1和天线2用于发射和接收电磁波信号。移动终端100中的每个天线可用于覆盖单个或多个通信频带。不同的天线还可以复用,以提高天线的利用率。例如:可以将天线1复用为无线局域网的分集天线。在另外一些实施例中,天线可以和调谐开关结合使用。
移动通信模块150可以提供应用在移动终端100上的包括2G/3G/4G/5G等无线通信的解决方案。移动通信模块150可以包括至少一个滤波器,开关,功率放大器,低噪声放大器(low noise amplifier,LNA)等。移动通信模块150可以由天线1接收电磁波,并对接收的电磁波进行滤波,放大等处理,传送至调制解调处理器进行解调。移动通信模块150还可以对经调制解调处理器调制后的信号放大,经天线1转为电磁波辐射出去。在一些实施例中,移动通信模块150的至少部分功能模块可以被设置于处理器110中。在一些实施例中,移动通信模块150的至少部分功能模块可以与处理器110的至少部分模块被设置在同一个器件中。
调制解调处理器可以包括调制器和解调器。其中,调制器用于将待发送的低频基带信号调制成中高频信号。解调器用于将接收的电磁波信号解调为低频基带信号。随后解调器将解调得到的低频基带信号传送至基带处理器处理。低频基带信号经基带处理器处理后,被传递给应用处理器。应用处理器通过音频设备(不限于扬声器170A,受话器170B等)输出声音信号,或通过显示屏194显示图像或视频。在一些实施例中,调制解调处理器可以是独立的器件。在另一些实施例中,调制解调处理器可以独立于处理器110,与移动通信模块150或其他功能模块设置在同一个器件中。
无线通信模块160可以提供应用在移动终端100上的包括无线局域网(wireless local area networks,WLAN)(如无线保真(wireless fidelity,Wi-Fi)网络),蓝牙(bluetooth,BT),全球导航卫星系统(global navigation satellite system,GNSS),调频(frequency modulation,FM),近距离无线通信技术(near field communication,NFC),红外技术(infrared,IR)等无线通信的解决方案。无线通信模块160可以是集成至少一个通信处理模块的一个或多个器件。无线通信模块160经由天线2接收电磁波,将电磁波信号调频以及滤波处理,将处理后的信号发送到处理器110。无线通信模块160还可以从处理器110接收待发送的信号,对其进行调频,放大,经天线2转为电磁波辐射出去。
在一些实施例中,移动终端100的天线1和移动通信模块150耦合,天线2和无线通信模块160耦合,使得移动终端100可以通过无线通信技术与网络以及其他设备通信。所述无线通信技术可以包括全球移动通讯系统(global system for mobile communications,GSM),通用分组无线服务(general packet radio service,GPRS),码分多址接入(code division multiple access,CDMA),宽带码分多址(wideband code division multiple access,WCDMA),时分码分多址(time-division code division multiple access,TD-SCDMA),长期演进(long term evolution,LTE),BT,GNSS,WLAN,NFC,FM,和/或IR技术等。所述GNSS可以包括全球卫星定位系统(global positioning system,GPS),全球导航卫星系统(global navigation satellite system,GLONASS),北斗卫星导航系统(beidou navigation satellite system,BDS),准天顶卫星系统(quasi-zenith satellite system,QZSS)和/或星基增强系统(satellite based augmentation systems,SBAS)。
移动终端100通过GPU,显示屏194,以及应用处理器等实现显示功能。GPU为图像处理的微处理器,连接显示屏194和应用处理器。GPU用于执行数学和几何计算,用于图形渲染。处理器110可包括一个或多个GPU,其执行程序指令以生成或改变显示信息。
显示屏194用于显示图像,视频等。显示屏194包括显示面板。显示面板可以采用液晶显示屏(liquid crystal display,LCD),有机发光二极管(organic light-emitting diode,OLED),有源矩阵有机发光二极体或主动矩阵有机发光二极体(active-matrix organic light emitting diode的,AMOLED),柔性发光二极管(flex light-emitting diode,FLED),Miniled,MicroLed,Micro-oLed,量子点发光二极管(quantum dot light emitting diodes,QLED)等。在一些实施例中,移动终端100可以包括1个或N个显示屏194,N为大于1的正整数。
移动终端100可以通过ISP,摄像头193,视频编解码器,GPU,显示屏194以及应用处理器等实现拍摄功能。
ISP用于处理摄像头193反馈的数据。例如,拍照时,打开快门,光线通过镜头被传递到摄像头感光元件上,光信号转换为电信号,摄像头感光元件将所述电信号传递给ISP处理,转化为肉眼可见的图像。ISP还可以对图像的噪点,亮度,肤色进行算法优化。ISP还可以对拍摄场景的曝光,色温等参数优化。在一些实施例中,ISP可以设置在摄像头193中。
摄像头193用于捕获静态图像或视频。物体通过镜头生成光学图像投射到感光元件。感光元件可以是电荷耦合器件(charge coupled device,CCD)或互补金属氧化物半导体(complementary metal-oxide-semiconductor,CMOS)光电晶体管。感光元件把光信 号转换成电信号,之后将电信号传递给ISP转换成数字图像信号。ISP将数字图像信号输出到DSP加工处理。DSP将数字图像信号转换成标准的RGB,YUV等格式的图像信号。在一些实施例中,移动终端100可以包括1个或N个摄像头193,N为大于1的正整数。
在本申请的一些实施例中,上述1个或N个摄像头193中可以包括至少一个长焦摄像头,可用于捕获距离移动终端100位置较远的拍摄对象。处理器110(例如具体可以是ISP、CPU、DSP、NPU中的一个或多个)可以对长焦摄像头采集到的图像进行细节增强和背景虚化的处理,提升移动终端长焦拍摄场景下的图像质量,提升用户的视觉体验。
数字信号处理器用于处理数字信号,除了可以处理数字图像信号,还可以处理其他数字信号。例如,当移动终端100在频点选择时,数字信号处理器用于对频点能量进行傅里叶变换等。
视频编解码器用于对数字视频压缩或解压缩。移动终端100可以支持一种或多种视频编解码器。这样,移动终端100可以播放或录制多种编码格式的视频,例如:动态图像专家组(moving picture experts group,MPEG)1,MPEG2,MPEG3,MPEG4等。
NPU为神经网络(neural-network,NN)计算处理器,通过借鉴生物神经网络结构,例如借鉴人脑神经元之间传递模式,对输入信息快速处理,还可以不断的自学习。通过NPU可以实现移动终端100的智能认知等应用,例如:图像识别,人脸识别,语音识别,文本理解等。
外部存储器接口120可以用于连接外部存储卡,例如Micro SD卡,实现扩展移动终端100的存储能力。外部存储卡通过外部存储器接口120与处理器110通信,实现数据存储功能。例如将音乐,视频等文件保存在外部存储卡中。
内部存储器121可以用于存储计算机可执行程序代码,所述可执行程序代码包括指令。内部存储器121可以包括存储程序区和存储数据区。其中,存储程序区可存储操作系统,至少一个功能所需的应用程序(比如声音播放功能,图像播放功能等)等。存储数据区可存储移动终端100使用过程中所创建的数据(比如音频数据,电话本等)等。此外,内部存储器121可以包括高速随机存取存储器,还可以包括非易失性存储器,例如至少一个磁盘存储器件,闪存器件,通用闪存存储器(universal flash storage,UFS)等。处理器110通过运行存储在内部存储器121的指令,和/或存储在设置于处理器中的存储器的指令,执行移动终端100的各种功能应用以及数据处理。
移动终端100可以通过音频模块170,扬声器170A,受话器170B,麦克风170C,耳机接口170D,以及应用处理器等实现音频功能。例如音乐播放,录音等。
音频模块170用于将数字音频信息转换成模拟音频信号输出,也用于将模拟音频输入转换为数字音频信号。音频模块170还可以用于对音频信号编码和解码。在一些实施例中,音频模块170可以设置于处理器110中,或将音频模块170的部分功能模块设置于处理器110中。
扬声器170A,也称“喇叭”,用于将音频电信号转换为声音信号。移动终端100可以通过扬声器170A收听音乐,或收听免提通话。
受话器170B,也称“听筒”,用于将音频电信号转换成声音信号。当移动终端100 接听电话或语音信息时,可以通过将受话器170B靠近人耳接听语音。
麦克风170C,也称“话筒”,“传声器”,用于将声音信号转换为电信号。当拨打电话或发送语音信息时,用户可以通过人嘴靠近麦克风170C发声,将声音信号输入到麦克风170C。移动终端100可以设置至少一个麦克风170C。在另一些实施例中,移动终端100可以设置两个麦克风170C,除了采集声音信号,还可以实现降噪功能。在另一些实施例中,移动终端100还可以设置三个,四个或更多麦克风170C,实现采集声音信号,降噪,还可以识别声音来源,实现定向录音功能等。
耳机接口170D用于连接有线耳机。耳机接口170D可以是USB接口130,也可以是3.5mm的开放移动移动终端平台(open mobile terminal platform,OMTP)标准接口,美国蜂窝电信工业协会(cellular telecommunications industry association of the USA,CTIA)标准接口。
按键190包括开机键,音量键等。按键190可以是机械按键。也可以是触摸式按键。移动终端100可以接收按键输入,产生与移动终端100的用户设置以及功能控制有关的键信号输入。
马达191可以产生振动提示。马达191可以用于来电振动提示,也可以用于触摸振动反馈。例如,作用于不同应用(例如拍照,音频播放等)的触摸操作,可以对应不同的振动反馈效果。作用于显示屏194不同区域的触摸操作,马达191也可对应不同的振动反馈效果。不同的应用场景(例如:时间提醒,接收信息,闹钟,游戏等)也可以对应不同的振动反馈效果。触摸振动反馈效果还可以支持自定义。
指示器192可以是指示灯,可以用于指示充电状态,电量变化,也可以用于指示消息,未接来电,通知等。
SIM卡接口195用于连接SIM卡。SIM卡可以通过插入SIM卡接口195,或从SIM卡接口195拔出,实现和移动终端100的接触和分离。移动终端100可以支持1个或N个SIM卡接口,N为大于1的正整数。SIM卡接口195可以支持Nano SIM卡,Micro SIM卡,SIM卡等。同一个SIM卡接口195可以同时插入多张卡。所述多张卡的类型可以相同,也可以不同。SIM卡接口195也可以兼容不同类型的SIM卡。SIM卡接口195也可以兼容外部存储卡。移动终端100通过SIM卡和网络交互,实现通话以及数据通信等功能。在一些实施例中,移动终端100采用eSIM,即:嵌入式SIM卡。eSIM卡可以嵌在移动终端100中,不能和移动终端100分离。
移动终端100的软件系统可以采用分层架构,事件驱动架构,微核架构,微服务架构,或云架构。本发明实施例以分层架构的Android系统为例,示例性说明移动终端100的软件结构。
图2是本发明实施例的移动终端100的软件结构框图。
分层架构将软件分成若干个层,每一层都有清晰的角色和分工。层与层之间通过软件接口通信。在一些实施例中,将Android系统分为四层,从上至下分别为应用程序层,应用程序框架层,安卓运行时(Android runtime)和系统库,以及内核层。
如图2所示,应用程序层可以包括一系列应用程序包,包括移动终端出厂前预置的应用程序,或者移动终端出厂后,用户通过例如应用市场或其他方式安装的应用程序。这些应用程序包括但不限于包括相机,图库,日历,通话,地图,导航,WLAN, 蓝牙,音乐,视频,短信息、浏览器、微信、淘宝、支付宝等应用程序(图中仅示出部分)。
应用程序框架层为应用程序层的应用程序提供应用编程接口(application programming interface,API)和编程框架。应用程序框架层包括一些预先定义的函数。
如图2所示,应用程序框架层可以包括窗口管理器,内容提供器,视图系统,电话管理器,资源管理器,通知管理器等。窗口管理器用于管理窗口程序。窗口管理器可以获取显示屏大小,判断是否有状态栏,锁定屏幕,截取屏幕等。内容提供器用来存放和获取数据,并使这些数据可以被应用程序访问。所述数据可以包括视频,图像,音频,拨打和接听的电话,浏览历史和书签,电话簿等。视图系统包括可视控件,例如显示文字的控件,显示图片的控件等。视图系统可用于构建应用程序。显示界面可以由一个或多个视图组成的。例如,包括短信通知图标的显示界面,可以包括显示文字的视图以及显示图片的视图。电话管理器用于提供移动终端100的通信功能。例如通话状态的管理(包括接通,挂断等)。资源管理器为应用程序提供各种资源,比如本地化字符串,图标,图片,布局文件,视频文件等等。通知管理器使应用程序可以在状态栏中显示通知信息,可以用于传达告知类型的消息,可以短暂停留后自动消失,无需用户交互。比如通知管理器被用于告知下载完成,消息提醒等。通知管理器还可以是以图表或者滚动条文本形式出现在系统顶部状态栏的通知,例如后台运行的应用程序的通知,还可以是以对话窗口形式出现在屏幕上的通知。例如在状态栏提示文本信息,发出提示音,移动终端振动,指示灯闪烁等。
Android Runtime包括核心库和虚拟机。Android runtime负责安卓系统的调度和管理。
核心库包含两部分:一部分是java语言需要调用的功能函数,另一部分是安卓的核心库。
应用程序层和应用程序框架层运行在虚拟机中。虚拟机将应用程序层和应用程序框架层的java文件执行为二进制文件。虚拟机用于执行对象生命周期的管理,堆栈管理,线程管理,安全和异常的管理,以及垃圾回收等功能。
系统库可以包括多个功能模块。例如:表面管理器(surface manager),媒体库(Media Libraries),三维图形处理库(例如:OpenGL ES),2D图形引擎(例如:SGL)等。
表面管理器用于对显示子系统进行管理,并且为多个应用程序提供了2D和3D图层的融合。
媒体库支持多种常用的音频,视频格式回放和录制,以及静态图像文件等。媒体库可以支持多种音视频编码格式,例如:MPEG4,H.264,MP3,AAC,AMR,JPG,PNG等。
三维图形处理库用于实现三维图形绘图,图像渲染,合成,和图层处理等。
2D图形引擎是2D绘图的绘图引擎。
在本申请的一些实施例中,系统库中还可以包括第一模块,可用于对移动终端100中长焦摄像头采集到的图像进行细节增强和背景虚化的处理,提升移动终端长焦拍摄场景下的图像质量,提升用户的视觉体验。
内核层是硬件和软件之间的层。内核层至少包含显示驱动,摄像头驱动,音频驱动,传感器驱动。
下面结合捕获拍照场景,示例性说明移动终端100软件以及硬件的工作流程。
当触摸传感器180K接收到触摸操作,相应的硬件中断被发给内核层。内核层将触摸操作加工成原始输入事件(包括触摸坐标,触摸操作的时间戳等信息)。原始输入事件被存储在内核层。应用程序框架层从内核层获取原始输入事件,识别该输入事件所对应的控件。以该触摸操作是触摸单击操作,该单击操作所对应的控件为相机应用图标的控件为例,相机应用调用应用框架层的接口,启动相机应用,进而通过调用内核层启动摄像头驱动,通过摄像头193捕获静态图像或视频。
以下实施例中所涉及的技术方案均可以在具有上述硬件架构和软件架构的移动终端100中实现。
本文以移动终端100是手机为例,结合附图对本申请实施例提供的技术方案进行详细说明。
示例的,用户可以通过触摸手机屏幕上特定的控件、按压特定的物理按键或按键组合、输入语音、隔空手势等方式,指示手机开启相机应用。响应于接收到用户开启相机的指示后,手机启动相机,显示拍摄界面。
例如:如图3中(1)所示,用户可以通过在手机桌面上点击“相机”应用图标301,指示手机开启相机应用,手机显示如图3中(2)所示的拍摄界面。
再例如:在手机处于锁屏状态时,用户也可以通过在手机屏幕上向右滑动的手势,指示手机开启相机应用,手机也可以显示如图3中(2)所示的拍摄界面。
或者,手机处于锁屏状态时,用户可以通过在锁屏界面上点击“相机”应用的快捷图标,指示手机开启相机应用,手机也可以显示如图3中(2)所示的拍摄界面。
又例如:在手机运行其他应用时,用户也可以通过点击相应的控件使得手机开启相机应用进行拍照。比如:用户正在使用即时通信类应用(例如:微信应用)时,用户也可以通过选择相机功能的控件,指示手机开启相机应用进行拍照和拍摄视频。
如图3中(2)所示,相机的拍摄界面一般包括有取景框302、拍摄控件以及其他功能控件(“大光圈”、“人像”、“拍照”、“录像”等)。其中,取景框302可用于预览摄像头采集到的图像(或画面),用户可以基于取景框302中的图像(或画面),决定指示手机执行拍摄操作的时机。其中,用户指示手机执行拍摄操作例如可以是用户点击拍摄控件的操作,或者用户按压音量键的操作等。在一些实施例中,拍摄界面中还可以包括变焦倍率指示303。通常,手机默认的变焦倍率为基本倍率,为“1×”。
举例说明,以手机集成有短焦(广角)摄像头、中焦摄像头和长焦摄像头三个摄像头为例进行说明。通常,用户使用中焦摄像头的场景最多,因此,一般将中焦摄像头设置为主摄像头。主摄像头的焦距设置为基准焦距,变焦倍率为“1×”。在一些实施例中,可以对主摄像头捕获的图像进行数码变焦(digital zoom)(或称为数字变焦),即,通过手机中的ISP或其他处理器将主摄像头捕获的“1×”的图像的每个像素面积放大,并相应缩小图像的取景范围,使得处理后的图像呈现等效于主摄像头采用其他变焦倍率(例如“2×”)拍摄的图像。也就是说,使用主摄像头拍摄的图像可以对应一个区间的变焦倍率,例如:“1×”至“5×”。
类似的,长焦摄像头的焦距与主摄像头的焦距的倍数,可作为长焦摄像头的变焦倍率。例如,长焦摄像头的焦距可以为主摄像头的焦距的5倍,即长焦摄像头的变焦 倍率为“5×”。同理,也可以对长焦摄像头捕获的图像进行数码变焦。即,使用长焦摄像头拍摄的图像可以对应另一个区间的变焦倍率,例如:“5×”至“50×”。
类似的,短焦(广角)摄像头的焦距与主摄像头的焦距的倍数,可作为短焦(广角)摄像头的变焦倍率。例如,短焦摄像头的焦距可以为主摄像头的焦距的0.5倍,即长焦摄像头的变焦倍率为“0.5×”。同理,也可以对短焦(广角)摄像头捕获的图像进行数码变焦。即,使用长焦摄像头拍摄的图像可以对应另一个区间的变焦倍率,例如:“0.5×”至“1×”。
在实际的拍摄场景中,短焦(广角)摄像头和中焦摄像头多用于拍摄距离手机位置较近的拍摄对象。长焦摄像头一般用于拍摄距离手机位置较远的拍摄对象。当然,在一些场景中,也可以用中焦摄像头的高变焦倍率拍摄距离手机位置较远的拍摄对象。本申请实施例中长焦拍摄场景可以理解为手机的变焦倍率大于预设倍率(例如:“5×”,“10×”,“20×”)的拍摄场景。在长焦拍摄场景下,手机可以使用中焦摄像头拍摄的图像,也可以使用长焦摄像头拍摄的图像,本申请实施例对此不做具体限定。
需要注意的是,在长焦拍摄场景中,由于拍摄对象距离手机的位置较远,手机摄像头采集到的原始图像清晰度不高。而且,当拍摄对象距离手机位置较远,手机使用长焦高倍率拍摄时,由于图像的景深很浅,手机无法通过大光圈等手段实现背景虚化的拍照效果。因此,本申请实施例提供了一种长焦拍摄场景下拍摄方法,可以对手机摄像头采集到的原始图像进行处理。具体的,智能识别出拍摄对象,并对拍摄对象进行细节增强,提升原始图像中拍摄对象部分图像的清晰度。还能根据识别出的拍摄对象自动识别图像的前后景,虚化拍摄对象外的背景,提升长焦拍摄场景下的图像的艺术感,增强用户的视觉体验。
其中,清晰度是指图像上各个部位的纹理细节及边界的清晰程度。例如,图像1为手机摄像头采集到的原始图像。若采用本申请实施例提供的方法,对图像1进行处理后,得到图像2。图像2中拍摄对象的清晰度高于图像1中拍摄对象的清晰度。若将图像1和图像2放大相同倍数,对比放大后的图像1和图像2,会发现放大后的图像2中拍摄对象的纹理细节更丰富,拍摄对象的各个部位的分界也更清晰。
下文,从手机切换到长焦拍摄场景,对摄像头采集的原始图像进行处理,预览显示处理后的图像,到拍摄生成的照片或视频的整个过程进行详细说明。
1、手机切换到长焦拍摄场景
在一些实施例中,用户可以手动调整手机拍摄时使用的变焦倍率。
例如:如图3中(2)所示,用户可以通过在拍摄界面中操作变焦倍率指示303,调整手机使用的变焦倍率。比如:当前手机使用的变焦倍率为“1×”时,用户可以通过点击一次或多次变焦倍率指示303,使得手机使用的变焦倍率变更为“5×”,即手机显示如图4中(1)所示的拍摄界面。在图4中(1)所示的拍摄界面中,取景框302中预览到的图像的取景范围显然比图3中(2)中取景框302中的取景范围小,但取景框302中预览到的被拍摄对象(例如:鸟)的尺寸大于取景框302中预览到的被拍摄对象。在一些示例中,图4中(1)所示的拍摄界面中还可以继续显示变焦倍率指示303,此时显示当前变焦倍率为“5×”,以便用户知晓当前的变焦倍率。
又例如:如图3中(3)所示,用户可以通过在拍摄界面中双指(或三指)捏合的 手势,减小手机使用的变焦倍率,或者双指(或三指)向外滑动的手势(与捏合相反的方向)增大手机使用的变焦倍率。
又例如:如图3中(4)所示,用户也可以在拍摄界面中通过拖动变焦标尺304来改变手机使用的变焦倍率。
又例如:用户也可以在拍摄界面中或拍摄设置界面中,通过切换当前所使用的摄像头来改变手机的变焦倍率。例如,用户选择切换到长焦摄像头,则手机自动增大变焦倍率。
又例如:用户也可以在拍摄界面中或拍摄设置界面中,通过选择长焦拍摄场景的选项或者远距离的拍摄场景的选项等,来改变手机的变焦倍率。
在另一些实施例中,手机也可以自动识别摄像头捕获的图像的具体场景,根据识别的具体场景自动调整变焦倍率。比如:若手机识别出摄像头捕获的图像为视野范围较大的景物,例如大海、高山、树林等,可以自动减小变焦倍数。又比如:若手机识别出摄像头捕获的图像为远距离的物体时,例如远处的鸟、运动场上的运动员等,可以自动增大变焦倍数,本申请对此不做限定。
当手机的变焦倍率被调整到大于或等于预设倍率时,手机进入长焦拍摄场景。
2、长焦拍摄场景下的图像预览
在本申请的一些实施例中,当手机的变焦倍率被调整到大于或等于预设倍率,例如“10×”,手机显示如图4中(2)所示的拍摄界面。在如图4中(2)所示的拍摄界面中,取景框302中包括有标记框401,用于标记手机识别出的拍摄对象。例如,拍摄对象为鸟。需要说明的是,图中示出的拍摄对象处于运动状态。本申请实施例提供的方法不限于拍摄对象是否处于运动状态或静止状态。在一些示例中,手机还可以显示提示框402,用于向用户推荐该拍摄对象类别对应的图像处理模式。例如,识别出的拍摄对象为鸟类,询问用户是否开启鸟类对应的图像处理模式(或称为鸟类模式、鸟类效果)。响应于用户选择开启相应的图像处理模式后,手机对采集到的图像进行相应的图像处理,并显示如图4中(3)所示的拍摄界面。在另一些示例中,在手机识别出拍摄对象的类别和位置后,也可以自动使用拍摄对象的类别对应的图像处理模式进行处理。即,当手机的变焦倍率被调整到大于或等于预设倍率时,手机直接显示如图4中(3)所示的拍摄界面。
在图4中(3)所示的拍摄界面中,取景框302中显示的图像为进行相应图像处理后的图像。可以看到,识别出的拍摄对象更清晰,清晰度更高。可选的,手机还可以显示识别出的拍摄对象对应的标签403。可选的,手机还可以对识别出的拍摄对象的背景进行虚化,使得处理后的图像具有艺术感,提升图像的视觉效果。
当然,在又一些示例中,在手机默认选择使用拍摄对象类别对应的图像处理模式后,或者用户手动开启拍摄对象类别对应的图像处理模式后,用户也可以手动关闭拍摄对象类别对应的图像处理模式,或者手机在识别出的拍摄对象的类别未对应有相应的图像处理模式时,自动关闭拍摄对象类别对应的图像处理模式。
在一种具体实现方式中,手机的ISP或其他处理器对摄像头采集到图像进行图像分析,从图像中识别出用户意图拍摄的拍摄对象,并对拍摄对象的类别和位置进行进一步的识别。其中,拍摄对象的类别可以基于用户通常拍摄到的物体进行的任意分类, 可以是一级分类,也可以是多级分类。例如:拍摄对象的类别可以包括:人、动物、植物等。其中,动物类别下还可以包括第二级分类,例如:狗、猫、松鼠、鸟等。进一步的,鸟的类别下还可以包括第三级分类,例如:麻雀、喜鹊、海鸥、海鸟、鸽子、大雁等。
当摄像头采集的图像中包括多个物体时,可以默认面积最大的物体为拍摄对象,或者位于图像中心区域的物体为拍摄对象,或者用户选择的物体(例如通过点击操作选择的物体)为拍摄对象。当然,也可以默认或选择多个物体作为拍摄对象,本申请实施例对此不做限定。
在一些示例中,手机的ISP或其他处理器可以采用基于深度学习的目标检测技术,例如:区域卷积神经网络(Region-Convolutional Neural Networks,R-CNN)、快速区域卷积神经网络(Fast R-CNN)、更快速区域卷积神经网络(Faster R-CNN)等,对拍摄对象的类别和位置的识别。可选的,可以在手机取景框中标记出拍摄对象的类别和位置,便于用户确认标记的是否为意图拍摄的拍摄对象,以及手机对拍摄对象的类别和位置是否识别准确。若确认手机自动识别的物体类别和位置正确,可以使用手机推荐的相应的图像处理模式。否则,用户也可以选择不使用手机推荐的图像处理模式,或者变更拍摄对象等,避免进行错误的图像处理。
在识别出拍摄对象的类别和位置后,可以对拍摄对象部分的图像进行处理。例如,可以基于图像分割技术,将拍摄对象对应的部分图像从采集的图像中分割出来,然后对分割出来的拍摄对象进行细节增强。
可以理解,手机可以预先训练出基于拍摄对象的类别,对拍摄对象进行细节增强的AI模型。当然,手机也可以直接从服务器处获取已训练好的AI模型。需要说明的是,在训练基于拍摄对象的类别,对拍摄对象进行细节增强的AI模型时,需要输入大量拍摄对象的训练样本,包括各个类别的训练样本。例如,针对鸟类中的喜鹊的训练样本,包括不同品种的喜鹊的图像,同一品种喜鹊的不同体积或颜色的图像,同一品种喜鹊的不同姿态的图像等,同一品种喜鹊在不同环境中的图像等。其中,对拍摄对象进行细节增强,可以包括对拍摄对象中模糊的区域,采用AI模型进行智能的像素填充等,以提升图像的清晰度。还包括:对拍摄对象中缺失的部位,采用AI模型进行智能的像素填充等,以修复拍摄对象的缺失部分等。还可以包括:提升拍摄对象部分的图像的整体清晰度等。
在另一些示例中,在手机对拍摄对象进行细节增强之前,手机还可以基于识别出的拍摄对象,自动对焦(automatic focus,AF)和自动曝光(automatic exposure,AE),以突出拍摄对象,提升图像的质量。可以注意的是,现有技术中一般基于固定位置进行对焦和曝光的,或者基于用户选择的位置(例如点击操作的触碰位置)进行对焦和曝光。然而,本申请实施例中是基于手机自动识别出的拍摄对象的位置进行对焦和曝光,以提升拍摄对象整体的清晰度。可以注意到,现有技术中,对焦框的大小是固定的,通常为手机默认设置的。然而,在本申请实施例基于拍摄对象的自动对焦过程中,对焦框的大小会随着手机识别出的拍摄对象的大小而自动调整。一些示例中,对焦框可以与图4中(2)所示的标记框401相同。
可选的,由于本申请实施例提供的方法是运用于长焦拍摄场景中,故可以设置手 机自动对焦时的最近搜索距离,例如10米。也就是说,手机在自动对焦时,搜索的距离范围为距离手机10米外至无穷远。换言之,手机不用搜索在距离手机10米至微距的范围。由此,可以减少手机自动对焦的时间,提升图像处理的效率。
在又一些示例中,在手机对拍摄对象进行细节增强之前,手机还可以基于自动对焦和自动曝光时确定的参数,采集多张(例如:6张)包含拍摄对象的图像。针对多张包含拍摄对象的图像,采用超分辨(super resolution)技术,合成一张高清图像。然后使用合成后的高清图像进行拍摄对象的细节增强。由此,进一步提升了图像的整体清晰度。
在又一些示例中,还可以采用背景提取算法,基于识别出的拍摄对象,识别出拍摄对象的背景,并对背景进行虚化。这样,更加突显拍摄对象,提升图像整体的艺术美感。其中,背景提取算法包括但不限于帧间差分法、背景差分法、环境算法(ViBe算法,ViBe+算法)等,本申请实施例对此不做限定。其中,对背景进行虚化,可以采用模糊算法对背景所在区域的图像进行处理,得到添加虚化效果的背景图像。模糊算法包括高斯滤波(Gaussian Filtering)、圆形滤波、导向滤波(Guided Filtering)、域滤波(Domain Filtering)中任一项或任几项。
在本申请的另一些实施例中,手机还可以将识别的拍摄对象的位置进行调整。在具体实现时,可以基于识别出的拍摄对象在图像中的位置,对图像中拍摄对象的背景进行裁剪或填充,使得拍摄对象的位置位于图像的中心位置上。或者,拍摄对象的位置也可以位于图像的其他位置,例如用于中心位置偏左或偏右一定距离的位置,本申请实施例对此不做限定。
例如,如图4中(4)所示的拍摄界面,取景框302中显示的图像中拍摄对象位于取景框的中心区域。这样,可以突出拍摄对象,进一步提升图像的视觉效果。
在本申请的又一些实施例中,用户在预览时,还可以选择在拍摄对象的类别对应的图像处理效果和不进行处理(即无效果)之间进行切换。例如,如图5中(1)所示的界面,手机还可以显示选项502。若用户选择选项502中的鸟类效果的选项,则取景框302中显示的图像为按照鸟类效果模式处理后的图像。若用户现在选项502中无效果的选项,如图5中(2)所示的界面,取景框302中显示的图像为未按照鸟类效果模式处理的图像。对比可以看出图5中(1)所示的界面中,取景框302中标记框501的图像(即拍摄对象)的清晰度,比图5中(2)中标记框501中图像的清晰度大。在另一个示例中,图5中(1)所示的界面中,取景框302中标记框501之外的图像(即背景)的清晰度,比图5中(2)中的标记框501之外图像的清晰度小。在又一个示例中,图5中(1)所示的界面中标记框501在取景框302中的位置,与图5中(2)所示的界面中标记框501位置也不同。由此可见,用户可以通过选项502切换图像处理前后的效果,以便用户进行对比,选择更喜欢的方法进行拍摄,提升用户的体验。
在本申请的又一些实施例中,在用户选择使用拍摄对象类别对应的图像处理后,还可以进一步选择是否调整拍摄对象的位置。例如,如图5中(3)所示的界面中,还可以显示选项控件503。当用户未选中选项控件503时,手机保留拍摄对象原始位置。若用户选中选项控件503时,手机显示如图5中(4)所示的界面,标记框501位于取景框302的中心区域。即,手机将拍摄对象的位置调整到图像中心区域。
在本申请的又一些实施例中,用户也可以手动开启本申请实施例提供的基于拍摄对象类别的图像处理功能。例如,如图6中(1)所示的界面,响应于检测到用户操作设置控件601,手机显示如图6中(2)所示的拍摄设置界面602。该拍摄设置界面602中显示有控件603,用于开启基于拍摄对象类别的图像处理功能。也就是说,当用户开启该功能后,手机处于长焦拍摄场景时会自动采用本申请实施例提供的方法,对采集的图像进行处理。当然,用户也可以通过该控件603,手动关闭长焦场景下基于拍摄对象类别的图像处理功能。
可选的,如图6中(3)所示的拍摄设置界面604中,可以显示有一些常用的类别控件605。用户可以通过该类别控件605,开启相应类别的图像处理功能。也就是说,当用户开启某个类别对应的图像处理功能后,手机处于长焦拍摄场景时会自动采用相应的图像处理方法,对采集的图像进行处理。例如:用户选择开启鸟类效果的功能,则手机会将采集到图像按照鸟类的图像处理模式进行处理。
或者,如图6中(4)所示的设置界面606,显示有多个类别的选项。用户可以根据实际场景选择相应的类别,使得手机按照相应的图像处理模式对采集的图像进行处理。
可见,当用户选择具体类别的处理模式后,手机可以不用再自动识别拍摄对象的类别,直接使用相应的图像处理模式,可以加速图像处理效果。或者,在手机识别错了拍摄对象的类别后,用户也可以通过该手动的方式选择正确的类别对应的图像处理模式。
3、拍照或拍摄视频
用户可以基于取景框中的预览,决定拍摄的时机。在检测到用户执行拍摄的操作后,手机基于当前时刻采集的图像,进行相应的图像处理,得到处理后的图像。其中,图像处理的方法可参考预览时的图像处理,这里不再赘述。
手机可以在相册中保存经过基于拍摄对象类别的图像处理后的图像。在一些示例中,手机还可以在相册中同时保存未经过基于拍摄对象类别的图像处理后的图像。
在手机处于连拍、拍摄慢动作、录制视频等其他模式下,也可以采用本申请提供的方法对采集的图像进行相应的处理。例如:手机连拍时,摄像头采集得到多张图像,手机可以对每张图像都进行基于拍摄对象的图像处理,得到多张处理后的图像。又例如:手机拍摄慢动作或录制视频时,慢动作和视频也都是有一帧帧图像组成,可以对其中每一帧图像进行基于拍摄对象的图像处理,再将处理后的图像组成新的慢动作和新的视频。
对于手机中保存的图像或者从其他设备接收的图像,该图像为场景拍摄场景下的图像时,也可以采用本申请提供的方法,对图像进行基于拍摄对象的图像处理。
本申请实施例还提供一种芯片系统,如图7所示,该芯片系统包括至少一个处理器1101和至少一个接口电路1102。处理器1101和接口电路1102可通过线路互联。例如,接口电路1102可用于从其它装置(例如移动终端100的存储器)接收信号。又例如,接口电路1102可用于向其它装置(例如处理器1101)发送信号。示例性的,接口电路1102可读取存储器中存储的指令,并将该指令发送给处理器1101。当所述指令被处理器1101执行时,可使得移动终端执行上述实施例中的移动终端100(比如, 手机)执行的各个步骤。当然,该芯片系统还可以包含其他分立器件,本申请实施例对此不作具体限定。
本申请实施例还提供一种装置,该装置包含在移动终端中,该装置具有实现上述实施例中任一方法中移动终端行为的功能。该功能可以通过硬件实现,也可以通过硬件执行相应的软件实现。硬件或软件包括至少一个与上述功能相对应的模块或单元。例如,检测模块或单元、以及确定模块或单元等。
本申请实施例还提供一种计算机可读存储介质,包括计算机指令,当计算机指令在移动终端上运行时,使得移动终端执行如上述实施例中任一方法。
本申请实施例还提供一种计算机程序产品,当计算机程序产品在计算机上运行时,使得计算机执行如上述实施例中任一方法。
可以理解的是,上述终端等为了实现上述功能,其包含了执行各个功能相应的硬件结构和/或软件模块。本领域技术人员应该很容易意识到,结合本文中所公开的实施例描述的各示例的单元及算法步骤,本申请实施例能够以硬件或硬件和计算机软件的结合形式来实现。某个功能究竟以硬件还是计算机软件驱动硬件的方式来执行,取决于技术方案的特定应用和设计约束条件。专业技术人员可以对每个特定的应用来使用不同方法来实现所描述的功能,但是这种实现不应认为超出本发明实施例的范围。
本申请实施例可以根据上述方法示例对上述终端等进行功能模块的划分,例如,可以对应各个功能划分各个功能模块,也可以将两个或两个以上的功能集成在一个处理模块中。上述集成的模块既可以采用硬件的形式实现,也可以采用软件功能模块的形式实现。需要说明的是,本发明实施例中对模块的划分是示意性的,仅仅为一种逻辑功能划分,实际实现时可以有另外的划分方式。
通过以上的实施方式的描述,所属领域的技术人员可以清楚地了解到,为描述的方便和简洁,仅以上述各功能模块的划分进行举例说明,实际应用中,可以根据需要而将上述功能分配由不同的功能模块完成,即将装置的内部结构划分成不同的功能模块,以完成以上描述的全部或者部分功能。上述描述的系统,装置和单元的具体工作过程,可以参考前述方法实施例中的对应过程,在此不再赘述。
在本申请实施例各个实施例中的各功能单元可以集成在一个处理单元中,也可以是各个单元单独物理存在,也可以两个或两个以上单元集成在一个单元中。上述集成的单元既可以采用硬件的形式实现,也可以采用软件功能单元的形式实现。
所述集成的单元如果以软件功能单元的形式实现并作为独立的产品销售或使用时,可以存储在一个计算机可读取存储介质中。基于这样的理解,本申请实施例的技术方案本质上或者说对现有技术做出贡献的部分或者该技术方案的全部或部分可以以软件产品的形式体现出来,该计算机软件产品存储在一个存储介质中,包括若干指令用以使得一台计算机设备(可以是个人计算机,服务器,或者网络设备等)或处理器执行本申请各个实施例所述方法的全部或部分步骤。而前述的存储介质包括:快闪存储器、移动硬盘、只读存储器、随机存取存储器、磁碟或者光盘等各种可以存储程序代码的介质。
以上所述,仅为本申请的具体实施方式,但本申请的保护范围并不局限于此,任何在本申请揭露的技术范围内的变化或替换,都应涵盖在本申请的保护范围之内。因 此,本申请的保护范围应以所述权利要求的保护范围为准。
Claims (16)
- 一种长焦场景下的拍摄方法,其特征在于,应用于包含相机的移动终端中,所述方法包括:所述移动终端启动所述相机,所述移动终端显示取景框,所述取景框用于显示第一预览画面,所述第一预览画面对应的所述相机的变焦倍率为第一倍率;接收用户输入的增大所述相机变焦倍率的第一操作;响应于所述第一操作,所述取景框显示第二预览画面,所述第二预览画面对应的所述相机的变焦倍率为第二倍率,所述第二倍率大于所述第一倍率;若所述第二倍率大于或等于预设倍率,则在所述移动终端显示所述第二预览画面之前,所述移动终端对所述相机采集的图像进行第一图像处理,生成所述第二预览画面;其中,所述第一图像处理包括:根据所述相机采集的图像识别出目标拍摄对象,以及所述目标拍摄对象的背景;提升所述目标拍摄对象所在区域的图像清晰度,为所述目标拍摄对象背景所在区域的图像添加虚化效果。
- 根据权利要求1所述的方法,其特征在于,所述方法还包括:所述移动终端在所述第二预览画面上显示有标记框,用于标记所述目标拍摄对象所在的区域。
- 根据权利要求1或2所述的方法,其特征在于,所述方法还包括:所述移动终端在所述第二预览画面上显示提示信息,用于推荐所述目标拍摄对象对应的所述第一图像处理。
- 根据权利要求3所述的方法,其特征在于,所述移动终端对所述相机采集的图像进行第一图像处理,包括:所述移动终端接收用户输入的选择所述第一图像处理的第二操作;响应于所述第二操作,所述移动终端对所述相机采集的图像进行所述第一图像处理。
- 根据权利要求1-4任一项所述的方法,其特征在于,所述第一图像处理,还包括:调整所述第二预览画面,以使所述目标拍摄对象位于所述取景框的中心区域。
- 根据权利要求1-5任一项所述的方法,其特征在于,在所述移动终端显示所述第二预览画面之后,所述方法还包括:接收用户输入的第三操作,所述第三操作用于指示所述移动终端关闭所述目标拍摄对象对应的所述第一图像处理功能;响应于所述第三操作,所述移动终端确定不对所述相机采集的图像进行所述第一图像处理。
- 根据权利要求1-6任一项所述的方法,其特征在于,所述第一图像处理中所述提升所述目标拍摄对象所在区域的图像清晰度,具体为:识别出所述目标拍摄对象的类别,从所述相机采集到的图像中分割出所述目标拍摄对象的第一图像;将所述第一图像输入到所述目标拍摄对象的类别对应的神经网络模型中,输出所述目标拍摄对象的第二图像,其中,所述第二图像的清晰度大于所述第一图像的清晰度。
- 根据权利要求7所述的方法,其特征在于,所述目标拍摄对象的类别对应的神经网络模型,为根据所述目标拍摄对象的类别下的多个拍摄对象的图像训练得到。
- 根据权利要求1-8任一项所述的方法,其特征在于,所述第一图像处理中识别出所述目标拍摄对象的背景,具体为:采用背景提取算法从所述相机采集的图像中识别出所述目标拍摄对象的背景,其中,所述背景提取算法包括:帧间差分法、背景差分法、环境算法中的任一项或任几项。
- 根据权利要求1-9任一项所述的方法,其特征在于,所述第一图像处理,还包括:在所述相机采集包含所述目标拍摄对象的图像时,自动基于所述目标拍摄对象进行对焦。
- 根据权利要求1-10任一项所述的方法,其特征在于,所述第一图像处理,还包括:控制所述相机连续采集N张包含所述目标拍摄对象的图像,采用超分辨率技术将所述N张包含所述目标拍摄对象的图像合成为第三图像;基于所述第三图像提升所述目标拍摄对象所在区域的图像清晰度,为所述目标拍摄对象背景所在区域的图像添加虚化效果。
- 根据权利要求1-11任一项所述的方法,其特征在于,所述第一图像处理中为所述目标拍摄对象背景所在区域的图像添加虚化效果,具体为:采用模糊算法对所述目标拍摄对象背景所在区域的图像进行处理,所述模糊算法包括高斯滤波、圆形滤波、导向滤波、域滤波中任一项或任几项。
- 一种移动终端,其特征在于,包括:处理器、存储器和触摸屏,所述存储器、所述触摸屏与所述处理器耦合,所述存储器用于存储计算机程序代码,所述计算机程序代码包括计算机指令,当所述处理器从所述存储器中读取所述计算机指令,以使得所述移动终端执行如权利要求1-12中任一项所述长焦场景下的拍摄方法。
- 一种计算机可读存储介质,其特征在于,包括计算机指令,当所述计算机指令在终端上运行时,使得所述终端执行如权利要求1-12中任一项所述长焦场景下的拍摄方法。
- 一种芯片系统,其特征在于,包括一个或多个处理器,当所述一个或多个处理器执行指令时,所述一个或多个处理器执行如权利要求1-12中任一项所述长焦场景下的拍摄方法。
- 一种移动终端上的图形用户界面,其特征在于,所述移动终端具有显示屏、摄像头、存储器、以及一个或多个处理器,所述一个或多个处理器用于执行存储在所述存储器中的一个或多个计算机程序,所述图形用户界面包括所述移动终端执行如权利要求1-12中任意一项所述的方法时显示的图形用户界面。
Priority Applications (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US17/779,876 US12096120B2 (en) | 2020-01-14 | 2020-10-28 | Photographing method in telephoto scenario and mobile terminal |
EP20914044.1A EP4020967B1 (en) | 2020-01-14 | 2020-10-28 | Photographic method in long focal length scenario, and mobile terminal |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202010038444.XA CN113194242B (zh) | 2020-01-14 | 2020-01-14 | 一种长焦场景下的拍摄方法及移动终端 |
CN202010038444.X | 2020-01-14 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2021143269A1 true WO2021143269A1 (zh) | 2021-07-22 |
Family
ID=76863529
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/CN2020/124545 WO2021143269A1 (zh) | 2020-01-14 | 2020-10-28 | 一种长焦场景下的拍摄方法及移动终端 |
Country Status (4)
Country | Link |
---|---|
US (1) | US12096120B2 (zh) |
EP (1) | EP4020967B1 (zh) |
CN (1) | CN113194242B (zh) |
WO (1) | WO2021143269A1 (zh) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2023015959A1 (zh) * | 2021-08-12 | 2023-02-16 | 荣耀终端有限公司 | 拍摄方法及电子设备 |
Families Citing this family (11)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9854156B1 (en) | 2016-06-12 | 2017-12-26 | Apple Inc. | User interface for camera effects |
US11112964B2 (en) | 2018-02-09 | 2021-09-07 | Apple Inc. | Media capture lock affordance for graphical user interface |
US11039074B1 (en) | 2020-06-01 | 2021-06-15 | Apple Inc. | User interfaces for managing media |
US11212449B1 (en) * | 2020-09-25 | 2021-12-28 | Apple Inc. | User interfaces for media capture and management |
US11778339B2 (en) | 2021-04-30 | 2023-10-03 | Apple Inc. | User interfaces for altering visual media |
US12112024B2 (en) | 2021-06-01 | 2024-10-08 | Apple Inc. | User interfaces for managing media styles |
CN113645408B (zh) * | 2021-08-12 | 2023-04-14 | 荣耀终端有限公司 | 拍摄方法、设备及存储介质 |
CN114710624B (zh) * | 2022-04-24 | 2024-09-03 | 维沃移动通信有限公司 | 拍摄方法和拍摄装置 |
CN115529413A (zh) * | 2022-08-26 | 2022-12-27 | 华为技术有限公司 | 拍摄方法及相关装置 |
CN115802144B (zh) * | 2023-01-04 | 2023-09-05 | 荣耀终端有限公司 | 视频拍摄方法及相关设备 |
CN116582743A (zh) * | 2023-07-10 | 2023-08-11 | 荣耀终端有限公司 | 一种拍摄方法、电子设备及介质 |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20120063697A1 (en) * | 2010-09-13 | 2012-03-15 | Samsung Electronics Co., Ltd. | Image processing method and image processing apparatus |
CN103871051A (zh) * | 2014-02-19 | 2014-06-18 | 小米科技有限责任公司 | 图像处理方法、装置和电子设备 |
CN108833768A (zh) * | 2018-05-10 | 2018-11-16 | 信利光电股份有限公司 | 一种多摄像头的拍摄方法、拍摄终端和可读存储介质 |
CN110248081A (zh) * | 2018-10-12 | 2019-09-17 | 华为技术有限公司 | 图像捕捉方法及电子设备 |
CN110572581A (zh) * | 2019-10-14 | 2019-12-13 | Oppo广东移动通信有限公司 | 基于终端设备的变焦虚化图像获取方法和装置 |
Family Cites Families (11)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2009193193A (ja) | 2008-02-13 | 2009-08-27 | Seiko Epson Corp | 画像再生装置、画像再生装置の制御方法および制御プログラム |
JP5036599B2 (ja) * | 2008-03-05 | 2012-09-26 | 株式会社リコー | 撮像装置 |
US8134597B2 (en) * | 2008-12-05 | 2012-03-13 | Sony Ericsson Mobile Communications Ab | Camera system with touch focus and method |
EP3101889A3 (en) * | 2015-06-02 | 2017-03-08 | LG Electronics Inc. | Mobile terminal and controlling method thereof |
KR102560780B1 (ko) * | 2016-10-05 | 2023-07-28 | 삼성전자주식회사 | 복수의 이미지 센서를 포함하는 이미지 처리 시스템 및 그것을 포함하는 전자 장치 |
CN107948516A (zh) * | 2017-11-30 | 2018-04-20 | 维沃移动通信有限公司 | 一种图像处理方法、装置及移动终端 |
CN108024056B (zh) * | 2017-11-30 | 2019-10-29 | Oppo广东移动通信有限公司 | 基于双摄像头的成像方法和装置 |
CN108307106B (zh) * | 2017-12-29 | 2021-01-26 | 维沃移动通信有限公司 | 一种图像处理方法、装置及移动终端 |
CN109951633B (zh) * | 2019-02-18 | 2022-01-11 | 华为技术有限公司 | 一种拍摄月亮的方法和电子设备 |
CN110445978B (zh) * | 2019-06-24 | 2020-12-15 | 华为技术有限公司 | 一种拍摄方法及设备 |
CN115914826A (zh) * | 2020-05-30 | 2023-04-04 | 华为技术有限公司 | 一种图像内容的去除方法及相关装置 |
-
2020
- 2020-01-14 CN CN202010038444.XA patent/CN113194242B/zh active Active
- 2020-10-28 EP EP20914044.1A patent/EP4020967B1/en active Active
- 2020-10-28 US US17/779,876 patent/US12096120B2/en active Active
- 2020-10-28 WO PCT/CN2020/124545 patent/WO2021143269A1/zh unknown
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20120063697A1 (en) * | 2010-09-13 | 2012-03-15 | Samsung Electronics Co., Ltd. | Image processing method and image processing apparatus |
CN103871051A (zh) * | 2014-02-19 | 2014-06-18 | 小米科技有限责任公司 | 图像处理方法、装置和电子设备 |
CN108833768A (zh) * | 2018-05-10 | 2018-11-16 | 信利光电股份有限公司 | 一种多摄像头的拍摄方法、拍摄终端和可读存储介质 |
CN110248081A (zh) * | 2018-10-12 | 2019-09-17 | 华为技术有限公司 | 图像捕捉方法及电子设备 |
CN110572581A (zh) * | 2019-10-14 | 2019-12-13 | Oppo广东移动通信有限公司 | 基于终端设备的变焦虚化图像获取方法和装置 |
Non-Patent Citations (1)
Title |
---|
See also references of EP4020967A4 |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2023015959A1 (zh) * | 2021-08-12 | 2023-02-16 | 荣耀终端有限公司 | 拍摄方法及电子设备 |
Also Published As
Publication number | Publication date |
---|---|
US12096120B2 (en) | 2024-09-17 |
EP4020967A1 (en) | 2022-06-29 |
CN113194242B (zh) | 2022-09-20 |
CN113194242A (zh) | 2021-07-30 |
EP4020967A4 (en) | 2022-11-16 |
EP4020967B1 (en) | 2024-04-10 |
US20220417416A1 (en) | 2022-12-29 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2021143269A1 (zh) | 一种长焦场景下的拍摄方法及移动终端 | |
WO2021093793A1 (zh) | 一种拍摄方法及电子设备 | |
CN113132620B (zh) | 一种图像拍摄方法及相关装置 | |
WO2021052232A1 (zh) | 一种延时摄影的拍摄方法及设备 | |
WO2020168956A1 (zh) | 一种拍摄月亮的方法和电子设备 | |
WO2020073959A1 (zh) | 图像捕捉方法及电子设备 | |
WO2021129198A1 (zh) | 一种长焦场景下的拍摄方法及终端 | |
CN112887583B (zh) | 一种拍摄方法及电子设备 | |
CN113497881B (zh) | 图像处理方法及装置 | |
WO2022143921A1 (zh) | 一种图像重建方法、相关装置及系统 | |
CN115967851A (zh) | 快速拍照方法、电子设备及计算机可读存储介质 | |
CN112150499A (zh) | 图像处理方法及相关装置 | |
CN114466101B (zh) | 显示方法及电子设备 | |
WO2022228010A1 (zh) | 一种生成封面的方法及电子设备 | |
CN116709018B (zh) | 一种变焦条分割方法及电子设备 | |
CN117479008B (zh) | 一种视频处理方法、电子设备及芯片系统 | |
CN116055872B (zh) | 图像获取方法、电子设备和计算机可读存储介质 | |
CN115802144B (zh) | 视频拍摄方法及相关设备 | |
WO2023160224A9 (zh) | 一种拍摄方法及相关设备 | |
CN118450269A (zh) | 图像处理方法和电子设备 | |
CN118678238A (zh) | 一种拍摄方法及电子设备 | |
CN118555469A (zh) | 一种拍摄方法 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 20914044 Country of ref document: EP Kind code of ref document: A1 |
|
ENP | Entry into the national phase |
Ref document number: 2020914044 Country of ref document: EP Effective date: 20220324 |
|
NENP | Non-entry into the national phase |
Ref country code: DE |