WO2022203169A1 - 전자 장치의 촬영 방법 및 그 전자 장치 - Google Patents
전자 장치의 촬영 방법 및 그 전자 장치 Download PDFInfo
- Publication number
- WO2022203169A1 WO2022203169A1 PCT/KR2022/000170 KR2022000170W WO2022203169A1 WO 2022203169 A1 WO2022203169 A1 WO 2022203169A1 KR 2022000170 W KR2022000170 W KR 2022000170W WO 2022203169 A1 WO2022203169 A1 WO 2022203169A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- image
- electronic device
- light emission
- processor
- area
- Prior art date
Links
- 238000000034 method Methods 0.000 title claims description 39
- 230000004044 response Effects 0.000 claims description 10
- 230000002194 synthesizing effect Effects 0.000 claims description 4
- 238000004891 communication Methods 0.000 description 48
- 238000012545 processing Methods 0.000 description 16
- 238000010586 diagram Methods 0.000 description 12
- 230000006870 function Effects 0.000 description 12
- 238000013528 artificial neural network Methods 0.000 description 8
- 238000013473 artificial intelligence Methods 0.000 description 7
- 238000005516 engineering process Methods 0.000 description 7
- 239000003381 stabilizer Substances 0.000 description 6
- 230000008569 process Effects 0.000 description 5
- 238000004590 computer program Methods 0.000 description 4
- 230000009977 dual effect Effects 0.000 description 4
- 238000001308 synthesis method Methods 0.000 description 4
- 241000593989 Scardinius erythrophthalmus Species 0.000 description 3
- 201000005111 ocular hyperemia Diseases 0.000 description 3
- 230000001133 acceleration Effects 0.000 description 2
- 238000000429 assembly Methods 0.000 description 2
- 230000000712 assembly Effects 0.000 description 2
- 239000002131 composite material Substances 0.000 description 2
- 238000013527 convolutional neural network Methods 0.000 description 2
- 230000000694 effects Effects 0.000 description 2
- 230000004313 glare Effects 0.000 description 2
- 238000010801 machine learning Methods 0.000 description 2
- 230000003287 optical effect Effects 0.000 description 2
- 230000002093 peripheral effect Effects 0.000 description 2
- 230000000306 recurrent effect Effects 0.000 description 2
- 229920006395 saturated elastomer Polymers 0.000 description 2
- 230000006978 adaptation Effects 0.000 description 1
- 238000003491 array Methods 0.000 description 1
- 230000002457 bidirectional effect Effects 0.000 description 1
- 230000005540 biological transmission Effects 0.000 description 1
- 230000015572 biosynthetic process Effects 0.000 description 1
- 230000010267 cellular communication Effects 0.000 description 1
- 230000001413 cellular effect Effects 0.000 description 1
- 230000000295 complement effect Effects 0.000 description 1
- 239000004020 conductor Substances 0.000 description 1
- 230000003111 delayed effect Effects 0.000 description 1
- 230000007613 environmental effect Effects 0.000 description 1
- 238000000605 extraction Methods 0.000 description 1
- 239000000446 fuel Substances 0.000 description 1
- 230000036541 health Effects 0.000 description 1
- 238000010191 image analysis Methods 0.000 description 1
- 238000003384 imaging method Methods 0.000 description 1
- 230000000415 inactivating effect Effects 0.000 description 1
- 230000010354 integration Effects 0.000 description 1
- 230000003155 kinesthetic effect Effects 0.000 description 1
- 229910044991 metal oxide Inorganic materials 0.000 description 1
- 150000004706 metal oxides Chemical class 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 230000001537 neural effect Effects 0.000 description 1
- 238000011017 operating method Methods 0.000 description 1
- 230000009467 reduction Effects 0.000 description 1
- 230000002787 reinforcement Effects 0.000 description 1
- 239000004065 semiconductor Substances 0.000 description 1
- 230000005236 sound signal Effects 0.000 description 1
- 230000000638 stimulation Effects 0.000 description 1
- 238000006467 substitution reaction Methods 0.000 description 1
- 239000000758 substrate Substances 0.000 description 1
- 238000003786 synthesis reaction Methods 0.000 description 1
- 229910052724 xenon Inorganic materials 0.000 description 1
- FHNFHKCVQCLJFQ-UHFFFAOYSA-N xenon atom Chemical compound [Xe] FHNFHKCVQCLJFQ-UHFFFAOYSA-N 0.000 description 1
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/95—Computational photography systems, e.g. light-field imaging systems
- H04N23/958—Computational photography systems, e.g. light-field imaging systems for extended depth of field imaging
- H04N23/959—Computational photography systems, e.g. light-field imaging systems for extended depth of field imaging by adjusting depth of field during image capture, e.g. maximising or setting range based on scene characteristics
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/70—Circuitry for compensating brightness variation in the scene
- H04N23/74—Circuitry for compensating brightness variation in the scene by influencing the scene brightness using illuminating means
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/56—Cameras or camera modules comprising electronic image sensors; Control thereof provided with illuminating means
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/60—Control of cameras or camera modules
- H04N23/61—Control of cameras or camera modules based on recognised objects
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/60—Control of cameras or camera modules
- H04N23/61—Control of cameras or camera modules based on recognised objects
- H04N23/611—Control of cameras or camera modules based on recognised objects where the recognised objects include parts of the human body
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/60—Control of cameras or camera modules
- H04N23/63—Control of cameras or camera modules by using electronic viewfinders
- H04N23/631—Graphical user interfaces [GUI] specially adapted for controlling image capture or setting capture parameters
- H04N23/632—Graphical user interfaces [GUI] specially adapted for controlling image capture or setting capture parameters for displaying or modifying preview images prior to image capturing, e.g. variety of image resolutions or capturing parameters
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/60—Control of cameras or camera modules
- H04N23/667—Camera operation mode switching, e.g. between still and video, sport and normal or high- and low-resolution modes
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/70—Circuitry for compensating brightness variation in the scene
- H04N23/71—Circuitry for evaluating the brightness variation
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/70—Circuitry for compensating brightness variation in the scene
- H04N23/741—Circuitry for compensating brightness variation in the scene by increasing the dynamic range of the image compared to the dynamic range of the electronic image sensors
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/95—Computational photography systems, e.g. light-field imaging systems
Definitions
- Various embodiments according to the present disclosure relate to a technique for selecting a photographing method based on a preview image analysis.
- the photographing state is determined according to the luminance state of the surrounding environment.
- a flash device that creates a flash that has almost the same properties as the sun's rays with a large amount of light and a high color temperature for a very short time. It is improved so that you can take photos with the appropriate exposure.
- the camera performs photographing using the auxiliary light emission and/or the main light emission.
- the auxiliary light emission uses a lower current than the main light emission to take pictures with low brightness
- the main light emission uses a strong current to take a picture with high brightness for a short time. shoot
- the reason for taking pictures with sub-flash and main flash is that, in the case of main flash, a strong current is used, so the flash cannot be turned on for a long time, and the flash lasts only for a short time due to severe glare.
- the exposure degree and focus cannot be predicted during a very short light emission time during flash photography of the camera, the exposure level and focus of the main light emission can be predicted in advance with the auxiliary light emission. In this case, if both auxiliary and main flashes are always used, the flash shooting time becomes longer due to two flash shootings, and glare may occur. Also, since the main light emits light for a very short time, the user's eye adaptation time is insufficient, and a red-eye phenomenon may appear.
- the electronic device includes a camera, a sensor capable of detecting a distance between the electronic device and an object, a display, and light emission that generates a flash corresponding to first and/or second light emission when photographing using the camera a module and at least one processor electrically connected to the camera, the sensor, the light emitting module, and the display, wherein the at least one processor displays an image obtained by driving the camera as a preview image on the display ( display), a main object included in the preview image is determined, and when the distance between the main object detected through the sensor and the electronic device is less than a threshold value, the image is captured using the first light emission when the distance between the main object sensed through the sensor and the electronic device is equal to or greater than the threshold value, capturing using the first light emission and the second light emission subsequent to the first light emission
- the second mode can be executed.
- An operation method of an electronic device includes an operation of displaying an image obtained by driving a camera as a preview image on a display, an operation of determining a main object included in the preview image, and a sensor When the distance between the main object detected through the and the electronic device is less than a threshold value, executing a first mode of photographing using a first light emission, the main object detected through the sensor and the electronic device and executing a second mode of photographing using the first light emission and a second light emission subsequent to the first light emission when the distance of is equal to or greater than the threshold value.
- the electronic device includes a camera, a sensor capable of detecting a distance between the electronic device and an object, a display, and light emission that generates a flash corresponding to first and/or second light emission when photographing using the camera a module and at least one processor electrically connected to the camera, the sensor, the light emitting module, and the display, wherein the at least one processor displays an image obtained by driving the camera as a preview image on the display ( display), dividing the preview image into a plurality of regions, and determining at least a portion of the plurality of regions as a foreground region based on distance information between an external object and the electronic device obtained through the sensor, A first mode in which a main object included in the foreground area is determined, and when the distance between the main object detected through the sensor and the electronic device is less than a threshold value, the first mode is used to capture images using the first light emission and when the distance between the main object sensed through the sensor and the electronic device is equal to or greater than the threshold value, a second mode of photographing
- the electronic device and method according to various embodiments of the present disclosure may reduce a photographing time and alleviate a red-eye phenomenon by analyzing a preview image and selecting a photographing method.
- FIG. 1 is a block diagram of an electronic device in a network environment, according to an embodiment.
- FIG. 2 is a block diagram illustrating a camera module, according to an embodiment.
- FIG. 3 is a block diagram of an electronic device according to an exemplary embodiment.
- FIG. 4 is a flowchart illustrating an operation of determining a photographing method based on a distance to a subject in an electronic device according to an exemplary embodiment.
- FIG. 5 is a diagram illustrating identification of an object included in a preview image according to distance information in an electronic device according to an exemplary embodiment.
- FIG. 6 is a diagram illustrating dividing a preview image into a plurality of regions according to distance information in an electronic device according to an exemplary embodiment.
- FIG. 7 is a flowchart illustrating an operation of determining a photographing mode and a synthesis method based on distance information and brightness information of a preview image in an electronic device according to an exemplary embodiment.
- FIG. 8 is a diagram illustrating obtaining an HDR image in an electronic device according to an exemplary embodiment.
- FIG. 1 is a block diagram of an electronic device 101 in a network environment 100 according to various embodiments of the present disclosure.
- an electronic device 101 communicates with an electronic device 102 through a first network 198 (eg, a short-range wireless communication network) or a second network 199 . It may communicate with at least one of the electronic device 104 and the server 108 through (eg, a long-distance wireless communication network). According to an embodiment, the electronic device 101 may communicate with the electronic device 104 through the server 108 .
- a first network 198 eg, a short-range wireless communication network
- a second network 199 e.g., a second network 199
- the electronic device 101 may communicate with the electronic device 104 through the server 108 .
- the electronic device 101 includes a processor 120 , a memory 130 , an input module 150 , a sound output module 155 , a display module 160 , an audio module 170 , and a sensor module ( 176), interface 177, connection terminal 178, haptic module 179, camera module 180, power management module 188, battery 189, communication module 190, subscriber identification module 196 , or an antenna module 197 .
- at least one of these components eg, the connection terminal 178
- some of these components are integrated into one component (eg, display module 160 ). can be
- the processor 120 for example, executes software (eg, a program 140) to execute at least one other component (eg, a hardware or software component) of the electronic device 101 connected to the processor 120. It can control and perform various data processing or operations. According to an embodiment, as at least part of data processing or operation, the processor 120 stores a command or data received from another component (eg, the sensor module 176 or the communication module 190 ) into the volatile memory 132 . may be stored in , process commands or data stored in the volatile memory 132 , and store the result data in the non-volatile memory 134 .
- software eg, a program 140
- the processor 120 stores a command or data received from another component (eg, the sensor module 176 or the communication module 190 ) into the volatile memory 132 .
- the processor 120 stores a command or data received from another component (eg, the sensor module 176 or the communication module 190 ) into the volatile memory 132 .
- the processor 120 is a main processor 121 (eg, a central processing unit or an application processor) or a secondary processor 123 (eg, a graphic processing unit, a neural network processing unit) a neural processing unit (NPU), an image signal processor, a sensor hub processor, or a communication processor).
- a main processor 121 eg, a central processing unit or an application processor
- a secondary processor 123 eg, a graphic processing unit, a neural network processing unit
- NPU neural processing unit
- an image signal processor e.g., a sensor hub processor, or a communication processor.
- the secondary processor 123 may, for example, act on behalf of the main processor 121 while the main processor 121 is in an inactive (eg, sleep) state, or when the main processor 121 is active (eg, executing an application). ), together with the main processor 121, at least one of the components of the electronic device 101 (eg, the display module 160, the sensor module 176, or the communication module 190) It is possible to control at least some of the related functions or states.
- the auxiliary processor 123 eg, image signal processor or communication processor
- the auxiliary processor 123 may include a hardware structure specialized for processing an artificial intelligence model.
- Artificial intelligence models can be created through machine learning. Such learning may be performed, for example, in the electronic device 101 itself on which the artificial intelligence model is performed, or may be performed through a separate server (eg, the server 108).
- the learning algorithm may include, for example, supervised learning, unsupervised learning, semi-supervised learning, or reinforcement learning, but in the above example not limited
- the artificial intelligence model may include a plurality of artificial neural network layers.
- Artificial neural networks include deep neural networks (DNNs), convolutional neural networks (CNNs), recurrent neural networks (RNNs), restricted boltzmann machines (RBMs), deep belief networks (DBNs), bidirectional recurrent deep neural networks (BRDNNs), It may be one of deep Q-networks or a combination of two or more of the above, but is not limited to the above example.
- the artificial intelligence model may include, in addition to, or alternatively, a software structure in addition to the hardware structure.
- the memory 130 may store various data used by at least one component (eg, the processor 120 or the sensor module 176 ) of the electronic device 101 .
- the data may include, for example, input data or output data for software (eg, the program 140 ) and instructions related thereto.
- the memory 130 may include a volatile memory 132 or a non-volatile memory 134 .
- the program 140 may be stored as software in the memory 130 , and may include, for example, an operating system 142 , middleware 144 , or an application 146 .
- the input module 150 may receive a command or data to be used by a component (eg, the processor 120 ) of the electronic device 101 from the outside (eg, a user) of the electronic device 101 .
- the input module 150 may include, for example, a microphone, a mouse, a keyboard, a key (eg, a button), or a digital pen (eg, a stylus pen).
- the sound output module 155 may output a sound signal to the outside of the electronic device 101 .
- the sound output module 155 may include, for example, a speaker or a receiver.
- the speaker can be used for general purposes such as multimedia playback or recording playback.
- the receiver can be used to receive incoming calls. According to an embodiment, the receiver may be implemented separately from or as a part of the speaker.
- the display module 160 may visually provide information to the outside (eg, a user) of the electronic device 101 .
- the display module 160 may include, for example, a control circuit for controlling a display, a hologram device, or a projector and a corresponding device.
- the display module 160 may include a touch sensor configured to sense a touch or a pressure sensor configured to measure the intensity of a force generated by the touch.
- the audio module 170 may convert a sound into an electric signal or, conversely, convert an electric signal into a sound. According to an embodiment, the audio module 170 acquires a sound through the input module 150 , or an external electronic device (eg, a sound output module 155 ) connected directly or wirelessly with the electronic device 101 .
- the electronic device 102) eg, a speaker or headphones
- the electronic device 102 may output a sound.
- the sensor module 176 detects an operating state (eg, power or temperature) of the electronic device 101 or an external environmental state (eg, a user state), and generates an electrical signal or data value corresponding to the sensed state. can do.
- the sensor module 176 may include, for example, a gesture sensor, a gyro sensor, a barometric pressure sensor, a magnetic sensor, an acceleration sensor, a grip sensor, a proximity sensor, a color sensor, an IR (infrared) sensor, a biometric sensor, It may include a temperature sensor, a humidity sensor, or an illuminance sensor.
- the interface 177 may support one or more specified protocols that may be used by the electronic device 101 to directly or wirelessly connect with an external electronic device (eg, the electronic device 102 ).
- the interface 177 may include, for example, a high definition multimedia interface (HDMI), a universal serial bus (USB) interface, an SD card interface, or an audio interface.
- the connection terminal 178 may include a connector through which the electronic device 101 can be physically connected to an external electronic device (eg, the electronic device 102 ).
- the connection terminal 178 may include, for example, an HDMI connector, a USB connector, an SD card connector, or an audio connector (eg, a headphone connector).
- the haptic module 179 may convert an electrical signal into a mechanical stimulus (eg, vibration or movement) or an electrical stimulus that the user can perceive through tactile or kinesthetic sense.
- the haptic module 179 may include, for example, a motor, a piezoelectric element, or an electrical stimulation device.
- the camera module 180 may capture still images and moving images. According to an embodiment, the camera module 180 may include one or more lenses, image sensors, image signal processors, or flashes.
- the power management module 188 may manage power supplied to the electronic device 101 .
- the power management module 188 may be implemented as, for example, at least a part of a power management integrated circuit (PMIC).
- PMIC power management integrated circuit
- the battery 189 may supply power to at least one component of the electronic device 101 .
- the battery 189 may include, for example, a non-rechargeable primary cell, a rechargeable secondary cell, or a fuel cell.
- the communication module 190 is a direct (eg, wired) communication channel or a wireless communication channel between the electronic device 101 and an external electronic device (eg, the electronic device 102, the electronic device 104, or the server 108). It can support establishment and communication performance through the established communication channel.
- the communication module 190 may include one or more communication processors that operate independently of the processor 120 (eg, an application processor) and support direct (eg, wired) communication or wireless communication.
- the communication module 190 is a wireless communication module 192 (eg, a cellular communication module, a short-range wireless communication module, or a global navigation satellite system (GNSS) communication module) or a wired communication module 194 (eg, : It may include a local area network (LAN) communication module, or a power line communication module).
- a wireless communication module 192 eg, a cellular communication module, a short-range wireless communication module, or a global navigation satellite system (GNSS) communication module
- GNSS global navigation satellite system
- wired communication module 194 eg, : It may include a local area network (LAN) communication module, or a power line communication module.
- a corresponding communication module among these communication modules is a first network 198 (eg, a short-range communication network such as Bluetooth, wireless fidelity (WiFi) direct, or infrared data association (IrDA)) or a second network 199 (eg, legacy It may communicate with the external electronic device 104 through a cellular network, a 5G network, a next-generation communication network, the Internet, or a computer network (eg, a telecommunication network such as a LAN or a WAN).
- a first network 198 eg, a short-range communication network such as Bluetooth, wireless fidelity (WiFi) direct, or infrared data association (IrDA)
- a second network 199 eg, legacy It may communicate with the external electronic device 104 through a cellular network, a 5G network, a next-generation communication network, the Internet, or a computer network (eg, a telecommunication network such as a LAN or a WAN).
- a telecommunication network
- the wireless communication module 192 uses subscriber information (eg, International Mobile Subscriber Identifier (IMSI)) stored in the subscriber identification module 196 within a communication network such as the first network 198 or the second network 199 .
- subscriber information eg, International Mobile Subscriber Identifier (IMSI)
- IMSI International Mobile Subscriber Identifier
- the electronic device 101 may be identified or authenticated.
- the wireless communication module 192 may support a 5G network after a 4G network and a next-generation communication technology, for example, a new radio access technology (NR).
- NR access technology includes high-speed transmission of high-capacity data (eMBB (enhanced mobile broadband)), minimization of terminal power and access to multiple terminals (mMTC (massive machine type communications)), or high reliability and low latency (URLLC (ultra-reliable and low-latency) -latency communications)).
- eMBB enhanced mobile broadband
- mMTC massive machine type communications
- URLLC ultra-reliable and low-latency
- the wireless communication module 192 may support a high frequency band (eg, mmWave band) to achieve a high data rate, for example.
- a high frequency band eg, mmWave band
- the wireless communication module 192 uses various techniques for securing performance in a high-frequency band, for example, beamforming, massive multiple-input and multiple-output (MIMO), all-dimensional multiplexing. It may support technologies such as full dimensional MIMO (FD-MIMO), an array antenna, analog beam-forming, or a large scale antenna.
- the wireless communication module 192 may support various requirements defined in the electronic device 101 , an external electronic device (eg, the electronic device 104 ), or a network system (eg, the second network 199 ).
- the wireless communication module 192 includes a peak data rate (eg, 20 Gbps or more) for realizing eMBB, loss coverage (eg, 164 dB or less) for realizing mMTC, or U-plane latency for realizing URLLC ( Example: Downlink (DL) and uplink (UL) each 0.5 ms or less, or round trip 1 ms or less) can be supported.
- a peak data rate eg, 20 Gbps or more
- loss coverage eg, 164 dB or less
- U-plane latency for realizing URLLC
- the antenna module 197 may transmit or receive a signal or power to the outside (eg, an external electronic device).
- the antenna module 197 may include an antenna including a conductor formed on a substrate (eg, a PCB) or a radiator formed of a conductive pattern.
- the antenna module 197 may include a plurality of antennas (eg, an array antenna). In this case, at least one antenna suitable for a communication method used in a communication network such as the first network 198 or the second network 199 is connected from the plurality of antennas by, for example, the communication module 190 . can be selected. A signal or power may be transmitted or received between the communication module 190 and an external electronic device through the selected at least one antenna.
- other components eg, a radio frequency integrated circuit (RFIC)
- RFIC radio frequency integrated circuit
- the antenna module 197 may form a mmWave antenna module.
- the mmWave antenna module comprises a printed circuit board, an RFIC disposed on or adjacent to a first side (eg, bottom side) of the printed circuit board and capable of supporting a designated high frequency band (eg, mmWave band); and a plurality of antennas (eg, an array antenna) disposed on or adjacent to a second side (eg, top or side) of the printed circuit board and capable of transmitting or receiving signals of the designated high frequency band. can do.
- peripheral devices eg, a bus, general purpose input and output (GPIO), serial peripheral interface (SPI), or mobile industry processor interface (MIPI)
- GPIO general purpose input and output
- SPI serial peripheral interface
- MIPI mobile industry processor interface
- the command or data may be transmitted or received between the electronic device 101 and the external electronic device 104 through the server 108 connected to the second network 199 .
- Each of the external electronic devices 102 or 104 may be the same as or different from the electronic device 101 .
- all or part of the operations executed by the electronic device 101 may be executed by one or more external electronic devices 102 , 104 , or 108 .
- the electronic device 101 may perform the function or service itself instead of executing the function or service itself.
- one or more external electronic devices may be requested to perform at least a part of the function or the service.
- One or more external electronic devices that have received the request may execute at least a part of the requested function or service, or an additional function or service related to the request, and transmit a result of the execution to the electronic device 101 .
- the electronic device 101 may process the result as it is or additionally and provide it as at least a part of a response to the request.
- cloud computing, distributed computing, mobile edge computing (MEC), or client-server computing technology may be used.
- the electronic device 101 may provide an ultra-low latency service using, for example, distributed computing or mobile edge computing.
- the external electronic device 104 may include an Internet of things (IoT) device.
- the server 108 may be an intelligent server using machine learning and/or neural networks.
- the external electronic device 104 or the server 108 may be included in the second network 199 .
- the electronic device 101 may be applied to an intelligent service (eg, smart home, smart city, smart car, or health care) based on 5G communication technology and IoT-related technology.
- the electronic device may be a device of various types.
- the electronic device may include, for example, a portable communication device (eg, a smart phone), a computer device, a portable multimedia device, a portable medical device, a camera, a wearable device, or a home appliance device.
- a portable communication device eg, a smart phone
- a computer device e.g., a laptop, a desktop, a tablet, or a portable multimedia device
- portable medical device e.g., a portable medical device
- camera e.g., a camera
- a wearable device e.g., a smart watch
- a home appliance device e.g., a smart bracelet
- first, second, or first or second may simply be used to distinguish an element from other elements in question, and may refer elements to other aspects (e.g., importance or order) is not limited. It is said that one (eg, first) component is “coupled” or “connected” to another (eg, second) component, with or without the terms “functionally” or “communicatively”. When referenced, it means that one component can be connected to the other component directly (eg by wire), wirelessly, or through a third component.
- module used in various embodiments of the present document may include a unit implemented in hardware, software, or firmware, for example, and interchangeably with terms such as logic, logic block, component, or circuit.
- a module may be an integrally formed part or a minimum unit or a part of the part that performs one or more functions.
- the module may be implemented in the form of an application-specific integrated circuit (ASIC).
- ASIC application-specific integrated circuit
- one or more instructions stored in a storage medium may be implemented as software (eg, the program 140) including
- the processor eg, the processor 120
- the device eg, the electronic device 101
- the one or more instructions may include code generated by a compiler or code executable by an interpreter.
- the device-readable storage medium may be provided in the form of a non-transitory storage medium.
- 'non-transitory' only means that the storage medium is a tangible device and does not contain a signal (eg, electromagnetic wave), and this term is used in cases where data is semi-permanently stored in the storage medium and It does not distinguish between temporary storage cases.
- a signal eg, electromagnetic wave
- the method according to various embodiments disclosed in this document may be provided by being included in a computer program product.
- Computer program products may be traded between sellers and buyers as commodities.
- the computer program product is distributed in the form of a machine-readable storage medium (eg compact disc read only memory (CD-ROM)), or via an application store (eg Play Store TM ) or on two user devices ( It can be distributed (eg downloaded or uploaded) directly or online between smartphones (eg: smartphones).
- a portion of the computer program product may be temporarily stored or temporarily generated in a machine-readable storage medium such as a memory of a server of a manufacturer, a server of an application store, or a memory of a relay server.
- the camera module 180 includes a lens assembly 210 , a flash 220 , an image sensor 230 , an image stabilizer 240 , a memory 250 (eg, a buffer memory), or an image signal processor. (260).
- the lens assembly 210 may collect light emitted from a subject, which is an image to be captured.
- the lens assembly 210 may include one or more lenses.
- the camera module 180 may include a plurality of lens assemblies 210 . In this case, the camera module 180 may form, for example, a dual camera, a 360 degree camera, or a spherical camera.
- Some of the plurality of lens assemblies 210 may have the same lens properties (eg, angle of view, focal length, auto focus, f number, or optical zoom), or at least one lens assembly may be a different lens assembly. It may have one or more lens properties that are different from the lens properties of .
- the lens assembly 210 may include, for example, a wide-angle lens or a telephoto lens.
- the flash 220 may emit light used to enhance light emitted or reflected from the subject.
- the flash 220 may include one or more light emitting diodes (eg, a red-green-blue (RGB) LED, a white LED, an infrared LED, or an ultraviolet LED), or a xenon lamp.
- the image sensor 230 may acquire an image corresponding to the subject by converting light emitted or reflected from the subject and transmitted through the lens assembly 210 into an electrical signal.
- the image sensor 230 may include, for example, one image sensor selected from among image sensors having different properties, such as an RGB sensor, a black and white (BW) sensor, an IR sensor, or a UV sensor, the same It may include a plurality of image sensors having properties, or a plurality of image sensors having different properties.
- Each image sensor included in the image sensor 230 may be implemented using, for example, a charged coupled device (CCD) sensor or a complementary metal oxide semiconductor (CMOS) sensor.
- CCD charged coupled device
- CMOS complementary metal oxide semiconductor
- the image stabilizer 240 moves at least one lens or the image sensor 230 included in the lens assembly 210 in a specific direction or Operation characteristics of the image sensor 230 may be controlled (eg, read-out timing may be adjusted, etc.). This makes it possible to compensate for at least some of the negative effects of the movement on the image being taken.
- the image stabilizer 240 is, according to an embodiment, the image stabilizer 240 is a gyro sensor (not shown) or an acceleration sensor (not shown) disposed inside or outside the camera module 180 . can be used to detect such a movement of the camera module 180 or the electronic device 101 .
- the image stabilizer 240 may be implemented as, for example, an optical image stabilizer.
- the memory 250 may temporarily store at least a portion of the image acquired through the image sensor 230 for a next image processing operation. For example, when image acquisition is delayed according to the shutter or a plurality of images are acquired at high speed, the acquired original image (eg, a Bayer-patterned image or a high-resolution image) is stored in the memory 250 and , a copy image corresponding thereto (eg, a low-resolution image) may be previewed through the display module 160 .
- the acquired original image eg, a Bayer-patterned image or a high-resolution image
- a copy image corresponding thereto eg, a low-resolution image
- the memory 250 may be configured as at least a part of the memory 130 or as a separate memory operated independently of the memory 130 .
- the image signal processor 260 may perform one or more image processing on an image acquired through the image sensor 230 or an image stored in the memory 250 .
- the one or more image processes may include, for example, depth map generation, three-dimensional modeling, panorama generation, feature point extraction, image synthesis, or image compensation (eg, noise reduction, resolution adjustment, brightness adjustment, blurring ( blurring), sharpening (sharpening), or softening (softening)
- the image signal processor 260 may include at least one of the components included in the camera module 180 (eg, an image sensor). 230), for example, exposure time control, readout timing control, etc.
- the image processed by the image signal processor 260 is stored back in the memory 250 for further processing.
- the image signal processor 260 may be configured as at least a part of the processor 120 or as a separate processor operated independently of the processor 120.
- the image signal processor 260 may be configured as the processor 120 and a separate processor, the at least one image processed by the image signal processor 260 may be displayed through the display module 160 as it is by the processor 120 or after additional image processing.
- the electronic device 101 may include a plurality of camera modules 180 each having different properties or functions.
- at least one of the plurality of camera modules 180 may be a wide-angle camera, and at least the other may be a telephoto camera.
- at least one of the plurality of camera modules 180 may be a front camera, and at least the other may be a rear camera.
- FIG. 3 is a block diagram of an electronic device according to an exemplary embodiment.
- the electronic device 101 may include a processor 310 , a camera 320 , a sensor 330 , a display 340 , and a memory 350 .
- the electronic device 101 may include additional components in addition to the components illustrated in FIG. 3 , or may omit at least one of the components illustrated in FIG. 3 .
- the processor 310 may execute an operation or data processing related to control and/or communication of at least one other component of the electronic device 101 using instructions stored in the memory 350 .
- the processor 310 includes a central processing unit (CPU), a graphics processing unit (GPU), a micro controller unit (MCU), a sensor hub, a supplementary processor, a communication processor, and an application. It may include at least one of a processor (application processor), application specific integrated circuit (ASIC), and field programmable gate arrays (FPGA), and may have a plurality of cores.
- the processor 310 may execute an application (eg, a camera application) stored in the memory 350 .
- the processor 310 may acquire an image by using the camera 320 while the camera application is running.
- the application described herein may be any application using the camera 320 .
- the processor 310 may display an image acquired using the camera 320 as a preview image on the display 340 . According to an embodiment, the processor 310 may obtain distance information from the object included in the preview image by using the sensor 330 . According to an embodiment, the processor 310 may divide the preview image into a plurality of regions based on a depth map of the preview image and perform image processing for each region. Specific details related to the operation of the processor 310 will be described later with reference to FIG. 4 .
- the camera 320 may acquire (or capture) an image (eg, a still image and/or a moving image).
- an image signal processor (not shown) electrically connected to the camera 320 may distinguish an object (eg, a person) and a background included in an image (eg, a preview image or an image stored in the memory 350 ).
- the image signal processor may be separated from the camera 320 or implemented as a part of the processor 310 .
- the camera 320 may include an image sensor.
- the image sensor may acquire and process color information.
- the senor 330 may include at least one of a depth sensor, a ToF sensor, and an image sensor (eg, a dual pixel image).
- the depth sensor may measure a depth of an external object and generate depth information corresponding to the external object using the measured depth.
- the sensor 330 is operatively connected to at least one of the processor 310, the camera 320, and the memory 350 to process color information, 3D information, distance information, location information, and the like. can do.
- the display 340 may display an image obtained through the camera 320 .
- the processor 310 may display an image acquired through the camera 320 on the display 340 as a preview image.
- the processor 310 may display at least a portion of an image acquired through the camera 320 as a preview on the display 340 .
- the electronic device 101 may obtain a user's input through the display 340 and transmit the user's input to the processor 310 .
- the memory 350 may mean one or more memory sets. According to an embodiment, the memory 350 is received from or generated by other components (eg, the processor 310 , the camera 320 , the sensor 330 , and the display 340 ). It can store data and/or commands. In various embodiments, the memory 350 may store an application using the camera 320 .
- FIG. 4 is a flowchart illustrating an operation of determining a photographing method based on a distance to a subject in an electronic device according to an exemplary embodiment.
- the processor 310 may display an image obtained by driving the camera 320 in operation 410 as a preview image on the display 340 .
- the processor 310 may display a preview image on the display 340 of an image acquired through the camera 320 while the camera application stored in the memory 350 is running.
- the processor 310 may determine a main object included in the preview image in operation 420 .
- the processor 310 may determine a main object included in an image corresponding to a preview image displayed on the display 340 from among images acquired by driving the camera 320 .
- the processor 310 may obtain depth information and/or distance information on at least one object included in the preview image by using the sensor 330 (eg, a depth sensor or a ToF sensor).
- the processor 310 may acquire distance information on at least one object included in the preview image using artificial intelligence (AI) and/or a dual pixel image sensor.
- AI artificial intelligence
- the processor 310 may divide the preview image into a plurality of regions based on the acquired depth information and/or distance information. For example, the processor 310 may separate an object closest to the electronic device 101 from the background image based on the distance information. According to an embodiment, the processor 310 may determine an object closest to the electronic device 101 as a main object. However, the object determined as the main object is not limited thereto.
- the processor 310 uses the first light emission (eg, auxiliary light emission) to A first mode of photographing may be executed. For example, when the distance between the electronic device 101 and the object closest to the electronic device 101 included in the preview image is less than a specified distance, the processor 310 may take a photograph using only the first light emission.
- the first light emission eg, auxiliary light emission
- the processor 310 is configured to perform auto focus (AF), auto white balance (AWB), and auto exposure (AE) when shooting using the first light emission, according to an embodiment. At least one operation may be performed. According to an embodiment, the processor 310 may calculate AF data for focus adjustment, AWB data for white balance adjustment, and AE data for exposure adjustment by performing the above operation.
- the processor 310 may adjust a focus on a subject (or a main object). According to an embodiment, the processor 310 may adjust the focus on the main object using distance information obtained using the sensor 330 (eg, a ToF sensor).
- the sensor 330 eg, a ToF sensor
- the processor 310 may adjust the position of the focus on the main object using an actuator. According to an embodiment, the processor 310 may deactivate the first light emission in response to completion of photographing using the first light emission.
- the processor 310 performs the first light emission and the second light emission subsequent to the first light emission.
- a threshold value can be used to execute the second mode of photographing.
- the processor 310 may take a picture using the first light emission and the second light emission.
- the second light emission may have a greater amount of light than the first light emission.
- the processor 310 when shooting using the first light emission, performs an AF operation and an AWB operation using AF data for focus adjustment, AWB data for white balance adjustment, or AE data for exposure adjustment. and at least one of an AE operation. According to an embodiment, the processor 310 may deactivate the first light emission in response to completion of photographing using the first light emission.
- FIG. 5 is a diagram illustrating identification of an object included in a preview image according to distance information in an electronic device according to an exemplary embodiment.
- the processor 310 may divide an image acquired through the camera 320 into a plurality of regions. For example, the processor 310 may divide the image into a plurality of regions based on depth information and/or distance information of an object included in the image.
- the preview image 510 displayed on the display 340 may include a plurality of objects.
- the preview image 510 may include a first object 501a, a second object 502a, and a third object 503a.
- the processor 310 may generate a depth map 520 corresponding to the preview image 510 .
- the processor 310 may divide the preview image 510 into a plurality of regions using the generated depth map.
- the first object 501a , the second object 501b , and the third object 501c of the preview image 510 include a first region 501b and a second object 501c included in the depth map 520 , respectively. It may correspond to the second area 502b and the third area 503b.
- the processor 310 may map a number to the plurality of regions based on a depth value obtained with respect to at least one object included in the preview image 510 .
- the processor 310 may map different numbers to the first area 501b, the second area 502b, and the third area 503b included in the depth map 520 . Specifically, for example, the processor 310 maps the first number to the first area 501b corresponding to the first object 501a closest to the electronic device 101 and communicates with the electronic device 101 . A third number greater than the first number may be mapped to the third area 503b corresponding to the third object 503a having the longest distance. Also, for example, in the second area 502b corresponding to the second object 502a that is farther than the first object 501a and closer than the third object 503a, the distance from the electronic device 101 is greater than the first number. A second number greater than the third number may be mapped.
- FIG. 6 is a diagram illustrating dividing a preview image into a plurality of regions according to distance information in an electronic device according to an exemplary embodiment.
- the processor 310 performs a preview image (eg, FIG. 6 ) based on depth information and/or distance information sensed by a sensor 330 (eg, a depth sensor or a ToF sensor).
- the preview image 510 of FIG. 5 may be divided into a foreground area and a rear view area.
- the processor 310 generates the preview image 510 based on the distance information in a first region 601 (eg, a foreground region or a region of interest), a second region 602 (eg, a rearview region or a region of interest). background area) and a third area 603 (eg, a background area or a background area).
- the processor 310 sets an area corresponding to an object closest to the electronic device 101 (eg, the main object of FIG. 4 or the first object 501a of FIG. 5 ) to the first area 601 .
- the processor 310 may determine an area corresponding to the object having the longest distance from the electronic device 101 as the third area 603 .
- the processor 310 may determine an area excluding the first area 610 and the third area 603 of the preview image 510 as the second area 602 .
- the processor 310 uses the sensor 330 (eg, an image sensor) to obtain brightness information corresponding to the first area 601 , the second area 602 , and the third area 603 . can be obtained. According to an embodiment, the processor 310 may obtain brightness difference information between the foreground area and the rear view area. For example, the processor 310 may obtain brightness difference information between the first area 601 and the second area 602 . Also, for example, the processor 310 may obtain brightness difference information between the first area 601 and the second area 602 .
- the sensor 330 eg, an image sensor
- FIG. 7 is a flowchart illustrating an operation of determining a photographing mode and a synthesis method based on distance information and brightness information of a preview image in an electronic device according to an exemplary embodiment.
- the processor 310 may determine whether the distance between the main object and the electronic device 101 is less than a threshold value. According to an embodiment, the processor 310 may obtain distance information between the electronic device 101 and an object closest to the electronic device 101 among objects included in the preview image. According to an embodiment, the processor 310 may set a shooting mode based on the obtained distance information. For example, the processor 310 may determine whether to perform imaging using only auxiliary light emission based on the acquired distance information.
- the processor 310 may take a photograph using the auxiliary light emission.
- the processor 310 may take a picture using only the auxiliary light emission. have.
- the processor 310 may perform at least one of an AF operation, an AWB operation, and an AE operation when photographing using auxiliary light emission.
- the processor 310 may deactivate the auxiliary light emission in response to the end of photographing using the auxiliary light emission.
- the electronic device 101 may improve the brightness and visibility of the entire image by photographing it for a long exposure time with weak light emission.
- the electronic device 101 when it is determined that the distance to the subject is close, the electronic device 101 can reduce the photographing time and alleviate the red-eye effect by using the auxiliary light emission without the main light emission.
- the processor 310 may take a picture using both the secondary light emission and the main light emission. According to an embodiment, when it is determined that the distance between the electronic device 101 and the closest object to the electronic device 101 among the objects included in the preview image is equal to or greater than a specified distance, the processor 310 performs both auxiliary light emission and main light emission. You can use it to take pictures. According to an embodiment, the processor 310 may perform photographing using the main light emission after completing photographing using the auxiliary light emission.
- the processor 310 may determine whether a difference in brightness between the main object and the remaining area is equal to or greater than a threshold value. According to an embodiment, the processor 310 may obtain brightness information of an area (or foreground area) corresponding to the main object and the remaining area (or background area) in the preview image by using the image sensor. According to an embodiment, the processor 310 may determine whether a difference between the brightness of the foreground area and the brightness of the rear view area is equal to or greater than a specified value. According to an embodiment, the processor 310 may determine the exposure time based on information on the difference in brightness between the foreground area and the background area in the preview image.
- the processor 310 may synthesize the images in the first manner in operation 709 .
- the processor 310 may determine an image synthesis method based on brightness information of the foreground area and/or the background area. For example, the processor 310 may set the exposure time based on the brightness of one of the foreground area and the background area.
- the processor 310 may obtain a plurality of images having different brightnesses from the image sensor.
- the processor 310 determines that the foreground region among a plurality of images acquired from the image sensor is Properly exposed images can be synthesized. According to an embodiment, when it is determined that the difference in brightness between the foreground region and the background region is greater than or equal to a threshold value and the background region is bright or saturated, the processor 310 determines that the background region among the plurality of images obtained from the image sensor is appropriate. You can composite the exposed images.
- the processor 310 determines whether the foreground region is properly exposed among a plurality of images obtained from the image sensor. Images can be composited. According to an embodiment, when it is determined that the difference in brightness between the foreground region and the rearview region is greater than or equal to a threshold value and the background region is dark, the processor 310 determines whether the rearview region is properly exposed among a plurality of images obtained from the image sensor. Images can be composited.
- the processor 310 determines that the difference in brightness between the foreground region and the background region is equal to or greater than a threshold, and the foreground region is dark and the background region is bright, the processor 310 generates an image in which the foreground region is properly exposed. and an image in which the rear view region is properly exposed may be synthesized.
- the processor 310 determines that the difference in brightness between the foreground area and the background area is equal to or greater than a threshold value, and the foreground area is bright and the background area is dark, the processor 310 generates an image in which the foreground area is properly exposed. and an image in which the rear view region is properly exposed may be synthesized.
- the processor 310 may synthesize the images in the second method in operation 711 .
- the processor 310 may obtain images having the same brightness from the image sensor.
- the processor 310 may not synthesize overexposed or underexposed images among a plurality of images.
- the electronic device 101 may improve image quality by determining a suitable synthesis method based on distance information and/or brightness information obtained from the preview image.
- FIG. 8 is a diagram illustrating obtaining an HDR image in an electronic device according to an exemplary embodiment.
- the processor 310 may generate a composite image (or HDR image) by using at least some of a plurality of images obtained from an image sensor.
- the processor 310 may generate an HDR image using a plurality of images having different exposure values.
- the processor 310 may generate a third image 803 (eg, an HDR image) by using the first image 801 and the second image 802 .
- an exposure value may mean an exposure amount or brightness of an image.
- the processor 310 may acquire a plurality of images having different exposure values from the image sensor. For example, the processor 310 may acquire the first image 801 and the second image 802 from the image sensor. The processor 310 may acquire the first image 801 and the second image 802 having different exposure times of the image sensor.
- the processor 310 may analyze HDR histograms of the first image 801 and the second image 802 .
- the histogram for image brightness represents the exposure amount as a class and the number of pixels as the frequency, and vice versa. can be all day.
- the processor 310 may determine the exposure amount of the acquired image by analyzing the histogram of the image brightness.
- the processor 310 may determine whether the acquired image is underexposure or overexposure. According to an embodiment, the processor 310 analyzes the HDR histogram to determine whether the first image 801 or the second image 802 corresponds to at least one of underexposure and overexposure. have. According to an embodiment, by analyzing the HDR histogram, the processor 310 may determine that a large number of pixels with an exposure amount smaller than a predetermined exposure amount is underexposure, and if the number of pixels with a greater exposure amount than a predetermined exposure amount is large, it may be determined as overexposure.
- a pixel of an image may correspond to a pixel of an image sensor.
- the processor 310 sets the exposure time of at least one short exposure pixel included in the image sensor corresponding to the overexposed area in the image to be short to prevent overexposure.
- a relaxed image can be obtained.
- the processor 310 may obtain the second image 802 obtained by setting the exposure time to be short.
- the processor 310 when the underexposed area is included in the image, the processor 310 sets the exposure time of at least one long exposure pixel included in the image sensor corresponding to the underexposed area in the image to be longer to prevent underexposure. A relaxed image can be obtained.
- the processor 310 when both the overexposed area and the underexposed area are included in the image, the processor 310 reduces at least one overexposed image and underexposed area included in the image sensor corresponding to the overexposed area. All images in which at least one underexposure included in the image sensor corresponding to .
- the processor 310 may generate an HDR image using the image including the overexposed area and the overexposed image. According to an embodiment, the processor 310 may generate an HDR image using an image including an underexposed area and an image from which underexposure is alleviated. According to an embodiment, the processor 310 may generate an HDR image by using both an image including both an overexposed area and an underexposed area, an overexposed image, and an underexposed image.
- the first image 801 may include an image having an overexposed area
- the second image 802 may include an image in which overexposure is reduced.
- the processor 310 may include an image photographed so that the foreground region (or the region of interest or the region corresponding to the main object) is properly exposed as the first image 801
- the second image Reference numeral 802 may include an image captured so that an area other than the foreground area is properly exposed.
- the processor 310 may generate a third image 803 (eg, an HDR image) by using the first image 801 and the second image 802 .
- each component eg, a module or a program of the above-described components may include a singular or a plurality of entities, and some of the plurality of entities may be separately disposed in other components. have.
- one or more components or operations among the above-described corresponding components may be omitted, or one or more other components or operations may be added.
- a plurality of components eg, a module or a program
- the integrated component may perform one or more functions of each component of the plurality of components identically or similarly to those performed by the corresponding component among the plurality of components prior to the integration. .
- operations performed by a module, program, or other component are executed sequentially, in parallel, repeatedly, or heuristically, or one or more of the operations are executed in a different order, or omitted. , or one or more other operations may be added.
- the electronic device may include a camera (eg, the camera 320 of FIG. 3 ) and a device capable of detecting a distance between the electronic device and an object.
- a sensor eg, the sensor 330 of FIG. 3
- a display eg, the display 340 of FIG.
- a light emitting module that generates a flash corresponding to the first light emission and/or the second light emission when shooting using the camera and at least one processor electrically connected to the camera, the sensor, the light emitting module, and the display, wherein the at least one processor displays an image obtained by driving the camera as a preview image on the display ( display), a main object included in the preview image is determined, and when the distance between the main object sensed through the sensor and the electronic device is less than a threshold value, the image is captured using the first light emission when the distance between the main object sensed through the sensor and the electronic device is equal to or greater than the threshold value, capturing using the first light emission and the second light emission subsequent to the first light emission
- the second mode can be executed.
- the at least one processor may deactivate the first light emission in response to the completion of photographing using the first light emission.
- the second light emission may have a greater amount of light than the first light emission.
- the at least one processor performs an auto focus (AF) operation, an auto white balance (AWB) operation, and an auto exposure (AE) operation when the at least one processor captures images using the first light emission.
- AF auto focus
- ALB auto white balance
- AE auto exposure
- the at least one processor determines a region corresponding to the main object as a first region, determines a region excluding the first region in the preview image as a second region, and the first region and obtaining brightness difference information of the second region, and determining an exposure time based on the obtained brightness difference information.
- the electronic device includes an image sensor, and the at least one processor acquires a plurality of images having different brightnesses from the image sensor when the obtained brightness difference is equal to or greater than a threshold value to obtain high dynamic range (HDR) images. ) to create an image.
- HDR high dynamic range
- the at least one processor may generate an HDR image by synthesizing an image having a low exposure.
- the at least one processor divides the preview image into a plurality of regions, divides the region based on distance information of each of the plurality of regions, and the brightness difference between the divided regions information can be obtained.
- an image obtained by driving a camera eg, the camera 320 of FIG. 3
- Example: an operation of displaying a preview image on the display 340 of FIG. 3 , an operation of determining a main object included in the preview image, and a sensor eg, the sensor 330 of FIG.
- the method of operating an electronic device may include an operation of inactivating the first light emission in response to completion of photographing using the first light emission.
- the operation of executing the first mode may include an operation of performing at least one of an AF operation, an AWB operation, and an AE operation.
- the method of operating an electronic device may include determining a region corresponding to the main object as a first region, determining a region excluding the first region from the preview image as a second region, and The method may include obtaining brightness difference information between the first area and the second area and determining an exposure time based on the obtained brightness difference information.
- the method of operating the electronic device may include generating an HDR image by acquiring a plurality of images having different brightnesses from an image sensor when the obtained brightness difference is equal to or greater than a threshold value.
- the operating method of the electronic device may include generating an HDR image by synthesizing an image with a low exposure when it is determined that there is overexposure in the first area or the second area.
- the method of operating the electronic device includes dividing the preview image into a plurality of regions and dividing the region based on distance information of each of the plurality of regions, and the acquired between the divided regions. It may include an operation of obtaining brightness difference information.
- the electronic device (eg, the electronic device 101 of FIG. 1 ) according to an embodiment includes a camera, a sensor capable of detecting a distance between the electronic device and an object, a display, and a display device when photographing using the camera.
- a light emitting module that generates a flash corresponding to first light emission and/or second light emission
- at least one processor electrically connected to the camera, the sensor, the light emitting module, and the display, wherein the at least one processor comprises the An image obtained by driving a camera is displayed as a preview image on the display, the preview image is divided into a plurality of regions, and based on distance information between an external object and the electronic device obtained through the sensor to determine at least a portion of the plurality of regions as a foreground region, determine a main object included in the foreground region, and determine a distance between the main object sensed through the sensor and the electronic device is critical
- a first mode of photographing using the first light emission is executed, and when the distance between the main object and the electronic device sensed through the sensor is equal to or greater than the threshold value, the first light emission and the first light emission are performed.
- a second mode of photographing using the second light emission subsequent to light emission may be executed.
- the at least one processor may obtain the distance information using at least one of an artificial intelligence (AI), a dual pixel image sensor, and a time of flight (TOF) sensor. .
- AI artificial intelligence
- TOF time of flight
- the at least one processor may map a number corresponding to distance information of each area to a foreground area and an area other than the foreground area among the plurality of areas.
- the at least one processor may obtain brightness difference information of the foreground area and an area other than the foreground area, and determine an exposure time based on the obtained brightness difference information.
- the electronic device may include an image sensor, and the at least one processor may generate an HDR image by acquiring a plurality of images having different brightnesses from the image sensor when the obtained brightness difference is equal to or greater than a threshold value. .
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Computing Systems (AREA)
- Theoretical Computer Science (AREA)
- Human Computer Interaction (AREA)
- Studio Devices (AREA)
Abstract
Description
Claims (15)
- 전자 장치에 있어서,카메라;상기 전자 장치와 객체 간 거리를 감지할 수 있는 센서;디스플레이;상기 카메라를 이용한 촬영 시 제1 발광 및/또는 제2 발광에 해당하는 플래시를 발생시키는 발광 모듈; 및상기 카메라, 상기 센서, 상기 발광 모듈, 및 상기 디스플레이와 전기적으로 연결된 적어도 하나의 프로세서를 포함하고,상기 적어도 하나의 프로세서는:상기 카메라를 구동하여 획득된 이미지를 상기 디스플레이에 프리뷰 이미지로 표시(display)하고,상기 프리뷰 이미지에 포함된 주요 객체(main object)를 결정하고,상기 센서를 통해 감지된, 상기 주요 객체와 상기 전자 장치와의 거리가 임계 값 미만인 경우, 상기 제1 발광을 이용하여 촬영하는 제1 모드를 실행하고,상기 센서를 통해 감지된, 상기 주요 객체와 상기 전자 장치와의 거리가 상기 임계 값 이상인 경우, 상기 제1 발광 및 상기 제1 발광에 후속하는 상기 제2 발광을 이용하여 촬영하는 제2 모드를 실행하는, 전자 장치.
- 청구항 1에 있어서,상기 적어도 하나의 프로세서는 상기 제1 발광을 이용한 촬영이 완료되는 것에 응답하여 상기 제1 발광을 비활성화 하는, 전자 장치.
- 청구항 1에 있어서,상기 제2 발광은 상기 제1 발광보다 광량이 많은, 전자 장치.
- 청구항 1에 있어서,상기 적어도 하나의 프로세서는 상기 제1 발광을 이용하여 촬영하는 경우, AF(auto focus) 동작, AWB(auto white balance) 동작 및 AE(auto exposure) 동작 중 적어도 하나를 수행하는, 전자 장치.
- 청구항 1에 있어서,상기 카메라는 이미지 센서를 포함하고,상기 적어도 하나의 프로세서는:상기 프리뷰 이미지 중 상기 주요 객체에 대응되는 영역을 제1 영역으로 결정하고,상기 프리뷰 이미지 중 상기 제1 영역을 제외한 영역을 제2 영역으로 결정하고,상기 제1 영역 및 상기 제2 영역의 밝기 차이 정보를 획득하고,상기 획득한 밝기 차이 정보에 기반하여 상기 이미지 센서의 노출 시간을 결정하는, 전자 장치.
- 청구항 5에 있어서,상기 적어도 하나의 프로세서는:상기 획득한 밝기 차이가 임계 값 이상인 경우, 상기 이미지 센서로부터 상이한 밝기를 가지는 복수의 이미지들을 획득하고,상기 복수의 이미지들을 기반으로 HDR(high dynamic range) 이미지를 생성하는, 전자 장치.
- 청구항 6에 있어서,상기 적어도 하나의 프로세서는 상기 제1 영역 또는 상기 제2 영역에 과대 노출이 있는 것으로 판단되는 경우 노출이 작은 이미지를 합성하여 HDR 이미지를 생성하는, 전자 장치.
- 청구항 5에 있어서,상기 적어도 하나의 프로세서는, 상기 프리뷰 이미지를 복수의 영역들로 나누고 상기 복수의 영역들 각각의 거리 정보에 기반하여 영역을 구분하고, 상기 구분된 영역들 사이의 상기 밝기 차이 정보를 획득하는, 전자 장치.
- 전자 장치의 동작 방법에 있어서,카메라를 구동하여 획득된 이미지를 디스플레이에 프리뷰 이미지로 표시(display)하는 동작;상기 프리뷰 이미지에 포함된 주요 객체(main object)를 결정하는 동작;센서를 통해 감지된 상기 주요 객체와 상기 전자 장치와의 거리가 임계 값 미만인 경우, 제1 발광을 이용하여 촬영하는 제1 모드를 실행하는 동작;상기 센서를 통해 감지된 상기 주요 객체와 상기 전자 장치와의 거리가 상기 임계 값 이상인 경우, 상기 제1 발광 및 상기 제1 발광에 후속하는 제2 발광을 이용하여 촬영하는 제2 모드를 실행하는 동작을 포함하는, 전자 장치의 동작 방법.
- 청구항 9에 있어서,상기 제1 발광을 이용한 촬영이 완료되는 것에 응답하여 상기 제1 발광을 비활성화 하는 동작을 포함하는, 전자 장치의 동작 방법.
- 청구항 9에 있어서,상기 제1 모드를 실행하는 동작은, AF 동작, AWB 동작 및 AE 동작 중 적어도 하나를 수행하는 동작을 포함하는, 전자 장치의 동작 방법.
- 청구항 9에 있어서,상기 프리뷰 이미지 중 상기 주요 객체에 대응되는 영역을 제1 영역으로 결정하는 동작;상기 프리뷰 이미지 중 상기 제1 영역을 제외한 영역을 제2 영역으로 결정하는 동작;상기 제1 영역 및 상기 제2 영역의 밝기 차이 정보를 획득하는 동작; 및상기 획득한 밝기 차이 정보에 기반하여, 상기 카메라에 포함된 이미지 센서의 노출 시간을 결정하는 동작을 포함하는, 전자 장치의 동작 방법.
- 청구항 12에 있어서,상기 획득한 밝기 차이가 임계 값 이상인 경우 상기 이미지 센서로부터 상이한 밝기의 복수의 이미지들을 획득하여 HDR 이미지를 생성하는 동작을 포함하는, 전자 장치의 동작 방법.
- 청구항 13에 있어서,상기 제1 영역 또는 상기 제2 영역에 과대 노출이 있는 것으로 판단되는 경우 노출이 작은 이미지를 합성하여 HDR 이미지를 생성하는 동작을 포함하는, 전자 장치의 동작 방법.
- 청구항12에 있어서,상기 프리뷰 이미지를 복수의 영역들로 나누고 상기 복수의 영역들 각각의 거리 정보에 기반하여 영역을 구분하는 동작 및 상기 구분된 영역들 사이의 상기 획득한 밝기 차이 정보를 획득하는 동작을 포함하는, 전자 장치의 동작 방법.
Priority Applications (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202280023529.8A CN117083870A (zh) | 2021-03-25 | 2022-01-05 | 电子装置的图像捕获方法及其电子装置 |
EP22775874.5A EP4319136A4 (en) | 2021-03-25 | 2022-01-05 | IMAGE CAPTURE METHOD FOR ELECTRONIC DEVICE AND RELATED ELECTRONIC DEVICE |
US18/448,504 US20230396889A1 (en) | 2021-03-25 | 2023-08-11 | Image capturing method for electronic device, and electronic device therefor |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
KR1020210038984A KR20220133637A (ko) | 2021-03-25 | 2021-03-25 | 전자 장치의 촬영 방법 및 그 전자 장치 |
KR10-2021-0038984 | 2021-03-25 |
Related Child Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US18/448,504 Continuation US20230396889A1 (en) | 2021-03-25 | 2023-08-11 | Image capturing method for electronic device, and electronic device therefor |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2022203169A1 true WO2022203169A1 (ko) | 2022-09-29 |
Family
ID=83397564
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/KR2022/000170 WO2022203169A1 (ko) | 2021-03-25 | 2022-01-05 | 전자 장치의 촬영 방법 및 그 전자 장치 |
Country Status (5)
Country | Link |
---|---|
US (1) | US20230396889A1 (ko) |
EP (1) | EP4319136A4 (ko) |
KR (1) | KR20220133637A (ko) |
CN (1) | CN117083870A (ko) |
WO (1) | WO2022203169A1 (ko) |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2005308929A (ja) * | 2004-04-20 | 2005-11-04 | Nikon Corp | カメラシステム |
KR20120071968A (ko) * | 2010-12-23 | 2012-07-03 | 삼성전자주식회사 | 디지털 영상 촬영 장치 및 이의 제어 방법 |
KR20160031173A (ko) * | 2014-09-12 | 2016-03-22 | 서울바이오시스 주식회사 | 카메라용 조명장치 및 조명방법 |
KR101716282B1 (ko) * | 2015-09-14 | 2017-03-14 | 주식회사 토비스 | 이중 카메라 플래시를 구비한 이동통신단말기 |
KR20180023785A (ko) * | 2016-08-24 | 2018-03-07 | 삼성전자주식회사 | 복수의 발광소자를 포함하는 전자 장치 및 전자 장치의 동작 방법 |
Family Cites Families (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111182199B (zh) * | 2018-11-13 | 2022-02-11 | 深圳富泰宏精密工业有限公司 | 电子装置及拍照方法 |
KR20210016984A (ko) * | 2019-08-06 | 2021-02-17 | 삼성전자주식회사 | 얼굴이 포함된 영상의 노출 값을 조절하는 전자 장치 |
-
2021
- 2021-03-25 KR KR1020210038984A patent/KR20220133637A/ko active Search and Examination
-
2022
- 2022-01-05 WO PCT/KR2022/000170 patent/WO2022203169A1/ko active Application Filing
- 2022-01-05 CN CN202280023529.8A patent/CN117083870A/zh active Pending
- 2022-01-05 EP EP22775874.5A patent/EP4319136A4/en active Pending
-
2023
- 2023-08-11 US US18/448,504 patent/US20230396889A1/en active Pending
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2005308929A (ja) * | 2004-04-20 | 2005-11-04 | Nikon Corp | カメラシステム |
KR20120071968A (ko) * | 2010-12-23 | 2012-07-03 | 삼성전자주식회사 | 디지털 영상 촬영 장치 및 이의 제어 방법 |
KR20160031173A (ko) * | 2014-09-12 | 2016-03-22 | 서울바이오시스 주식회사 | 카메라용 조명장치 및 조명방법 |
KR101716282B1 (ko) * | 2015-09-14 | 2017-03-14 | 주식회사 토비스 | 이중 카메라 플래시를 구비한 이동통신단말기 |
KR20180023785A (ko) * | 2016-08-24 | 2018-03-07 | 삼성전자주식회사 | 복수의 발광소자를 포함하는 전자 장치 및 전자 장치의 동작 방법 |
Non-Patent Citations (1)
Title |
---|
See also references of EP4319136A4 * |
Also Published As
Publication number | Publication date |
---|---|
CN117083870A (zh) | 2023-11-17 |
EP4319136A4 (en) | 2024-09-18 |
EP4319136A1 (en) | 2024-02-07 |
US20230396889A1 (en) | 2023-12-07 |
KR20220133637A (ko) | 2022-10-05 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2021162307A1 (ko) | 전자 장치 및 그의 hdr 영상 생성 방법 | |
WO2022119372A1 (ko) | 영상 처리를 수행하는 전자 장치 및 그 동작 방법 | |
WO2021215795A1 (ko) | 전자 장치의 컬러 필터, 및 그 전자 장치 | |
WO2019066370A1 (ko) | 외부 광에 기반하여 카메라를 제어하는 전자 장치 및 제어 방법 | |
WO2022244970A1 (ko) | 전자 장치의 촬영 방법 및 그 전자 장치 | |
WO2022196993A1 (ko) | 카메라 모듈의 화각을 이용하여 이미지를 촬영하는 전자 장치 및 방법 | |
WO2022235075A1 (ko) | 전자 장치 및 전자 장치의 동작 방법 | |
WO2022191585A1 (ko) | 전자 장치의 촬영 방법 및 그 전자 장치 | |
WO2021235884A1 (ko) | Awb를 수행하여 이미지를 생성하는 전자 장치 및 방법 | |
WO2022203169A1 (ko) | 전자 장치의 촬영 방법 및 그 전자 장치 | |
WO2022245148A1 (ko) | 이미지 처리 방법 및 그 전자 장치 | |
WO2024085487A1 (ko) | 카메라의 설정을 변경하기 위한 전자 장치, 방법, 및 비일시적 컴퓨터 판독가능 저장 매체 | |
WO2024106746A1 (ko) | 디지털 보케 이미지의 해상도를 높이기 위한 전자 장치 및 방법 | |
WO2021230580A1 (ko) | 멀티 카메라를 이용하여 화질 개선하는 전자 장치 및 방법 | |
WO2022240183A1 (ko) | 이미지 데이터와 모션 데이터를 포함하는 파일을 생성하는 방법 및 그 전자 장치 | |
WO2023080520A1 (ko) | 이미지의 화질을 개선하기 위한 전자 장치 | |
WO2023080495A1 (ko) | 하이 다이나믹 레인지 이미지들을 촬영하는 전자 장치와 이의 동작 방법 | |
WO2022025574A1 (ko) | 이미지 센서와 이미지 시그널 프로세서를 포함하는 전자 장치 및 그의 방법 | |
WO2022240186A1 (ko) | 이미지 왜곡을 보정하는 방법 및 그 전자 장치 | |
WO2024071639A1 (ko) | 플리커링 보정 방법 및 그 전자 장치 | |
WO2024154949A1 (ko) | 레이저 센서를 이용하여 획득한 신뢰도를 나타내는 데이터에 기반하여 렌즈 어셈블리의 위치를 조절하기 위한 전자 장치, 방법, 및 비-일시적 컴퓨터 판독 가능 저장 매체 | |
WO2022220590A1 (ko) | 영상을 처리하는 방법 및 이를 수행하는 전자 장치 | |
WO2024096308A1 (ko) | 광원 객체 이미지 처리 방법 및 장치 | |
WO2022250344A1 (ko) | 이미지 센서 및 동적 비전 센서를 포함하는 전자 장치 및 그 동작 방법 | |
WO2024085673A1 (ko) | 다중 노출 영상을 획득하는 전자 장치 및 그 동작 방법 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 22775874 Country of ref document: EP Kind code of ref document: A1 |
|
WWE | Wipo information: entry into national phase |
Ref document number: 202280023529.8 Country of ref document: CN |
|
WWE | Wipo information: entry into national phase |
Ref document number: 202317072269 Country of ref document: IN |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2022775874 Country of ref document: EP |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
ENP | Entry into the national phase |
Ref document number: 2022775874 Country of ref document: EP Effective date: 20231024 |