WO2018182282A1 - Dispositif électronique et procédé de traitement d'image associé - Google Patents

Dispositif électronique et procédé de traitement d'image associé Download PDF

Info

Publication number
WO2018182282A1
WO2018182282A1 PCT/KR2018/003591 KR2018003591W WO2018182282A1 WO 2018182282 A1 WO2018182282 A1 WO 2018182282A1 KR 2018003591 W KR2018003591 W KR 2018003591W WO 2018182282 A1 WO2018182282 A1 WO 2018182282A1
Authority
WO
WIPO (PCT)
Prior art keywords
image
electronic device
pixel
pixel included
exposure time
Prior art date
Application number
PCT/KR2018/003591
Other languages
English (en)
Korean (ko)
Inventor
이규봉
박정용
신정수
유형진
이우용
이인표
Original Assignee
삼성전자 주식회사
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 삼성전자 주식회사 filed Critical 삼성전자 주식회사
Priority to US16/498,809 priority Critical patent/US11070736B2/en
Publication of WO2018182282A1 publication Critical patent/WO2018182282A1/fr

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/58Means for changing the camera field of view without moving the camera body, e.g. nutating or panning of optics or image sensors
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/60Control of cameras or camera modules
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/70Circuitry for compensating brightness variation in the scene
    • H04N23/73Circuitry for compensating brightness variation in the scene by influencing the exposure time
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/45Cameras or camera modules comprising electronic image sensors; Control thereof for generating image signals from two or more image sensors being of different type or operating in different modes, e.g. with a CMOS sensor for moving images in combination with a charge-coupled device [CCD] for still images
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/60Control of cameras or camera modules
    • H04N23/62Control of parameters via user interfaces
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/60Control of cameras or camera modules
    • H04N23/63Control of cameras or camera modules by using electronic viewfinders
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/60Control of cameras or camera modules
    • H04N23/68Control of cameras or camera modules for stable pick-up of the scene, e.g. compensating for camera body vibrations
    • H04N23/681Motion detection
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/60Control of cameras or camera modules
    • H04N23/68Control of cameras or camera modules for stable pick-up of the scene, e.g. compensating for camera body vibrations
    • H04N23/681Motion detection
    • H04N23/6811Motion detection based on the image signal
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/60Control of cameras or camera modules
    • H04N23/68Control of cameras or camera modules for stable pick-up of the scene, e.g. compensating for camera body vibrations
    • H04N23/681Motion detection
    • H04N23/6812Motion detection based on additional sensors, e.g. acceleration sensors
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N25/00Circuitry of solid-state image sensors [SSIS]; Control thereof
    • H04N25/50Control of the SSIS exposure
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N25/00Circuitry of solid-state image sensors [SSIS]; Control thereof
    • H04N25/50Control of the SSIS exposure
    • H04N25/53Control of the integration time
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N25/00Circuitry of solid-state image sensors [SSIS]; Control thereof
    • H04N25/50Control of the SSIS exposure
    • H04N25/57Control of the dynamic range
    • H04N25/58Control of the dynamic range involving two or more exposures
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N25/00Circuitry of solid-state image sensors [SSIS]; Control thereof
    • H04N25/40Extracting pixel data from image sensors by controlling scanning circuits, e.g. by modifying the number of pixels sampled or to be sampled
    • H04N25/44Extracting pixel data from image sensors by controlling scanning circuits, e.g. by modifying the number of pixels sampled or to be sampled by partially reading an SSIS array

Definitions

  • Various embodiments of the present disclosure relate to an electronic device and an image processing method thereof.
  • the present disclosure relates to an apparatus and a method for processing at least one image by using two or more images acquired by one shooting.
  • an electronic device may include a camera module to provide a user with various functions related to the camera module.
  • an image having a panning shot effect is obtained by moving an electronic device having a long exposure time of the camera module at a speed of a moving subject.
  • the panning shot may refer to a photographing method or an effect in which the moving subject is clearly expressed as if it is stationary, and the background is blurred to make the moving subject stand out.
  • the user when photographing a subject while moving the electronic device at the speed of a moving subject, the user may have to move the electronic device at the speed of the subject, and the electronic device may be moved by the user's movement (eg, hand shaking). If it is shaken, there may be a problem that it is difficult for the user to obtain the desired image.
  • the user's movement eg, hand shaking
  • an image having a panning shot may be obtained by continuously photographing a moving subject at regular time intervals. For example, by calculating a motion vector between the positions of the subjects included in the plurality of images, it is possible to determine whether an object included in the image is an object corresponding to a moving subject or an object corresponding to a background.
  • the image having the panning shot effect may be obtained by processing the image so that the panning shot effect is applied to the object corresponding to the moving subject or the object corresponding to the background.
  • An electronic device may include a camera module; And a processor, the processor receiving a signal related to photographing and, in response to the signal, via the camera module, a first image associated with a first exposure time and a second exposure shorter than the first exposure time. Acquire a second image related to time by one shot, and include at least one pixel included in the first image and at least one included in the second image corresponding to at least one pixel included in the first image Compare the pixels, based on the comparison result, divide at least one pixel included in the second image into a first group and a second group, determine one of the first group and the second group, and At least one pixel included in the determined group may be processed.
  • an image processing method of an electronic device may include: receiving a signal related to photographing; In response to the signal, acquiring, via a camera module, a first image related to a first exposure time and a second image related to a second exposure time shorter than the first exposure time in one shot; Comparing at least one pixel included in the first image with at least one pixel included in the second image corresponding to at least one pixel included in the first image; Dividing at least one pixel included in the second image into a first group and a second group based on the comparison result; Determining one of the first group and the second group; And processing at least one pixel included in the determined group.
  • An electronic device and an image processing method thereof may provide an image to which various high quality image effects are applied by using a plurality of images acquired by one shooting.
  • FIG. 1 is a block diagram illustrating a network environment including an electronic device according to various embodiments of the present disclosure.
  • FIG. 2 is a block diagram of an electronic device according to various embodiments of the present disclosure.
  • FIG. 3 is a block diagram of a program module according to various embodiments of the present disclosure.
  • FIG. 4 is a flowchart illustrating an image processing method of an electronic device according to various embodiments of the present disclosure.
  • FIG. 5 is a flowchart illustrating an image processing method of an electronic device according to various embodiments of the present disclosure.
  • FIG. 6 is a flowchart illustrating an image processing method of an electronic device according to various embodiments of the present disclosure.
  • FIG. 7 is an exemplary diagram illustrating an electronic device including a plurality of camera modules according to various embodiments of the present disclosure.
  • 8A and 8B are exemplary views illustrating patterns of elements included in an image sensor according to various embodiments of the present disclosure.
  • FIG. 9A is an exemplary view conceptually illustrating a first image according to various embodiments of the present disclosure
  • FIG. 9B is an exemplary view conceptually illustrating a second image according to various embodiments of the present disclosure
  • FIG. 9C is a conceptual diagram of the present invention. According to various embodiments of the present disclosure, an example conceptually showing that a subject included in a first image and a subject included in a second image are matched.
  • 10A is an exemplary view conceptually illustrating a first image associated with a first exposure time according to various embodiments of the present disclosure
  • FIG. 10B illustrates a second exposure time shorter than the first exposure time according to various embodiments of the present disclosure
  • 10C is an exemplary view conceptually illustrating a related second image
  • FIG. 10C is an exemplary view conceptually illustrating a second image processed with a panning shot effect according to various embodiments of the present disclosure.
  • 11A is an exemplary view conceptually illustrating a first image related to a first exposure time according to various embodiments of the present disclosure
  • FIG. 11B illustrates a second exposure time shorter than the first exposure time according to various embodiments of the present disclosure
  • 11C is an exemplary view conceptually illustrating a related second image
  • FIG. 11C is an exemplary view conceptually illustrating a second image processed with a panning shot effect according to various embodiments of the present disclosure.
  • the expression “device configured to” may mean that the device “can” together with other devices or components.
  • processor configured (or configured to) perform A, B, and C may be implemented by executing a dedicated processor (eg, an embedded processor) to perform its operation, or one or more software programs stored in a memory device. It may mean a general purpose processor (eg, a CPU or an application processor) capable of performing the corresponding operations.
  • An electronic device may be, for example, a smartphone, a tablet PC, a mobile phone, a video phone, an e-book reader, a desktop PC, a laptop PC, a netbook computer, a workstation, a server, a PDA, a PMP. (portable multimediaplayer), an MP3 player, a medical device, a camera, or a wearable device.
  • Wearable devices may be accessory (e.g. watches, rings, bracelets, anklets, necklaces, eyeglasses, contact lenses, or head-mounted-devices (HMDs), textiles or clothing integrated (e.g.
  • an electronic device may comprise, for example, a television, a digital video disk (DVD) player, Audio, Refrigerator, Air Conditioner, Cleaner, Oven, Microwave Oven, Washing Machine, Air Purifier, Set Top Box, Home Automation Control Panel, Security Control Panel, Media Box (e.g. Samsung HomeSync TM , Apple TV TM , or Google TV TM ) , A game console (eg, Xbox TM , PlayStation TM ), an electronic dictionary, an electronic key, a camcorder, or an electronic picture frame.
  • DVD digital video disk
  • the electronic device may include a variety of medical devices (e.g., various portable medical measuring devices such as blood glucose meters, heart rate monitors, blood pressure meters, or body temperature meters), magnetic resonance angiography (MRA), magnetic resonance imaging (MRI), Computed tomography (CT), cameras or ultrasounds), navigation devices, global navigation satellite systems (GNSS), event data recorders (EDRs), flight data recorders (FDRs), automotive infotainment devices, ship electronics (E.g., various portable medical measuring devices such as blood glucose meters, heart rate monitors, blood pressure meters, or body temperature meters), magnetic resonance angiography (MRA), magnetic resonance imaging (MRI), Computed tomography (CT), cameras or ultrasounds), navigation devices, global navigation satellite systems (GNSS), event data recorders (EDRs), flight data recorders (FDRs), automotive infotainment devices, ship electronics (E.g.
  • various portable medical measuring devices such as blood glucose meters, heart rate monitors, blood pressure meters, or body temperature meters
  • MRA magnetic resonance angiography
  • an electronic device may be a part of a furniture, building / structure or automobile, an electronic board, an electronic signature receiving device, a projector, or various measuring devices (eg, water, electricity, Gas, or a radio wave measuring instrument).
  • the electronic device may be flexible or a combination of two or more of the aforementioned various devices.
  • Electronic devices according to embodiments of the present disclosure are not limited to the above-described devices.
  • the term user may refer to a person who uses an electronic device or a device (eg, an artificial intelligence electronic device) that uses an electronic device.
  • the electronic device 101 may include a bus 110, a processor 120, a memory 130, an input / output interface 150, a display 160, and a communication interface 170.
  • the electronic device 101 may omit at least one of the components or additionally include other components.
  • the bus 110 may include circuitry that connects the components 110-170 to each other and transfers communication (eg, control messages or data) between the components.
  • the processor 120 may include one or more of a central processing unit, an application processor, or a communication processor (CP).
  • the processor 120 may execute, for example, an operation or data processing related to control and / or communication of at least one other component of the electronic device 101.
  • the memory 130 may include volatile and / or nonvolatile memory.
  • the memory 130 may store, for example, commands or data related to at least one other element of the electronic device 101.
  • the memory 130 may store software and / or a program 140.
  • the program 140 may include, for example, a kernel 141, middleware 143, an application programming interface (API) 145, an application program (or “application”) 147, or the like.
  • API application programming interface
  • application or “application”
  • At least a portion of kernel 141, middleware 143, or API 145 may be referred to as an operating system.
  • the kernel 141 may be a system resource (eg, used to execute an action or function implemented in, for example, other programs (eg, middleware 143, API 145, or application program 147).
  • the bus 110, the processor 120, or the memory 130 may be controlled or managed.
  • the kernel 141 may provide an interface for controlling or managing system resources by accessing individual components of the electronic device 101 from the middleware 143, the API 145, or the application program 147. Can be.
  • the middleware 143 may serve as an intermediary for allowing the API 145 or the application program 147 to communicate with the kernel 141 to exchange data.
  • the middleware 143 may process one or more work requests received from the application program 147 according to priority.
  • the middleware 143 may use system resources (eg, the bus 110, the processor 120, or the memory 130, etc.) of the electronic device 101 for at least one of the application programs 147. Prioritize and process the one or more work requests.
  • the API 145 is an interface for the application 147 to control functions provided by the kernel 141 or the middleware 143.
  • the API 145 may include at least the following: file control, window control, image processing, or character control. It can contain one interface or function (eg command).
  • the input / output interface 150 may transmit, for example, a command or data input from a user or another external device to other component (s) of the electronic device 101, or other components of the electronic device 101 ( Commands or data received from the device) can be output to the user or other external device.
  • Display 160 may be, for example, a liquid crystal display (LCD), a light emitting diode (LED) display, an organic light emitting diode (OLED) display, or a microelectromechanical system (MEMS) display, or an electronic paper display. It may include.
  • the display 160 may display, for example, various types of content (eg, text, images, videos, icons, and / or symbols, etc.) to the user.
  • the display 160 may include a touch screen and may receive, for example, a touch, gesture, proximity, or hovering input using an electronic pen or a part of a user's body.
  • the communication interface 170 may establish communication between the electronic device 101 and an external device (eg, the first external electronic device 102, the second external electronic device 104, or the server 106). Can be.
  • the communication interface 170 may be connected to the network 162 through wireless or wired communication to communicate with an external device (eg, the second external electronic device 104 or the server 106).
  • Wireless communication includes, for example, long-term evolution (LTE), LTE Advance (LTE-A), code division multiple access (CDMA), wideband CDMA (WCDMA), universal mobile telecommunications system (UMTS), and wireless broadband (WBro). Or cellular communication using at least one of the Global System for Mobile Communications (GSM) and the like.
  • wireless communication may include, for example, near field communication 164.
  • Local area communication 164 includes, for example, wireless fidelity (WiFi), Bluetooth, Bluetooth Low Power (BLE), Zigbee, near field communication (NFC), magnetic secure transmission, and radio. It may include at least one of a frequency (frequency) or a body area network (BAN).
  • the wireless communication may include GNSS.
  • the GNSS may be, for example, a Global Positioning System (GPS), a Global Navigation Satellite System (Glonass), a Beidou Navigation Satellite System (hereinafter referred to as "Beidou”), or a Galileo, the Europeanglobal satellite-based navigation system.
  • GPS Global Positioning System
  • Galileo Beidou Navigation Satellite System
  • the wired communication may include, for example, at least one of a universal serial bus (USB), a high definition multimedia interface (HDMI), a standard standard232 (RS-232), a power line communication, a plain old telephone service (POTS), and the like.
  • the network 162 may comprise a telecommunications network, for example at least one of a computer network (eg, LAN or WAN), the Internet, or a telephone network.
  • Each of the first and second external electronic devices 102 and 104 may be a device of the same or different type as the electronic device 101. According to various embodiments of the present disclosure, all or part of operations executed in the electronic device 101 may be executed in another or a plurality of electronic devices (for example, the electronic devices 102 and 104 or the server 106). According to this, when the electronic device 101 needs to perform a function or service automatically or by request, the electronic device 101 may instead execute or execute the function or service by itself, or at least some function associated therewith.
  • the other electronic device may request the requested function or The additional function may be executed and the result may be transmitted to the electronic device 101.
  • the electronic device 101 may provide the requested function or service by processing the received result as it is or additionally.
  • Cloud computing distributed computing, or client-server computing techniques can be used.
  • the electronic device 201 may include, for example, all or part of the electronic device 101 illustrated in FIG. 1.
  • the electronic device 201 may include one or more processors (eg, an AP) 210, a communication module 220, a subscriber identification module 224, a memory 230, a sensor module 240, an input device 250, and a display. 260, an interface 270, an audio module 280, a camera module 291, a power management module 295, a battery 296, an indicator 297, and a motor 298.
  • the 210 may control a plurality of hardware or software components connected to the processor 210 by running an operating system or an application program, and may perform various data processing and operations.
  • the processor 210 may further include a graphic processing unit (GPU) and / or an image signal processor. 210 may include at least some of the components shown in FIG. 2 (eg, cellular module 221). The processor 210 other components: processing by loading the command or data received from at least one (e.g., non-volatile memory) in the volatile memory) and can store the result data into the nonvolatile memory.
  • a graphic processing unit GPU
  • an image signal processor may include at least some of the components shown in FIG. 2 (eg, cellular module 221).
  • the processor 210 other components: processing by loading the command or data received from at least one (e.g., non-volatile memory) in the volatile memory) and can store the result data into the nonvolatile memory.
  • the communication module 220 may include, for example, a cellular module 221, a WiFi module 223, a Bluetooth module 225, a GNSS module 227, an NFC module 228, and an RF module 229. have.
  • the cellular module 221 may provide, for example, a voice call, a video call, a text service, or an internet service through a communication network.
  • the cellular module 221 may perform identification and authentication of the electronic device 201 in a communication network by using a subscriber identification module (eg, a SIM card) 224.
  • the cellular module 221 may perform at least some of the functions that the processor 210 may provide.
  • the cellular module 221 may include a communication processor (CP).
  • CP communication processor
  • at least some (eg, two or more) of the cellular module 221, the WiFi module 223, the Bluetooth module 225, the GNSS module 227, or the NFC module 228 may be one integrated chip. (IC) or in an IC package.
  • the RF module 229 may transmit / receive a communication signal (for example, an RF signal), for example.
  • the RF module 229 may include, for example, a transceiver, a power amp module (PAM), a frequency filter, a low noise amplifier (LNA), an antenna, or the like.
  • PAM power amp module
  • LNA low noise amplifier
  • At least one of the cellular module 221, the WiFi module 223, the Bluetooth module 225, the GNSS module 227, or the NFC module 228 may transmit and receive an RF signal through a separate RF module.
  • Subscriber identification module 224 may include, for example, a card or embedded SIM that includes a subscriber identification module, and may include unique identification information (eg, integrated circuit card identifier (ICCID)) or subscriber information (eg, IMSI). (international mobile subscriberidentity)).
  • ICCID integrated circuit card identifier
  • IMSI international mobile subscriberidentity
  • the memory 230 may include, for example, an internal memory 232 or an external memory 234.
  • the internal memory 232 may include, for example, volatile memory (for example, DRAM, SRAM, or SDRAM), nonvolatile memory (for example, one time programmable ROM (OTPROM), PROM, EPROM, EEPROM, mask ROM, flash ROM).
  • the flash memory may include at least one of a flash memory, a hard drive, or a solid state drive (SSD)
  • the external memory 234 may be a flash drive, for example, a compact flash (CF) or a secure digital (SD). ), Micro-SD, Mini-SD, extreme digital (xD), multi-media card (MMC), memory stick, etc.
  • the external memory 234 may be functionally connected to the electronic device 201 through various interfaces. Or physically connected.
  • the sensor module 240 may measure, for example, a physical quantity or detect an operation state of the electronic device 201 and convert the measured or detected information into an electrical signal.
  • the sensor module 240 includes, for example, a gesture sensor 240A, a gyro sensor 240B, an air pressure sensor 240C, a magnetic sensor 240D, an acceleration sensor 240E, a grip sensor 240F, and a proximity sensor ( 240G), color sensor 240H (e.g., red (green, blue) sensor), biometric sensor 240I, temperature / humidity sensor 240J, illuminance sensor 240K, or UV (ultra violet) ) May include at least one of the sensors 240M.
  • sensor module 240 may include, for example, an e-nose sensor, an electromyography (EMG) sensor, an electrocardiogram (EEG) sensor, an electrocardiogram (ECG) sensor, Infrared (IR) sensors, iris sensors and / or fingerprint sensors.
  • the sensor module 240 may further include a control circuit for controlling at least one or more sensors belonging therein.
  • the electronic device 201 further includes a processor configured to control the sensor module 240 as part of or separately from the processor 210, while the processor 210 is in a sleep state. The sensor module 240 may be controlled.
  • the input device 250 may include, for example, a touch panel 252, a (digital) pen sensor 254, a key 256, or an ultrasonic input device 258.
  • the touch panel 252 may use at least one of capacitive, resistive, infrared, or ultrasonic methods, for example.
  • the touch panel 252 may further include a control circuit.
  • the touch panel 252 may further include a tactile layer to provide a tactile response to the user.
  • the (digital) pen sensor 254 may be, for example, part of a touch panel or may include a separate recognition sheet.
  • the key 256 may include, for example, a physical button, an optical key, or a keypad.
  • the ultrasonic input device 258 may detect ultrasonic waves generated by an input tool through a microphone (for example, the microphone 288) and check data corresponding to the detected ultrasonic waves.
  • Display 260 may include panel 262, hologram device 264, projector 266, and / or control circuitry to control them.
  • the panel 262 may be implemented to be, for example, flexible, transparent, or wearable.
  • the panel 262 may be configured with the touch panel 252 and one or more modules.
  • panel 262 may include a pressure sensor (or force sensor) capable of measuring the strength of the pressure on the user's touch.
  • the pressure sensor may be integrally implemented with the touch panel 252 or one or more sensors separate from the touch panel 252.
  • the hologram 264 may show a stereoscopic image in the air by using interference of light.
  • the projector 266 may display an image by projecting light onto a screen.
  • the screen may be located inside or outside the electronic device 201.
  • the interface 270 may include, for example, an HDMI 272, a USB 274, an optical interface 276, or a D-subminiature 278.
  • the interface 270 may be included in, for example, the communication interface 170 illustrated in FIG. 1.
  • interface 270 may include, for example, a mobile high-definition link (MHL) interface, an SD card / multi-media card (MMC) interface, or an infrared data association (IrDA) compliant interface. have.
  • MHL mobile high-definition link
  • MMC Secure Digital Card
  • IrDA infrared data association
  • the audio module 280 may bidirectionally convert, for example, a sound and an electrical signal. At least some components of the audio module 280 may be included in, for example, the input / output interface 145 illustrated in FIG. 1.
  • the audio module 280 may process sound information input or output through, for example, a speaker 282, a receiver 284, an earphone 286, a microphone 288, or the like.
  • the camera module 291 is, for example, a device capable of capturing still images (hereinafter referred to as “images”) and moving images.
  • the camera module 291 may include one or more image sensors (eg, a front sensor or a rear sensor) and a lens. , Image signal processor (ISP), or flash (eg, LED or xenon lamp, etc.).
  • the image sensor may include an integrated circuit photoelectric conversion element using a semiconductor device manufacturing technology.
  • the image sensor included in the camera module 291 may, for example, detect photons of light received through the lens and convert the photon into an electrical signal.
  • the camera module 291 may include an image sensor in which photodiodes for detecting light and generating charges are arranged in one or two dimensions.
  • the processor 210 independently of each device, for example, the time (hereinafter, "exposure time") that the device corresponding to each pixel of the image included in the image sensor is exposed to light Can also be set.
  • the camera module 291 may include two or more lenses, for example.
  • the second lens may be disposed on the rear surface of the electronic device.
  • the first lens may be disposed on the same surface as the front or rear surface of the electronic device on which the second lens is disposed.
  • light received through the first lens may be detected through the first image sensor, and light received through the second lens may be detected through the second image sensor.
  • the processor 210 may independently set the exposure time of each image sensor.
  • the processor 210 may determine the exposure time based on, for example, data (eg, angular velocity, illuminance, etc.) acquired through the sensor module 240.
  • data eg, angular velocity, illuminance, etc.
  • one pixel included in the image corresponds to one element included in the image sensor, but the present invention is not limited to the above exemplary embodiment.
  • the power management module 295 may manage power of the electronic device 201, for example.
  • the power management module 295 may include a power management integrated circuit (PMIC), a charger IC, or a battery or fuel gauge.
  • the PMIC may have a wired and / or wireless charging scheme.
  • the wireless charging method may include, for example, a magnetic resonance method, a magnetic induction method, an electromagnetic wave method, or the like, and may further include additional circuits for wireless charging, such as a coil loop, a resonance circuit, a rectifier, and the like. have.
  • the battery gauge may measure, for example, the remaining amount of the battery 296, the voltage, the current, or the temperature during charging.
  • the battery 296 may include, for example, a rechargeable cell and / or a solar cell.
  • the indicator 297 may display a specific state of the electronic device 201 or a part thereof (for example, the processor 210), for example, a booting state, a message state, or a charging state.
  • the motor 298 may convert electrical signals into mechanical vibrations, and may generate vibrations or haptic effects.
  • the electronic device 201 may be, for example, a mobile TV support device capable of processing media data according to a standard such as digital multimedia broadcasting (DMB), digital video broadcasting (DVB), or mediaFlo TM . GPU).
  • DMB digital multimedia broadcasting
  • DVD digital video broadcasting
  • mediaFlo TM mediaFlo TM . GPU
  • Each of the components described in this document may be composed of one or more components, and the names of the corresponding components may vary depending on the type of electronic device.
  • the electronic device eg, the electronic device 201) may include some components, omit additional components, or combine some of the components to form a single entity. It is possible to perform the same function of the previous corresponding components.
  • the program module 310 may include an operating system and / or various applications running on the operating system for controlling resources related to the electronic device (eg, the electronic device 101).
  • the application program 147 may be included.
  • the operating system may include, for example, Android TM , iOS TM , Windows TM , Symbian TM , Tizen TM , or Bada TM .
  • the program module 310 may include the kernel 320 (eg, the kernel 141), the middleware 330 (eg, the middleware 143), and the API 360 (eg, the API 145).
  • At least a portion of the program module 310 may be preloaded on the electronic device or may be an external electronic device (eg, an electronic device ( 102, 104, server 106, etc.).
  • the kernel 320 may include, for example, a system resource manager 321 and / or a device driver 323.
  • the system resource manager 321 may perform control, allocation, or retrieval of system resources.
  • the system resource manager 321 may include a process manager, a memory manager, or a file system manager.
  • the device driver 323 may include, for example, a display driver, a camera driver, a Bluetooth driver, a shared memory driver, a USB driver, a keypad driver, a WiFi driver, an audio driver, or an inter-process communication (IPC) driver.
  • the middleware 330 may provide various functions through the API 360, for example, to provide functions commonly required by the application 370, or to allow the application 370 to use limited system resources inside the electronic device.
  • the middleware 330 may include a runtime library 335, an application manager 341, a window manager 342, a multimedia manager 343, a resource manager 344, a power manager 345, and a database manager ( 346, a package manager 347, a connectivity manager 348, a notification manager 349, a location manager 350, a graphic manager 351, or a security manager 352.
  • the runtime library 335 may include, for example, a library module that the compiler uses to add new functionality through the programming language while the application 370 is running.
  • the runtime library 335 may perform input / output management, memory management, or arithmetic function processing.
  • the application manager 341 may manage, for example, the life cycle of the application 370.
  • the window manager 342 may manage GUI resources used on the screen.
  • the multimedia manager 343 may identify a format necessary for playing the media files, and may encode or decode the media file using a codec suitable for the format.
  • the resource manager 344 may manage space of source code or memory of the application 370.
  • the power manager 345 may manage, for example, the capacity or power of the battery and provide power information necessary for the operation of the electronic device.
  • the power manager 345 may interwork with a basic input / output system (BIOS).
  • the database manager 346 may create, retrieve, or change a database to be used, for example, in the application 370.
  • the package manager 347 may manage installation or update of an application distributed in the form of a package file.
  • the connectivity manager 348 may manage, for example, a wireless connection.
  • the notification manager 349 may provide the user with events such as, for example, an arrival message, an appointment, a proximity notification, and the like.
  • the location manager 350 may manage location information of the electronic device, for example.
  • the graphic manager 351 may manage, for example, graphic effects to be provided to the user or a user interface related thereto.
  • the security manager 352 may provide system security or user authentication, for example.
  • the middleware 330 may include a telephony manager for managing a voice or video call function of the electronic device or a middleware module capable of forming a combination of functions of the above-described components. .
  • the middleware 330 may provide a module specialized for each type of operating system.
  • the middleware 330 may dynamically delete some of the existing components or add new components.
  • API 360 is, for example, a set of API programming functions, which may be provided in different configurations depending on the operating system. For example, in the case of Android or iOS, one API set may be provided for each platform, and in Tizen, two or more API sets may be provided for each platform.
  • the application 370 is, for example, a home 371, a dialer 372, an SMS / MMS 373, an instant message (IM) 374, a browser 375, a camera 376, an alarm 377. , Contacts 378, voice dials 379, emails 380, calendars 381, media players 382, albums 383, watches 384, health care (e.g., measures exercise or blood sugar, etc.) Or an application for providing environmental information (eg, barometric pressure, humidity, or temperature information).
  • the application 370 may include an information exchange application capable of supporting information exchange between the electronic device and the external electronic device.
  • the information exchange application may include, for example, a notification relay application for delivering specific information to the external electronic device, or a device management application for managing the external electronic device.
  • the notification delivery application may deliver notification information generated by another application of the electronic device to the external electronic device, or receive notification information from the external electronic device and provide the notification information to the user.
  • the device management application may be, for example, the ability of an external electronic device to communicate with the electronic device (e.g. turn-on / turn-off of the external electronic device itself (or some component) or the brightness (or resolution) of the display). Control), or install, delete, or update an application running on the external electronic device.
  • the application 370 may include an application (eg, a health care application of a mobile medical device) designated according to an attribute of the external electronic device.
  • the application 370 may include an application received from an external electronic device.
  • At least a portion of the program module 310 may be implemented (eg, executed) in software, firmware, hardware (eg, the processor 210), or a combination of at least two or more thereof, and a module for performing one or more functions; It can include a program, routine, instruction set, or process.
  • module includes a unit composed of hardware, software, or firmware, and may be used interchangeably with terms such as logic, logic blocks, components, or circuits.
  • the module may be an integrally formed part or a minimum unit or part of performing one or more functions.
  • Modules may be implemented mechanically or electronically, for example, application-specific integrated circuit (ASIC) chips, field-programmable gate arrays (FPGAs), or known or future developments that perform certain operations. It can include a programmable logic device.
  • ASIC application-specific integrated circuit
  • FPGAs field-programmable gate arrays
  • At least a portion of an apparatus (eg, modules or functions thereof) or method (eg, operations) according to various embodiments may be stored on a computer-readable storage medium (eg, memory 130) in the form of a program module. It can be implemented as.
  • Computer-readable recording media include hard disks, floppy disks, magnetic media (such as magnetic tape), optical recording media (such as CD-ROM, DVD, magnetic-optical media (such as floppy disks), internal memory, and the like.
  • Instructions may include code generated by a compiler or code executable by an interpreter Modules or program modules according to various embodiments may include at least one or more of the above-described components. In some embodiments, operations performed by a module, a program module, or another component may be executed sequentially, in parallel, repeatedly, or heuristically, or at least, or may include other components. Some operations may be executed in a different order, omitted, or other operations may be added.
  • FIG. 4 is a flowchart illustrating an image processing method of an electronic device according to various embodiments of the present disclosure.
  • the electronic device may drive a camera module (eg, the camera module 291).
  • a camera module eg, the camera module 291.
  • the electronic device receives, for example, a user input through an input device (eg, input device 250) for executing an application capable of providing a camera function (eg, camera 376)
  • the camera On the other hand, when the camera module is driven, the electronic device displays, for example, a user interface for a function related to the camera module (for example, a shooting function, a setting function for shooting, etc.).
  • the electronic device may display a user interface related to a function for photographing a moving subject on the display, wherein the user moves through the interface displayed on the display.
  • a function for photographing a subject may be selected, for example, an electronic device may have an ISO (international organization for standardization) value, adjustment related to light sensitivity. Dogs may expose an interface that can be adjusted to levels such as to display the other hand, the parameter value related to the photographing (for example: ISO value, the aperture value, etc.) may be set to a variety of processors.
  • ISO international organization for standardization
  • the electronic device may acquire a preview image through the camera module.
  • the preview image may mean, for example, an image acquired by the electronic device using light received through the lens of the camera module after the camera module is driven and before photographing of the subject is determined.
  • the electronic device may temporarily store the obtained preview image in a memory (eg, the memory 230).
  • the electronic device may output the obtained preview image to the display.
  • the electronic device may acquire the preview image at a preset time interval.
  • the electronic device may determine whether to photograph the subject through the camera module. For example, when a user selects a photographing function using a user interface displayed on a display, the electronic device may determine photographing of a subject.
  • the electronic device may determine whether to photograph the subject based on the preview image. For example, the electronic device may compare the first preview image and the second preview image acquired immediately before acquiring the first preview image, and subject the subject according to a result of comparing the first preview image and the second preview image. You can decide to shoot. For example, the electronic device may compare at least one object included in the first preview image and at least one object included in the second preview image, and check the movement of the subject based on the comparison result. In this case, when the movement of the subject is confirmed, the electronic device may determine the photographing of the subject.
  • the electronic device may acquire an image of the subject.
  • the electronic device may store the acquired image in a memory (eg, the memory 230).
  • the electronic device may receive light through the lens and acquire an image of the subject by exposing the image sensor to light.
  • the electronic device may acquire an image of the subject by adjusting the image sensor to be exposed to light for an exposure time set for the image sensor from the time when the user selects the photographing function.
  • the electronic device may, for example, include a mechanical configuration (eg, a mechanical shutter) that adjusts the amount of light directed to the image sensor, so that the image sensor is exposed to light during the exposure time, and the electrical signal from the image sensor may be adjusted.
  • the processor may adjust the time for extracting the same to obtain the same result as the image sensor is exposed to light during the set exposure time.
  • the electronic device may acquire a first image related to the first exposure time and a second image related to the second exposure time, and the second exposure time will be described as an example.
  • the invention is not limited to the above exemplary embodiment.
  • the electronic device may include a camera module including a first camera module including a first lens and a second camera module including a second lens, and include a first corresponding to each camera module. It may include an image sensor and a second image sensor.
  • the electronic device may independently set, for each image sensor, a time period during which the first image sensor and the second image sensor are exposed to light.
  • the first image sensor may be exposed to light received through the first lens, and the electronic device may acquire the first image based on the light detected by the first image sensor.
  • the second image sensor may be exposed to light received through the second lens, and the electronic device may acquire the second image based on the light detected by the second image sensor.
  • the electronic device may adjust the first image sensor and the second image sensor to be exposed to light for an exposure time set for the first image sensor and the second image sensor, respectively, from the time when the user selects the photographing function. And a first image and a second image, respectively. This will be described with reference to FIG. 7.
  • the first lens 710 of the first camera module and the second lens 720 of the second camera module may be disposed on the rear surface of the electronic device 700, and the electronic device 700 may include: The first image may be acquired based on the light received through the first lens 710, and the second image may be obtained based on the light received through the second lens 720.
  • the first lens 710 and the second lens 720 are disposed on the rear surface of the electronic device as an example, but the present invention is not limited to the above exemplary embodiment.
  • the electronic device may independently set, for each device, a time at which each device included in the image sensor is exposed to light. For example, the electronic device may determine at least one device set as the first exposure time as the first device group, and determine at least one device set as the second exposure time as the second device group. In this case, for example, the electronic device may acquire the first image based on the light detected through the first device group, and obtain the second image based on the light detected through the second device group. have. On the other hand, for example, when the first exposure time is longer than the second exposure time, the device set as the first exposure time may be determined as the long exposure device and the device set as the second exposure time as the short exposure device. This will be described with reference to FIGS. 8A and 8B.
  • 8A and 8B are exemplary views illustrating patterns of elements included in an image sensor according to various embodiments of the present disclosure.
  • the long exposure element 810 and the short exposure element 820 included in the image sensor 800 may be arranged to cross each other in a horizontal stripe pattern.
  • FIG. 8B when the elements included in the image sensor 800 are divided into two elements, one element may be a long exposure element 810, and the other element may be disposed as a short exposure element 820.
  • the long exposure element 810 and the short exposure element 820 are arranged to be intersected or bisected in a horizontal stripe pattern, but the present invention is limited to the above exemplary embodiment.
  • the long exposure element 810 and the short exposure element 820 may be disposed to form various types of regular patterns or irregular patterns.
  • the electronic device may extract a feature of at least one pixel included in the image acquired in operation 430.
  • the characteristic of the pixel may include, for example, hue, brightness, saturation, whether motion blur occurs, the intensity of the motion blur, the direction of the motion blur, and the like.
  • the motion blur is included in the image, for example, at the moment of acquiring the image of the subject in the apparatus for acquiring the image, afterimage of the subject caused by the shaking of the image capturing apparatus or the subject moving rapidly. This may mean that the boundary of the object is blurred.
  • the image sensor when photographing a moving subject with a camera having a short exposure time, the image sensor is exposed to light for a short time, thereby obtaining an image having low brightness and almost no motion blur.
  • the image sensor when photographing a moving subject with a camera set to a long exposure time, since the image sensor is exposed to light for a long time, the brightness is bright and the signal to noise ratio (SNR) is good, but a lot of motion blur is included.
  • SNR signal to noise ratio
  • An image can be obtained.
  • the subject corresponding to the background is fixed, less motion blur may occur even when shooting a background with a camera having a long exposure time, but when shooting a moving subject, the motion blur may be reduced depending on the exposure time of the camera. May occur.
  • the electronic device may extract at least one pixel feature by, for example, comparing at least one pixel included in an image and at least one peripheral pixel adjacent to a predetermined range with the at least one pixel. .
  • the electronic device may identify motion blur for at least one pixel by comparing at least one pixel included in an image and at least one peripheral pixel adjacent to a predetermined range with the at least one pixel.
  • the electronic device may calculate, for example, at least one of the intensity and the direction of the motion blur for at least one pixel.
  • the electronic device may use various known methods to calculate at least one of an intensity and a direction of motion blur for at least one pixel.
  • the electronic device may divide the image into at least one area including at least one pixel based on the similarity of features between pixels adjacent to each other. For example, the electronic device may divide the image into at least one area by using a superpixel segmentation method. The electronic device may calculate, for example, at least one of the strength and the direction of the motion blur for at least one area.
  • the electronic device may compare at least one pixel included in the first image with at least one pixel included in the second image.
  • the electronic device may determine, for example, at least one pixel included in the second image corresponding to the at least one pixel included in the first image, the characteristics of the at least one pixel included in the first image, Features of at least one pixel included in a second image corresponding to at least one pixel included in a first image may be compared.
  • the electronic device may include an intensity and a direction of the motion blur for at least one pixel included in the first image, and at least one included in the second image corresponding to the at least one pixel included in the first image. The intensity and direction of motion blur for a pixel can be compared.
  • the electronic device may include, for example, at least one of an intensity and a direction of the motion blur for at least one pixel included in the first image, and at least one included in the second image corresponding to the at least one pixel included in the first image.
  • at least one pixel included in the second image may be determined as a pixel associated with a moving subject, and if less than or equal to the predetermined criterion, The at least one pixel included in the two images may be determined as a pixel not related to the moving subject.
  • the preset reference may mean an error tolerance that can determine the feature of at least one pixel included in the first image and the feature of the at least one pixel included in the second image as the same. Can be.
  • the electronic device may include at least one area including at least one pixel included in the first image and at least one included in the second image corresponding to at least one pixel included in the first image. It is also possible to compare at least one region containing pixels of.
  • the electronic device includes a first lens and a second lens, acquires a first image by using a first image sensor corresponding to the first lens, and a second image sensor corresponding to the second lens.
  • the position of the subject included in the first image may be different from the position of the subject included in the second image.
  • the electronic device may, for example, calculate a motion vector between the position of the subject included in the first image and the position of the subject included in the second image, and based on the calculated motion vector, the first image may be calculated.
  • the subject included in and the subject included in the second image may be aligned.
  • the motion vector may mean, for example, a direction and a distance between the position of the subject included in the first image and the position of the subject included in the second image. This will be described with reference to FIGS. 9A to 9C.
  • FIG. 9A is an exemplary view conceptually illustrating a first image according to various embodiments of the present disclosure
  • FIG. 9B is an exemplary view conceptually illustrating a second image according to various embodiments of the present disclosure
  • FIG. 9C is a conceptual diagram of the present invention. According to various embodiments of the present disclosure, an example conceptually showing that a subject included in a first image and a subject included in a second image are matched.
  • the subject 915 included in the first image 910 may be a first object.
  • the subject 925 included in the second image and located in the left side of the image may be positioned in the right side of the second image.
  • the bicycle included in the first image 910 The positions of at least one pixel associated with the bicycle and at least one pixel associated with the bicycle included in the second image 920 corresponding to the at least one pixel associated with the bicycle included in the first image 910 may be different. .
  • the electronic device may include the first image based on the feature of at least one pixel associated with the bicycle included in the first image 910 and the feature of the at least one pixel associated with the bicycle included in the second image 920. At least one pixel associated with the bicycle included in the second image 920 corresponding to the at least one pixel associated with the bicycle included in 910 may be determined. In addition, the electronic device may, for example, between a position of at least one pixel associated with a bicycle included in the first image 910 and a position of at least one pixel associated with a bicycle included in the second image 920. The motion vector may be calculated, and at least one pixel included in the first image 910 and at least one pixel included in the second image 920 may be matched based on the calculated motion vector.
  • the electronic device may process a second image related to a second exposure time shorter than the first exposure time.
  • the electronic device may determine a mode related to image processing. For example, when the electronic device processes at least one pixel associated with a moving subject, the electronic device determines the mode as a mode for processing at least one pixel associated with a moving subject (hereinafter, referred to as a “subject processing mode”). And process at least one pixel not associated with a moving subject (eg, a pixel associated with a background) when processing at least one pixel not associated with a moving subject (e.g., a pixel associated with a background). Can be determined as the mode (hereinafter, "background processing mode").
  • the electronic device may include, for example, at least one pixel associated with a moving subject included in the second image and at least one pixel not associated with the moving subject (eg, a pixel associated with a background) according to a mode related to image processing. ) Can be processed. Meanwhile, the electronic device may display, for example, a user interface for selecting a mode related to image processing on the display, and may process the second image based on the mode selected by the user.
  • the electronic device may process at least one pixel included in the second image based on an intensity and a direction of the motion blur with respect to the at least one pixel included in the first image. For example, the intensity of the motion blur for at least one pixel included in the first image and the intensity of the motion blur for at least one pixel included in the second image corresponding to the at least one pixel included in the first image.
  • the panning shot effect may be applied to at least one pixel included in the second image in proportion to the difference therebetween.
  • the electronic device applies a panning shot effect to at least one pixel included in the second image
  • the electronic device includes at least one of the at least one image included in the first image corresponding to the at least one pixel included in the second image.
  • the panning shot effect may be applied to at least one pixel included in the second image in the same direction or in the opposite direction to the direction of the motion blur.
  • FIG. 5 is a flowchart illustrating an image processing method of an electronic device according to various embodiments of the present disclosure. Detailed descriptions of contents overlapping with those described in FIG. 4 will be omitted.
  • the electronic device may drive a camera module.
  • the electronic device may drive the camera module.
  • the electronic device may display, for example, a user interface for a function related to the camera module on the display.
  • the electronic device may acquire a preview image through the camera module.
  • the electronic device may acquire data (hereinafter, “sensor data”) from the sensor module.
  • the electronic device may acquire data related to the movement of the electronic device through the sensor module.
  • the electronic device may acquire data related to the angular velocity of the movement of the electronic device using the gyro sensor.
  • the electronic device may acquire data related to the surrounding environment of the electronic device through the sensor module.
  • the electronic device may calculate the illuminance of the surrounding environment using the illuminance sensor.
  • the electronic device may determine an exposure time of the image sensor based on the sensor data.
  • the electronic device may determine an exposure time of the image sensor based on data acquired through the gyro sensor.
  • the angle of view of the lens included in the electronic device is 90 degrees in the horizontal direction and the number of pixels in the horizontal direction of the image is 1800, one pixel and the angle of view of 0.05 degrees in the horizontal direction of the image may correspond. Can be.
  • the electronic device may move while the object moves 0.1 degrees with respect to the horizontal direction. The time of may be determined as the maximum exposure time at which motion blur does not occur in the image.
  • the electronic device may determine the maximum exposure time at which the motion blur does not occur in the image to 1/3600 seconds. For example, when it is determined that the maximum exposure time without motion blur in the image is 1/3600 seconds, motion blur occurs in the first image related to the first exposure time, and the second exposure time is shorter than the first exposure time. In order to prevent motion blur from occurring in the related second image, the electronic device may set the first exposure time to exceed 1/3600 seconds and the second exposure time to 1/3600 seconds or less. For example, the electronic device may set the first exposure time to exceed 1/1200 seconds so that colors to be displayed on one pixel are displayed on six or more pixels of the first image.
  • the electronic device may determine an exposure time of the image sensor based on the preview image. For example, the electronic device may compare the first preview image with the second preview image acquired immediately before acquiring the first preview image before determining photographing of the subject. In this case, the electronic device may calculate information related to the movement of the subject (for example, a direction in which the subject moves, a speed, etc.) based on the position of the objects included in the preview images and the acquisition time of the preview images. The exposure time for the image sensor may be determined based on the obtained information. For example, the electronic device may determine that motion blur occurs when a color to be displayed on one pixel included in an image exceeds two pixels.
  • the electronic device may determine a maximum exposure time at which the motion blur does not occur in the image to 1/3600 seconds.
  • the electronic device may determine whether to photograph the subject through the camera module. For example, when a user selects a photographing function using a user interface displayed on a display, the electronic device may determine photographing of a subject.
  • the electronic device may branch to operation 520 to obtain sensor data, and set an exposure time based on the acquired sensor data. For example, if the user does not select a shooting function for a certain time, the electronic device may branch to operation 520.
  • the electronic device may acquire an image of the subject.
  • the electronic device may store the acquired image in a memory.
  • the electronic device may receive light through the lens and acquire an image of the subject by exposing the image sensor to light.
  • the electronic device may acquire an image of the subject by adjusting the image sensor to be exposed to light for an exposure time set for the image sensor from the time when the user selects the photographing function.
  • the electronic device may include a first lens and a second lens, and may include a first image sensor and a second image sensor corresponding to each lens.
  • the first image sensor may be exposed to light received through the first lens, and the electronic device may acquire the first image based on the light detected by the first image sensor.
  • the second image sensor may be exposed to light received through the second lens, and the electronic device may acquire the second image based on the light detected by the second image sensor.
  • the electronic device may adjust the first image sensor and the second image sensor to be exposed to light for an exposure time set for the first image sensor and the second image sensor, respectively, from the time when the user selects the photographing function. And a first image and a second image, respectively.
  • the electronic device may independently set, for each device, a time at which each device included in the image sensor is exposed to light. For example, the electronic device may determine at least one device set as the first exposure time as the first device group, and determine at least one device set as the second exposure time as the second device group. In this case, for example, the electronic device may acquire the first image based on the light detected through the first device group, and obtain the second image based on the light detected through the second device group. have. On the other hand, for example, when the first exposure time is longer than the second exposure time, the device set as the first exposure time may be determined as the long exposure device and the device set as the second exposure time as the short exposure device.
  • the electronic device may extract features of at least one pixel included in each image acquired in operation 550. Meanwhile, the electronic device may extract at least one pixel feature by, for example, comparing at least one pixel included in an image and at least one peripheral pixel adjacent to a predetermined range with the at least one pixel. . For example, the electronic device may identify motion blur for at least one pixel by comparing at least one pixel included in an image and at least one peripheral pixel adjacent to a predetermined range with the at least one pixel. The electronic device may calculate, for example, at least one of the intensity and the direction of the motion blur for at least one pixel.
  • the electronic device may divide the image into at least one area including at least one pixel based on the similarity of features between pixels adjacent to each other.
  • the electronic device may compare at least one pixel included in the first image and at least one pixel included in the second image.
  • the electronic device may determine, for example, at least one pixel included in the second image corresponding to the at least one pixel included in the first image, the characteristics of the at least one pixel included in the first image, Features of at least one pixel included in a second image corresponding to at least one pixel included in a first image may be compared.
  • the electronic device may include an intensity and a direction of the motion blur for at least one pixel included in the first image, and at least one included in the second image corresponding to the at least one pixel included in the first image. The intensity and direction of motion blur for a pixel can be compared.
  • the electronic device may include, for example, at least one of an intensity and a direction of the motion blur for at least one pixel included in the first image, and at least one included in the second image corresponding to the at least one pixel included in the first image.
  • at least one pixel included in the second image may be determined as a pixel associated with a moving subject, and if less than or equal to the predetermined criterion, The at least one pixel included in the two images may be determined as a pixel not related to the moving subject.
  • the electronic device may determine a mode related to image processing based on the sensor data.
  • the electronic device may determine the mode as a subject processing mode and the electronic organ is not fixed. If it is determined that the mode can be determined as the background processing mode. Meanwhile, the electronic device may display, for example, a user interface for selecting a mode related to image processing on the display, and may determine a mode related to image processing based on the mode selected by the user.
  • the electronic device may process a second image related to a second exposure time shorter than the first exposure time.
  • the electronic device may process at least one of at least one pixel related to the moving subject included in the second image and at least one pixel not related to the moving subject, for example, according to a mode related to image processing. have.
  • the electronic device may process at least one pixel included in the second image based on an intensity and a direction of the motion blur with respect to the at least one pixel included in the first image. For example, the intensity of the motion blur for at least one pixel included in the first image and the motion blur for at least one pixel included in the second image corresponding to the at least one pixel included in the first image. Based on the difference between the intensities, a panning shot effect may be applied to at least one pixel included in the second image. For example, when the electronic device applies a panning shot effect to at least one pixel included in the second image, the electronic device includes at least one of the at least one image included in the first image corresponding to the at least one pixel included in the second image. A panning shot effect may be applied to at least one pixel included in the second image based on the direction of motion blur with respect to the pixel.
  • FIG. 6 is a flowchart illustrating an image processing method of an electronic device according to various embodiments of the present disclosure. Detailed descriptions of contents overlapping with those described in FIGS. 4 and 5 will be omitted.
  • the electronic device may acquire an image of the subject through the camera module (eg, the camera module 291).
  • the electronic device may receive light through a lens included in a camera module, and acquire an image of a subject by exposing the image sensor to light.
  • the electronic device may include, for example, a first camera module including a first lens and a second camera module including a second lens, and include a first image sensor and a second image sensor corresponding to each camera module. It may include.
  • the first image sensor may be exposed to light received through the first lens, and the electronic device may acquire the first image based on the light detected by the first image sensor.
  • the second image sensor may be exposed to light received through the second lens, and the electronic device may acquire the second image based on the light detected by the second image sensor.
  • the electronic device may independently set, for each device, a time at which each device included in the image sensor is exposed to light. For example, the electronic device may determine at least one device set as the first exposure time as the first device group, and determine at least one device set as the second exposure time as the second device group. In this case, for example, the electronic device may acquire the first image based on the light detected through the first device group, and obtain the second image based on the light detected through the second device group. have.
  • the electronic device may compare at least one pixel included in the first image and at least one pixel included in the second image.
  • the electronic device may extract features of at least one pixel included in the image acquired in operation 610.
  • the electronic device may extract features of at least one pixel by comparing at least one pixel included in an image and at least one peripheral pixel adjacent to a predetermined range with the at least one pixel.
  • the electronic device may determine, for example, at least one pixel included in the second image corresponding to the at least one pixel included in the first image, the characteristics of the at least one pixel included in the first image, Features of at least one pixel included in a second image corresponding to at least one pixel included in a first image may be compared.
  • the electronic device may divide at least one pixel included in the second image into a plurality of groups.
  • the electronic device may include at least one associated with a moving subject included in the second image based on a result of comparing the at least one pixel included in the first image with the at least one pixel included in the second image.
  • the pixels of may be classified into a first group, and at least one pixel (eg, a pixel related to a background) not related to the moving subject included in the second image may be divided into a second group.
  • the electronic device may determine one of the first group and the second group. For example, when the mode related to image processing is a subject processing mode, the electronic device may determine a first group among the first group and the second group, and when the mode related to image processing is a background processing mode, A second group of the group and the second group may be determined.
  • the electronic device may process pixels included in the group determined in operation 640. For example, the electronic device may apply a panning shot effect to at least one pixel included in the first group or the second group.
  • 10A is an exemplary view conceptually illustrating a first image associated with a first exposure time according to various embodiments of the present disclosure
  • FIG. 10B illustrates a second exposure time shorter than the first exposure time according to various embodiments of the present disclosure
  • 10C is an exemplary view conceptually illustrating a related second image
  • FIG. 10C is an exemplary view conceptually illustrating a second image processed with a panning shot effect according to various embodiments of the present disclosure.
  • the first image 1010 may be included in the first image 1010.
  • a moving subject 1015 such as a bicycle
  • a stationary subject such as a background
  • the electronic device when the electronic device acquires the second image 1020 by using a second image sensor that is set to a second exposure time that is sufficiently short so that no motion blur occurs, the second image 1020 may be applied to the second image 1020.
  • the included moving subject 1025 eg, a bicycle
  • the background may be clearly displayed because no motion blur occurs.
  • the electronic device may process a panning shot effect on a pixel related to a background included in the second image 1020 according to a mode related to image processing.
  • the electronic device may determine a mode related to image processing as a background processing mode, and the second image 1020 may be included.
  • a panning shot effect may be applied to at least one pixel associated with a background.
  • the electronic device may determine a mode related to image processing as a background processing mode, and The panning shot effect may be applied to pixels related to the background included in the image 1020.
  • 11A is an exemplary view conceptually illustrating a first image related to a first exposure time according to various embodiments of the present disclosure
  • FIG. 11B illustrates a second exposure time shorter than the first exposure time according to various embodiments of the present disclosure
  • 11C is an exemplary view conceptually illustrating a related second image
  • FIG. 11C is an exemplary view conceptually illustrating a second image processed with a panning shot effect according to various embodiments of the present disclosure.
  • the first image 1110 when an electronic device acquires a first image 1110 using a first image sensor set to a first exposure time long enough to cause motion blur, the first image 1110 may be included in the first image 1110.
  • a moving subject 1115 eg, a stream of water
  • a stationary subject eg, a background
  • the moving subject 1115 may be displayed clearly because no motion blur occurs. have.
  • the electronic device when the electronic device acquires the second image 1120 using the second image sensor that is set to a second exposure time that is sufficiently short so that motion blur does not occur, the second image 1120 may be applied to the second image 1120.
  • the included moving subject 1125 eg, water stream
  • the background may be clearly displayed because no motion blur occurs.
  • the electronic device may perform a panning shot effect on at least one pixel associated with the moving subject 1125 included in the second image 1120 (eg, a pixel associated with a stream of water) according to a mode related to image processing. Can be processed.
  • a panning shot effect on at least one pixel associated with the moving subject 1125 included in the second image 1120 (eg, a pixel associated with a stream of water) according to a mode related to image processing. Can be processed.
  • the electronic device may determine a mode related to image processing as a subject processing mode, which is included in the second image 1120.
  • the panning shot effect may be applied to at least one pixel associated with the moving subject 1125.
  • the electronic device may determine a mode related to image processing as a subject processing mode, and the second image.
  • the panning shot effect may be processed on at least one pixel associated with the moving subject 1125 included in 1120.

Abstract

La présente invention concerne un dispositif électronique qui, selon divers modes de réalisation, comprend un module de prise de vues et un processeur, le processeur pouvant : recevoir un signal connexe à la photographie; acquérir, en réponse au signal, en prenant une seule photographie, une première image connexe à un premier temps d'exposition et une seconde image connexe à un second temps d'exposition plus court que le premier temps d'exposition, par l'intermédiaire du module de prise de vues; comparer au moins un pixel inclus dans la première image et au moins un pixel inclus dans la seconde image, qui correspond au ou aux pixels inclus dans la première image; diviser le ou les pixels inclus dans la seconde image en un premier groupe et un second groupe, sur la base d'un résultat de la comparaison; déterminer l'un parmi le premier groupe et le second groupe; et traiter au moins un pixel inclus dans le groupe déterminé. En outre, divers modes de réalisation sont possibles.
PCT/KR2018/003591 2017-03-27 2018-03-27 Dispositif électronique et procédé de traitement d'image associé WO2018182282A1 (fr)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US16/498,809 US11070736B2 (en) 2017-03-27 2018-03-27 Electronic device and image processing method thereof

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
KR10-2017-0038452 2017-03-27
KR1020170038452A KR102317624B1 (ko) 2017-03-27 2017-03-27 전자 장치 및 그의 이미지 처리 방법

Publications (1)

Publication Number Publication Date
WO2018182282A1 true WO2018182282A1 (fr) 2018-10-04

Family

ID=63677995

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/KR2018/003591 WO2018182282A1 (fr) 2017-03-27 2018-03-27 Dispositif électronique et procédé de traitement d'image associé

Country Status (3)

Country Link
US (1) US11070736B2 (fr)
KR (1) KR102317624B1 (fr)
WO (1) WO2018182282A1 (fr)

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2019040068A1 (fr) * 2017-08-24 2019-02-28 Sony Mobile Communications Inc. Dispositifs de traitement d'image à détection de flou de vitesse efficace et leurs procédés de fonctionnement
JP2019213020A (ja) * 2018-06-04 2019-12-12 ソニーセミコンダクタソリューションズ株式会社 撮像装置および撮像制御方法

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2010166304A (ja) * 2009-01-15 2010-07-29 Fujifilm Corp 撮影装置および撮影装置の制御方法並びにプログラム
KR101573131B1 (ko) * 2011-05-31 2015-11-30 모바일 이미징 인 스웨덴 아베 이미지 촬상 방법 및 장치
US20160301868A1 (en) * 2015-04-10 2016-10-13 Qualcomm Incorporated Automated generation of panning shots
KR20160143138A (ko) * 2015-06-04 2016-12-14 엘지전자 주식회사 카메라 및 그 제어 방법

Family Cites Families (14)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPS5136664B2 (fr) 1972-01-14 1976-10-09
US6639626B1 (en) * 1998-06-18 2003-10-28 Minolta Co., Ltd. Photographing apparatus with two image sensors of different size
US7978239B2 (en) * 2007-03-01 2011-07-12 Eastman Kodak Company Digital camera using multiple image sensors to provide improved temporal sampling
JP4561845B2 (ja) 2008-02-29 2010-10-13 カシオ計算機株式会社 撮像装置と画像処理プログラム
JP4661922B2 (ja) * 2008-09-03 2011-03-30 ソニー株式会社 画像処理装置、撮像装置、固体撮像素子、画像処理方法およびプログラム
JP5136664B2 (ja) 2010-07-09 2013-02-06 カシオ計算機株式会社 画像処理装置、及びプログラム
US8224176B1 (en) * 2011-01-10 2012-07-17 Eastman Kodak Company Combined ambient and flash exposure for improved image quality
US8890986B2 (en) * 2011-04-27 2014-11-18 Aptina Imaging Corporation Method and apparatus for capturing high dynamic range images using multi-frame interlaced exposure images
US9251439B2 (en) * 2011-08-18 2016-02-02 Nikon Corporation Image sharpness classification system
JP5950664B2 (ja) * 2012-04-10 2016-07-13 キヤノン株式会社 撮像装置およびその制御方法
KR101871945B1 (ko) * 2013-01-17 2018-08-02 한화에어로스페이스 주식회사 영상 처리 장치 및 방법
JP6434963B2 (ja) * 2014-04-22 2018-12-05 ソニーセミコンダクタソリューションズ株式会社 画像処理装置、画像処理方法、電子機器、並びにプログラム
JP6467190B2 (ja) * 2014-10-20 2019-02-06 キヤノン株式会社 露出制御装置及びその制御方法、撮像装置、プログラム、記憶媒体
US10205885B2 (en) * 2016-01-26 2019-02-12 Canon Kabushiki Kaisha Image capturing apparatus for controlling shutter speed during panning shooting and method for controlling the same

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2010166304A (ja) * 2009-01-15 2010-07-29 Fujifilm Corp 撮影装置および撮影装置の制御方法並びにプログラム
KR101573131B1 (ko) * 2011-05-31 2015-11-30 모바일 이미징 인 스웨덴 아베 이미지 촬상 방법 및 장치
US20160301868A1 (en) * 2015-04-10 2016-10-13 Qualcomm Incorporated Automated generation of panning shots
KR20160143138A (ko) * 2015-06-04 2016-12-14 엘지전자 주식회사 카메라 및 그 제어 방법

Also Published As

Publication number Publication date
KR20180109204A (ko) 2018-10-08
US11070736B2 (en) 2021-07-20
KR102317624B1 (ko) 2021-10-26
US20200036883A1 (en) 2020-01-30

Similar Documents

Publication Publication Date Title
WO2017209560A1 (fr) Procédé de sortie d'écran et dispositif électronique le prenant en charge
US20180007315A1 (en) Electronic device and image capturing method thereof
WO2016163739A1 (fr) Appareil et procédé de réglage d'appareil de prise de vues
WO2018048183A1 (fr) Dispositif électronique pliable comprenant un afficheur souple
WO2018093060A1 (fr) Dispositif électronique et procédé de commande de dispositif électronique
WO2017188577A1 (fr) Procédé de commande de charge de batterie et dispositif électronique associé
WO2018074898A1 (fr) Procédé et dispositif électronique permettant de commander une puissance d'émission
WO2018048177A1 (fr) Dispositif électronique et procédé de traitement de multiples images
WO2018139752A1 (fr) Procédé et dispositif électronique pour commander une pluralité de caméras
WO2015122616A1 (fr) Procédé de photographie d'un dispositif électronique et son dispositif électronique
WO2018026145A1 (fr) Dispositif électronique et procédé de suivi du regard de dispositif électronique
KR102469426B1 (ko) 이미지 처리 장치 및 이의 동작 방법
WO2017171248A1 (fr) Dispositif électronique et procédé de traitement d'image
WO2018106064A1 (fr) Dispositif électronique permettant de commander un véhicule aérien sans pilote et son procédé de commande
WO2018143642A1 (fr) Dispositif électronique et procédé de détermination de coordonnées tactiles de celui-ci
WO2018128460A1 (fr) Procédé d'affichage d'image d'écran et son dispositif électronique
WO2018174674A1 (fr) Dispositif électronique et procédé d'authentification de données biométriques par l'intermédiaire de plusieurs caméras
WO2018117681A1 (fr) Procédé de traitement d'image et dispositif électronique le prenant en charge
WO2018143675A1 (fr) Procédé de commande de biocapteur et dispositif électronique
WO2018034544A1 (fr) Procédé de commande de connexion de réseau de communication, support d'informations et dispositif électronique associé
WO2018084649A1 (fr) Procédé et appareil permettant d'acquérir des informations par capture d'un œil
WO2018021726A1 (fr) Dispositif électronique et procédé de commande d'activation d'un module de caméra
WO2017175962A1 (fr) Dispositif électronique permettant l'affichage d'image et son procédé de commande
WO2018131852A1 (fr) Dispositif électronique utilisé pour exécuter un appel vidéo, et support d'enregistrement lisible par ordinateur
WO2017082554A1 (fr) Dispositif électronique pour détecter un dispositif accessoire et son procédé de fonctionnement

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 18777465

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 18777465

Country of ref document: EP

Kind code of ref document: A1