WO2017018043A1 - 電子機器、電子機器の動作方法及び制御プログラム - Google Patents
電子機器、電子機器の動作方法及び制御プログラム Download PDFInfo
- Publication number
- WO2017018043A1 WO2017018043A1 PCT/JP2016/065525 JP2016065525W WO2017018043A1 WO 2017018043 A1 WO2017018043 A1 WO 2017018043A1 JP 2016065525 W JP2016065525 W JP 2016065525W WO 2017018043 A1 WO2017018043 A1 WO 2017018043A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- imaging range
- moving body
- unit
- live view
- standard
- Prior art date
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/20—Analysis of motion
- G06T7/246—Analysis of motion using feature-based methods, e.g. the tracking of corners or segments
- G06T7/248—Analysis of motion using feature-based methods, e.g. the tracking of corners or segments involving reference images or patches
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/20—Scenes; Scene-specific elements in augmented reality scenes
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/45—Cameras or camera modules comprising electronic image sensors; Control thereof for generating image signals from two or more image sensors being of different type or operating in different modes, e.g. with a CMOS sensor for moving images in combination with a charge-coupled device [CCD] for still images
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/60—Control of cameras or camera modules
- H04N23/62—Control of parameters via user interfaces
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/60—Control of cameras or camera modules
- H04N23/63—Control of cameras or camera modules by using electronic viewfinders
- H04N23/631—Graphical user interfaces [GUI] specially adapted for controlling image capture or setting capture parameters
- H04N23/632—Graphical user interfaces [GUI] specially adapted for controlling image capture or setting capture parameters for displaying or modifying preview images prior to image capturing, e.g. variety of image resolutions or capturing parameters
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/60—Control of cameras or camera modules
- H04N23/63—Control of cameras or camera modules by using electronic viewfinders
- H04N23/633—Control of cameras or camera modules by using electronic viewfinders for displaying additional information relating to control or operation of the camera
- H04N23/635—Region indicators; Field of view indicators
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/60—Control of cameras or camera modules
- H04N23/64—Computer-aided capture of images, e.g. transfer from script file into camera, check of taken image quality, advice or proposal for image composition or decision on when to take image
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/60—Control of cameras or camera modules
- H04N23/667—Camera operation mode switching, e.g. between still and video, sport and normal or high- and low-resolution modes
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/60—Control of cameras or camera modules
- H04N23/698—Control of cameras or camera modules for achieving an enlarged field of view, e.g. panoramic image capture
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10016—Video; Image sequence
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10024—Color image
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30236—Traffic on road, railway or crossing
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V2201/00—Indexing scheme relating to image or video recognition or understanding
- G06V2201/07—Target detection
Definitions
- This disclosure relates to electronic equipment.
- Patent Document 1 As described in Patent Document 1, conventionally, a technique for imaging a moving body has been proposed.
- an object of the present invention is to provide a technique capable of easily imaging a moving body.
- the electronic device includes a first imaging unit that captures an image of the first imaging range, and a second angle that is wider than the first imaging range when the first imaging unit captures the first imaging range.
- a second imaging unit that images the imaging range, a display unit having a display screen, a detection unit, a determination unit, and an estimation unit are provided.
- the detection unit detects the position of the moving body that moves within the second imaging range based on the image signal from the second imaging unit.
- the determination unit determines whether the moving body exists outside the first imaging range and within the second imaging range based on the position of the moving body detected by the detection unit.
- the estimation unit determines the periphery of the first imaging range based on the position of the moving body detected by the detection unit.
- the approach area through which the moving body passes when it enters the first imaging range is estimated.
- the display unit displays, on the display screen, first notification information for notifying the entry area together with the first live view image captured by the first imaging unit.
- the operation method of the electronic device includes: a first imaging range that captures the first imaging range; and a first imaging range when the first imaging unit captures the first imaging range. It is an operating method of an electronic device provided with the 2nd image pick-up part which images the 2nd image pick-up range with a wider angle.
- the operation method of the electronic device includes a first step of detecting the position of the moving body that moves within the second imaging range based on the image signal from the second imaging unit, and the moving body based on the position of the moving body. A second step of determining whether the moving body is outside the first imaging range and within the second imaging range; and a case where it is determined in the second step that the moving body is outside the first imaging range and within the second imaging range.
- the control program includes a first imaging unit that captures the first imaging range, and a wider angle than the first imaging range when the first imaging unit captures the first imaging range. It is a control program for controlling an electronic device provided with the 2nd imaging part which images the 2nd imaging range.
- the control program includes a first step of detecting a position of the moving body that moves within the second imaging range based on an image signal from the second imaging unit, and a moving body based on the position of the moving body. Is determined to be present outside the first imaging range and in the second imaging range, and in the second step, it is determined that the moving object is outside the first imaging range and within the second imaging range.
- ⁇ Appearance of electronic equipment> 1 and 2 are a perspective view and a back view, respectively, schematically showing an example of the external appearance of the electronic device 1.
- the electronic device 1 is a mobile phone such as a smartphone, for example.
- the electronic device 1 can communicate with other communication devices through a base station, a server, and the like.
- the electronic device 1 includes a cover panel 2 located on the front surface 1a of the electronic device 1 and a device case 3 to which the cover panel 2 is attached.
- the cover panel 2 and the device case 3 form the exterior of the electronic device 1.
- the shape of the electronic device 1 is, for example, a substantially rectangular plate shape in plan view.
- the cover panel 2 is provided with a display screen (display area) 2a on which various information such as characters, symbols, and figures displayed on the display panel 120 described later is displayed.
- a display screen (display area) 2a on which various information such as characters, symbols, and figures displayed on the display panel 120 described later is displayed.
- Most of the peripheral edge 2b surrounding the display screen 2a in the cover panel 2 is black, for example, by applying a film or the like. Thereby, most of the peripheral edge 2b of the cover panel 2 is a non-display area where various information displayed on the display panel 120 is not displayed.
- a touch panel 130 described later is attached to the back surface of the cover panel 2.
- the display panel 120 is affixed on the main surface on the opposite side to the main surface by the side of the cover panel 2 in the touch panel 130.
- the display panel 120 is attached to the back surface of the cover panel 2 via the touch panel 130.
- a third lens transparent portion 20 is provided at the upper end of the cover panel 2 so that a lens included in a third imaging unit 200 described later can be visually recognized from the outside of the electronic device 1. Yes.
- a receiver hole 16 is provided at the upper end of the cover panel 2.
- a speaker hole 17 is provided at the lower end of the cover panel 2.
- a microphone hole 15 is provided on the bottom surface 1 c of the electronic device 1, that is, on the bottom surface (lower side surface) of the device case 3.
- an imaging lens of a first imaging unit 180 described later can be visually recognized from the outside of the electronic device 1 on the back surface 1 b of the electronic device 1, that is, the upper end portion of the back surface of the device case 3.
- the first lens transparent portion 18 is provided.
- a second lens transparent portion 19 is provided at the upper end of the back surface of the device case 3 so that an imaging lens of a second imaging portion 190 described later can be visually recognized from the outside of the electronic device 1.
- the first lens transparent portion 18 and the second lens transparent portion 19 are provided, for example, along the longitudinal direction of the device case 3 on the back surface of the device case 3.
- an operation key group 140 including a plurality of operation keys 141 is provided.
- Each operation key 141 is a hardware key such as a push button, and its surface is exposed from the lower end of the cover panel 2.
- the user can give various instructions to the electronic device 1 by pressing each operation key 141 with a finger or the like.
- the plurality of operation keys 141 include, for example, a home key, a back key, and a task key.
- the home key is an operation key for displaying a home screen (initial screen) on the display screen 2a.
- the back key is an operation key for switching the display on the display screen 2a to the previous display.
- the task key is an operation key for displaying a list of application programs running on the electronic device 1 on the display screen 2a.
- FIG. 3 is a block diagram illustrating an example of an electrical configuration of the electronic device 1.
- the electronic device 1 includes a control unit 100, a wireless communication unit 110, a display unit 121, a touch panel 130, an operation key group 140, a microphone 150, a receiver 160, an external speaker 170, and a first imaging unit 180.
- a second imaging unit 190, a third imaging unit 200, and a battery 210 are provided. Each of these components provided in the electronic device 1 is housed in the device case 3.
- the control unit 100 is a kind of computer, and includes, for example, a CPU (Central Processing Unit) 101, a DSP (Digital Signal Processor) 102, a storage unit 103, and the like. It can be said that the control unit 100 is a control circuit.
- the control unit 100 can comprehensively manage the operation of the electronic device 1 by controlling other components of the electronic device 1.
- the control unit 100 may further include a sub-processor (co-processor) such as a system-on-a-chip (SoC), a micro-control unit (MCU), and a field-programmable gate array (FPGA).
- SoC system-on-a-chip
- MCU micro-control unit
- FPGA field-programmable gate array
- the control unit 100 may perform various types of control by causing the CPU 101 and the sub-processing device to cooperate with each other, or may perform various types of control while switching one of the two.
- the storage unit 103 includes a non-transitory recording medium such as a ROM (Read Only Memory) and a RAM (Random Access Memory) that can be read by the CPU 101 and the DSP 102.
- the ROM included in the storage unit 103 is, for example, a flash ROM (flash memory) that is a nonvolatile memory.
- the storage unit 103 stores a plurality of control programs 103 a for controlling the electronic device 1.
- the plurality of control programs 103a include a main program, a plurality of application programs (hereinafter, simply referred to as an application or an application), and the like.
- Various functions of the control unit 100 are realized by the CPU 101 and the DSP 102 executing various control programs 103 a in the storage unit 103.
- an application program (hereinafter referred to as a camera application) for capturing a still image or a moving image using the first imaging unit 180, the second imaging unit 190, or the third imaging unit 200 is used. Is stored).
- the storage unit 103 may include a computer-readable non-transitory recording medium other than the ROM and RAM.
- the storage unit 103 may include, for example, a small hard disk drive and an SSD (Solid State Drive). Further, all the functions of the control unit 100 or a part of the functions of the control unit 100 may be realized by hardware that does not require software for realizing the function.
- the wireless communication unit 110 has an antenna 111.
- the wireless communication unit 110 may receive a signal from a mobile phone different from the electronic device 1 or a signal from a communication device such as a web server connected to the Internet via the base station by the antenna 111. Is possible.
- the wireless communication unit 110 can perform amplification processing and down-conversion on the signal received by the antenna 111 and output the signal to the control unit 100.
- the control unit 100 can perform demodulation processing or the like on the input reception signal and acquire information such as a sound signal indicating voice or music included in the reception signal.
- the wireless communication unit 110 can perform up-conversion and amplification processing on the transmission signal generated by the control unit 100 and wirelessly transmit the processed transmission signal from the antenna 111.
- the transmission signal from the antenna 111 is received by a communication device such as a mobile phone other than the electronic device 1 or a web server connected to the Internet through the base station.
- the display unit 121 includes a display panel 120 and a display screen 2a.
- the display panel 120 is, for example, a liquid crystal panel or an organic EL panel.
- the display panel 120 can display various types of information such as characters, symbols, and figures by being controlled by the control unit 100.
- Various types of information displayed on the display panel 120 are displayed on the display screen 2a.
- the touch panel 130 is, for example, a projected capacitive touch panel.
- the touch panel 130 can detect an operation with an operator such as a finger on the display screen 2a.
- an electrical signal corresponding to the operation is input from the touch panel 130 to the control unit 100.
- the control part 100 can specify the content of operation performed with respect to the display screen 2a based on the electrical signal from the touch panel 130, and can perform the process according to the content.
- the user can also give various instructions to the electronic device 1 by operating the display screen 2a with an operator other than a finger, for example, an electrostatic touch panel pen such as a stylus pen.
- each operation key 141 of the operation key group 140 When each operation key 141 of the operation key group 140 is pressed by the user, an operation signal indicating that the operation key 141 is operated is output to the control unit 100. Thereby, the control unit 100 can determine whether or not the operation key 141 is operated based on the operation signal from each operation key 141. The control unit 100 can perform processing according to the operated operation key 141.
- Each operation key 141 may be a software key displayed on the display screen 2a instead of a hardware key such as a push button. In this case, an operation on the software key is detected by the touch panel 130, and the control unit 100 can perform processing according to the operated software key.
- the microphone 150 can convert a sound input from the outside of the electronic device 1 into an electric sound signal and output it to the control unit 100. Sound from the outside of the electronic device 1 is taken into the electronic device 1 from the microphone hole 15 provided on the bottom surface (lower side surface) of the device case 3 and input to the microphone 150, for example.
- the external speaker 170 is, for example, a tymic speaker.
- the external speaker 170 can convert an electrical sound signal from the control unit 100 into sound and output the sound.
- the sound output from the external speaker 170 is output to the outside of the electronic device 1 from the speaker hole 17 provided at the lower end of the cover panel 2, for example.
- the sound output from the speaker hole 17 has a volume that can be heard at a place away from the electronic device 1.
- the receiver 160 can output a received sound and is, for example, a dynamic speaker.
- the receiver 160 can convert an electrical sound signal from the control unit 100 into a sound and output the sound.
- the sound output from the receiver 160 is output to the outside through, for example, the receiver hole 16 provided in the upper end portion of the cover panel 2.
- the volume of the sound output from the receiver hole 16 is, for example, smaller than the volume of the sound output from the external speaker 170 via the speaker hole 17.
- a piezoelectric vibration element may be provided.
- the piezoelectric vibration element can vibrate based on an audio signal from the control unit 100.
- the piezoelectric vibration element is provided on the back surface of the cover panel 2, for example, and can vibrate the cover panel 2 by its own vibration based on an audio signal. Then, when the user brings his / her ear close to the cover panel 2, the vibration of the cover panel 2 is transmitted to the user as sound.
- the receiver hole 16 is not necessary.
- the battery 210 can output the power of the electronic device 1.
- the battery 210 is a rechargeable battery such as a lithium ion secondary battery.
- the battery 210 can supply power to various electronic components such as the control unit 100 and the wireless communication unit 110 included in the electronic device 1.
- Each of the first imaging unit 180, the second imaging unit 190, and the third imaging unit 200 includes a lens, an imaging element, and the like. Each of the first imaging unit 180, the second imaging unit 190, and the third imaging unit 200 captures a subject under the control of the control unit 100, generates a still image or a moving image that indicates the captured subject, and sends it to the control unit 100. It is possible to output.
- the control unit 100 can store an input still image or moving image in a nonvolatile memory (flash memory) or a volatile memory (RAM) of the storage unit 103.
- the lens of the third imaging unit 200 is visible from the third lens transparent unit 20 provided on the cover panel 2. Therefore, the third imaging unit 200 can image an object present on the cover panel 2 side of the electronic device 1, that is, the front surface 1 a side of the electronic device 1.
- Such an imaging unit may be referred to as an “in camera”.
- the third imaging unit 200 may be referred to as an “in camera 200”.
- the lens of the first imaging unit 180 is visible from the first lens transparent unit 18 provided on the back surface 1b of the electronic device 1.
- the lens included in the second imaging unit 190 is visible from the second lens transparent portion 19 provided on the back surface 1 b of the electronic device 1. Therefore, the first imaging unit 180 and the second imaging unit 190 can image an object that is present on the back surface 1b side of the electronic device 1.
- the second imaging unit 190 can capture a second imaging range having a wider angle (a wider angle of view) than the first imaging range captured by the first imaging unit 180. Then, when the first imaging unit 180 is imaging the first imaging range, the second imaging unit 190 images a second imaging range that is wider in angle (wider in view angle) than the first imaging range. In other words, when the first imaging unit 180 and the second imaging unit 190 are imaging the first and second imaging ranges, respectively, the angle of view of the second imaging unit 190 is that of the first imaging unit 180. The angle of view is wider than the angle of view.
- FIG. 4 illustrates the first imaging range 185 and the second imaging range 195 when the first imaging unit 180 and the second imaging unit 190 are imaging the first imaging range 185 and the second imaging range 195, respectively. It is a figure which shows an example of a relationship typically with. As shown in FIG. 4, when the first imaging unit 180 is imaging the first imaging range 185, the second imaging range 195 captured by the second imaging unit 190 is more than the first imaging range 185. It is wide and includes the first imaging range 185.
- the first imaging unit 180 is referred to as “standard camera 180”, and the second imaging unit 190 is referred to as “wide-angle camera 190”.
- the first imaging range 185 captured by the standard camera 180 is referred to as “standard imaging range 185”, and the second imaging range 195 captured by the wide-angle camera 190 is referred to as “wide-angle imaging range 195”.
- each of the lenses of the standard camera 180, the wide-angle camera 190, and the in-camera 200 is a single focus lens. Note that at least one of the lenses of the standard camera 180, the wide-angle camera 190, and the in-camera 200 may be a zoom lens.
- the electronic apparatus 1 has a zoom function for each of the standard camera 180, the wide-angle camera 190, and the in-camera 200. That is, the electronic device 1 has a standard camera zoom function for zooming in on a subject imaged by the standard camera 180, a wide-angle camera zoom function for zooming in on a subject imaged by the wide-angle camera 190, and an in-zoom that zooms in on the subject imaged by the in-camera 200. It has a camera zoom function. When the subject to be imaged is zoomed in by the camera zoom function, the imaging range is narrowed. On the other hand, when the subject to be imaged is zoomed out by the camera zoom function, the imaging range is widened.
- each of the standard camera zoom function, the wide-angle camera zoom function, and the in-camera zoom function is a digital zoom function.
- at least one of the standard camera zoom function, the wide-angle camera zoom function, and the in-camera zoom function may be an optical zoom function realized by a zoom lens.
- the standard camera 180 is the standard.
- the wide-angle camera 190 captures a wide-angle imaging range 195 that is wider than the standard imaging range 185.
- the zoom magnification of each of the standard camera 180 and the wide-angle camera 190 is 1, the wide-angle imaging range 195 is wider than the standard imaging range 185.
- the wide-angle camera zoom function of the electronic device 1 is disabled.
- the zoom magnification of the wide-angle camera 190 is fixed to 1.
- the fixed field angle of the wide angle imaging range 195 is wider than the maximum field angle of the standard imaging range 185.
- the wide-angle camera zoom function of the electronic device 1 becomes effective.
- the minimum angle of view of the wide-angle camera 190 may be smaller than the maximum angle of view of the standard camera 180. That is, when the wide-angle camera zoom function is effective, the wide-angle imaging range 195 may have a narrower field angle than the standard imaging range 185.
- FIG. 5 is a flowchart illustrating an example of the operation of the electronic device 1 when the camera application is executed.
- the control unit 100 executes (activates) the camera application stored in the storage unit 103 in step S1, as shown in FIG.
- a home screen initial screen
- a plurality of figures hereinafter, sometimes referred to as application execution figures
- the application execution graphic may include a graphic called an icon.
- the selection operation for the application execution graphic displayed on the display screen 2a for example, an operation of moving the operation element away from the operation element after the user brings the operation element such as a finger close to the application execution graphic can be considered.
- an operation for selecting an application execution graphic displayed on the display screen 2a for example, an operation in which a user touches an operation element such as a finger on the application execution graphic and then moves the operation element away is considered. It is done.
- These operations are called so-called tap operations.
- the selection operation by the tap operation is used as a selection operation for various information displayed on the display screen 2a in addition to the selection operation for the application execution graphic.
- repeated description of the selection operation by the tap operation is omitted.
- the control unit 100 supplies power to the standard camera 180 and the wide-angle camera 190 among the standard camera 180, the wide-angle camera 190, and the in-camera 200 in step S2, and the standard camera 180 and The wide-angle camera 190 is activated.
- the standard camera 180 and the wide-angle camera 190 are activated, the standard camera 180 is a recording camera in which captured still images or moving images are recorded in a nonvolatile memory, and the wide-angle camera 190 is a moving body described later. This is a camera for performing the detection process.
- step S ⁇ b> 3 the control unit 100 controls the display panel 120 to display a live view image (also referred to as a through image, a preview image, or simply a preview) in which the standard camera 180 captures the standard imaging range 185. 2a is displayed. That is, the control unit 100 causes the display screen 2a to display captured images continuously captured at a predetermined frame rate by the standard camera 180 in real time.
- the live view image is an image that is displayed in order for the user to confirm captured images that are continuously captured in real time. It can be said that the plurality of live view images displayed continuously are a kind of moving image. Each live view image can be said to be each frame image of a moving image.
- the live view image is temporarily stored in the volatile memory in the storage unit 103 by the control unit 100. After being stored, it is displayed on the display screen 2a.
- the live view image captured by the standard camera 180 may be referred to as a “standard live view image”.
- FIG. 6 is a diagram showing an example of display on the display screen 2a displaying the standard live view image 300.
- the standard live view image 300 is displayed in a central area 420 (an area other than the upper end 400 and the lower end 410) of the display screen 2a. That is, the subject within the standard imaging range 185 is displayed in the center area 420 of the display screen 2a.
- the operation button 310 is displayed on the lower end 410 of the display screen 2a.
- a mode switching button 320, a camera switching button 330, and a display switching button 340 are displayed on the upper end 400 of the display screen 2a.
- the mode switching button 320 is an operation button for switching the imaging mode of the electronic device 1.
- the imaging mode of the electronic device 1 is the still image imaging mode
- the control unit 100 changes the imaging mode of the electronic device 1 to a still image. Switch from the imaging mode to the moving image imaging mode.
- the imaging mode of the electronic device 1 is the moving image imaging mode
- the control unit 100 changes the imaging mode of the electronic device 1 to the moving image imaging mode. To still image capture mode.
- the camera switching button 330 is an operation button for switching a recording camera for recording a still image or a moving image.
- the control unit 100 changes the recording camera from the standard camera 180 to the wide-angle camera 190, for example. Switch.
- the control unit 100 stops the power supply to the standard camera 180 and stops the operation of the standard camera 180.
- the display unit 121 displays a live view image (hereinafter referred to as “live view image”) in which the wide-angle camera 190 images the wide-angle imaging range 195 instead of the standard live-view image 300.
- live view image a live view image
- a wide-angle live view image is displayed on the display screen 2a.
- the control unit 100 switches the recording camera from the wide-angle camera 190 to the in-camera 200, for example.
- the control unit 100 supplies power to the in-camera 200 and activates the in-camera 200. Then, the control unit 100 stops the supply of power to the wide-angle camera 190 and stops the operation of the wide-angle camera 190.
- the display unit 121 displays a live view image captured by the in-camera 200 on the display screen 2a instead of the wide-angle live view image.
- the control unit 100 switches the recording camera from the in-camera 200 to, for example, the standard camera 180.
- the control unit 100 supplies power to the standard camera 180 and the wide-angle camera 190 and activates the standard camera 180 and the wide-angle camera 190. Then, the control unit 100 stops the supply of power to the in camera 200 and stops the operation of the in camera 200.
- the display unit 121 displays the standard live view image 300 on the display screen 2a instead of the live view image captured by the in-camera 200.
- the recording camera when the camera application is activated may be the wide-angle camera 190 or the in-camera 200 instead of the standard camera 180.
- the order in which the recording cameras are switched may be other than the above order.
- the recording camera when the operation to the camera switching button 330 is detected is the standard camera 180
- the recording camera is switched from the standard camera 180 to the in-camera 200
- the operation to the camera switching button 330 is performed.
- the recording camera at the time of detection is the in-camera 200
- the recording camera is switched from the in-camera 200 to the wide-angle camera 190
- the recording camera when the operation to the camera switching button 330 is detected is the wide-angle camera.
- the recording camera may be switched from the wide-angle camera 190 to the standard camera 180.
- the display unit 121 replaces the camera switching button 330 for sequentially switching the recording cameras, and 2 for switching each of the standard camera 180, the wide-angle camera 190, and the in-camera 200 to two cameras other than the recording camera.
- Two camera switching buttons may be displayed on the display screen 2a.
- the display unit 121 includes a camera switching button for switching the recording camera from the standard camera 180 to the wide-angle camera 190 instead of the camera switching button 330.
- a camera switching button for switching the recording camera from the standard camera 180 to the in-camera 200 may be displayed on the display screen 2a.
- the display unit 121 replaces the camera switching button 330 with a camera switching button for switching the recording camera from the wide-angle camera 190 to the in-camera 200, and a recording camera. And a camera switching button for switching from the wide-angle camera 190 to the standard camera 180 may be displayed on the display screen 2a.
- the display unit 121 replaces the camera switching button 330 with a camera switching button for switching the recording camera from the in-camera 200 to the standard camera 180, and a recording camera. And a camera switching button for switching from the in-camera 200 to the wide-angle camera 190 may be displayed on the display screen 2a.
- the control unit 100 switches the recording camera to the camera corresponding to the operated camera switching button.
- the display switching button 340 is an operation button for switching display / non-display of the wide-angle live view image when the standard camera 180 and the wide-angle camera 190 are activated.
- the display switching button 340 is displayed only when the standard camera 180 and the wide-angle camera 190 are activated.
- the touch panel 130 performs a predetermined operation (for example, a tap operation) on the display switching button 340.
- the display unit 121 displays the wide-angle live view image together with the standard live view image 300 on the display screen 2a.
- FIG. 7 is a diagram illustrating an example of display on the display screen 2a displaying the standard live view image 300 and the wide-angle live view image 350.
- the standard live view image 300 and the wide-angle live view image 350 are displayed in the upper part and the lower part of the central area 420 of the display screen 2a, respectively.
- the display position and display size of the standard live view image 300 and the wide-angle live view image 350 on the display screen 2a are not limited to the example of FIG.
- the standard live view image 300 and the wide-angle live view image 350 may be displayed side by side in the left-right direction of the display screen 2a.
- the standard live view image 300 and the wide-angle live view image 350 may be displayed in a partially overlapped state.
- the standard live view image 300 captured by the standard camera 180 and the wide-angle live view image 350 captured by the wide-angle camera 190 are displayed together on the display screen 2a. Both the subject in the standard imaging range 185 captured by the camera 180 and the subject in the wide-angle imaging range 195 captured by the wide-angle camera 190 can be confirmed.
- the display unit 121 displays the wide-angle live view image 350. Hide. Then, as shown in FIG. 6, the standard live view image 300 is displayed in the central area 420 of the display screen 2a.
- the wide-angle camera 190 outputs a captured image to the control unit 100 while the power is supplied, regardless of whether the wide-angle live view image 350 is displayed or not displayed on the display screen 2a.
- a captured image captured by the wide-angle camera 190 is stored in a volatile memory in the storage unit 103 by the control unit 100.
- the operation button 310 functions as a shutter button when the imaging mode of the electronic device 1 is the still image imaging mode.
- the operation button 310 functions as an operation button for starting and stopping moving image capturing when the image capturing mode of the electronic device 1 is the moving image capturing mode.
- the control unit 100 displays a recording camera (FIG. 5) when the operation button 310 is operated.
- a still image for recording different from the live view image captured by the standard camera 180 is stored in the nonvolatile memory in the storage unit 103, and the still image is displayed on the display screen 2a.
- the control unit 100 captures an image with the recording camera. Storage of a moving image for recording different from the live view image into the nonvolatile memory in the storage unit 103 is started. After that, when the touch panel 130 detects a predetermined operation on the operation button 310, the control unit 100 stops storing the recording moving image captured by the recording camera in the nonvolatile memory in the storage unit 103.
- a predetermined operation for example, a tap operation
- the operation mode of the recording camera is different when capturing a still image for recording, capturing a moving image for recording, and capturing a live view image. For example, between the operation mode for capturing a still image for recording, the operation mode for capturing a moving image for recording, and the operation mode for capturing a live view image, The number of pixels and the exposure time are different. For example, a recording still image has a larger number of pixels than a live view image.
- step S4 the control unit 100 determines whether or not there is a moving body that moves within the wide-angle imaging range 195. Specifically, the control unit 100 performs, for example, image processing such as moving object detection based on inter-frame differences on a series of input images continuously input from the wide-angle camera 190 at a predetermined frame rate. The position of the moving body in each input image is detected. As the position of the moving body, for example, the center coordinates of the area where the moving body exists in each input image are detected. For example, the wide-angle live view image 350 output from the wide-angle camera 190 and stored in the volatile memory in the storage unit 103 is used for the detection processing of the position of the moving object.
- image processing such as moving object detection based on inter-frame differences on a series of input images continuously input from the wide-angle camera 190 at a predetermined frame rate.
- the position of the moving body in each input image is detected.
- the position of the moving body for example, the center coordinates of the area where the moving body exists in each input image
- control unit 100 functions as a detection unit that detects the position of the moving body that moves within the wide-angle imaging range 195. Then, when the moving unit is detected in the wide-angle live view image 350, the control unit 100 determines that the moving body exists within the wide-angle imaging range 195. On the other hand, when a moving body is not detected in the wide-angle live view image 350, the control unit 100 determines that there is no moving body in the wide-angle imaging range 195.
- step S4 is executed again. That is, in step S ⁇ b> 4, the moving body detection process is executed every predetermined time until it is determined that the moving body exists in the wide-angle imaging range 195.
- step S5 is executed.
- the control unit 100 determines whether or not the moving object detected in step S4 is within the standard imaging range 185. Specifically, the control unit 100 detects the position of the moving body (for example, the center coordinates of the moving body) in the wide-angle live view image 350 detected in step S4 as the standard imaging range 185 in the wide-angle live view image 350. It is determined whether or not it is within the partial area corresponding to.
- the control unit 100 determines whether or not the position of the moving body in the wide-angle live view image 350 detected in step S4 is within a partial region in the wide-angle live view image 350 where the subject in the standard imaging range 185 is reflected. Determine. Then, when the position of the moving body in the wide-angle live view image 350 detected in step S4 is within the partial region corresponding to the standard imaging range 185 in the wide-angle live view image 350, the control unit 100 moves. It is determined that the body is within the standard imaging range 185. On the other hand, if the position of the moving body in the wide-angle live view image 350 detected in step S4 is not within the partial region corresponding to the standard imaging range 185 in the wide-angle live view image 350, the standard imaging is performed.
- control unit 100 functions as a determination unit that determines whether or not the moving object exists within the standard imaging range 185. Further, since it is determined in step S5 whether or not the moving body is present in the standard imaging range 185 for the moving body determined to be present in the wide-angle imaging range 195 in step S4, the control unit 100 Can also be said to function as a determination unit that determines whether or not the moving object exists outside the standard imaging range 185 and within the wide-angle imaging range 195.
- step S6 is executed.
- the control unit 100 estimates an approach area that passes when the moving body enters the standard imaging range 185 at the periphery of the standard imaging range 185 based on the position of the moving body detected in step S ⁇ b> 4. .
- a partial area corresponding to the standard imaging range 185 (a partial area in which the subject in the standard imaging range 185 is shown) in the wide-angle live view image 350 (an image in which the subject in the wide-angle imaging range 195 is shown).
- 351 is shown separated.
- the surrounding area other than the partial area 351 (the area corresponding to the outside of the standard imaging range and the inside of the wide-angle imaging range 195) in the wide-angle live view image 350 is divided into a plurality of areas. .
- the surroundings are represented by straight lines connecting the four vertices at the upper left, upper right, lower right and lower left of the wide-angle live view image 350 and the four vertices at the upper left, upper right, lower right and lower left of the partial area 351, respectively.
- This area is divided into an upper area 352, a lower area 353, a left area 354, and a right area 355.
- the upper edge 356a, the lower edge 356b, the left edge 356c, and the right edge 356d constituting the peripheral edge 356 of the partial area 351 are respectively an upper area 352, a lower area 353, a left area 354, and a right area 355 of the wide-angle live view image 350.
- the upper edge 356a, the lower edge 356b, the left edge 356c, and the right edge 356d of the partial area 351 correspond to the upper edge, the lower edge, the left edge, and the right edge that form the periphery of the standard imaging range 185, respectively.
- a moving body 500 for example, a train
- the right region 355 of the wide-angle live view image 350 is shown in the right region 355 of the wide-angle live view image 350.
- step S6 the control unit 100 determines in which region the moving body 500 detected in step S4 is located among the upper region 352, the lower region 353, the left region 354, and the right region 355 of the wide-angle live view image 350. Determine.
- the control unit 100 includes an area where the moving body 500 is determined to be located among the upper edge 356a, the lower edge 356b, the left edge 356c, and the right edge 356d of the partial area 351 in the wide-angle live view image 350. Identify the tangent edges.
- the moving body 500 performs standard imaging on an edge corresponding to the specified edge of the partial region 351 among the upper edge, the lower edge, the left edge, and the right edge constituting the periphery of the standard imaging range 185. It is estimated that the approach area passes when entering the range 185. In this way, the control unit 100 estimates the approach area that passes when the moving body 500 enters the standard imaging range 185 at the periphery of the standard imaging range 185 based on the detected position of the moving body 500. It functions as a part.
- the control unit 100 determines that the moving body 500 is located in the right region 355 of the wide-angle live view image 350. Then, the control unit 100 estimates the right edge of the standard imaging range 185 as an approach region that passes when the moving body 500 enters the standard imaging range 185.
- step S6 when the approach area through which the moving body 500 enters when entering the standard imaging range 185 is estimated, step S7 is executed.
- step S ⁇ b> 7 the display unit 121 displays first notification information for notifying the approach area estimated in step S ⁇ b> 6 on the display screen 2 a together with the standard live view image 300.
- FIG. 9 is a diagram illustrating an example of display on the display screen 2a displaying the first notification information.
- FIG. 9 shows an example of display on the display screen 2a when the right edge of the standard imaging range 185 is estimated to be the entry area.
- the first marker 360 as the first notification information is displayed on the end 420d so as to overlap the right end of the standard live view image 300.
- the first marker 360 is a bar-like figure extending along the vertical direction at the right end 420 d of the central region 420.
- the color of the first marker 360 is, for example, a color that can be easily distinguished from the standard live view image 300.
- the control unit 100 displays the wide-angle live view image 350. It is determined that the moving body 500 is located in the left area 354. Next, the control unit 100 estimates the left edge of the standard imaging range 185 as an approach area that passes when the moving object 500 enters the standard imaging range 185. Then, the display unit 121 displays the first notification information for notifying the estimated intrusion area together with the standard live view image 300 on the display screen 2a.
- FIG. 11 shows an example of the display of the display screen 2a displaying the first notification information when the left edge of the standard imaging range 185 is estimated as the approach area.
- the first marker 360 as the first notification information is displayed on the end 420c so as to overlap the left end of the standard live view image 300.
- the first marker 360 is a bar-shaped figure extending along the vertical direction at the left end 420 c of the central region 420.
- the control unit 100 determines that the moving object 510 is located in the upper area 352 of the live view image 350.
- the control unit 100 estimates the upper edge of the standard imaging range 185 as an approach area through which the moving body 510 passes when entering the standard imaging range 185.
- the display unit 121 displays the first notification information for notifying the estimated intrusion area together with the standard live view image 300 on the display screen 2a.
- FIG. 13 shows an example of display on the display screen 2a displaying the first notification information when the upper edge of the standard imaging range 185 is estimated as the approach area.
- the first marker 360 as the first notification information is displayed on the end 420 a so as to overlap the upper end of the standard live view image 300.
- the first marker 360 is a bar-shaped figure extending along the left-right direction at the upper end 420 a of the central region 420.
- the control unit 100 displays the wide-angle live view image 350. It is determined that the moving body 500 is located in the lower region 353 of the camera. Next, the control unit 100 estimates the lower edge of the standard imaging range 185 as an entry region that passes when the moving body 500 enters the standard imaging range 185. Then, the display unit 121 displays the first notification information for notifying the estimated intrusion area together with the standard live view image 300 on the display screen 2a.
- FIG. 15 shows an example of display on the display screen 2a that displays the first notification information when the lower edge of the standard imaging range 185 is estimated to be the approach area.
- the first marker 360 as the first notification information is displayed on the side end 420b so as to overlap the lower end of the standard live view image 300.
- the first marker 360 is a bar-shaped figure extending along the left-right direction at the lower end 420 b of the central region 420.
- the control unit 100 moves the entry area through which the moving body passes when entering the standard imaging range 185. Is estimated.
- the display unit 121 displays first notification information for notifying the estimated approach area on the display screen 2a together with the standard live view image 300.
- the user can determine whether the moving body exists outside the standard imaging range 185 and within the wide-angle imaging range 195 and from which area the moving body enters when entering the standard imaging range 185. I can know. Therefore, the user can easily image the moving body that enters the standard imaging range 185 by operating the operation button 310 while viewing the first notification information and the standard live view image 300.
- the display unit 121 displays first notification information on a portion of the display screen 2 a that displays the standard live view image 300 that corresponds to the approach area estimated to pass when the moving object enters the standard imaging range 185.
- the first marker 360 is displayed. Accordingly, the user can more intuitively know from which region of the standard imaging range 185 the moving body that enters the standard imaging range 185 from the wide-angle imaging range 195 enters.
- the first marker 360 is displayed superimposed on the end of the standard live view image 300, it is possible to suppress the standard live view image 300 from becoming difficult to see due to the first marker 360.
- the first marker 360 When the first marker 360 is displayed so as to be superimposed on the standard live view image 300, the first marker 360 is not a marker that makes the standard live view image 300 below it visible, but the standard below it. A marker that allows the live view image 300 to be seen through may be used.
- step S7 after the first notification information is displayed on the display screen 2a, the processing after step S4 is executed again.
- the display unit 121 displays the standard live view image 300 on the display screen 2a.
- the first marker 360 is continuously displayed on the right end 420d of the central area 420.
- FIG. 16 is a diagram illustrating an example of the wide-angle live view image 350 when the moving object 500 is located within the standard imaging range 185.
- the moving object 500 is located in a partial area 351 corresponding to the standard imaging range 185 in the wide-angle live view image 350.
- the control unit 100 determines that the moving body 500 exists in the standard imaging range 185.
- step S8 is executed.
- the display unit 121 displays second notification information indicating that the moving object 500 exists within the standard imaging range 185 on the display screen 2 a together with the standard live view image 300.
- FIG. 17 is a diagram illustrating an example of display on the display screen 2a displaying the second notification information.
- a frame-shaped second marker 370 that borders the peripheral edge of the central region 420 of the display screen 2a is displayed.
- the second marker 370 is displayed so as to overlap the peripheral end portion of the standard live view image 300.
- the color of the second marker 370 is, for example, a color that can be easily distinguished from the standard live view image 300.
- the second marker 370 is not a marker that makes the standard live view image 300 below it invisible, but the standard live view image below it. It may be a marker that allows the view image 300 to be seen through.
- the display unit 121 displays second notification information for notifying that the moving body 500 exists in the standard imaging range 185. It is displayed on the display screen 2a together with the standard live view image 300. Thereby, it becomes easy for the user to confirm that the moving object 500 exists in the standard imaging range 185. Further, when the electronic device 1 is operating in the still image capturing mode, the user operates the operation button 310 when visually recognizing the second notification information, thereby storing a still image on which the moving body 500 is reflected. 103 can be recorded. Therefore, the user can easily image the mobile object 500 at an appropriate timing when the mobile object 500 exists in the standard imaging range 185.
- step S8 after the second notification information is displayed, the processing after step S4 is executed again. Thereby, the display unit 121 continues to display the second notification information while the control unit 100 determines that the moving object 500 exists in the standard imaging range 185.
- FIG. 18 is a diagram illustrating an example of a wide-angle live view image 350 when the moving object 500 leaves the standard imaging range 185.
- the moving object 500 that moves to the left is located in the left region 354 of the wide-angle live view image 350.
- the control unit 100 determines that the moving object 500 does not exist within the standard imaging range 185.
- step S ⁇ b> 6 the control unit 100 determines that the moving object 500 is located in the left region 354 of the wide-angle live view image 350.
- the control unit 100 estimates the left edge of the standard imaging range 185 as an approach area that passes when the moving body 500 enters the standard imaging range 185.
- control unit 100 does not detect the moving direction of the moving body, but based on the detection result of the position of the moving body, the intrusion area that passes when the moving body enters the standard imaging range 185. Estimated. Therefore, as shown in FIG. 18, even when the moving body 500 is not moving toward the standard imaging range 185, it is assumed that the moving body 500 moves toward the standard imaging range 185 from the detected position. Estimate the region. Then, the display unit 121 displays the first notification information as illustrated in FIG. 11 on the display screen 2a together with the standard live view image 300.
- control unit 100 also detects the moving direction of the moving object, and estimates the approach area through which the moving object passes when entering the standard imaging range 185 based on the moving direction and position of the moving object. Good.
- the approach area can be estimated only for a moving body that moves from the wide-angle imaging range 195 toward the standard imaging range 185 among the moving bodies that move within the wide-angle imaging range 195.
- the operation of the electronic device 1 in this case will be described in detail in a later-described modification.
- step S ⁇ b> 4 shown in FIG. 5 the control unit 100 determines that the moving body 500 does not exist within the wide-angle imaging range 195. While the control unit 100 determines that the moving body 500 does not exist within the wide-angle imaging range 195, the display unit 121 does not display the first notification information and the second notification information on the display screen 2a.
- the display example of the first and second notification information when the standard live view image 300 is displayed on the display screen 2a and the wide-angle live view image 350 is not displayed has been described. As described above, even when the standard live view image 300 and the wide-angle live view image 350 are displayed on the display screen 2a, the first and second notification information are displayed.
- FIG. 19 is a diagram illustrating an example of display on the display screen 2a that displays the first notification information when the standard live view image 300 and the wide-angle live view image 350 are displayed together.
- first notification information for notifying that the right edge of the standard imaging range 185 is the entry area is displayed.
- the standard live view in the example of FIG. 19, in the area corresponding to the estimated approach area (the right edge of the standard imaging range 185) in the area around the standard live view image 300 on the display screen 2a, specifically, the standard live view.
- a first marker 360 as first notification information is displayed in a region 302d on the right side of the image 300.
- the first marker 360 is not displayed over the standard live view image 300 but is displayed outside the standard live view image 300.
- the display unit 121 displays the first notification information indicating the approach area estimated to pass when the moving body enters the standard imaging range 185 together with the standard live view together with the wide-angle live view image 350 that captures the moving body.
- the image 300 is displayed on the display screen 2a. Therefore, it becomes easier for the user to confirm the entry area through which the moving body passes when entering the standard imaging range 185 from the wide-angle imaging range 195.
- FIG. 20 is a diagram illustrating an example of display on the display screen 2a that displays the second notification information when the standard live view image 300 and the wide-angle live view image 350 are displayed together.
- a frame-shaped second marker 370 that surrounds the standard live view image 300 is displayed.
- the first notification information displayed on the display unit 121 may be other figures instead of the rod-shaped first marker 360.
- the first notification information may be an arrow-shaped graphic 361 displayed at the end of the standard live view image 300 as shown in FIG.
- the graphic 361 notifies from which area of the standard imaging range 185 the moving body enters depending on, for example, the displayed position and the direction of the arrow.
- a left-pointing arrow-shaped graphic 361 for notifying that the right edge of the standard imaging range 185 is the entry region is displayed so as to overlap the right end portion of the standard live view image 300.
- FIG. 22 is a diagram showing an example of display on the display screen 2a that displays the graphic 361 as the first notification information when the standard live view image 300 and the wide-angle live view image 350 are displayed together.
- a graphic 361 for notifying that the right edge of the standard imaging range 185 is the entry area is displayed.
- the graphic 361 as first notification information is displayed in the right area of the image 300.
- the graphic 361 is not displayed over the standard live view image 300 but is displayed outside the standard live view image 300.
- the first notification information may be characters indicating the estimated approach area.
- the second notification information is not a frame-shaped graphic that borders the peripheral edge of the standard live view image 300 or a frame-shaped graphic that surrounds the periphery of the standard live view image 300, but other graphics or characters. There may be.
- the first and second notification information may be displayed at a portion other than the end of the central region 420 or the periphery of the standard live view image 300. For example, a character as the first notification information or a character as the second notification information may be displayed so as to overlap the central portion of the standard live view image 300.
- FIG. 23 is a diagram illustrating an example of a wide-angle live view image 350 that reflects two moving bodies 500 and 510.
- the moving body 500 moving in the left direction is shown in the right region 355 of the wide-angle live view image 350.
- a moving body 510 that moves in the upper right direction is shown in the left region 354 of the wide-angle live view image 350.
- control unit 100 estimates the left edge of the standard imaging range 185 as an approach region that passes when the moving body 510 enters the standard imaging range 185. Then, the display unit 121 displays two pieces of first notification information for notifying the mobile bodies 500 and 510 of the estimated approach areas.
- FIG. 24 is a diagram illustrating an example of display on the display screen 2a that displays two pieces of first notification information.
- the first marker 360 that notifies that the right edge of the standard imaging range 185 is the entry area of the moving object 500 is displayed on the right end 420d of the central area 420 where the standard live view image 300 is displayed.
- the standard live view image 300 is displayed so as to overlap the right end portion.
- the first marker 362 for notifying that the left edge of the standard imaging range 185 is the entry area of the moving object 510 is displayed on the left end 420c of the central area 420 where the standard live view image 300 is displayed.
- the image 300 is displayed so as to overlap the left end portion.
- the first marker 360 and the first marker 362 are displayed so as to be distinguished from each other. For example, the first marker 360 and the first marker 362 are displayed in different colors.
- the standard imaging range 185 when the intrusion area estimated to pass when a plurality of moving objects respectively enter is the same part, a plurality of movements are made to the part corresponding to the part on the display screen 2a. A plurality of pieces of first notification information for the body may be displayed.
- the control unit 100 moves to the right area of the wide-angle live view image 350. It is determined that the bodies 500 and 510 are located. Next, the control unit 100 estimates the right edge of the standard imaging range 185 as an entry region that passes when the moving bodies 500 and 510 enter the standard imaging range 185, respectively. And the display part 121 displays the 1st notification information which notifies the approach area
- FIG. 26 shows an example of display on the display screen 2a displaying the first notification information when the right edge of the standard imaging range 185 is estimated to be the entry area of the moving bodies 500 and 510.
- a first marker 360 for notifying the entry region for the moving object 500 is displayed at the right end 420d of the central region 420 where the standard live view image 300 is displayed.
- a first marker 362 for notifying the entry area for the moving body 510 is displayed in an area 420e inside the right end 420d of the central area 420 where the standard live view image 300 is displayed on the display screen 2a.
- the approach area through which the moving object enters at the periphery of the standard imaging range 185 has been estimated from four parts obtained by dividing the periphery of the standard imaging range 185 into four. However, it may be estimated from a portion obtained by dividing the periphery of the standard imaging range 185 into more than four.
- FIG. 27 shows an example of a wide-angle live view image 350 that shows a region other than the partial region 351 corresponding to the standard imaging range 185 (region corresponding to the outside of the standard imaging range 185 and the inside of the wide-angle imaging range 195) divided into eight.
- FIG. 27 each of the upper area 352, the lower area 353, the left area 354, and the right area 355 of the wide-angle live view image 350 shown in FIG. 8 is further divided into two areas in the circumferential direction. Yes.
- the upper area 352, the lower area 353, the left area 354, and the right area 355 of the wide-angle live view image 350 include, for example, the midpoints of the upper edge, the lower edge, the left edge, and the right edge of the wide-angle live view image 350
- the region 351 is divided into two regions by straight lines connecting the middle points of the upper edge 356a, the lower edge 356b, the left edge 356c, and the right edge 356d, respectively.
- the upper area 352 and the lower area 353 of the wide-angle live view image 350 are divided into two areas by, for example, straight lines passing through the middle point of the upper edge 356a and the middle point of the lower edge 356b, respectively.
- the left area 354 and the right area 355 of the wide-angle live view image 350 are divided into two areas by a straight line passing through the middle point of the left edge 356c of the partial area 351 and the middle point of the right edge 356d, respectively. It may be.
- the regions other than the partial region 351 in the wide-angle live view image 350 are the upper left region 352a, the upper right region 352b, the lower left region 353a, the lower right region 353b, the upper left region 354a, and the lower left region. It is divided into eight areas, 354b, upper right area 355a, and lower right area 355b.
- the lower right edge portion 356db are the upper left area 352a, upper right area 352b, lower left area 353a, lower right area 353b, upper left area 354a, lower left area 354b, upper right area of the wide-angle live view image 350, respectively. It is in contact with the region 355a and the lower right region 355b.
- the lower part 356db is an upper edge left part, an upper edge right part, a lower edge left part, a lower edge right part, a left edge upper part, a left edge lower part, and a right edge upper part that constitute the periphery of the standard imaging range 185, respectively.
- the lower right edge part In the example of FIG. 27, a moving body 500 that moves toward the left is shown in the lower right region 355b of the wide-angle live view image 350.
- step S6 shown in FIG. 5 the control unit 100 detects that the moving body detected in step S4 is the upper left region 352a, upper right region 352b, lower left region 353a, and lower right region 353b of the wide-angle live view image 350.
- the left upper area 354a, the lower left area 354b, the upper right area 355a, and the lower right area 355b are determined to be located.
- the control unit 100 controls the upper left portion 356aa, the upper right portion 356ab, the lower left portion 356ba, the lower right portion 356bb, the left upper portion 356ca, the left edge of the partial region 351 in the wide-angle live view image 350.
- the control part 100 is the upper edge left part which comprises the periphery of the standard imaging range 185, an upper edge right part, a lower edge left part, a lower edge right part, a left edge upper part, a left edge lower part, a right edge upper part.
- a portion corresponding to the specified portion of the partial region 351 is estimated as an approach region that passes when the moving body enters the standard imaging range 185.
- the control unit 100 determines that the moving body 500 is located in the lower right region 355b of the wide-angle live view image 350. Then, when the moving body 500 enters the lower right edge portion of the standard imaging range 185 corresponding to the lower right edge portion 356db of the partial region 351 of the wide-angle live view image 350, the control unit 100 enters the standard imaging range 185. It is estimated that it is an approach area that passes through. And the display part 121 displays the 1st notification information which shows the estimated approach area
- FIG. 28 shows an example of display on the display screen 2a when the lower right portion of the standard imaging range 185 is estimated as the entry area.
- the portion corresponding to the lower right edge portion of the standard imaging range 185 in the display screen 2a specifically, the lower portion of the right end portion of the central region 420 displaying the standard live view image 300.
- the first marker 360 as the first notification information is displayed at 420 f so as to overlap the lower part of the right end portion of the standard live view image 300.
- the approach area that passes when the moving body enters is estimated, and the first notification information indicating the estimated approach area is displayed on the display screen 2a. Therefore, compared with the case where the approach area is estimated from a portion obtained by dividing the peripheral edge of the standard imaging range 185 into four parts, the user moves from the wide-angle imaging range 195 to the standard imaging range 185. It is possible to know more accurately from which area of the standard imaging range 185 500 enters.
- the number of divisions and the division method of the peripheral edge of the standard imaging range 185 when estimating the approach area through which the moving body enters the standard imaging range 185 are not limited to the above example.
- FIG. 29 is a diagram illustrating an example of a wide-angle live view image 350 in which a moving body 520 that is a person is reflected.
- the moving body 520 person moving in the left direction is located in the right region 355 of the wide-angle live view image 350.
- FIG. 30 is a diagram illustrating an example of a wide-angle live view image 350 in which a moving body 530 that is a dog is reflected.
- the moving body 530 dog that moves in the left direction is located in the right region 355 of the wide-angle live view image 350.
- the control unit 100 determines that the moving body 520 is located in the right region 355 of the wide-angle live view image 350, and estimates the right edge of the standard imaging range 185 as the approach region. Then, as shown in FIG. 9, the display unit 121 displays a first marker 360 as first notification information on the right end 420d of the central region 420 where the standard live view image 300 is displayed on the display screen 2a. The standard live view image 300 is displayed so as to overlap the right end portion. Since the process for the moving body 530 is the same as the process for the moving body 520, detailed description thereof is omitted.
- the control unit 100 does not detect the moving direction of the moving body, and the moving body enters the standard imaging range 185 at the periphery of the standard imaging range 185 based on the detection result of the position of the moving body.
- the approach area that passes when coming is estimated.
- the control unit 100 detects the moving direction of the moving body in addition to the position of the moving body. Then, based on the detected position and moving direction of the moving body, the control unit 100 estimates an approach area that passes when the moving body enters the standard imaging range 185 at the periphery of the standard imaging range 185.
- FIG. 31 is a flowchart showing an example of the operation of the electronic apparatus 1 according to this modification.
- the processing in steps S11 to S13 is the same as steps S1 to S3 shown in FIG.
- the control unit 100 detects, for example, moving object detection using a difference between frames for a series of input images continuously input from the wide-angle camera 190 at a predetermined frame rate. By performing this image processing, the position and moving direction of the moving body in each input image are detected.
- a wide-angle live view image 350 is used.
- the control unit 100 functions as a detection unit that detects the position and moving direction of a moving body that moves within the wide-angle imaging range 195. Then, when the moving unit is detected in the wide-angle live view image 350, the control unit 100 determines that the moving body exists within the wide-angle imaging range 195. On the other hand, when a moving body is not detected in the wide-angle live view image 350, the control unit 100 determines that there is no moving body in the wide-angle imaging range 195.
- step S14 If it is determined in step S14 that the moving body does not exist within the wide-angle imaging range 195, step S14 is executed again. On the other hand, if it is determined in step S14 that the moving body is within the wide-angle imaging range 195, step S15 is executed.
- step S18 is executed.
- the display unit 121 displays, on the display screen 2 a, the second notification information indicating that the moving body exists within the standard imaging range 185 as illustrated in FIG. 17 together with the standard live view image 300.
- step S16 is executed.
- the control unit 100 based on the position and moving direction of the moving body detected in step S14, enters the entrance area through which the moving body enters the standard imaging range 185 at the periphery of the standard imaging range 185. Is estimated. Specifically, the control unit 100 passes through any part of the periphery of the partial area 351 of the wide-angle live view image 350 when the moving body travels straight from the detected position along the detected moving direction. To determine whether the image enters the standard imaging range 185.
- the control unit 100 estimates the approach area at the periphery of the standard imaging range 185 based on the position and moving direction of the moving body will be described using the wide-angle live view image 350 shown in FIG.
- a moving body 500 that moves toward the left is shown in the right region 355 of the wide-angle live view image 350.
- a moving body 510 that moves in the upper right direction is shown in the left region 354 of the wide-angle live view image 350.
- step S14 the control unit 100 detects the position and moving direction 500a of the moving body 500 in the wide-angle live view image 350.
- step S ⁇ b> 16 the control unit 100 passes through the right edge 356 d of the partial area 351 and enters the partial area 351 when moving straight along the movement direction 500 a from the position where the moving body 500 is detected. Judge that it will come. Then, the control unit 100 estimates a portion corresponding to the right edge 356d of the partial region 351 in the standard imaging range 185 as an approach region that passes when the moving body 500 enters the standard imaging range 185.
- step S14 the control unit 100 detects the position and moving direction 510a of the moving body 510 in the wide-angle live view image 350.
- Step S ⁇ b> 16 the control unit 100 determines that the moving body 510 does not pass through the peripheral edge of the partial region 351 when it moves straight along the moving direction 510 a.
- the control unit 100 does not specify the entry region.
- the approach area is not estimated depending on the detected moving direction of the moving body. .
- step S17 the 1st notification information which shows the approach area
- FIG. When the wide-angle live view image 350 as shown in FIG. 32 is obtained, the first communication information for the moving body 510 is not displayed because the approach area is not estimated for the moving body 510.
- the control unit 100 determines the standard imaging range based on the detected position and moving direction of the moving body. An approach area through which a moving body moving toward 185 passes when entering the standard imaging range 185 is estimated. And the control part 100 displays the 1st notification information which shows the estimated approach area
- FIG. 33 is a flowchart illustrating an example of the operation of the electronic apparatus 1 including the normal imaging mode and the moving object detection mode.
- step S22 the control unit 100 supplies power to, for example, the standard camera 180 among the standard camera 180, the wide-angle camera 190, and the in-camera 200. Supply and start. That is, when the camera application is activated, the electronic device 1 operates in the normal imaging mode.
- step S23 the display unit 121 displays the standard live view image 300 on the display screen 2a.
- FIG. 34 is a diagram illustrating an example of display on the display screen 2 a that displays the standard live view image 300 when the electronic apparatus 1 operates in the normal imaging mode. As shown in FIG. 34, since the wide-angle camera 190 is not activated, the display switching button 340 shown in FIG. 6 is not displayed.
- a moving body detection switching button 380 for switching the operation mode of the electronic device 1 between the normal imaging mode and the moving body detection mode is displayed on the lower end portion 410 of the display screen 2a.
- the moving object detection switching button 380 is displayed only when the recording camera is the standard camera 180.
- step S ⁇ b> 24 when the operation mode of the electronic device 1 is the normal imaging mode, when the touch panel 130 detects a predetermined operation (for example, a tap operation) on the moving object detection switching button 380, the control unit 100 displays the electronic device 1. Is switched from the normal imaging mode to the moving object detection mode.
- the control unit 100 supplies power to the wide-angle camera 190 and activates the wide-angle camera 190 in step S25. Then, the control unit 100 starts the moving object detection process shown in steps S26 to S30. Since the series of processes in steps S26 to S30 is the same as the series of processes in steps S4 to S8 in FIG.
- the control unit 100 changes the operation mode of the electronic device 1 to the moving body. Switch from the detection mode to the normal imaging mode.
- the control unit 100 stops the power supply to the wide angle camera 190 and stops the operation of the wide angle camera 190. And the control part 100 stops the detection process of a moving body.
- the recording camera is the standard camera 180
- the wide-angle camera 190 is activated and the moving object is detected only when an operation by the user for operating the electronic apparatus 1 in the moving object detection mode is detected. Since the processing is performed, the power consumption of the electronic device 1 can be reduced.
- the control unit 100 performs the position or position and movement direction detection processing for all the detected moving objects, and performs the approach region estimation processing.
- the control unit 100 performs these processes only on a target mobile body (hereinafter also referred to as a target mobile body).
- a target mobile body for example, the process is performed only on a specific moving body (for example, a specific person) or on a specific type of moving body (for example, all of a plurality of moving bodies detected as a person).
- FIG. 35 is a flowchart showing an example of the operation of the electronic apparatus 1 according to this modification.
- the processing in steps S31 to S33 is the same as the processing in steps S1 to S3 shown in FIG.
- step S34 the control unit 100 performs image processing such as template matching on a series of input images continuously input from the wide-angle camera 190 at a predetermined frame rate.
- image processing such as template matching on a series of input images continuously input from the wide-angle camera 190 at a predetermined frame rate.
- the position of the target moving body in each input image is detected.
- the target moving body is a person, for example, a known face authentication technique or the like is used.
- the target moving body is set in advance by the user, and information indicating the target moving body is stored in the storage unit 103.
- a reference image for detecting the target moving body is captured in advance by the standard camera 180 or the like and stored in the nonvolatile memory in the storage unit 103.
- a wide-angle live view image 350 is used for the target moving object detection process.
- the control unit 100 detects the position of the target moving body by detecting the position of the partial area corresponding to the reference image indicating the target moving body in the wide-angle live view image 350.
- the control unit 100 functions as a detection unit that detects the position of the target moving body located within the wide-angle imaging range 195. Then, when the target moving body is detected in the wide-angle live view image 350, the control unit 100 determines that the target moving body exists within the wide-angle imaging range 195. On the other hand, when the target moving body is not detected in the wide-angle live view image 350, the control unit 100 determines that there is no target moving body in the wide-angle imaging range 195.
- step S34 If it is determined in step S34 that the target moving body is within the standard imaging range 185, step S34 is executed again. On the other hand, if it is determined in step S34 that the target moving body exists within the wide-angle imaging range 195, step S35 is executed.
- step S35 the control unit 100 determines whether or not the target moving body detected in step S34 is within the standard imaging range 185. Specifically, the control unit 100 detects that the position of the target moving body (for example, the center coordinates of the target moving body) in the wide-angle live view image 350 detected in step S34 is the partial region 351 in the wide-angle live view image 350. It is determined whether it is within. Then, when the position of the target moving body in the wide-angle live view image 350 detected in step S34 is within the partial area 351 in the wide-angle live view image 350, the control unit 100 determines that the target moving body is the standard imaging range. 185 is determined to exist.
- the position of the target moving body for example, the center coordinates of the target moving body
- the control unit 100 determines that the target moving body is the standard imaging range. 185 is determined to exist.
- control unit 100 moves the target within the standard imaging range 185. It is determined that the body does not exist. As described above, the control unit 100 functions as a determination unit that determines whether or not the target moving object exists within the standard imaging range 185.
- control unit 100 can also be said to function as a determination unit that determines whether or not the target moving body exists outside the standard imaging range 185 and within the wide-angle imaging range 195.
- step S35 If it is determined in step S35 that the target moving body does not exist within the standard imaging range 185, that is, the target moving body exists outside the standard imaging range 185 and within the wide-angle imaging range 195, step S36 is executed.
- step S ⁇ b> 36 the control unit 100 determines an approach region that passes when the target moving body enters the standard imaging range 185 at the periphery of the standard imaging range 185 based on the position of the target moving body detected in step S ⁇ b> 34. presume.
- the control unit 100 estimates the approach area at the periphery of the standard imaging range 185 with respect to the target moving body will be described using the wide-angle live view image 350 illustrated in FIG.
- the standard imaging range 185 is narrower than when the zoom magnification of the standard camera 180 is 1 by the zoom-in function of the standard camera 180.
- the range of the partial region 351 shown in FIG. 36 is narrower than the partial region 351 shown in FIG.
- the zoom magnification of the standard camera 180 may be 1 ⁇ .
- a moving body 520 and a moving body 521 that move toward the left are shown in the right area 355 of the wide-angle live view image 350.
- the moving body 520 and the moving body 521 are people.
- the face of the moving body 520 is set as the target moving body.
- a partial region 357 in which the face of the moving body 520 appears in the wide-angle live view image 350 is detected as a portion corresponding to the target moving body.
- step S36 the control unit 100 determines that the face of the moving body 520 that is the target moving body is the right region 355 of the wide-angle live view image 350. It is determined that it is located at. Then, the control unit 100 estimates the right edge of the standard imaging range 185 as an entry region that passes when the face of the moving object 520 enters the standard imaging range 185. On the other hand, since the moving body 521 is not a target moving body, the position detection process and the approach area estimation process are not performed on the moving body 521.
- step S36 when the approach area through which the target moving body enters when entering the standard imaging range 185 is estimated, step S37 is executed.
- step S37 the display unit 121 displays the display screen 2a as shown in FIG.
- the first marker 360 as the first notification information indicating the entry area with respect to the target moving body is displayed on the right end 420d of the central area 420 where the standard live view image 300 is displayed on the display screen 2a. Is done.
- the 1st notification information with respect to the mobile body 521 is not displayed on the display screen 2a.
- control unit 100 estimates an approach area that passes when entering the standard imaging range 185 with respect to a target moving body among a plurality of moving bodies, and indicates the estimated approach area.
- the first notification information is displayed on the display screen 2a. This makes it easier for the user to image the target moving body.
- FIG. 38 is a diagram illustrating an example of the wide-angle live view image 350 when the moving body 521 is positioned within the standard imaging range 185 and the moving body 520 is positioned within the right region 355 of the wide-angle live view image 350.
- the moving body 521 is located in the partial area 351 corresponding to the standard imaging range 185 in the wide-angle live view image 350.
- the display unit 121 displays the display screen 2a as shown in FIG.
- the moving object 521 is reflected in the standard live view image 300.
- the second notification information for the moving body 521 is not displayed on the display screen 2a.
- the face of the moving body 520 that is the target moving body remains located in the right region 355 of the wide-angle live view image 350, and therefore, as the first notification information for the face of the moving body 520
- the first marker 360 remains displayed on the right end 420d of the central area 420 where the standard live view image 300 is displayed on the display screen 2a.
- FIG. 40 is a diagram illustrating an example of the wide-angle live view image 350 when the moving body 520 is located within the standard imaging range 185 and the moving body 521 is located within the left region 354 of the wide-angle live view image 350. .
- the moving body 520 is located in a partial region 351 corresponding to the standard imaging range 185 in the wide-angle live view image 350.
- the control unit 100 determines that the target moving body (the face of the moving body 520) is present in the standard imaging range 185.
- step S38 is executed.
- the display unit 121 displays the display screen 2a as shown in FIG.
- the second marker 370 as the second notification information indicating that the target moving body exists within the standard imaging range 185 is displayed so as to border the peripheral edge of the central region 420 of the display screen 2a. Is done.
- the display unit 121 displays a third marker 390 for identifying the target moving body in a portion corresponding to the partial region 357 in the display screen 2a.
- the display unit 121 Second notification information for notifying that the moving body is present within the standard imaging range 185 is displayed on the display screen 2 a together with the standard live view image 300. This makes it easier for the user to image the target moving body.
- the control unit 100 determines that the target moving body exists in the standard imaging range 185, the focus of the standard camera 180 is set to the moving body. You may make it match. This makes it easier for the user to image the target moving body.
- the display unit 121 when the control unit 100 determines that the moving body is in the standard imaging range 185, the display unit 121 notifies that the moving body exists in the standard imaging range 185. Although the notification information is displayed on the display screen 2a together with the standard live view image 300, the display unit 121 displays the second notification information even when it is determined that the moving object is present in the standard imaging range 185. You don't have to. Even when the display unit 121 does not display the second notification information when it is determined that the moving body is within the standard imaging range 185, the user can move the moving body within the standard imaging range 185 as described above.
- the moving body exists outside the standard imaging range 185 and within the wide-angle imaging range 195, and the moving body enters the standard imaging range 185. You can know from which area you will come in. Even when the second notification information is not displayed, the user visually recognizes that the moving body is reflected in the standard live view image 300, so that the moving body is within the standard imaging range 185. I can confirm that. Therefore, even when the display unit 121 does not display the second notification information, the user can easily image the moving body using the first notification information.
- the technology of the present disclosure is applied to a mobile phone such as a smartphone has been described as an example.
- the technology of the present disclosure may be applied to other electronic devices including a plurality of imaging units having different angles of view. It can also be applied to equipment.
- the technology of the present disclosure can be applied to an electronic device such as a digital camera, a personal computer, or a tablet terminal.
- the electronic device 1 has been described in detail, but the above description is an example in all aspects, and the disclosure is not limited thereto.
- the various modifications described above can be applied in combination as long as they do not contradict each other. And it is understood that the countless modification which is not illustrated can be assumed without deviating from the scope of this disclosure.
Abstract
Description
図1及び図2は、それぞれ、電子機器1の外観の一例を概略的に示す斜視図及び裏面図である。電子機器1は、例えば、スマートフォン等の携帯電話機である。電子機器1は、基地局及びサーバ等を通じて他の通信装置と通信することが可能である。
図3は、電子機器1の電気的構成の一例を示すブロック図である。図3に示されるように、電子機器1には、制御部100、無線通信部110、表示部121、タッチパネル130、操作キー群140、マイク150、レシーバ160、外部スピーカ170、第1撮像部180、第2撮像部190、第3撮像部200、及び電池210が設けられている。電子機器1に設けられたこれらの構成要素のそれぞれは、機器ケース3内に収められている。
図5は、カメラアプリが実行される際の電子機器1の動作の一例を示すフローチャートである。表示画面2aに対して所定の操作が行われると、図5に示されるように、ステップS1において、制御部100は、記憶部103内に記憶されたカメラアプリを実行(起動)する。表示画面2aには、例えば、電子機器1が各種アプリを実行する前の初期状態において、ホーム画面(初期画面)が表示されている。また、ホーム画面には、各種アプリをそれぞれ実行するための複数の図形(以降、アプリ実行用図形と呼ぶことがある)が表示されている。アプリ実行用図形は、アイコンと呼ばれる図形を含んでいてもよい。制御部100は、表示画面2aに表示されるカメラアプリを実行するためのアプリ実行用図形に対するユーザの選択操作をタッチパネル130が検出すると、記憶部103内に記憶されたカメラアプリを実行する。
以下に各種変形例について説明する。
上記の例では、制御部100は、移動体の移動方向を検出せずに移動体の位置の検出結果に基づいて、標準撮像範囲185の周縁における、当該移動体が標準撮像範囲185に入ってくる際に通過する進入領域を推定している。本変形例では、制御部100は、移動体の位置に加えて移動体の移動方向を検出する。そして、制御部100は、検出した移動体の位置及び移動方向に基づいて、標準撮像範囲185の周縁における、当該移動体が標準撮像範囲185に入ってくる際に通過する進入領域を推定する。
上記の各例では、記録用カメラが標準カメラ180である場合には、制御部100は、常に広角カメラ190を動作させて移動体の検出処理を行っていた。これに対して、本変形例では、電子機器1は、記録用カメラが標準カメラ180である場合であっても広角カメラ190を動作させずに移動体の検出処理を行わない通常撮像モードと、記録用カメラが標準カメラ180である場合に広角カメラ190を動作させて移動体の検出処理を行う移動体検出モードとを備えている。図33は、通常撮像モード及び移動体検出モードを備える電子機器1の動作の一例を示すフローチャートである。
上記の各例では、制御部100は、検出された移動体全てに対して、位置あるいは位置及び移動方向の検出処理を行い、進入領域の推定処理を行っていた。これに対して、本変形例では、制御部100は、対象となる移動体(以降、対象移動体とも呼ぶことがある)に対してのみこれらの処理を行う。例えば、特定の移動体(例えば特定の人物)に対して、あるいは、特定の種類の移動体(例えば人物として検出される複数の移動体の全て)に対してのみ処理が行われる。
2a 表示画面
100 制御部
120 表示パネル
121 表示部
180 第1撮像部(標準カメラ)
185 第1撮像範囲(標準撮像範囲)
190 第2撮像部(広角カメラ)
195 第2撮像範囲(広角撮像範囲)
300 標準ライブビュー画像
350 広角ライブビュー画像
360,362 第1マーカー
370 第2マーカー
500,510,520,521,530 移動体
Claims (8)
- 第1撮像範囲を撮像する第1撮像部と、
前記第1撮像部が前記第1撮像範囲を撮像している場合に、前記第1撮像範囲より広角な第2撮像範囲を撮像する第2撮像部と、
表示画面を有し、前記第1撮像部が撮像する第1ライブビュー画像を前記表示画面に表示する表示部と、
前記第2撮像部からの画像信号に基づいて、前記第2撮像範囲内を移動する移動体の位置を検出する検出部と、
前記検出部が検出する前記位置に基づいて、前記移動体が前記第1撮像範囲外かつ前記第2撮像範囲内に存在するか否かを判定する判定部と、
前記判定部において前記移動体が前記第1撮像範囲外かつ前記第2撮像範囲内に存在すると判定される場合には、前記検出部が検出する前記位置に基づいて、前記第1撮像範囲の周縁における、前記移動体が前記第1撮像範囲に入ってくる際に通過する進入領域を推定する推定部と
を備え、
前記表示部は、前記進入領域を通知する第1通知情報を前記第1ライブビュー画像と共に前記表示画面に表示する、電子機器。 - 請求項1に記載の電子機器であって、
前記検出部は、前記第2撮像部からの画像信号に基づいて、前記第2撮像範囲内を移動する前記移動体の移動方向を検出し、
前記推定部は、前記判定部において前記移動体が前記第1撮像範囲外かつ前記第2撮像範囲内に存在すると判定される場合には、前記検出部が検出する前記位置及び前記移動方向に基づいて、前記進入領域を推定する、電子機器。 - 請求項1及び請求項2のいずれか一つに記載の電子機器であって、
前記表示部は、前記第1ライブビュー画像を表示する前記表示画面における、前記進入領域と対応する部分に、前記第1通知情報としての第1マーカーを表示する、電子機器。 - 請求項1乃至請求項3のいずれか一つに記載の電子機器であって、
前記判定部は、前記検出部が検出する前記位置に基づいて、前記移動体が前記第1撮像範囲内に存在するか否かを判定し、
前記表示部は、前記判定部において前記移動体が前記第1撮像範囲内に存在すると判定される場合には、前記移動体が前記第1撮像範囲内に存在することを通知する第2通知情報を前記第1ライブビュー画像と共に前記表示画面に表示する、電子機器。 - 請求項4に記載の電子機器であって、
前記表示部は、前記第1ライブビュー画像を表示する前記表示画面における、前記第1撮像範囲の周縁に対応する部分を縁取る第2マーカーを前記第2通知情報として表示する、電子機器。 - 請求項1乃至請求項5のいずれか一つに記載の電子機器であって、
前記表示部は、前記第2撮像部が撮像する第2ライブビュー画像を、前記第1ライブビュー画像に並べて前記表示画面に表示する、電子機器。 - 第1撮像範囲を撮像する第1撮像部と、前記第1撮像部が前記第1撮像範囲を撮像している場合に、前記第1撮像範囲より広角な第2撮像範囲を撮像する第2撮像部とを備える電子機器の動作方法であって、
前記第2撮像部からの画像信号に基づいて、前記第2撮像範囲内を移動する移動体の位置を検出する第1工程と、
前記位置に基づいて、前記移動体が前記第1撮像範囲外かつ前記第2撮像範囲内に存在するか否かを判定する第2工程と、
前記第2工程において前記移動体が前記第1撮像範囲外かつ前記第2撮像範囲内に存在すると判定される場合には、前記位置に基づいて、前記第1撮像範囲の周縁における、前記移動体が前記第1撮像範囲に入ってくる際に通過する進入領域を推定する第3工程と、
前記進入領域を通知する通知情報を、前記第1撮像部が撮像するライブビュー画像と共に表示する第4工程とを備える、電子機器の動作方法。 - 第1撮像範囲を撮像する第1撮像部と、前記第1撮像部が前記第1撮像範囲を撮像している場合に、前記第1撮像範囲より広角な第2撮像範囲を撮像する第2撮像部とを備える電子機器を制御するための制御プログラムであって、
前記電子機器に、
前記第2撮像部からの画像信号に基づいて、前記第2撮像範囲内を移動する移動体の位置を検出する第1工程と、
前記位置に基づいて、前記移動体が前記第1撮像範囲外かつ前記第2撮像範囲内に存在するか否かを判定する第2工程と、
前記第2工程において前記移動体が前記第1撮像範囲外かつ前記第2撮像範囲内に存在すると判定される場合には、前記位置に基づいて、前記第1撮像範囲の周縁における、前記移動体が前記第1撮像範囲に入ってくる際に通過する進入領域を推定する第3工程と、
前記進入領域を通知する通知情報を、前記第1撮像部が撮像するライブビュー画像と共に表示する第4工程と
を実行させるための制御プログラム。
Priority Applications (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2017531052A JPWO2017018043A1 (ja) | 2015-07-29 | 2016-05-26 | 電子機器、電子機器の動作方法及び制御プログラム |
US15/747,378 US20180220066A1 (en) | 2015-07-29 | 2016-05-26 | Electronic apparatus, operating method of electronic apparatus, and non-transitory computer-readable recording medium |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2015-149682 | 2015-07-29 | ||
JP2015149682 | 2015-07-29 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2017018043A1 true WO2017018043A1 (ja) | 2017-02-02 |
Family
ID=57885544
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/JP2016/065525 WO2017018043A1 (ja) | 2015-07-29 | 2016-05-26 | 電子機器、電子機器の動作方法及び制御プログラム |
Country Status (3)
Country | Link |
---|---|
US (1) | US20180220066A1 (ja) |
JP (1) | JPWO2017018043A1 (ja) |
WO (1) | WO2017018043A1 (ja) |
Families Citing this family (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP3496386A4 (en) * | 2016-11-24 | 2019-07-03 | Huawei Technologies Co., Ltd. | METHOD AND GUIDE DEVICE FOR PHOTOGRAPHY COMPOSITION |
US11468174B2 (en) * | 2017-08-11 | 2022-10-11 | Eys3D Microelectronics Co. | Surveillance camera system and related surveillance system thereof |
EP4024843A4 (en) * | 2019-08-29 | 2022-08-31 | SZ DJI Technology Co., Ltd. | DISPLAY METHOD, PHOTOGRAPHING METHOD AND RELATED DEVICE |
CN111698428B (zh) * | 2020-06-23 | 2021-07-16 | 广东小天才科技有限公司 | 一种文档拍摄的方法、装置、电子设备和存储介质 |
US20220109822A1 (en) * | 2020-10-02 | 2022-04-07 | Facebook Technologies, Llc | Multi-sensor camera systems, devices, and methods for providing image pan, tilt, and zoom functionality |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2008136024A (ja) * | 2006-11-29 | 2008-06-12 | Fujifilm Corp | 撮影装置、撮影システムおよび撮影方法 |
JP2010118984A (ja) * | 2008-11-14 | 2010-05-27 | Nikon Corp | 撮影装置 |
JP2012029245A (ja) * | 2010-07-27 | 2012-02-09 | Sanyo Electric Co Ltd | 撮像装置 |
JP2012042805A (ja) * | 2010-08-20 | 2012-03-01 | Olympus Imaging Corp | 撮像装置 |
Family Cites Families (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP4140567B2 (ja) * | 2004-07-14 | 2008-08-27 | 松下電器産業株式会社 | 物体追跡装置および物体追跡方法 |
RU2486707C1 (ru) * | 2009-03-17 | 2013-06-27 | Хуавэй Текнолоджиз Ко., Лтд. | Способ, устройство и система для распределения мощности нисходящей линии связи |
US8237771B2 (en) * | 2009-03-26 | 2012-08-07 | Eastman Kodak Company | Automated videography based communications |
JP2012049651A (ja) * | 2010-08-24 | 2012-03-08 | Ricoh Co Ltd | 撮像装置及び撮像方法 |
JP6004371B2 (ja) * | 2011-11-16 | 2016-10-05 | パナソニックIpマネジメント株式会社 | 撮像装置 |
-
2016
- 2016-05-26 US US15/747,378 patent/US20180220066A1/en not_active Abandoned
- 2016-05-26 WO PCT/JP2016/065525 patent/WO2017018043A1/ja active Application Filing
- 2016-05-26 JP JP2017531052A patent/JPWO2017018043A1/ja active Pending
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2008136024A (ja) * | 2006-11-29 | 2008-06-12 | Fujifilm Corp | 撮影装置、撮影システムおよび撮影方法 |
JP2010118984A (ja) * | 2008-11-14 | 2010-05-27 | Nikon Corp | 撮影装置 |
JP2012029245A (ja) * | 2010-07-27 | 2012-02-09 | Sanyo Electric Co Ltd | 撮像装置 |
JP2012042805A (ja) * | 2010-08-20 | 2012-03-01 | Olympus Imaging Corp | 撮像装置 |
Also Published As
Publication number | Publication date |
---|---|
US20180220066A1 (en) | 2018-08-02 |
JPWO2017018043A1 (ja) | 2018-04-12 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US9674395B2 (en) | Methods and apparatuses for generating photograph | |
WO2017018043A1 (ja) | 電子機器、電子機器の動作方法及び制御プログラム | |
US10452890B2 (en) | Fingerprint template input method, device and medium | |
JP6205073B2 (ja) | 撮像装置、撮像方法、及びプログラム | |
RU2612892C2 (ru) | Способ автоматической фокусировки и устройство автоматической фокусировки | |
US8319883B2 (en) | Image pickup apparatus and control method thereof | |
US10602063B2 (en) | Electronic apparatus, electronic apparatus operating method, and non-transitory computer readable recording medium | |
KR102018887B1 (ko) | 신체 부위 검출을 이용한 이미지 프리뷰 | |
US10270975B2 (en) | Preview image display method, apparatus and storage medium | |
WO2016038976A1 (ja) | マルチ撮像装置、マルチ撮像方法、プログラム、及び記録媒体 | |
CN110992327A (zh) | 镜头脏污状态的检测方法、装置、终端及存储介质 | |
JP2017518658A5 (ja) | ||
EP3381180B1 (en) | Photographing device and method of controlling the same | |
CN108900903B (zh) | 视频处理方法及装置、电子设备和存储介质 | |
JP6147829B2 (ja) | 電子機器および電子機器の録画制御方法 | |
JP6266577B2 (ja) | 電子機器 | |
WO2015072166A1 (ja) | 撮像装置、撮像アシスト方法及び撮像アシストプログラムを記録した記録媒体 | |
JP2017069618A (ja) | 電子機器及び撮像方法 | |
US11265529B2 (en) | Method and apparatus for controlling image display | |
JP6374535B2 (ja) | 操作装置、追尾システム、操作方法、及びプログラム | |
JP5880459B2 (ja) | 電子機器、表示制御方法及びプログラム | |
JP6679430B2 (ja) | 撮像装置、撮像装置の制御方法及びプログラム | |
CN111225202B (zh) | 一种图片故障诊断方法、装置及系统 | |
WO2015052975A1 (ja) | 情報処理装置、撮像装置、撮像システム、情報処理装置の制御方法、撮像装置の制御方法およびプログラム | |
JP6149952B2 (ja) | 電子機器、表示制御方法及びプログラム |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 16830148 Country of ref document: EP Kind code of ref document: A1 |
|
ENP | Entry into the national phase |
Ref document number: 2017531052 Country of ref document: JP Kind code of ref document: A |
|
WWE | Wipo information: entry into national phase |
Ref document number: 15747378 Country of ref document: US |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 16830148 Country of ref document: EP Kind code of ref document: A1 |