US20130002922A1 - Imaging apparatus and computer readable recording medium - Google Patents

Imaging apparatus and computer readable recording medium Download PDF

Info

Publication number
US20130002922A1
US20130002922A1 US13/610,277 US201213610277A US2013002922A1 US 20130002922 A1 US20130002922 A1 US 20130002922A1 US 201213610277 A US201213610277 A US 201213610277A US 2013002922 A1 US2013002922 A1 US 2013002922A1
Authority
US
United States
Prior art keywords
subject
frame
imaging apparatus
moving picture
section
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US13/610,277
Inventor
Akira Miyata
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Casio Computer Co Ltd
Original Assignee
Casio Computer Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Casio Computer Co Ltd filed Critical Casio Computer Co Ltd
Priority to US13/610,277 priority Critical patent/US20130002922A1/en
Publication of US20130002922A1 publication Critical patent/US20130002922A1/en
Abandoned legal-status Critical Current

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/60Control of cameras or camera modules
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/60Control of cameras or camera modules
    • H04N23/61Control of cameras or camera modules based on recognised objects
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/60Control of cameras or camera modules
    • H04N23/63Control of cameras or camera modules by using electronic viewfinders
    • H04N23/633Control of cameras or camera modules by using electronic viewfinders for displaying additional information relating to control or operation of the camera
    • H04N23/635Region indicators; Field of view indicators
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/60Control of cameras or camera modules
    • H04N23/67Focus control based on electronic image sensor signals

Definitions

  • the present invention relates to an imaging apparatus and a computer readable recording medium storing a computer program for operating an imaging apparatus, including, in particular, a moving picture capturing means for capturing a moving picture of a subject using an image capturing device and a shutter button that is able to be pressed both halfway and fully, wherein the imaging apparatus tracks a moving subject and recognizes a position of the subject.
  • the following technology is known as a technology for tracking a subject dynamically in an imaging apparatus such as a digital camera.
  • the zooming frame is controlled to track the moving target subject, and the moving image data within the zooming frame is stored.
  • This technology makes it possible to have the zooming frame track the moving target subject.
  • the case in which the moving target subject goes out of the frame is unexpected.
  • the zooming frame is located on the edge of the image. Therefore, even if a user slightly changes the camera angle so that the zooming frame is set to the target subject, the target subject will go out of the frame again in a moment. Meanwhile, once the tracking has started, the user can not change the tracking target without stopping recording mode.
  • this known technology has problems in usability.
  • an imaging apparatus includes: moving picture capturing means for capturing a moving picture of a subject using an image capturing device; first display controlling means for displaying frame data captured in sequence by the moving picture capturing means on a display and for displaying a remarkable point on a predetermined position of the displayed frame data; a shutter button that is adapted to be pressed both halfway and fully; second display controlling means for, when the shutter button is pressed halfway, updating a display position of the remarkable point to a position of the subject so as to track the subject with the remarkable point; and third display controlling means for resetting the display position of the remarkable point to the predetermined position when the halfway-press of the shutter button is released after the shutter button is pressed halfway.
  • the remarkable point may be a remarkable area.
  • a computer readable recording medium stores a computer program, which is executable by a computer to cause the computer to operate an imaging apparatus, wherein the imaging apparatus includes moving picture capturing means for capturing a moving picture of a subject using an image capturing device and a shutter button that is adapted to be pressed both halfway and fully.
  • the computer program causes the computer to execute functions including: displaying frame data captured in sequence by the moving picture capturing means on a display and displaying a remarkable point on a predetermined position of the displayed frame data; updating a display position of the remarkable point to a position of the subject so as to track the subject with the remarkable point when the shutter button is pressed halfway; and resetting the display position of the remarkable point when the halfway-press of the shutter button is released after the shutter button is pressed halfway.
  • FIG. 1 is a block diagram showing a digital camera as one embodiment of this invention.
  • FIG. 2 is a flowchart for explaining operation of a digital camera of this embodiment.
  • FIG. 3 is a diagram illustrating a tracking frame and a trimming frame displayed by the image display block.
  • FIG. 4 is a diagram illustrating a tracking frame and a trimming frame displayed by the image display block.
  • FIG. 5 is a diagram illustrating a tracking frame and a trimming frame displayed by the image display block.
  • FIG. 1 is an electrical block diagram of the digital camera 1 , which realizes one embodiment of an imaging apparatus of this invention.
  • the digital camera 1 includes a lens unit 2 , a lens driving block 3 , an aperture 4 , a Charge Coupled Device (CCD) 5 , a driver 6 , a Timing Generator (TG) 7 , a sampling converter 8 , an image signal generator 9 , a Central Processing Unit (CPU) 10 , a key input block 11 , a memory 12 , a Dynamic Random Access Memory (DRAM) 13 , a flash memory 14 , an image display block 15 , and a bus 16 .
  • CPU Central Processing Unit
  • DRAM Dynamic Random Access Memory
  • the lens unit 2 includes a focus lens group 2 a, and a zoom lens 2 b.
  • the lens driving block 3 is connected to the lens unit 2 .
  • the lens driving block 3 includes a focus motor which drives the focus lens group 2 a along the optical axis of the lens, a zoom motor which drives the zoom lens 2 b along the optical axis of the lens, and a focus motor driver and a zoom motor driver which drive the focus motor and the zoom motor respectively.
  • the aperture 4 includes a driving circuit, which drives the aperture 4 by a control signal sent from the CPU 10 .
  • the aperture 4 is a mechanism which controls the amount of light entering through the lens unit 2 .
  • the CCD 5 which is driven by the driver 6 , outputs a video signal to the sampling converter 8 , by executing photoelectric conversion with respect to the intensity of light for each RGB color of the target subject in a constant period.
  • the operating timing of the vertical driver 6 and the sampling converter 8 are controlled by the CPU 10 via the TG 7 .
  • the CCD 5 has Bayer Pattern Color Filter and a function as an electronic shutter. The shutter speed of the electronic shutter is controlled by the CPU 10 via the driver 6 and the TG 7 .
  • the sampling converter 8 is connected to the TG 7 and includes a Correlated Double Sampling (CDS) Circuit, which executes Correlated Double Sampling on the imaging signal outputted from the CCD 5 and stores the result, an Automatic Gain Control (AGC) Circuit, which executes Automatic Gain Control on the imaging signal after the sampling, an A/D Converter, which changes the analog imaging signal after the Automatic Gain Control into a digital signal.
  • CDS Correlated Double Sampling
  • AGC Automatic Gain Control
  • the imaging signal outputted from the CCD 5 is sent to the image signal generator 9 as a digital signal via the sampling converter 8 .
  • the image signal generator 9 executes gamma correction, a white balance process and other processes on the imaging signal sent from the sampling converter 8 and generates a luminance/chrominance signal (YUV data).
  • the image data of the generated luminance/chrominance signal is sent to CPU 10 .
  • the image signal generator 9 executes image processing on the image data outputted from the CCD 5 .
  • the CPU 10 executes compression/decompression processing (for example the compression/decompression processing of JPEG/MPEG format) for the image data sent from the image signal generator 9 , Auto Focus (AF) processing, and imaging processing.
  • the CPU 10 is a one-chip microcomputer which controls each part of the digital camera 1 and has functions of image-recognition processing for the main target subject, position detection processing for a main target subject and display processing for a tracking frame and a trimming frame.
  • the key input block 11 includes several operation keys such as a shutter button that can be pressed halfway (“halfway-press”) and fully (“full press”), a mode select key, a menu key, an arrow key, a SET key and other keys, and outputs a manipulation signal in accordance with a user's key operation to the CPU 10 .
  • the memory 12 stores a control program which is needed for the CPU 10 to control each part, as well as necessary data, and the CPU 10 works in accordance with the program.
  • the DRAM 13 is used as a buffer memory which temporarily stores the image data, which is sent from the CPU 10 after capturing by the CCD 5 .
  • the DRAM 13 is also used as a working memory of the CPU 10 .
  • the flash memory 14 is a recording medium which stores compressed image data.
  • the image display block 15 includes a color LCD and a driving circuit for the LCD.
  • the image display block 15 displays a target subject captured by CCD 5 as a through image in an imaging standby mode, and the image display block 15 displays a recorded image which is read out from the flash memory 14 and decompressed in a recorded image reproducing mode.
  • the CPU 10 starts a through image displaying process. That is, the CPU 10 starts a process in which: the CCD 5 captures the target subject at a predetermined frame rate, the CPU 10 makes the buffer memory (the DRAM 13 ) store frame image data of luminance and color-difference signal which is captured in sequence by the CCD 5 and processed by the image signal generator 9 , and the CPU 10 displays the sequentially stored image data using the image display block 15 (Step S 1 ). At that time the CPU 10 displays the tracking frame 21 (Observed Position) at a predetermined position of the image display block 15 (here, it is assumed that the predetermined position is the center position) based on predetermined position information which is stored in the memory 12 .
  • the tracking frame 21 Observed Position
  • the CPU 10 judges whether the shutter button is pressed halfway or not (Step S 2 ). The judgment depends on whether or not an operation signal corresponding to the halfway-press of the shutter button is sent from the key input block 11 .
  • the tracking frame tracks the main subject if a user moves the digital camera so that the main subject, which the user wants to capture, overlaps the tracking frame 21 , and presses the shutter button halfway when the tracking frame 21 overlaps the main subject.
  • step S 3 the CPU 10 displays the trimming frame 22 based on the current display position of the tracking frame (the same as the predetermined position at first) (Step S 3 ). In other words, the position of the tracking frame corresponds to the trimming position and therefore the trimming frame is displayed so that the tracking frame is in the center of the trimming frame. Meanwhile, the size of trimming (the size of the trimming frame) is determined previously here, however, it is possible to change the size arbitrarily by an operation of the zoom key.
  • FIG. 3A shows the screen displayed by the image display block 15 when the shutter button is pressed halfway.
  • the tracking frame 21 and the trimming frame 22 are displayed in the center of the screen, and the tracking frame 21 is displayed at the center of the trimming frame 22 . Since a fish is overlapped with the tracking frame 21 in this example, the fish is set as the main subject for capturing and the tracking frame tracks the fish automatically.
  • the CPU 10 recognizes the subject, which is overlapped with the tracking frame 21 , in the captured image based on the frame image data which is captured by the CCD 5 when the shutter button is pressed halfway (Step S 4 ).
  • the subject that is located on the predetermined position is recognized.
  • the recognition is performed with an image recognition process and the subject in the predetermined position is recognized.
  • the fish which is located on the tracking frame 21 is recognized when the shutter button is pressed halfway under the condition demonstrated in the FIG. 3A .
  • An explanation of the image recognition process is omitted since it is a publicly known technology.
  • the CPU 10 gets new frame image data captured by the CCD 5 (Step S 5 ) and performs a process for detecting where the recognized main subject (the fish in this example) is in the captured image (Step S 6 ).
  • the position is detected where the recognized main subject is.
  • the detected position data is stored into position data memory in the buffer memory (DRAM 13 ), and if a position data is already stored in the position data memory, the stored data is refreshed by replacing the stored data with the new detected position data.
  • the image recognition process for the newly captured frame image data it is possible to apply the process to the whole image data, or it is also possible to apply the process only to image data surrounding the previous position of the tracking frame (the position of the main subject). Thus, it is possible to reduce the process load of the image recognition process.
  • CPU 10 judges whether the location of the main target subject is detected or not, that is to say, judges whether a new position data is stored into the position data memory (Step S 7 ). If a new location of the main target subject is detected, CPU 10 updates the display positions of the tracking frame 21 and the trimming frame 22 based on the stored position data. Then the frame image data is displayed on the image display block 15 together with the tracking frame 21 and the trimming frame 22 (Step S 8 ). If the position data of the main target subject is not changed before-and-after the update, the display positions of the tracking frame 21 and the trimming frame 22 are not changed.
  • the trimming frame 22 is controlled to move within a field angle of the captured image (sometimes referred to herein as a captured image angle or image angle), and never runs off the edge of the field angle.
  • Step S 7 if it is determined in Step S 7 that the location of the main target subject cannot be detected, CPU 10 displays a message that the target subject has gone out of the frame, along with the frame image data, and goes on to step S 10 (Step S 9 ).
  • the stored data in the position data memory is not updated, the display of the tracking frame 21 and the trimming frame 22 is not updated, and the tracking frame 21 and the trimming frame 22 are displayed on the same position as before.
  • step S 10 it is judged whether the shutter button is fully pressed or not. If the shutter button isn't fully pressed, it is judged whether the halfway-press of the shutter button is released or not (Step S 11 ). If the halfway-press is not released, the process performed by the CPU 10 goes back to step S 5 , in which the newly captured frame image data is acquired and the above mentioned process (Steps S 5 to S 11 ) is repeated.
  • FIG. 4 and FIG. 5 show examples of displaying the tracking frame 21 and the trimming frame 22 during the loop from step S 5 to step S 11 .
  • FIGS. 4 and 5 show the positioning of the tracking frame 21 and trimming frame 22 which are displayed when the fish, swimming toward the lower left, is captured.
  • the tracking frame 21 and the trimming frame 22 are moved toward the lower left together with the moving of the main subject, the fish, from the image shown in FIG. 3A , in which the shutter button is pressed halfway, to FIG. 5C via FIG. 4A and FIG. 4B .
  • the left side of the trimming frame 22 is bounded by the captured image angle. Therefore, the trimming frame 22 doesn't move any farther to the left even if the fish moves still more left.
  • the trimming frame 22 moves within the range of the image angle, following the position of the tracking frame 21 and the main target subject. For example, the trimming frame 22 doesn't move upward any more if the upper side of the trimming frame 22 is bounded by the image angle.
  • Step S 7 the notice that the target subject has gone out of the frame is displayed and the tracking frame 21 and the trimming frame 22 are displayed on the same position as before. In other words, when the target subject leaves the frame, the tracking frame 21 (and the trimming frame 22 ) does not move and is displayed without change.
  • step S 11 if it is judged that the halfway-press is released in step S 11 , the CPU 10 displays the tracking frame 21 on the predetermined position (the center position) based on the predetermined position information which is stored in the memory 12 , and goes back to step S 2 (Step S 12 ). In this case, the detected position information which is stored in the position data memory in the buffer memory is cleared.
  • FIG. 3B shows the screen image that is displayed on the image display block 15 when the halfway-press of the shutter button is released.
  • the tracking frame 21 doesn't track the target subject and is displayed in the center of the screen, as shown in FIG. 3B .
  • CPU 10 stops displaying the trimming frame 22 . If a user wants to change the main target subject or the main target subject goes out of the frame, the user can simply release the halfway-press of the shutter button. As a result of this operation, the track frame 21 returns to its normal location (e.g., the center of the image in this embodiment). Therefore, the user can restart the operation to track the main target subject from the beginning.
  • the tracking frame 21 is left displayed on the edge of the image when the main target subject goes out of the frame ( FIG. 5E ).
  • the main target subject is likely to go out of the frame again quickly.
  • step S 10 If it is judged that the shutter button is fully pressed in step S 10 , the CPU 10 performs AF processing within a given AF area based on the position of the tracking frame (the position information stored in the position data memory) (Step S 13 ). Next the CPU 10 performs the capturing process of a still image (Step S 14 ), and performs the trimming process on the still image which is captured in the capturing process using the position of the tracking frame 21 just before the full press of the shutter button (the position information stored in the position data memory) and the predetermined trimming size (Step S 15 ).
  • the CPU 10 displays, as a preview display, the image data trimmed by the trimming process (Trimming Image Data) using the image display block 15 (Step S 16 ), and judges whether the trimming image data is to be stored or not (Step S 17 ).
  • a screen display for assigning whether the trimming image data is to be stored or not is displayed together with the preview display. If the user assigns the image to be stored by an operation of the key input block 11 , it is determined that the image will be stored. If the user assigns not to store the image, it is determined that the image will not be stored. If it is judged in step S 17 that the image will be stored, the trimming image data is compressed and stored into the flash memory 14 (Step S 18 ).
  • the through image is displayed with a tracking frame 21 on the predetermined position.
  • a shutter button is pressed halfway when the tracking frame 21 overlaps the main subject which a user wants to capture, the subject, which is overlapped with the tracking frame 21 , is recognized.
  • the tracking frame 21 is displayed and tracked to the recognized main target subject. If the halfway-press of the shutter button is released, the tracking frame 21 is displayed on the predetermined position.
  • the present invention allows the present invention to deal with the case in which the main target subject goes out of the frame or the case in which a user wants to change the main target subject, by releasing the halfway-press of the shutter button, thereby releasing the tracking of the subject with the tracking frame 21 .
  • the tracking frame is still displayed if the halfway-press of the shutter button is released, and is moved to the predetermined position (e.g., the center of the screen). Therefore, the tracking frame 21 can be overlapped on the main target subject while the tracking frame is in the predetermined position, and it is unnecessary to overlap the tracking frame 21 to the main target subject while the tracking frame 21 is on the side of the screen.
  • the AF processing and the trimming processing are performed based on the position of the tracking frame 21 .
  • the AF area and/or the trimming area, which is displayed as a substitution for the tracking frame 21 are generically named “Remarkable Area”.
  • the AF processing is performed after the full press of the shutter button.
  • Step S 9 a message that the target subject has gone out of the frame (framing out) is displayed (Step S 9 ).
  • the user can recognize the framing out of the target subject by watching the image display block 15 .
  • the main subject which is overlapped with the tracking frame 21 , is recognized by using an image recognition process in the step S 4 and it is detected where the recognized main subject is in the newly captured frame image data.
  • the digital camera 1 described above is not limited to the above mentioned embodiment.
  • This invention may be used in various devices such as mobile terminal devices with cameras, personal computers with cameras, IC recorders with cameras, and/or digital video cameras.
  • the present invention may be practiced in any device having a function for capturing an image of a subject.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Studio Devices (AREA)
  • Focusing (AREA)
  • Automatic Focus Adjustment (AREA)

Abstract

Once an imaging mode is set, an imaging apparatus displays a through image with a tracking frame on the predetermined position. If a shutter button is pressed halfway, an object that is overlapped by the tracking frame is recognized as a main subject that a user wants to capture. Until the shutter button is fully pressed or until the halfway-press is released, it is detected where the recognized subject is in frame image data that is captured in sequence, and the tracking frame is displayed based on the detected position. If the halfway press of the shutter button is released before the shutter button is fully pressed, the tracking frame is displayed on the predetermined position. Then if the shutter button is pressed halfway once again, the process is repeated.

Description

    CROSS-REFERENCE TO RELATED APPLICATIONS
  • This is a Divisional of U.S. application Ser. No. 11/770,114, filed Jun. 28, 2007, which is based on Japanese Patent Application No. 2006-181470, filed Jun. 30, 2006, the entire contents of both of which are incorporated herein by reference.
  • BACKGROUND OF THE INVENTION
  • 1. Field of the Invention
  • The present invention relates to an imaging apparatus and a computer readable recording medium storing a computer program for operating an imaging apparatus, including, in particular, a moving picture capturing means for capturing a moving picture of a subject using an image capturing device and a shutter button that is able to be pressed both halfway and fully, wherein the imaging apparatus tracks a moving subject and recognizes a position of the subject.
  • 2. Description of the Related Art
  • The following technology is known as a technology for tracking a subject dynamically in an imaging apparatus such as a digital camera. In this technology, if a moving target subject overlaps the zooming frame when capturing a moving image, the zooming frame is controlled to track the moving target subject, and the moving image data within the zooming frame is stored.
  • This technology makes it possible to have the zooming frame track the moving target subject. However, the case in which the moving target subject goes out of the frame is unexpected. When the moving target subject goes out of the frame, the zooming frame is located on the edge of the image. Therefore, even if a user slightly changes the camera angle so that the zooming frame is set to the target subject, the target subject will go out of the frame again in a moment. Meanwhile, once the tracking has started, the user can not change the tracking target without stopping recording mode. Thus, this known technology has problems in usability.
  • SUMMARY OF THE INVENTION
  • According to one aspect of the invention, an imaging apparatus includes: moving picture capturing means for capturing a moving picture of a subject using an image capturing device; first display controlling means for displaying frame data captured in sequence by the moving picture capturing means on a display and for displaying a remarkable point on a predetermined position of the displayed frame data; a shutter button that is adapted to be pressed both halfway and fully; second display controlling means for, when the shutter button is pressed halfway, updating a display position of the remarkable point to a position of the subject so as to track the subject with the remarkable point; and third display controlling means for resetting the display position of the remarkable point to the predetermined position when the halfway-press of the shutter button is released after the shutter button is pressed halfway.
  • According to the present invention, the remarkable point may be a remarkable area.
  • According to a further aspect of the invention, a computer readable recording medium stores a computer program, which is executable by a computer to cause the computer to operate an imaging apparatus, wherein the imaging apparatus includes moving picture capturing means for capturing a moving picture of a subject using an image capturing device and a shutter button that is adapted to be pressed both halfway and fully. The computer program causes the computer to execute functions including: displaying frame data captured in sequence by the moving picture capturing means on a display and displaying a remarkable point on a predetermined position of the displayed frame data; updating a display position of the remarkable point to a position of the subject so as to track the subject with the remarkable point when the shutter button is pressed halfway; and resetting the display position of the remarkable point when the halfway-press of the shutter button is released after the shutter button is pressed halfway.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • The accompanying drawings illustrate preferred embodiments of an imaging apparatus according to the invention, and together with the detailed description of the preferred embodiments given below serve to explain the principles of the invention.
  • FIG. 1 is a block diagram showing a digital camera as one embodiment of this invention.
  • FIG. 2 is a flowchart for explaining operation of a digital camera of this embodiment.
  • FIG. 3 is a diagram illustrating a tracking frame and a trimming frame displayed by the image display block.
  • FIG. 4 is a diagram illustrating a tracking frame and a trimming frame displayed by the image display block.
  • FIG. 5 is a diagram illustrating a tracking frame and a trimming frame displayed by the image display block.
  • DETAILED DESCRIPTION
  • Now, an embodiment of the invention will be described in detail with reference to the accompanying drawings using an example in which an imaging apparatus of this invention is applied to a digital camera.
  • A. The Structure of the Digital Camera
  • FIG. 1 is an electrical block diagram of the digital camera 1, which realizes one embodiment of an imaging apparatus of this invention. The digital camera 1 includes a lens unit 2, a lens driving block 3, an aperture 4, a Charge Coupled Device (CCD) 5, a driver 6, a Timing Generator (TG) 7, a sampling converter 8, an image signal generator 9, a Central Processing Unit (CPU) 10, a key input block 11, a memory 12, a Dynamic Random Access Memory (DRAM) 13, a flash memory 14, an image display block 15, and a bus 16.
  • The lens unit 2 includes a focus lens group 2 a, and a zoom lens 2 b. The lens driving block 3 is connected to the lens unit 2. The lens driving block 3 includes a focus motor which drives the focus lens group 2 a along the optical axis of the lens, a zoom motor which drives the zoom lens 2 b along the optical axis of the lens, and a focus motor driver and a zoom motor driver which drive the focus motor and the zoom motor respectively.
  • The aperture 4 includes a driving circuit, which drives the aperture 4 by a control signal sent from the CPU 10. The aperture 4 is a mechanism which controls the amount of light entering through the lens unit 2.
  • The CCD 5, which is driven by the driver 6, outputs a video signal to the sampling converter 8, by executing photoelectric conversion with respect to the intensity of light for each RGB color of the target subject in a constant period. The operating timing of the vertical driver 6 and the sampling converter 8 are controlled by the CPU 10 via the TG 7. Meanwhile, the CCD 5 has Bayer Pattern Color Filter and a function as an electronic shutter. The shutter speed of the electronic shutter is controlled by the CPU 10 via the driver 6 and the TG 7.
  • The sampling converter 8 is connected to the TG 7 and includes a Correlated Double Sampling (CDS) Circuit, which executes Correlated Double Sampling on the imaging signal outputted from the CCD 5 and stores the result, an Automatic Gain Control (AGC) Circuit, which executes Automatic Gain Control on the imaging signal after the sampling, an A/D Converter, which changes the analog imaging signal after the Automatic Gain Control into a digital signal. The imaging signal outputted from the CCD 5 is sent to the image signal generator 9 as a digital signal via the sampling converter 8.
  • The image signal generator 9 executes gamma correction, a white balance process and other processes on the imaging signal sent from the sampling converter 8 and generates a luminance/chrominance signal (YUV data). The image data of the generated luminance/chrominance signal is sent to CPU 10. In other words, the image signal generator 9 executes image processing on the image data outputted from the CCD 5.
  • The CPU 10 executes compression/decompression processing (for example the compression/decompression processing of JPEG/MPEG format) for the image data sent from the image signal generator 9, Auto Focus (AF) processing, and imaging processing. The CPU 10 is a one-chip microcomputer which controls each part of the digital camera 1 and has functions of image-recognition processing for the main target subject, position detection processing for a main target subject and display processing for a tracking frame and a trimming frame.
  • The key input block 11 includes several operation keys such as a shutter button that can be pressed halfway (“halfway-press”) and fully (“full press”), a mode select key, a menu key, an arrow key, a SET key and other keys, and outputs a manipulation signal in accordance with a user's key operation to the CPU 10. The memory 12 stores a control program which is needed for the CPU 10 to control each part, as well as necessary data, and the CPU 10 works in accordance with the program.
  • The DRAM 13 is used as a buffer memory which temporarily stores the image data, which is sent from the CPU 10 after capturing by the CCD 5. The DRAM 13 is also used as a working memory of the CPU 10. The flash memory 14 is a recording medium which stores compressed image data.
  • The image display block 15 includes a color LCD and a driving circuit for the LCD. The image display block 15 displays a target subject captured by CCD 5 as a through image in an imaging standby mode, and the image display block 15 displays a recorded image which is read out from the flash memory 14 and decompressed in a recorded image reproducing mode.
  • B. The Operation of the Digital Camera
  • The operation of the digital camera 1 in this embodiment will be described with reference to FIG. 2.
  • Once the imaging mode is set by the user's operation of the mode select key in the key input block 11, the CPU 10 starts a through image displaying process. That is, the CPU 10 starts a process in which: the CCD 5 captures the target subject at a predetermined frame rate, the CPU 10 makes the buffer memory (the DRAM 13) store frame image data of luminance and color-difference signal which is captured in sequence by the CCD 5 and processed by the image signal generator 9, and the CPU 10 displays the sequentially stored image data using the image display block 15 (Step S1). At that time the CPU 10 displays the tracking frame 21 (Observed Position) at a predetermined position of the image display block 15 (here, it is assumed that the predetermined position is the center position) based on predetermined position information which is stored in the memory 12.
  • Then the CPU 10 judges whether the shutter button is pressed halfway or not (Step S2). The judgment depends on whether or not an operation signal corresponding to the halfway-press of the shutter button is sent from the key input block 11. According to this embodiment, the tracking frame tracks the main subject if a user moves the digital camera so that the main subject, which the user wants to capture, overlaps the tracking frame 21, and presses the shutter button halfway when the tracking frame 21 overlaps the main subject.
  • If the CPU 10 judges that the shutter button isn't pressed halfway in step S2, it remains in step S2 until the shutter button is pressed halfway. If the CPU 10 judges that the shutter button is pressed halfway, the CPU 10 displays the trimming frame 22 based on the current display position of the tracking frame (the same as the predetermined position at first) (Step S3). In other words, the position of the tracking frame corresponds to the trimming position and therefore the trimming frame is displayed so that the tracking frame is in the center of the trimming frame. Meanwhile, the size of trimming (the size of the trimming frame) is determined previously here, however, it is possible to change the size arbitrarily by an operation of the zoom key.
  • FIG. 3A shows the screen displayed by the image display block 15 when the shutter button is pressed halfway. As shown in FIG. 3A, the tracking frame 21 and the trimming frame 22 are displayed in the center of the screen, and the tracking frame 21 is displayed at the center of the trimming frame 22. Since a fish is overlapped with the tracking frame 21 in this example, the fish is set as the main subject for capturing and the tracking frame tracks the fish automatically.
  • The CPU 10 recognizes the subject, which is overlapped with the tracking frame 21, in the captured image based on the frame image data which is captured by the CCD 5 when the shutter button is pressed halfway (Step S4). In this example, since the tracking frame 21 is located on the predetermined position, the subject that is located on the predetermined position is recognized. The recognition is performed with an image recognition process and the subject in the predetermined position is recognized. For example, the fish which is located on the tracking frame 21 is recognized when the shutter button is pressed halfway under the condition demonstrated in the FIG. 3A. An explanation of the image recognition process is omitted since it is a publicly known technology.
  • Then the CPU 10 gets new frame image data captured by the CCD 5 (Step S5) and performs a process for detecting where the recognized main subject (the fish in this example) is in the captured image (Step S6). In other words, by performing image recognition of newly captured frame image data, the position is detected where the recognized main subject is. When the position is detected in this step, the detected position data is stored into position data memory in the buffer memory (DRAM 13), and if a position data is already stored in the position data memory, the stored data is refreshed by replacing the stored data with the new detected position data.
  • Meanwhile, with regard to the image recognition process for the newly captured frame image data, it is possible to apply the process to the whole image data, or it is also possible to apply the process only to image data surrounding the previous position of the tracking frame (the position of the main subject). Thus, it is possible to reduce the process load of the image recognition process.
  • CPU 10 judges whether the location of the main target subject is detected or not, that is to say, judges whether a new position data is stored into the position data memory (Step S7). If a new location of the main target subject is detected, CPU 10 updates the display positions of the tracking frame 21 and the trimming frame 22 based on the stored position data. Then the frame image data is displayed on the image display block 15 together with the tracking frame 21 and the trimming frame 22 (Step S8). If the position data of the main target subject is not changed before-and-after the update, the display positions of the tracking frame 21 and the trimming frame 22 are not changed. Additionally, when updating the display positions of the tracking frame 21 and the trimming frame 22, the trimming frame 22 is controlled to move within a field angle of the captured image (sometimes referred to herein as a captured image angle or image angle), and never runs off the edge of the field angle.
  • On the other hand, if it is determined in Step S7 that the location of the main target subject cannot be detected, CPU 10 displays a message that the target subject has gone out of the frame, along with the frame image data, and goes on to step S10 (Step S9). In this case, the stored data in the position data memory is not updated, the display of the tracking frame 21 and the trimming frame 22 is not updated, and the tracking frame 21 and the trimming frame 22 are displayed on the same position as before.
  • In step S10, it is judged whether the shutter button is fully pressed or not. If the shutter button isn't fully pressed, it is judged whether the halfway-press of the shutter button is released or not (Step S11). If the halfway-press is not released, the process performed by the CPU 10 goes back to step S5, in which the newly captured frame image data is acquired and the above mentioned process (Steps S5 to S11) is repeated.
  • FIG. 4 and FIG. 5 show examples of displaying the tracking frame 21 and the trimming frame 22 during the loop from step S5 to step S11. Specifically, FIGS. 4 and 5 show the positioning of the tracking frame 21 and trimming frame 22 which are displayed when the fish, swimming toward the lower left, is captured.
  • The tracking frame 21 and the trimming frame 22 are moved toward the lower left together with the moving of the main subject, the fish, from the image shown in FIG. 3A, in which the shutter button is pressed halfway, to FIG. 5C via FIG. 4A and FIG. 4B. In FIG. 5C, the left side of the trimming frame 22 is bounded by the captured image angle. Therefore, the trimming frame 22 doesn't move any farther to the left even if the fish moves still more left. Thus, the trimming frame 22 moves within the range of the image angle, following the position of the tracking frame 21 and the main target subject. For example, the trimming frame 22 doesn't move upward any more if the upper side of the trimming frame 22 is bounded by the image angle.
  • Additionally, if the fish moves still more lower left as shown in FIG. 5D, the trimming frame 22 only moves downward and the tracking frame 21 moves together with the moving of the main target subject. And then if the main target subject, the fish, goes out of the frame (a judgment is made in Step S7 that the position of the target subject cannot be detected), the notice that the target subject has gone out of the frame is displayed and the tracking frame 21 and the trimming frame 22 are displayed on the same position as before. In other words, when the target subject leaves the frame, the tracking frame 21 (and the trimming frame 22) does not move and is displayed without change.
  • Going back to the flowchart shown in FIG. 2, if it is judged that the halfway-press is released in step S11, the CPU 10 displays the tracking frame 21 on the predetermined position (the center position) based on the predetermined position information which is stored in the memory 12, and goes back to step S2 (Step S12). In this case, the detected position information which is stored in the position data memory in the buffer memory is cleared. FIG. 3B shows the screen image that is displayed on the image display block 15 when the halfway-press of the shutter button is released.
  • When the halfway-press of the shutter button is released, the tracking frame 21 doesn't track the target subject and is displayed in the center of the screen, as shown in FIG. 3B. In addition, when the shutter button is released from the halfway-press, CPU 10 stops displaying the trimming frame 22. If a user wants to change the main target subject or the main target subject goes out of the frame, the user can simply release the halfway-press of the shutter button. As a result of this operation, the track frame 21 returns to its normal location (e.g., the center of the image in this embodiment). Therefore, the user can restart the operation to track the main target subject from the beginning.
  • By contrast, if, for example the position of the tracking frame 21 is not changed when the halfway-press of the shutter button is released, the tracking frame 21 is left displayed on the edge of the image when the main target subject goes out of the frame (FIG. 5E). Thus, if a user tries to make the tracking frame 21 track the main target subject by overlapping the tracking frame 21, which is displayed on edge, with the main target subject and pressing the shutter button halfway, the main target subject is likely to go out of the frame again quickly.
  • In addition, if the position of the tracking frame 21 is not changed when the halfway-press of the shutter button is released, if a user wants to change the main target subject, it is necessary for the user to release the halfway-press when the tracking frame 21 is displayed not on the edge of the image but rather when the tracking frame is displayed on a position which is convenient for capturing and tracking for the main target subject so as to eliminate the above problem. Thus, it is impossible to deal with the case in which the user wants to capture the image immediately. Therefore it is desirable to solve the above problem. In the aspect of this invention described above, it is possible to cope well with the above problem by displaying the tracking frame 21 on the predetermined position (e.g., the center of the image) when the halfway-press of the shutter button is released.
  • If it is judged that the shutter button is fully pressed in step S10, the CPU 10 performs AF processing within a given AF area based on the position of the tracking frame (the position information stored in the position data memory) (Step S13). Next the CPU 10 performs the capturing process of a still image (Step S14), and performs the trimming process on the still image which is captured in the capturing process using the position of the tracking frame 21 just before the full press of the shutter button (the position information stored in the position data memory) and the predetermined trimming size (Step S15).
  • Next the CPU 10 displays, as a preview display, the image data trimmed by the trimming process (Trimming Image Data) using the image display block 15 (Step S16), and judges whether the trimming image data is to be stored or not (Step S17). At that time, a screen display for assigning whether the trimming image data is to be stored or not is displayed together with the preview display. If the user assigns the image to be stored by an operation of the key input block 11, it is determined that the image will be stored. If the user assigns not to store the image, it is determined that the image will not be stored. If it is judged in step S17 that the image will be stored, the trimming image data is compressed and stored into the flash memory 14 (Step S18).
  • C. Summary of Operation of the Above-Described Embodiment
  • As described above, in this embodiment, once the imaging mode is set, the through image is displayed with a tracking frame 21 on the predetermined position. And if a shutter button is pressed halfway when the tracking frame 21 overlaps the main subject which a user wants to capture, the subject, which is overlapped with the tracking frame 21, is recognized. As long as the shutter button is pressed halfway, the tracking frame 21 is displayed and tracked to the recognized main target subject. If the halfway-press of the shutter button is released, the tracking frame 21 is displayed on the predetermined position. This allows the present invention to deal with the case in which the main target subject goes out of the frame or the case in which a user wants to change the main target subject, by releasing the halfway-press of the shutter button, thereby releasing the tracking of the subject with the tracking frame 21. For example, if the main target subject goes out of the frame, the tracking frame is still displayed if the halfway-press of the shutter button is released, and is moved to the predetermined position (e.g., the center of the screen). Therefore, the tracking frame 21 can be overlapped on the main target subject while the tracking frame is in the predetermined position, and it is unnecessary to overlap the tracking frame 21 to the main target subject while the tracking frame 21 is on the side of the screen.
  • D. Other Embodiments
  • It is possible to change the above mentioned embodiment using the following modifications, for example.
  • (1) In the above mentioned embodiment, the AF processing and the trimming processing are performed based on the position of the tracking frame 21. However, it is possible to select one of the processing, the AF processing and the trimming processing. In this case, it is possible to display either the AF area or the trimming area only instead of the tracking frame. The AF area and/or the trimming area, which is displayed as a substitution for the tracking frame 21, are generically named “Remarkable Area”.
  • (2) In the above mentioned embodiment, the AF processing is performed after the full press of the shutter button. However, it is possible to perform the AF processing after the halfway-press of the shutter button using the AF area based on the position of the tracking frame (the predetermined position).
  • (3) In the above mentioned embodiment, if the target subject goes out of the frame (it is judged that the location of the main target subject can't be detected in step S7), a message that the target subject has gone out of the frame (framing out) is displayed (Step S9). However, it is possible not to display the message of the framing out of the target subject, because the user can recognize the framing out of the target subject by watching the image display block 15.
  • (4) In the above mentioned embodiment, the main subject, which is overlapped with the tracking frame 21, is recognized by using an image recognition process in the step S4 and it is detected where the recognized main subject is in the newly captured frame image data. However, it is possible to detect the position of the main subject in the image by using a technique which detects the motion vector of the main subject, instead of the image recognition process. It is sufficient to practice this aspect of the present invention using a method to be able to detect the position of the main subject in the frame image data.
  • (5) It is possible to use a dot, a circle or an x-mark instead of the tracking frame 21 in the above embodiment. That is, it is sufficient to provide a method for user to be able to catch the position for detecting the main subject and/or the current position of the recognized main subject. The display of a point, which makes a user catch the position for detecting the main subject and/or the current position of the recognized main subject, for example the tracking frame 21, a dot, a circle and/or an x-mark, are generically called “Remarkable Point”.
  • (6) Meanwhile, if the shutter button is fully pressed suddenly without staying in the halfway-pressed stage, it is possible to process the AF processing based on the tracking frame on the predetermined position, process the capturing process of a still image, and then store the still image data captured by the image capturing process. In such case, it is also possible to generate a trimming image data of the still image data captured by the image capturing process using the tracking frame, which is in the predetermined position, and store it.
  • (7) In addition, it is possible for a user to change the size of the trimming area (trimming size) arbitrarily.
  • (8) It is also possible to combine an arbitrary number of the above modifications (1) to (7).
  • The digital camera 1 described above is not limited to the above mentioned embodiment. This invention may be used in various devices such as mobile terminal devices with cameras, personal computers with cameras, IC recorders with cameras, and/or digital video cameras. The present invention may be practiced in any device having a function for capturing an image of a subject.

Claims (14)

1. An imaging apparatus comprising:
a moving picture capturing section for capturing a moving picture of a subject using an image capturing device;
a first display controlling section for displaying frame data captured in sequence by the moving picture capturing section on a display; and
a second display controlling section for displaying a remarkable point on a position of the subject so as to track the subject with the remarkable point;
wherein when the subject goes out of the a frame of the moving picture, the second display controlling section continuously displays the remarkable point on a position at which the subject was located immediately before going out of the frame.
2. The imaging apparatus according to claim 1, wherein when the subject goes out of the frame of the moving picture, the second display controlling section displays a message notifying of the framing out of the subject.
3. The imaging apparatus according to claim 1, further comprising an auto-focus controlling section for executing an auto-focus process with respect to the subject located on the position of the remarkable point.
4. The imaging apparatus according to claim 3, further comprising a still image capturing section for capturing a still image of the subject using the image capturing device after the auto-focus process is executed by the auto-focus controlling section.
5. The imaging apparatus according to claim 4, further comprising:
a trimming frame displaying section for displaying a trimming frame on the display and for displacing a display position of the trimming frame such that the trimming frame tracks the subject in accordance with the tracking performed by the second display controlling section; and
a trimming section for trimming the still image captured by the still image capturing section based on the position of the trimming frame.
6. The imaging apparatus according to claim 1, further comprising a shutter button that is adapted to be pressed both halfway and fully;
wherein the second display controlling section displays the remarkable point on the position of the subject so as to track the subject with the remarkable point when the shutter button is pressed halfway.
7. An imaging apparatus comprising:
a moving picture capturing section for capturing a moving picture of a subject using an image capturing device;
a first display controlling section for displaying frame data captured in sequence by the moving picture capturing section on a display; and
a second display controlling section for displaying a remarkable area on a position of the subject so as to track the subject with the remarkable area;
wherein when the subject goes out of a frame of the moving picture, the second display controlling section continuously displays the remarkable area on a position at which the subject was located immediately before going out of the frame.
8. The imaging apparatus according to claim 7, wherein when the subject goes out of the frame of the moving picture, the second display controlling section displays a message notifying of the framing out of the subject.
9. The imaging apparatus according to claim 1, further comprising an auto-focus controlling section for executing an auto-focus process with respect to the subject located on the position of the remarkable area.
10. The imaging apparatus according to claim 9, further comprising a still image capturing section for capturing a still image of the subject using the image capturing device after the auto-focus process is executed by the auto-focus controlling section.
11. The imaging apparatus according to claim 10, further comprising:
a trimming frame displaying section for displaying a trimming frame on the display and for displacing a display position of the trimming frame such that the trimming frame tracks the subject in accordance with the tracking performed by the second display controlling section; and
a trimming section for trimming the still image captured by the still image capturing section based on the position of the trimming frame.
12. The imaging apparatus according to claim 7, further comprising a shutter button that is adapted to be pressed both halfway and fully;
wherein the second display controlling section displays the remarkable area on the position of the subject so as to track the subject with the remarkable area when the shutter button is pressed halfway.
13. A non-transitory computer readable recording medium storing a computer program which is executable by a computer to cause the computer to operate an imaging apparatus, the imaging apparatus including a moving picture capturing section for capturing a moving picture of a subject using an image capturing device, the program causing the computer to execute functions comprising:
displaying frame data captured in sequence by the moving picture capturing section on a display; and
displaying a remarkable point on a position of the subject so as to track the subject with the remarkable point;
wherein when the subject goes out of a frame of the moving picture, the remarkable point is continuously displayed on a position at which the subject was located immediately before going out of the frame.
14. A non-transitory computer readable recording medium storing a computer program which is executable by a computer to cause the computer to operate an imaging apparatus, the imaging apparatus including a moving picture capturing section for capturing a moving picture of a subject using an image capturing device, the program causing the computer to execute functions comprising:
displaying frame data captured in sequence by the moving picture capturing section on a display; and
displaying a remarkable area on a position of the subject so as to track the subject with the remarkable area;
wherein when the subject goes out of a frame of the moving picture, the remarkable area is continuously displayed on a position at which the subject was located immediately before going out of the frame.
US13/610,277 2006-06-30 2012-09-11 Imaging apparatus and computer readable recording medium Abandoned US20130002922A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US13/610,277 US20130002922A1 (en) 2006-06-30 2012-09-11 Imaging apparatus and computer readable recording medium

Applications Claiming Priority (4)

Application Number Priority Date Filing Date Title
JP2006-181470 2006-06-30
JP2006181470A JP4761146B2 (en) 2006-06-30 2006-06-30 Imaging apparatus and program thereof
US11/770,114 US8284256B2 (en) 2006-06-30 2007-06-28 Imaging apparatus and computer readable recording medium
US13/610,277 US20130002922A1 (en) 2006-06-30 2012-09-11 Imaging apparatus and computer readable recording medium

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
US11/770,114 Division US8284256B2 (en) 2006-06-30 2007-06-28 Imaging apparatus and computer readable recording medium

Publications (1)

Publication Number Publication Date
US20130002922A1 true US20130002922A1 (en) 2013-01-03

Family

ID=38876171

Family Applications (2)

Application Number Title Priority Date Filing Date
US11/770,114 Active 2030-12-05 US8284256B2 (en) 2006-06-30 2007-06-28 Imaging apparatus and computer readable recording medium
US13/610,277 Abandoned US20130002922A1 (en) 2006-06-30 2012-09-11 Imaging apparatus and computer readable recording medium

Family Applications Before (1)

Application Number Title Priority Date Filing Date
US11/770,114 Active 2030-12-05 US8284256B2 (en) 2006-06-30 2007-06-28 Imaging apparatus and computer readable recording medium

Country Status (6)

Country Link
US (2) US8284256B2 (en)
JP (1) JP4761146B2 (en)
KR (1) KR100847614B1 (en)
CN (1) CN100591099C (en)
HK (1) HK1111032A1 (en)
TW (1) TWI341684B (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20130044251A1 (en) * 2008-02-04 2013-02-21 Samsung Electronics Co., Ltd. Digital image processing apparatus and method of controlling the same
US20190297265A1 (en) * 2018-03-21 2019-09-26 Sawah Innovations Inc. User-feedback video stabilization device and method

Families Citing this family (32)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2006044476A2 (en) 2004-10-12 2006-04-27 Robert Vernon Vanman Method of and system for mobile surveillance and event recording
JP2008187591A (en) * 2007-01-31 2008-08-14 Fujifilm Corp Imaging apparatus and imaging method
US8599368B1 (en) 2008-01-29 2013-12-03 Enforcement Video, Llc Laser-based speed determination device for use in a moving vehicle
JP4492697B2 (en) * 2007-12-28 2010-06-30 カシオ計算機株式会社 Imaging apparatus and program
JP4958805B2 (en) * 2008-01-22 2012-06-20 キヤノン株式会社 Imaging apparatus and display control method thereof
CA2714362A1 (en) 2008-01-29 2009-08-06 Enforcement Video, Llc Omnidirectional camera for use in police car event recording
KR101445606B1 (en) * 2008-02-05 2014-09-29 삼성전자주식회사 Digital photographing apparatus, method for controlling the same, and recording medium storing program to implement the method
US20090213218A1 (en) 2008-02-15 2009-08-27 Andrew Cilia System and method for multi-resolution storage of images
KR101436326B1 (en) * 2008-07-31 2014-09-01 삼성전자주식회사 Digital photographing apparatus, method for controlling the same, and recording medium storing program to implement the method
US8237847B2 (en) * 2008-10-16 2012-08-07 Fujinon Corporation Auto focus system having AF frame auto-tracking function
JP5210841B2 (en) * 2008-12-12 2013-06-12 パナソニック株式会社 Imaging device
US7995909B2 (en) 2008-12-12 2011-08-09 Samsung Electro-Mechanics Co., Ltd. Auto-focusing method
KR100975928B1 (en) * 2008-12-12 2010-08-13 삼성전기주식회사 Auto-focusing method
KR100975929B1 (en) * 2008-12-12 2010-08-13 삼성전기주식회사 Auto-focusing method
JP2011139231A (en) * 2009-12-28 2011-07-14 Sony Corp Image processing apparatus, image processing method and program
JP2012049651A (en) 2010-08-24 2012-03-08 Ricoh Co Ltd Imaging apparatus and imaging method
KR101694421B1 (en) * 2010-11-11 2017-01-09 삼성전자주식회사 Method and apparatus for controlling continuous auto focus of mobile terminal comprising camera
TWI584647B (en) * 2012-07-12 2017-05-21 Chi Lin Hong A method and apparatus for preventing defocusing
US10178298B2 (en) 2012-12-20 2019-01-08 Sony Corporation Image processing device, image processing method, and recording medium for optimal trimming of a captured image
US9165364B1 (en) * 2012-12-26 2015-10-20 Canon Kabushiki Kaisha Automatic tracking image pickup system
KR20140104807A (en) * 2013-02-21 2014-08-29 삼성전자주식회사 Method, apparatus and recording medium for guiding photography shooting using subject-related attributes
JP6103526B2 (en) * 2013-03-15 2017-03-29 オリンパス株式会社 Imaging device, image display device, and display control method for image display device
JP6223170B2 (en) * 2013-12-19 2017-11-01 キヤノン株式会社 Imaging device, control method thereof, and control program
KR101446781B1 (en) * 2014-04-21 2014-10-01 삼성전자주식회사 Electronic device and controlling method
JP6641365B2 (en) * 2015-06-22 2020-02-05 富士フイルム株式会社 Radiation irradiation apparatus, control method of radiation irradiation apparatus, and program
US10341605B1 (en) 2016-04-07 2019-07-02 WatchGuard, Inc. Systems and methods for multiple-resolution storage of media streams
JP6862202B2 (en) 2017-02-08 2021-04-21 キヤノン株式会社 Image processing device, imaging device and control method
WO2019061419A1 (en) * 2017-09-30 2019-04-04 深圳市大疆创新科技有限公司 Photographing apparatus control method, photographing apparatus and storage medium
JP6506440B2 (en) * 2018-04-03 2019-04-24 オリンパス株式会社 Imaging device, imaging method, and program
WO2020067323A1 (en) * 2018-09-27 2020-04-02 富士フイルム株式会社 Radiation imaging device
US11523060B2 (en) * 2018-11-29 2022-12-06 Ricoh Company, Ltd. Display device, imaging device, object moving method, and recording medium
JP2021180379A (en) 2020-05-12 2021-11-18 キヤノン株式会社 Imaging device and control method for imaging device

Family Cites Families (20)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP4391637B2 (en) * 1999-10-04 2009-12-24 オリンパス株式会社 camera
JP2004101632A (en) * 2002-09-05 2004-04-02 Canon Inc Camera equipped with automatic focusing device
JP3949067B2 (en) * 2003-02-14 2007-07-25 三洋電機株式会社 Auto focus camera
DE602004032488D1 (en) * 2003-02-18 2011-06-16 Canon Kk Photographing device with means for receiving radio information and control method therefor
US20040207743A1 (en) * 2003-04-15 2004-10-21 Nikon Corporation Digital camera system
US7466356B2 (en) * 2003-07-25 2008-12-16 Hewlett-Packard Development Company, L.P. Method and apparatus for setting a marker on an object and tracking the position of the object
JP3918788B2 (en) * 2003-08-06 2007-05-23 コニカミノルタフォトイメージング株式会社 Imaging apparatus and program
JP4352916B2 (en) * 2004-02-04 2009-10-28 ソニー株式会社 Imaging apparatus and imaging method
JP4130641B2 (en) * 2004-03-31 2008-08-06 富士フイルム株式会社 Digital still camera and control method thereof
JP2005347885A (en) * 2004-05-31 2005-12-15 Casio Comput Co Ltd Imaging unit, photographing range adjustment method, and program
JP2005341449A (en) * 2004-05-31 2005-12-08 Toshiba Corp Digital still camera
JP2006058405A (en) * 2004-08-18 2006-03-02 Casio Comput Co Ltd Camera apparatus and automatic focusing control method
JP2006098676A (en) * 2004-09-29 2006-04-13 Nikon Corp Autofocus camera
KR101120027B1 (en) * 2004-11-01 2012-03-23 삼성전자주식회사 Apparatus and method for processing on screen display when shutter is pressed half state in digital camera
WO2006082967A1 (en) * 2005-02-07 2006-08-10 Matsushita Electric Industrial Co., Ltd. Imaging device
JP2006254413A (en) * 2005-02-08 2006-09-21 Nikon Corp Imaging apparatus and camera body
US7945938B2 (en) * 2005-05-11 2011-05-17 Canon Kabushiki Kaisha Network camera system and control method therefore
JP4572815B2 (en) * 2005-11-18 2010-11-04 富士フイルム株式会社 Imaging apparatus and imaging method
JP5045125B2 (en) * 2006-03-15 2012-10-10 株式会社ニコン Subject tracking device and optical device
US20080278589A1 (en) * 2007-05-11 2008-11-13 Karl Ola Thorn Methods for identifying a target subject to automatically focus a digital camera and related systems, and computer program products

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
English Machine Translation of Toyoichi et al. (JP 2005341449 A) *

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20130044251A1 (en) * 2008-02-04 2013-02-21 Samsung Electronics Co., Ltd. Digital image processing apparatus and method of controlling the same
US9094599B2 (en) * 2008-02-04 2015-07-28 Samsung Electronics Co., Ltd. Digital image processing apparatus and method of controlling the same
US20150271387A1 (en) * 2008-02-04 2015-09-24 Samsung Electronics Co., Ltd. Digital image processing apparatus and method of controlling the same
US9374520B2 (en) * 2008-02-04 2016-06-21 Samsung Electronics Co., Ltd. Digital image processing apparatus and method of controlling the same
US20190297265A1 (en) * 2018-03-21 2019-09-26 Sawah Innovations Inc. User-feedback video stabilization device and method

Also Published As

Publication number Publication date
US20080002028A1 (en) 2008-01-03
TW200808044A (en) 2008-02-01
HK1111032A1 (en) 2008-07-25
KR20080002667A (en) 2008-01-04
JP2008009263A (en) 2008-01-17
KR100847614B1 (en) 2008-07-21
CN101098403A (en) 2008-01-02
CN100591099C (en) 2010-02-17
US8284256B2 (en) 2012-10-09
TWI341684B (en) 2011-05-01
JP4761146B2 (en) 2011-08-31

Similar Documents

Publication Publication Date Title
US8284256B2 (en) Imaging apparatus and computer readable recording medium
JP4919160B2 (en) Imaging apparatus and program thereof
TWI326553B (en) Imaging apparatus, data extraction method, and data extraction program
JP4943721B2 (en) Color noise removal method for image data and imaging apparatus using the method
JP5782813B2 (en) Imaging apparatus and image display method
US10334336B2 (en) Method of controlling digital photographing apparatus and digital photographing apparatus using the same
TWI492618B (en) Image pickup device and computer readable recording medium
JP2013098792A (en) Imaging device and control method for imaging device
JP2003262786A (en) Imaging device and its automatic focusing method
JP4740074B2 (en) Imaging apparatus and imaging method
JP2005072963A (en) Imaging device and its control method
JP4948011B2 (en) Imaging apparatus, control method therefor, computer program, and storage medium
JP2005278003A (en) Image processing apparatus
JP2006094200A (en) Imaging apparatus, focusing display method, and program
WO2022181087A1 (en) Image processing device and method, and program
JP2009010746A (en) Image recording and reproducing apparatus
JP4398346B2 (en) Camera, photographing sensitivity control method, shutter speed control method, and photographing sensitivity control program
JP2010028273A (en) Imaging apparatus, display control method, and program
JP2008085688A (en) Image pickup device, method and program
JP2007166447A (en) Imaging apparatus, zoom display method, and program
JP2002281374A (en) Electronic camera
JP2005215206A (en) Imaging unit, imaging method, program, and storage medium
JP2006253939A (en) Imaging apparatus
JP2000324433A (en) Electronic camera

Legal Events

Date Code Title Description
STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION