US20130002922A1 - Imaging apparatus and computer readable recording medium - Google Patents
Imaging apparatus and computer readable recording medium Download PDFInfo
- Publication number
- US20130002922A1 US20130002922A1 US13/610,277 US201213610277A US2013002922A1 US 20130002922 A1 US20130002922 A1 US 20130002922A1 US 201213610277 A US201213610277 A US 201213610277A US 2013002922 A1 US2013002922 A1 US 2013002922A1
- Authority
- US
- United States
- Prior art keywords
- subject
- frame
- imaging apparatus
- moving picture
- section
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
- 238000003384 imaging method Methods 0.000 title claims abstract description 38
- 238000000034 method Methods 0.000 claims abstract description 33
- 230000008569 process Effects 0.000 claims abstract description 30
- 238000009966 trimming Methods 0.000 claims description 59
- 230000006870 function Effects 0.000 claims description 6
- 238000004590 computer program Methods 0.000 claims description 5
- 238000009432 framing Methods 0.000 claims description 5
- 230000015654 memory Effects 0.000 description 19
- 241000251468 Actinopterygii Species 0.000 description 10
- 238000005070 sampling Methods 0.000 description 9
- 230000008859 change Effects 0.000 description 8
- 238000010586 diagram Methods 0.000 description 5
- 230000001276 controlling effect Effects 0.000 description 3
- 230000006835 compression Effects 0.000 description 2
- 238000007906 compression Methods 0.000 description 2
- 230000002596 correlated effect Effects 0.000 description 2
- 230000006837 decompression Effects 0.000 description 2
- 238000012986 modification Methods 0.000 description 2
- 230000004048 modification Effects 0.000 description 2
- 230000003287 optical effect Effects 0.000 description 2
- 238000001454 recorded image Methods 0.000 description 2
- 238000006243 chemical reaction Methods 0.000 description 1
- 230000000875 corresponding effect Effects 0.000 description 1
- 238000001514 detection method Methods 0.000 description 1
- 230000007246 mechanism Effects 0.000 description 1
- 238000006467 substitution reaction Methods 0.000 description 1
- 230000009182 swimming Effects 0.000 description 1
- 230000003936 working memory Effects 0.000 description 1
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/60—Control of cameras or camera modules
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/60—Control of cameras or camera modules
- H04N23/61—Control of cameras or camera modules based on recognised objects
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/60—Control of cameras or camera modules
- H04N23/63—Control of cameras or camera modules by using electronic viewfinders
- H04N23/633—Control of cameras or camera modules by using electronic viewfinders for displaying additional information relating to control or operation of the camera
- H04N23/635—Region indicators; Field of view indicators
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/60—Control of cameras or camera modules
- H04N23/67—Focus control based on electronic image sensor signals
Definitions
- the present invention relates to an imaging apparatus and a computer readable recording medium storing a computer program for operating an imaging apparatus, including, in particular, a moving picture capturing means for capturing a moving picture of a subject using an image capturing device and a shutter button that is able to be pressed both halfway and fully, wherein the imaging apparatus tracks a moving subject and recognizes a position of the subject.
- the following technology is known as a technology for tracking a subject dynamically in an imaging apparatus such as a digital camera.
- the zooming frame is controlled to track the moving target subject, and the moving image data within the zooming frame is stored.
- This technology makes it possible to have the zooming frame track the moving target subject.
- the case in which the moving target subject goes out of the frame is unexpected.
- the zooming frame is located on the edge of the image. Therefore, even if a user slightly changes the camera angle so that the zooming frame is set to the target subject, the target subject will go out of the frame again in a moment. Meanwhile, once the tracking has started, the user can not change the tracking target without stopping recording mode.
- this known technology has problems in usability.
- an imaging apparatus includes: moving picture capturing means for capturing a moving picture of a subject using an image capturing device; first display controlling means for displaying frame data captured in sequence by the moving picture capturing means on a display and for displaying a remarkable point on a predetermined position of the displayed frame data; a shutter button that is adapted to be pressed both halfway and fully; second display controlling means for, when the shutter button is pressed halfway, updating a display position of the remarkable point to a position of the subject so as to track the subject with the remarkable point; and third display controlling means for resetting the display position of the remarkable point to the predetermined position when the halfway-press of the shutter button is released after the shutter button is pressed halfway.
- the remarkable point may be a remarkable area.
- a computer readable recording medium stores a computer program, which is executable by a computer to cause the computer to operate an imaging apparatus, wherein the imaging apparatus includes moving picture capturing means for capturing a moving picture of a subject using an image capturing device and a shutter button that is adapted to be pressed both halfway and fully.
- the computer program causes the computer to execute functions including: displaying frame data captured in sequence by the moving picture capturing means on a display and displaying a remarkable point on a predetermined position of the displayed frame data; updating a display position of the remarkable point to a position of the subject so as to track the subject with the remarkable point when the shutter button is pressed halfway; and resetting the display position of the remarkable point when the halfway-press of the shutter button is released after the shutter button is pressed halfway.
- FIG. 1 is a block diagram showing a digital camera as one embodiment of this invention.
- FIG. 2 is a flowchart for explaining operation of a digital camera of this embodiment.
- FIG. 3 is a diagram illustrating a tracking frame and a trimming frame displayed by the image display block.
- FIG. 4 is a diagram illustrating a tracking frame and a trimming frame displayed by the image display block.
- FIG. 5 is a diagram illustrating a tracking frame and a trimming frame displayed by the image display block.
- FIG. 1 is an electrical block diagram of the digital camera 1 , which realizes one embodiment of an imaging apparatus of this invention.
- the digital camera 1 includes a lens unit 2 , a lens driving block 3 , an aperture 4 , a Charge Coupled Device (CCD) 5 , a driver 6 , a Timing Generator (TG) 7 , a sampling converter 8 , an image signal generator 9 , a Central Processing Unit (CPU) 10 , a key input block 11 , a memory 12 , a Dynamic Random Access Memory (DRAM) 13 , a flash memory 14 , an image display block 15 , and a bus 16 .
- CPU Central Processing Unit
- DRAM Dynamic Random Access Memory
- the lens unit 2 includes a focus lens group 2 a, and a zoom lens 2 b.
- the lens driving block 3 is connected to the lens unit 2 .
- the lens driving block 3 includes a focus motor which drives the focus lens group 2 a along the optical axis of the lens, a zoom motor which drives the zoom lens 2 b along the optical axis of the lens, and a focus motor driver and a zoom motor driver which drive the focus motor and the zoom motor respectively.
- the aperture 4 includes a driving circuit, which drives the aperture 4 by a control signal sent from the CPU 10 .
- the aperture 4 is a mechanism which controls the amount of light entering through the lens unit 2 .
- the CCD 5 which is driven by the driver 6 , outputs a video signal to the sampling converter 8 , by executing photoelectric conversion with respect to the intensity of light for each RGB color of the target subject in a constant period.
- the operating timing of the vertical driver 6 and the sampling converter 8 are controlled by the CPU 10 via the TG 7 .
- the CCD 5 has Bayer Pattern Color Filter and a function as an electronic shutter. The shutter speed of the electronic shutter is controlled by the CPU 10 via the driver 6 and the TG 7 .
- the sampling converter 8 is connected to the TG 7 and includes a Correlated Double Sampling (CDS) Circuit, which executes Correlated Double Sampling on the imaging signal outputted from the CCD 5 and stores the result, an Automatic Gain Control (AGC) Circuit, which executes Automatic Gain Control on the imaging signal after the sampling, an A/D Converter, which changes the analog imaging signal after the Automatic Gain Control into a digital signal.
- CDS Correlated Double Sampling
- AGC Automatic Gain Control
- the imaging signal outputted from the CCD 5 is sent to the image signal generator 9 as a digital signal via the sampling converter 8 .
- the image signal generator 9 executes gamma correction, a white balance process and other processes on the imaging signal sent from the sampling converter 8 and generates a luminance/chrominance signal (YUV data).
- the image data of the generated luminance/chrominance signal is sent to CPU 10 .
- the image signal generator 9 executes image processing on the image data outputted from the CCD 5 .
- the CPU 10 executes compression/decompression processing (for example the compression/decompression processing of JPEG/MPEG format) for the image data sent from the image signal generator 9 , Auto Focus (AF) processing, and imaging processing.
- the CPU 10 is a one-chip microcomputer which controls each part of the digital camera 1 and has functions of image-recognition processing for the main target subject, position detection processing for a main target subject and display processing for a tracking frame and a trimming frame.
- the key input block 11 includes several operation keys such as a shutter button that can be pressed halfway (“halfway-press”) and fully (“full press”), a mode select key, a menu key, an arrow key, a SET key and other keys, and outputs a manipulation signal in accordance with a user's key operation to the CPU 10 .
- the memory 12 stores a control program which is needed for the CPU 10 to control each part, as well as necessary data, and the CPU 10 works in accordance with the program.
- the DRAM 13 is used as a buffer memory which temporarily stores the image data, which is sent from the CPU 10 after capturing by the CCD 5 .
- the DRAM 13 is also used as a working memory of the CPU 10 .
- the flash memory 14 is a recording medium which stores compressed image data.
- the image display block 15 includes a color LCD and a driving circuit for the LCD.
- the image display block 15 displays a target subject captured by CCD 5 as a through image in an imaging standby mode, and the image display block 15 displays a recorded image which is read out from the flash memory 14 and decompressed in a recorded image reproducing mode.
- the CPU 10 starts a through image displaying process. That is, the CPU 10 starts a process in which: the CCD 5 captures the target subject at a predetermined frame rate, the CPU 10 makes the buffer memory (the DRAM 13 ) store frame image data of luminance and color-difference signal which is captured in sequence by the CCD 5 and processed by the image signal generator 9 , and the CPU 10 displays the sequentially stored image data using the image display block 15 (Step S 1 ). At that time the CPU 10 displays the tracking frame 21 (Observed Position) at a predetermined position of the image display block 15 (here, it is assumed that the predetermined position is the center position) based on predetermined position information which is stored in the memory 12 .
- the tracking frame 21 Observed Position
- the CPU 10 judges whether the shutter button is pressed halfway or not (Step S 2 ). The judgment depends on whether or not an operation signal corresponding to the halfway-press of the shutter button is sent from the key input block 11 .
- the tracking frame tracks the main subject if a user moves the digital camera so that the main subject, which the user wants to capture, overlaps the tracking frame 21 , and presses the shutter button halfway when the tracking frame 21 overlaps the main subject.
- step S 3 the CPU 10 displays the trimming frame 22 based on the current display position of the tracking frame (the same as the predetermined position at first) (Step S 3 ). In other words, the position of the tracking frame corresponds to the trimming position and therefore the trimming frame is displayed so that the tracking frame is in the center of the trimming frame. Meanwhile, the size of trimming (the size of the trimming frame) is determined previously here, however, it is possible to change the size arbitrarily by an operation of the zoom key.
- FIG. 3A shows the screen displayed by the image display block 15 when the shutter button is pressed halfway.
- the tracking frame 21 and the trimming frame 22 are displayed in the center of the screen, and the tracking frame 21 is displayed at the center of the trimming frame 22 . Since a fish is overlapped with the tracking frame 21 in this example, the fish is set as the main subject for capturing and the tracking frame tracks the fish automatically.
- the CPU 10 recognizes the subject, which is overlapped with the tracking frame 21 , in the captured image based on the frame image data which is captured by the CCD 5 when the shutter button is pressed halfway (Step S 4 ).
- the subject that is located on the predetermined position is recognized.
- the recognition is performed with an image recognition process and the subject in the predetermined position is recognized.
- the fish which is located on the tracking frame 21 is recognized when the shutter button is pressed halfway under the condition demonstrated in the FIG. 3A .
- An explanation of the image recognition process is omitted since it is a publicly known technology.
- the CPU 10 gets new frame image data captured by the CCD 5 (Step S 5 ) and performs a process for detecting where the recognized main subject (the fish in this example) is in the captured image (Step S 6 ).
- the position is detected where the recognized main subject is.
- the detected position data is stored into position data memory in the buffer memory (DRAM 13 ), and if a position data is already stored in the position data memory, the stored data is refreshed by replacing the stored data with the new detected position data.
- the image recognition process for the newly captured frame image data it is possible to apply the process to the whole image data, or it is also possible to apply the process only to image data surrounding the previous position of the tracking frame (the position of the main subject). Thus, it is possible to reduce the process load of the image recognition process.
- CPU 10 judges whether the location of the main target subject is detected or not, that is to say, judges whether a new position data is stored into the position data memory (Step S 7 ). If a new location of the main target subject is detected, CPU 10 updates the display positions of the tracking frame 21 and the trimming frame 22 based on the stored position data. Then the frame image data is displayed on the image display block 15 together with the tracking frame 21 and the trimming frame 22 (Step S 8 ). If the position data of the main target subject is not changed before-and-after the update, the display positions of the tracking frame 21 and the trimming frame 22 are not changed.
- the trimming frame 22 is controlled to move within a field angle of the captured image (sometimes referred to herein as a captured image angle or image angle), and never runs off the edge of the field angle.
- Step S 7 if it is determined in Step S 7 that the location of the main target subject cannot be detected, CPU 10 displays a message that the target subject has gone out of the frame, along with the frame image data, and goes on to step S 10 (Step S 9 ).
- the stored data in the position data memory is not updated, the display of the tracking frame 21 and the trimming frame 22 is not updated, and the tracking frame 21 and the trimming frame 22 are displayed on the same position as before.
- step S 10 it is judged whether the shutter button is fully pressed or not. If the shutter button isn't fully pressed, it is judged whether the halfway-press of the shutter button is released or not (Step S 11 ). If the halfway-press is not released, the process performed by the CPU 10 goes back to step S 5 , in which the newly captured frame image data is acquired and the above mentioned process (Steps S 5 to S 11 ) is repeated.
- FIG. 4 and FIG. 5 show examples of displaying the tracking frame 21 and the trimming frame 22 during the loop from step S 5 to step S 11 .
- FIGS. 4 and 5 show the positioning of the tracking frame 21 and trimming frame 22 which are displayed when the fish, swimming toward the lower left, is captured.
- the tracking frame 21 and the trimming frame 22 are moved toward the lower left together with the moving of the main subject, the fish, from the image shown in FIG. 3A , in which the shutter button is pressed halfway, to FIG. 5C via FIG. 4A and FIG. 4B .
- the left side of the trimming frame 22 is bounded by the captured image angle. Therefore, the trimming frame 22 doesn't move any farther to the left even if the fish moves still more left.
- the trimming frame 22 moves within the range of the image angle, following the position of the tracking frame 21 and the main target subject. For example, the trimming frame 22 doesn't move upward any more if the upper side of the trimming frame 22 is bounded by the image angle.
- Step S 7 the notice that the target subject has gone out of the frame is displayed and the tracking frame 21 and the trimming frame 22 are displayed on the same position as before. In other words, when the target subject leaves the frame, the tracking frame 21 (and the trimming frame 22 ) does not move and is displayed without change.
- step S 11 if it is judged that the halfway-press is released in step S 11 , the CPU 10 displays the tracking frame 21 on the predetermined position (the center position) based on the predetermined position information which is stored in the memory 12 , and goes back to step S 2 (Step S 12 ). In this case, the detected position information which is stored in the position data memory in the buffer memory is cleared.
- FIG. 3B shows the screen image that is displayed on the image display block 15 when the halfway-press of the shutter button is released.
- the tracking frame 21 doesn't track the target subject and is displayed in the center of the screen, as shown in FIG. 3B .
- CPU 10 stops displaying the trimming frame 22 . If a user wants to change the main target subject or the main target subject goes out of the frame, the user can simply release the halfway-press of the shutter button. As a result of this operation, the track frame 21 returns to its normal location (e.g., the center of the image in this embodiment). Therefore, the user can restart the operation to track the main target subject from the beginning.
- the tracking frame 21 is left displayed on the edge of the image when the main target subject goes out of the frame ( FIG. 5E ).
- the main target subject is likely to go out of the frame again quickly.
- step S 10 If it is judged that the shutter button is fully pressed in step S 10 , the CPU 10 performs AF processing within a given AF area based on the position of the tracking frame (the position information stored in the position data memory) (Step S 13 ). Next the CPU 10 performs the capturing process of a still image (Step S 14 ), and performs the trimming process on the still image which is captured in the capturing process using the position of the tracking frame 21 just before the full press of the shutter button (the position information stored in the position data memory) and the predetermined trimming size (Step S 15 ).
- the CPU 10 displays, as a preview display, the image data trimmed by the trimming process (Trimming Image Data) using the image display block 15 (Step S 16 ), and judges whether the trimming image data is to be stored or not (Step S 17 ).
- a screen display for assigning whether the trimming image data is to be stored or not is displayed together with the preview display. If the user assigns the image to be stored by an operation of the key input block 11 , it is determined that the image will be stored. If the user assigns not to store the image, it is determined that the image will not be stored. If it is judged in step S 17 that the image will be stored, the trimming image data is compressed and stored into the flash memory 14 (Step S 18 ).
- the through image is displayed with a tracking frame 21 on the predetermined position.
- a shutter button is pressed halfway when the tracking frame 21 overlaps the main subject which a user wants to capture, the subject, which is overlapped with the tracking frame 21 , is recognized.
- the tracking frame 21 is displayed and tracked to the recognized main target subject. If the halfway-press of the shutter button is released, the tracking frame 21 is displayed on the predetermined position.
- the present invention allows the present invention to deal with the case in which the main target subject goes out of the frame or the case in which a user wants to change the main target subject, by releasing the halfway-press of the shutter button, thereby releasing the tracking of the subject with the tracking frame 21 .
- the tracking frame is still displayed if the halfway-press of the shutter button is released, and is moved to the predetermined position (e.g., the center of the screen). Therefore, the tracking frame 21 can be overlapped on the main target subject while the tracking frame is in the predetermined position, and it is unnecessary to overlap the tracking frame 21 to the main target subject while the tracking frame 21 is on the side of the screen.
- the AF processing and the trimming processing are performed based on the position of the tracking frame 21 .
- the AF area and/or the trimming area, which is displayed as a substitution for the tracking frame 21 are generically named “Remarkable Area”.
- the AF processing is performed after the full press of the shutter button.
- Step S 9 a message that the target subject has gone out of the frame (framing out) is displayed (Step S 9 ).
- the user can recognize the framing out of the target subject by watching the image display block 15 .
- the main subject which is overlapped with the tracking frame 21 , is recognized by using an image recognition process in the step S 4 and it is detected where the recognized main subject is in the newly captured frame image data.
- the digital camera 1 described above is not limited to the above mentioned embodiment.
- This invention may be used in various devices such as mobile terminal devices with cameras, personal computers with cameras, IC recorders with cameras, and/or digital video cameras.
- the present invention may be practiced in any device having a function for capturing an image of a subject.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Studio Devices (AREA)
- Focusing (AREA)
- Automatic Focus Adjustment (AREA)
Abstract
Once an imaging mode is set, an imaging apparatus displays a through image with a tracking frame on the predetermined position. If a shutter button is pressed halfway, an object that is overlapped by the tracking frame is recognized as a main subject that a user wants to capture. Until the shutter button is fully pressed or until the halfway-press is released, it is detected where the recognized subject is in frame image data that is captured in sequence, and the tracking frame is displayed based on the detected position. If the halfway press of the shutter button is released before the shutter button is fully pressed, the tracking frame is displayed on the predetermined position. Then if the shutter button is pressed halfway once again, the process is repeated.
Description
- This is a Divisional of U.S. application Ser. No. 11/770,114, filed Jun. 28, 2007, which is based on Japanese Patent Application No. 2006-181470, filed Jun. 30, 2006, the entire contents of both of which are incorporated herein by reference.
- 1. Field of the Invention
- The present invention relates to an imaging apparatus and a computer readable recording medium storing a computer program for operating an imaging apparatus, including, in particular, a moving picture capturing means for capturing a moving picture of a subject using an image capturing device and a shutter button that is able to be pressed both halfway and fully, wherein the imaging apparatus tracks a moving subject and recognizes a position of the subject.
- 2. Description of the Related Art
- The following technology is known as a technology for tracking a subject dynamically in an imaging apparatus such as a digital camera. In this technology, if a moving target subject overlaps the zooming frame when capturing a moving image, the zooming frame is controlled to track the moving target subject, and the moving image data within the zooming frame is stored.
- This technology makes it possible to have the zooming frame track the moving target subject. However, the case in which the moving target subject goes out of the frame is unexpected. When the moving target subject goes out of the frame, the zooming frame is located on the edge of the image. Therefore, even if a user slightly changes the camera angle so that the zooming frame is set to the target subject, the target subject will go out of the frame again in a moment. Meanwhile, once the tracking has started, the user can not change the tracking target without stopping recording mode. Thus, this known technology has problems in usability.
- According to one aspect of the invention, an imaging apparatus includes: moving picture capturing means for capturing a moving picture of a subject using an image capturing device; first display controlling means for displaying frame data captured in sequence by the moving picture capturing means on a display and for displaying a remarkable point on a predetermined position of the displayed frame data; a shutter button that is adapted to be pressed both halfway and fully; second display controlling means for, when the shutter button is pressed halfway, updating a display position of the remarkable point to a position of the subject so as to track the subject with the remarkable point; and third display controlling means for resetting the display position of the remarkable point to the predetermined position when the halfway-press of the shutter button is released after the shutter button is pressed halfway.
- According to the present invention, the remarkable point may be a remarkable area.
- According to a further aspect of the invention, a computer readable recording medium stores a computer program, which is executable by a computer to cause the computer to operate an imaging apparatus, wherein the imaging apparatus includes moving picture capturing means for capturing a moving picture of a subject using an image capturing device and a shutter button that is adapted to be pressed both halfway and fully. The computer program causes the computer to execute functions including: displaying frame data captured in sequence by the moving picture capturing means on a display and displaying a remarkable point on a predetermined position of the displayed frame data; updating a display position of the remarkable point to a position of the subject so as to track the subject with the remarkable point when the shutter button is pressed halfway; and resetting the display position of the remarkable point when the halfway-press of the shutter button is released after the shutter button is pressed halfway.
- The accompanying drawings illustrate preferred embodiments of an imaging apparatus according to the invention, and together with the detailed description of the preferred embodiments given below serve to explain the principles of the invention.
-
FIG. 1 is a block diagram showing a digital camera as one embodiment of this invention. -
FIG. 2 is a flowchart for explaining operation of a digital camera of this embodiment. -
FIG. 3 is a diagram illustrating a tracking frame and a trimming frame displayed by the image display block. -
FIG. 4 is a diagram illustrating a tracking frame and a trimming frame displayed by the image display block. -
FIG. 5 is a diagram illustrating a tracking frame and a trimming frame displayed by the image display block. - Now, an embodiment of the invention will be described in detail with reference to the accompanying drawings using an example in which an imaging apparatus of this invention is applied to a digital camera.
-
FIG. 1 is an electrical block diagram of the digital camera 1, which realizes one embodiment of an imaging apparatus of this invention. The digital camera 1 includes alens unit 2, alens driving block 3, anaperture 4, a Charge Coupled Device (CCD) 5, adriver 6, a Timing Generator (TG) 7, asampling converter 8, animage signal generator 9, a Central Processing Unit (CPU) 10, akey input block 11, amemory 12, a Dynamic Random Access Memory (DRAM) 13, aflash memory 14, animage display block 15, and abus 16. - The
lens unit 2 includes afocus lens group 2 a, and azoom lens 2 b. Thelens driving block 3 is connected to thelens unit 2. Thelens driving block 3 includes a focus motor which drives thefocus lens group 2 a along the optical axis of the lens, a zoom motor which drives thezoom lens 2 b along the optical axis of the lens, and a focus motor driver and a zoom motor driver which drive the focus motor and the zoom motor respectively. - The
aperture 4 includes a driving circuit, which drives theaperture 4 by a control signal sent from theCPU 10. Theaperture 4 is a mechanism which controls the amount of light entering through thelens unit 2. - The
CCD 5, which is driven by thedriver 6, outputs a video signal to thesampling converter 8, by executing photoelectric conversion with respect to the intensity of light for each RGB color of the target subject in a constant period. The operating timing of thevertical driver 6 and thesampling converter 8 are controlled by theCPU 10 via theTG 7. Meanwhile, theCCD 5 has Bayer Pattern Color Filter and a function as an electronic shutter. The shutter speed of the electronic shutter is controlled by theCPU 10 via thedriver 6 and theTG 7. - The
sampling converter 8 is connected to theTG 7 and includes a Correlated Double Sampling (CDS) Circuit, which executes Correlated Double Sampling on the imaging signal outputted from theCCD 5 and stores the result, an Automatic Gain Control (AGC) Circuit, which executes Automatic Gain Control on the imaging signal after the sampling, an A/D Converter, which changes the analog imaging signal after the Automatic Gain Control into a digital signal. The imaging signal outputted from theCCD 5 is sent to theimage signal generator 9 as a digital signal via thesampling converter 8. - The
image signal generator 9 executes gamma correction, a white balance process and other processes on the imaging signal sent from thesampling converter 8 and generates a luminance/chrominance signal (YUV data). The image data of the generated luminance/chrominance signal is sent toCPU 10. In other words, theimage signal generator 9 executes image processing on the image data outputted from theCCD 5. - The
CPU 10 executes compression/decompression processing (for example the compression/decompression processing of JPEG/MPEG format) for the image data sent from theimage signal generator 9, Auto Focus (AF) processing, and imaging processing. TheCPU 10 is a one-chip microcomputer which controls each part of the digital camera 1 and has functions of image-recognition processing for the main target subject, position detection processing for a main target subject and display processing for a tracking frame and a trimming frame. - The
key input block 11 includes several operation keys such as a shutter button that can be pressed halfway (“halfway-press”) and fully (“full press”), a mode select key, a menu key, an arrow key, a SET key and other keys, and outputs a manipulation signal in accordance with a user's key operation to theCPU 10. Thememory 12 stores a control program which is needed for theCPU 10 to control each part, as well as necessary data, and theCPU 10 works in accordance with the program. - The
DRAM 13 is used as a buffer memory which temporarily stores the image data, which is sent from theCPU 10 after capturing by theCCD 5. TheDRAM 13 is also used as a working memory of theCPU 10. Theflash memory 14 is a recording medium which stores compressed image data. - The
image display block 15 includes a color LCD and a driving circuit for the LCD. Theimage display block 15 displays a target subject captured byCCD 5 as a through image in an imaging standby mode, and theimage display block 15 displays a recorded image which is read out from theflash memory 14 and decompressed in a recorded image reproducing mode. - The operation of the digital camera 1 in this embodiment will be described with reference to
FIG. 2 . - Once the imaging mode is set by the user's operation of the mode select key in the
key input block 11, theCPU 10 starts a through image displaying process. That is, theCPU 10 starts a process in which: theCCD 5 captures the target subject at a predetermined frame rate, theCPU 10 makes the buffer memory (the DRAM 13) store frame image data of luminance and color-difference signal which is captured in sequence by theCCD 5 and processed by theimage signal generator 9, and theCPU 10 displays the sequentially stored image data using the image display block 15 (Step S1). At that time theCPU 10 displays the tracking frame 21 (Observed Position) at a predetermined position of the image display block 15 (here, it is assumed that the predetermined position is the center position) based on predetermined position information which is stored in thememory 12. - Then the
CPU 10 judges whether the shutter button is pressed halfway or not (Step S2). The judgment depends on whether or not an operation signal corresponding to the halfway-press of the shutter button is sent from thekey input block 11. According to this embodiment, the tracking frame tracks the main subject if a user moves the digital camera so that the main subject, which the user wants to capture, overlaps thetracking frame 21, and presses the shutter button halfway when thetracking frame 21 overlaps the main subject. - If the
CPU 10 judges that the shutter button isn't pressed halfway in step S2, it remains in step S2 until the shutter button is pressed halfway. If theCPU 10 judges that the shutter button is pressed halfway, theCPU 10 displays the trimmingframe 22 based on the current display position of the tracking frame (the same as the predetermined position at first) (Step S3). In other words, the position of the tracking frame corresponds to the trimming position and therefore the trimming frame is displayed so that the tracking frame is in the center of the trimming frame. Meanwhile, the size of trimming (the size of the trimming frame) is determined previously here, however, it is possible to change the size arbitrarily by an operation of the zoom key. -
FIG. 3A shows the screen displayed by theimage display block 15 when the shutter button is pressed halfway. As shown inFIG. 3A , thetracking frame 21 and the trimmingframe 22 are displayed in the center of the screen, and thetracking frame 21 is displayed at the center of the trimmingframe 22. Since a fish is overlapped with thetracking frame 21 in this example, the fish is set as the main subject for capturing and the tracking frame tracks the fish automatically. - The
CPU 10 recognizes the subject, which is overlapped with thetracking frame 21, in the captured image based on the frame image data which is captured by theCCD 5 when the shutter button is pressed halfway (Step S4). In this example, since thetracking frame 21 is located on the predetermined position, the subject that is located on the predetermined position is recognized. The recognition is performed with an image recognition process and the subject in the predetermined position is recognized. For example, the fish which is located on thetracking frame 21 is recognized when the shutter button is pressed halfway under the condition demonstrated in theFIG. 3A . An explanation of the image recognition process is omitted since it is a publicly known technology. - Then the
CPU 10 gets new frame image data captured by the CCD 5 (Step S5) and performs a process for detecting where the recognized main subject (the fish in this example) is in the captured image (Step S6). In other words, by performing image recognition of newly captured frame image data, the position is detected where the recognized main subject is. When the position is detected in this step, the detected position data is stored into position data memory in the buffer memory (DRAM 13), and if a position data is already stored in the position data memory, the stored data is refreshed by replacing the stored data with the new detected position data. - Meanwhile, with regard to the image recognition process for the newly captured frame image data, it is possible to apply the process to the whole image data, or it is also possible to apply the process only to image data surrounding the previous position of the tracking frame (the position of the main subject). Thus, it is possible to reduce the process load of the image recognition process.
-
CPU 10 judges whether the location of the main target subject is detected or not, that is to say, judges whether a new position data is stored into the position data memory (Step S7). If a new location of the main target subject is detected,CPU 10 updates the display positions of thetracking frame 21 and the trimmingframe 22 based on the stored position data. Then the frame image data is displayed on theimage display block 15 together with thetracking frame 21 and the trimming frame 22 (Step S8). If the position data of the main target subject is not changed before-and-after the update, the display positions of thetracking frame 21 and the trimmingframe 22 are not changed. Additionally, when updating the display positions of thetracking frame 21 and the trimmingframe 22, the trimmingframe 22 is controlled to move within a field angle of the captured image (sometimes referred to herein as a captured image angle or image angle), and never runs off the edge of the field angle. - On the other hand, if it is determined in Step S7 that the location of the main target subject cannot be detected,
CPU 10 displays a message that the target subject has gone out of the frame, along with the frame image data, and goes on to step S10 (Step S9). In this case, the stored data in the position data memory is not updated, the display of thetracking frame 21 and the trimmingframe 22 is not updated, and thetracking frame 21 and the trimmingframe 22 are displayed on the same position as before. - In step S10, it is judged whether the shutter button is fully pressed or not. If the shutter button isn't fully pressed, it is judged whether the halfway-press of the shutter button is released or not (Step S11). If the halfway-press is not released, the process performed by the
CPU 10 goes back to step S5, in which the newly captured frame image data is acquired and the above mentioned process (Steps S5 to S11) is repeated. -
FIG. 4 andFIG. 5 show examples of displaying thetracking frame 21 and the trimmingframe 22 during the loop from step S5 to step S11. Specifically,FIGS. 4 and 5 show the positioning of thetracking frame 21 and trimmingframe 22 which are displayed when the fish, swimming toward the lower left, is captured. - The
tracking frame 21 and the trimmingframe 22 are moved toward the lower left together with the moving of the main subject, the fish, from the image shown inFIG. 3A , in which the shutter button is pressed halfway, toFIG. 5C viaFIG. 4A andFIG. 4B . InFIG. 5C , the left side of the trimmingframe 22 is bounded by the captured image angle. Therefore, the trimmingframe 22 doesn't move any farther to the left even if the fish moves still more left. Thus, the trimmingframe 22 moves within the range of the image angle, following the position of thetracking frame 21 and the main target subject. For example, the trimmingframe 22 doesn't move upward any more if the upper side of the trimmingframe 22 is bounded by the image angle. - Additionally, if the fish moves still more lower left as shown in
FIG. 5D , the trimmingframe 22 only moves downward and thetracking frame 21 moves together with the moving of the main target subject. And then if the main target subject, the fish, goes out of the frame (a judgment is made in Step S7 that the position of the target subject cannot be detected), the notice that the target subject has gone out of the frame is displayed and thetracking frame 21 and the trimmingframe 22 are displayed on the same position as before. In other words, when the target subject leaves the frame, the tracking frame 21 (and the trimming frame 22) does not move and is displayed without change. - Going back to the flowchart shown in
FIG. 2 , if it is judged that the halfway-press is released in step S11, theCPU 10 displays thetracking frame 21 on the predetermined position (the center position) based on the predetermined position information which is stored in thememory 12, and goes back to step S2 (Step S12). In this case, the detected position information which is stored in the position data memory in the buffer memory is cleared.FIG. 3B shows the screen image that is displayed on theimage display block 15 when the halfway-press of the shutter button is released. - When the halfway-press of the shutter button is released, the
tracking frame 21 doesn't track the target subject and is displayed in the center of the screen, as shown inFIG. 3B . In addition, when the shutter button is released from the halfway-press,CPU 10 stops displaying the trimmingframe 22. If a user wants to change the main target subject or the main target subject goes out of the frame, the user can simply release the halfway-press of the shutter button. As a result of this operation, thetrack frame 21 returns to its normal location (e.g., the center of the image in this embodiment). Therefore, the user can restart the operation to track the main target subject from the beginning. - By contrast, if, for example the position of the
tracking frame 21 is not changed when the halfway-press of the shutter button is released, thetracking frame 21 is left displayed on the edge of the image when the main target subject goes out of the frame (FIG. 5E ). Thus, if a user tries to make thetracking frame 21 track the main target subject by overlapping thetracking frame 21, which is displayed on edge, with the main target subject and pressing the shutter button halfway, the main target subject is likely to go out of the frame again quickly. - In addition, if the position of the
tracking frame 21 is not changed when the halfway-press of the shutter button is released, if a user wants to change the main target subject, it is necessary for the user to release the halfway-press when thetracking frame 21 is displayed not on the edge of the image but rather when the tracking frame is displayed on a position which is convenient for capturing and tracking for the main target subject so as to eliminate the above problem. Thus, it is impossible to deal with the case in which the user wants to capture the image immediately. Therefore it is desirable to solve the above problem. In the aspect of this invention described above, it is possible to cope well with the above problem by displaying thetracking frame 21 on the predetermined position (e.g., the center of the image) when the halfway-press of the shutter button is released. - If it is judged that the shutter button is fully pressed in step S10, the
CPU 10 performs AF processing within a given AF area based on the position of the tracking frame (the position information stored in the position data memory) (Step S13). Next theCPU 10 performs the capturing process of a still image (Step S14), and performs the trimming process on the still image which is captured in the capturing process using the position of thetracking frame 21 just before the full press of the shutter button (the position information stored in the position data memory) and the predetermined trimming size (Step S15). - Next the
CPU 10 displays, as a preview display, the image data trimmed by the trimming process (Trimming Image Data) using the image display block 15 (Step S16), and judges whether the trimming image data is to be stored or not (Step S17). At that time, a screen display for assigning whether the trimming image data is to be stored or not is displayed together with the preview display. If the user assigns the image to be stored by an operation of thekey input block 11, it is determined that the image will be stored. If the user assigns not to store the image, it is determined that the image will not be stored. If it is judged in step S17 that the image will be stored, the trimming image data is compressed and stored into the flash memory 14 (Step S18). - As described above, in this embodiment, once the imaging mode is set, the through image is displayed with a
tracking frame 21 on the predetermined position. And if a shutter button is pressed halfway when thetracking frame 21 overlaps the main subject which a user wants to capture, the subject, which is overlapped with thetracking frame 21, is recognized. As long as the shutter button is pressed halfway, thetracking frame 21 is displayed and tracked to the recognized main target subject. If the halfway-press of the shutter button is released, thetracking frame 21 is displayed on the predetermined position. This allows the present invention to deal with the case in which the main target subject goes out of the frame or the case in which a user wants to change the main target subject, by releasing the halfway-press of the shutter button, thereby releasing the tracking of the subject with thetracking frame 21. For example, if the main target subject goes out of the frame, the tracking frame is still displayed if the halfway-press of the shutter button is released, and is moved to the predetermined position (e.g., the center of the screen). Therefore, thetracking frame 21 can be overlapped on the main target subject while the tracking frame is in the predetermined position, and it is unnecessary to overlap thetracking frame 21 to the main target subject while thetracking frame 21 is on the side of the screen. - It is possible to change the above mentioned embodiment using the following modifications, for example.
- (1) In the above mentioned embodiment, the AF processing and the trimming processing are performed based on the position of the
tracking frame 21. However, it is possible to select one of the processing, the AF processing and the trimming processing. In this case, it is possible to display either the AF area or the trimming area only instead of the tracking frame. The AF area and/or the trimming area, which is displayed as a substitution for thetracking frame 21, are generically named “Remarkable Area”. - (2) In the above mentioned embodiment, the AF processing is performed after the full press of the shutter button. However, it is possible to perform the AF processing after the halfway-press of the shutter button using the AF area based on the position of the tracking frame (the predetermined position).
- (3) In the above mentioned embodiment, if the target subject goes out of the frame (it is judged that the location of the main target subject can't be detected in step S7), a message that the target subject has gone out of the frame (framing out) is displayed (Step S9). However, it is possible not to display the message of the framing out of the target subject, because the user can recognize the framing out of the target subject by watching the
image display block 15. - (4) In the above mentioned embodiment, the main subject, which is overlapped with the
tracking frame 21, is recognized by using an image recognition process in the step S4 and it is detected where the recognized main subject is in the newly captured frame image data. However, it is possible to detect the position of the main subject in the image by using a technique which detects the motion vector of the main subject, instead of the image recognition process. It is sufficient to practice this aspect of the present invention using a method to be able to detect the position of the main subject in the frame image data. - (5) It is possible to use a dot, a circle or an x-mark instead of the
tracking frame 21 in the above embodiment. That is, it is sufficient to provide a method for user to be able to catch the position for detecting the main subject and/or the current position of the recognized main subject. The display of a point, which makes a user catch the position for detecting the main subject and/or the current position of the recognized main subject, for example thetracking frame 21, a dot, a circle and/or an x-mark, are generically called “Remarkable Point”. - (6) Meanwhile, if the shutter button is fully pressed suddenly without staying in the halfway-pressed stage, it is possible to process the AF processing based on the tracking frame on the predetermined position, process the capturing process of a still image, and then store the still image data captured by the image capturing process. In such case, it is also possible to generate a trimming image data of the still image data captured by the image capturing process using the tracking frame, which is in the predetermined position, and store it.
- (7) In addition, it is possible for a user to change the size of the trimming area (trimming size) arbitrarily.
- (8) It is also possible to combine an arbitrary number of the above modifications (1) to (7).
- The digital camera 1 described above is not limited to the above mentioned embodiment. This invention may be used in various devices such as mobile terminal devices with cameras, personal computers with cameras, IC recorders with cameras, and/or digital video cameras. The present invention may be practiced in any device having a function for capturing an image of a subject.
Claims (14)
1. An imaging apparatus comprising:
a moving picture capturing section for capturing a moving picture of a subject using an image capturing device;
a first display controlling section for displaying frame data captured in sequence by the moving picture capturing section on a display; and
a second display controlling section for displaying a remarkable point on a position of the subject so as to track the subject with the remarkable point;
wherein when the subject goes out of the a frame of the moving picture, the second display controlling section continuously displays the remarkable point on a position at which the subject was located immediately before going out of the frame.
2. The imaging apparatus according to claim 1 , wherein when the subject goes out of the frame of the moving picture, the second display controlling section displays a message notifying of the framing out of the subject.
3. The imaging apparatus according to claim 1 , further comprising an auto-focus controlling section for executing an auto-focus process with respect to the subject located on the position of the remarkable point.
4. The imaging apparatus according to claim 3 , further comprising a still image capturing section for capturing a still image of the subject using the image capturing device after the auto-focus process is executed by the auto-focus controlling section.
5. The imaging apparatus according to claim 4 , further comprising:
a trimming frame displaying section for displaying a trimming frame on the display and for displacing a display position of the trimming frame such that the trimming frame tracks the subject in accordance with the tracking performed by the second display controlling section; and
a trimming section for trimming the still image captured by the still image capturing section based on the position of the trimming frame.
6. The imaging apparatus according to claim 1 , further comprising a shutter button that is adapted to be pressed both halfway and fully;
wherein the second display controlling section displays the remarkable point on the position of the subject so as to track the subject with the remarkable point when the shutter button is pressed halfway.
7. An imaging apparatus comprising:
a moving picture capturing section for capturing a moving picture of a subject using an image capturing device;
a first display controlling section for displaying frame data captured in sequence by the moving picture capturing section on a display; and
a second display controlling section for displaying a remarkable area on a position of the subject so as to track the subject with the remarkable area;
wherein when the subject goes out of a frame of the moving picture, the second display controlling section continuously displays the remarkable area on a position at which the subject was located immediately before going out of the frame.
8. The imaging apparatus according to claim 7 , wherein when the subject goes out of the frame of the moving picture, the second display controlling section displays a message notifying of the framing out of the subject.
9. The imaging apparatus according to claim 1 , further comprising an auto-focus controlling section for executing an auto-focus process with respect to the subject located on the position of the remarkable area.
10. The imaging apparatus according to claim 9 , further comprising a still image capturing section for capturing a still image of the subject using the image capturing device after the auto-focus process is executed by the auto-focus controlling section.
11. The imaging apparatus according to claim 10 , further comprising:
a trimming frame displaying section for displaying a trimming frame on the display and for displacing a display position of the trimming frame such that the trimming frame tracks the subject in accordance with the tracking performed by the second display controlling section; and
a trimming section for trimming the still image captured by the still image capturing section based on the position of the trimming frame.
12. The imaging apparatus according to claim 7 , further comprising a shutter button that is adapted to be pressed both halfway and fully;
wherein the second display controlling section displays the remarkable area on the position of the subject so as to track the subject with the remarkable area when the shutter button is pressed halfway.
13. A non-transitory computer readable recording medium storing a computer program which is executable by a computer to cause the computer to operate an imaging apparatus, the imaging apparatus including a moving picture capturing section for capturing a moving picture of a subject using an image capturing device, the program causing the computer to execute functions comprising:
displaying frame data captured in sequence by the moving picture capturing section on a display; and
displaying a remarkable point on a position of the subject so as to track the subject with the remarkable point;
wherein when the subject goes out of a frame of the moving picture, the remarkable point is continuously displayed on a position at which the subject was located immediately before going out of the frame.
14. A non-transitory computer readable recording medium storing a computer program which is executable by a computer to cause the computer to operate an imaging apparatus, the imaging apparatus including a moving picture capturing section for capturing a moving picture of a subject using an image capturing device, the program causing the computer to execute functions comprising:
displaying frame data captured in sequence by the moving picture capturing section on a display; and
displaying a remarkable area on a position of the subject so as to track the subject with the remarkable area;
wherein when the subject goes out of a frame of the moving picture, the remarkable area is continuously displayed on a position at which the subject was located immediately before going out of the frame.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US13/610,277 US20130002922A1 (en) | 2006-06-30 | 2012-09-11 | Imaging apparatus and computer readable recording medium |
Applications Claiming Priority (4)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2006-181470 | 2006-06-30 | ||
JP2006181470A JP4761146B2 (en) | 2006-06-30 | 2006-06-30 | Imaging apparatus and program thereof |
US11/770,114 US8284256B2 (en) | 2006-06-30 | 2007-06-28 | Imaging apparatus and computer readable recording medium |
US13/610,277 US20130002922A1 (en) | 2006-06-30 | 2012-09-11 | Imaging apparatus and computer readable recording medium |
Related Parent Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US11/770,114 Division US8284256B2 (en) | 2006-06-30 | 2007-06-28 | Imaging apparatus and computer readable recording medium |
Publications (1)
Publication Number | Publication Date |
---|---|
US20130002922A1 true US20130002922A1 (en) | 2013-01-03 |
Family
ID=38876171
Family Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US11/770,114 Active 2030-12-05 US8284256B2 (en) | 2006-06-30 | 2007-06-28 | Imaging apparatus and computer readable recording medium |
US13/610,277 Abandoned US20130002922A1 (en) | 2006-06-30 | 2012-09-11 | Imaging apparatus and computer readable recording medium |
Family Applications Before (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US11/770,114 Active 2030-12-05 US8284256B2 (en) | 2006-06-30 | 2007-06-28 | Imaging apparatus and computer readable recording medium |
Country Status (6)
Country | Link |
---|---|
US (2) | US8284256B2 (en) |
JP (1) | JP4761146B2 (en) |
KR (1) | KR100847614B1 (en) |
CN (1) | CN100591099C (en) |
HK (1) | HK1111032A1 (en) |
TW (1) | TWI341684B (en) |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20130044251A1 (en) * | 2008-02-04 | 2013-02-21 | Samsung Electronics Co., Ltd. | Digital image processing apparatus and method of controlling the same |
US20190297265A1 (en) * | 2018-03-21 | 2019-09-26 | Sawah Innovations Inc. | User-feedback video stabilization device and method |
Families Citing this family (32)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2006044476A2 (en) | 2004-10-12 | 2006-04-27 | Robert Vernon Vanman | Method of and system for mobile surveillance and event recording |
JP2008187591A (en) * | 2007-01-31 | 2008-08-14 | Fujifilm Corp | Imaging apparatus and imaging method |
US8599368B1 (en) | 2008-01-29 | 2013-12-03 | Enforcement Video, Llc | Laser-based speed determination device for use in a moving vehicle |
JP4492697B2 (en) * | 2007-12-28 | 2010-06-30 | カシオ計算機株式会社 | Imaging apparatus and program |
JP4958805B2 (en) * | 2008-01-22 | 2012-06-20 | キヤノン株式会社 | Imaging apparatus and display control method thereof |
CA2714362A1 (en) | 2008-01-29 | 2009-08-06 | Enforcement Video, Llc | Omnidirectional camera for use in police car event recording |
KR101445606B1 (en) * | 2008-02-05 | 2014-09-29 | 삼성전자주식회사 | Digital photographing apparatus, method for controlling the same, and recording medium storing program to implement the method |
US20090213218A1 (en) | 2008-02-15 | 2009-08-27 | Andrew Cilia | System and method for multi-resolution storage of images |
KR101436326B1 (en) * | 2008-07-31 | 2014-09-01 | 삼성전자주식회사 | Digital photographing apparatus, method for controlling the same, and recording medium storing program to implement the method |
US8237847B2 (en) * | 2008-10-16 | 2012-08-07 | Fujinon Corporation | Auto focus system having AF frame auto-tracking function |
JP5210841B2 (en) * | 2008-12-12 | 2013-06-12 | パナソニック株式会社 | Imaging device |
US7995909B2 (en) | 2008-12-12 | 2011-08-09 | Samsung Electro-Mechanics Co., Ltd. | Auto-focusing method |
KR100975928B1 (en) * | 2008-12-12 | 2010-08-13 | 삼성전기주식회사 | Auto-focusing method |
KR100975929B1 (en) * | 2008-12-12 | 2010-08-13 | 삼성전기주식회사 | Auto-focusing method |
JP2011139231A (en) * | 2009-12-28 | 2011-07-14 | Sony Corp | Image processing apparatus, image processing method and program |
JP2012049651A (en) | 2010-08-24 | 2012-03-08 | Ricoh Co Ltd | Imaging apparatus and imaging method |
KR101694421B1 (en) * | 2010-11-11 | 2017-01-09 | 삼성전자주식회사 | Method and apparatus for controlling continuous auto focus of mobile terminal comprising camera |
TWI584647B (en) * | 2012-07-12 | 2017-05-21 | Chi Lin Hong | A method and apparatus for preventing defocusing |
US10178298B2 (en) | 2012-12-20 | 2019-01-08 | Sony Corporation | Image processing device, image processing method, and recording medium for optimal trimming of a captured image |
US9165364B1 (en) * | 2012-12-26 | 2015-10-20 | Canon Kabushiki Kaisha | Automatic tracking image pickup system |
KR20140104807A (en) * | 2013-02-21 | 2014-08-29 | 삼성전자주식회사 | Method, apparatus and recording medium for guiding photography shooting using subject-related attributes |
JP6103526B2 (en) * | 2013-03-15 | 2017-03-29 | オリンパス株式会社 | Imaging device, image display device, and display control method for image display device |
JP6223170B2 (en) * | 2013-12-19 | 2017-11-01 | キヤノン株式会社 | Imaging device, control method thereof, and control program |
KR101446781B1 (en) * | 2014-04-21 | 2014-10-01 | 삼성전자주식회사 | Electronic device and controlling method |
JP6641365B2 (en) * | 2015-06-22 | 2020-02-05 | 富士フイルム株式会社 | Radiation irradiation apparatus, control method of radiation irradiation apparatus, and program |
US10341605B1 (en) | 2016-04-07 | 2019-07-02 | WatchGuard, Inc. | Systems and methods for multiple-resolution storage of media streams |
JP6862202B2 (en) | 2017-02-08 | 2021-04-21 | キヤノン株式会社 | Image processing device, imaging device and control method |
WO2019061419A1 (en) * | 2017-09-30 | 2019-04-04 | 深圳市大疆创新科技有限公司 | Photographing apparatus control method, photographing apparatus and storage medium |
JP6506440B2 (en) * | 2018-04-03 | 2019-04-24 | オリンパス株式会社 | Imaging device, imaging method, and program |
WO2020067323A1 (en) * | 2018-09-27 | 2020-04-02 | 富士フイルム株式会社 | Radiation imaging device |
US11523060B2 (en) * | 2018-11-29 | 2022-12-06 | Ricoh Company, Ltd. | Display device, imaging device, object moving method, and recording medium |
JP2021180379A (en) | 2020-05-12 | 2021-11-18 | キヤノン株式会社 | Imaging device and control method for imaging device |
Family Cites Families (20)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP4391637B2 (en) * | 1999-10-04 | 2009-12-24 | オリンパス株式会社 | camera |
JP2004101632A (en) * | 2002-09-05 | 2004-04-02 | Canon Inc | Camera equipped with automatic focusing device |
JP3949067B2 (en) * | 2003-02-14 | 2007-07-25 | 三洋電機株式会社 | Auto focus camera |
DE602004032488D1 (en) * | 2003-02-18 | 2011-06-16 | Canon Kk | Photographing device with means for receiving radio information and control method therefor |
US20040207743A1 (en) * | 2003-04-15 | 2004-10-21 | Nikon Corporation | Digital camera system |
US7466356B2 (en) * | 2003-07-25 | 2008-12-16 | Hewlett-Packard Development Company, L.P. | Method and apparatus for setting a marker on an object and tracking the position of the object |
JP3918788B2 (en) * | 2003-08-06 | 2007-05-23 | コニカミノルタフォトイメージング株式会社 | Imaging apparatus and program |
JP4352916B2 (en) * | 2004-02-04 | 2009-10-28 | ソニー株式会社 | Imaging apparatus and imaging method |
JP4130641B2 (en) * | 2004-03-31 | 2008-08-06 | 富士フイルム株式会社 | Digital still camera and control method thereof |
JP2005347885A (en) * | 2004-05-31 | 2005-12-15 | Casio Comput Co Ltd | Imaging unit, photographing range adjustment method, and program |
JP2005341449A (en) * | 2004-05-31 | 2005-12-08 | Toshiba Corp | Digital still camera |
JP2006058405A (en) * | 2004-08-18 | 2006-03-02 | Casio Comput Co Ltd | Camera apparatus and automatic focusing control method |
JP2006098676A (en) * | 2004-09-29 | 2006-04-13 | Nikon Corp | Autofocus camera |
KR101120027B1 (en) * | 2004-11-01 | 2012-03-23 | 삼성전자주식회사 | Apparatus and method for processing on screen display when shutter is pressed half state in digital camera |
WO2006082967A1 (en) * | 2005-02-07 | 2006-08-10 | Matsushita Electric Industrial Co., Ltd. | Imaging device |
JP2006254413A (en) * | 2005-02-08 | 2006-09-21 | Nikon Corp | Imaging apparatus and camera body |
US7945938B2 (en) * | 2005-05-11 | 2011-05-17 | Canon Kabushiki Kaisha | Network camera system and control method therefore |
JP4572815B2 (en) * | 2005-11-18 | 2010-11-04 | 富士フイルム株式会社 | Imaging apparatus and imaging method |
JP5045125B2 (en) * | 2006-03-15 | 2012-10-10 | 株式会社ニコン | Subject tracking device and optical device |
US20080278589A1 (en) * | 2007-05-11 | 2008-11-13 | Karl Ola Thorn | Methods for identifying a target subject to automatically focus a digital camera and related systems, and computer program products |
-
2006
- 2006-06-30 JP JP2006181470A patent/JP4761146B2/en not_active Expired - Fee Related
-
2007
- 2007-06-28 US US11/770,114 patent/US8284256B2/en active Active
- 2007-06-28 TW TW096123421A patent/TWI341684B/en not_active IP Right Cessation
- 2007-06-29 KR KR1020070065092A patent/KR100847614B1/en active IP Right Grant
- 2007-07-02 CN CN200710127317A patent/CN100591099C/en not_active Expired - Fee Related
-
2008
- 2008-05-23 HK HK08105713.7A patent/HK1111032A1/en not_active IP Right Cessation
-
2012
- 2012-09-11 US US13/610,277 patent/US20130002922A1/en not_active Abandoned
Non-Patent Citations (1)
Title |
---|
English Machine Translation of Toyoichi et al. (JP 2005341449 A) * |
Cited By (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20130044251A1 (en) * | 2008-02-04 | 2013-02-21 | Samsung Electronics Co., Ltd. | Digital image processing apparatus and method of controlling the same |
US9094599B2 (en) * | 2008-02-04 | 2015-07-28 | Samsung Electronics Co., Ltd. | Digital image processing apparatus and method of controlling the same |
US20150271387A1 (en) * | 2008-02-04 | 2015-09-24 | Samsung Electronics Co., Ltd. | Digital image processing apparatus and method of controlling the same |
US9374520B2 (en) * | 2008-02-04 | 2016-06-21 | Samsung Electronics Co., Ltd. | Digital image processing apparatus and method of controlling the same |
US20190297265A1 (en) * | 2018-03-21 | 2019-09-26 | Sawah Innovations Inc. | User-feedback video stabilization device and method |
Also Published As
Publication number | Publication date |
---|---|
US20080002028A1 (en) | 2008-01-03 |
TW200808044A (en) | 2008-02-01 |
HK1111032A1 (en) | 2008-07-25 |
KR20080002667A (en) | 2008-01-04 |
JP2008009263A (en) | 2008-01-17 |
KR100847614B1 (en) | 2008-07-21 |
CN101098403A (en) | 2008-01-02 |
CN100591099C (en) | 2010-02-17 |
US8284256B2 (en) | 2012-10-09 |
TWI341684B (en) | 2011-05-01 |
JP4761146B2 (en) | 2011-08-31 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US8284256B2 (en) | Imaging apparatus and computer readable recording medium | |
JP4919160B2 (en) | Imaging apparatus and program thereof | |
TWI326553B (en) | Imaging apparatus, data extraction method, and data extraction program | |
JP4943721B2 (en) | Color noise removal method for image data and imaging apparatus using the method | |
JP5782813B2 (en) | Imaging apparatus and image display method | |
US10334336B2 (en) | Method of controlling digital photographing apparatus and digital photographing apparatus using the same | |
TWI492618B (en) | Image pickup device and computer readable recording medium | |
JP2013098792A (en) | Imaging device and control method for imaging device | |
JP2003262786A (en) | Imaging device and its automatic focusing method | |
JP4740074B2 (en) | Imaging apparatus and imaging method | |
JP2005072963A (en) | Imaging device and its control method | |
JP4948011B2 (en) | Imaging apparatus, control method therefor, computer program, and storage medium | |
JP2005278003A (en) | Image processing apparatus | |
JP2006094200A (en) | Imaging apparatus, focusing display method, and program | |
WO2022181087A1 (en) | Image processing device and method, and program | |
JP2009010746A (en) | Image recording and reproducing apparatus | |
JP4398346B2 (en) | Camera, photographing sensitivity control method, shutter speed control method, and photographing sensitivity control program | |
JP2010028273A (en) | Imaging apparatus, display control method, and program | |
JP2008085688A (en) | Image pickup device, method and program | |
JP2007166447A (en) | Imaging apparatus, zoom display method, and program | |
JP2002281374A (en) | Electronic camera | |
JP2005215206A (en) | Imaging unit, imaging method, program, and storage medium | |
JP2006253939A (en) | Imaging apparatus | |
JP2000324433A (en) | Electronic camera |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
STCB | Information on status: application discontinuation |
Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION |