US9338436B2 - Imaging device and imaging method - Google Patents

Imaging device and imaging method Download PDF

Info

Publication number
US9338436B2
US9338436B2 US13/486,456 US201213486456A US9338436B2 US 9338436 B2 US9338436 B2 US 9338436B2 US 201213486456 A US201213486456 A US 201213486456A US 9338436 B2 US9338436 B2 US 9338436B2
Authority
US
United States
Prior art keywords
imaging
data
timing
imaging device
unit
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Expired - Fee Related, expires
Application number
US13/486,456
Other languages
English (en)
Other versions
US20120314101A1 (en
Inventor
Yuuji OOBA
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Sony Corp
Original Assignee
Sony Corp
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Sony Corp filed Critical Sony Corp
Assigned to SONY CORPORATION reassignment SONY CORPORATION ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: OOBA, YUUJI
Publication of US20120314101A1 publication Critical patent/US20120314101A1/en
Priority to US15/076,893 priority Critical patent/US10045009B2/en
Application granted granted Critical
Publication of US9338436B2 publication Critical patent/US9338436B2/en
Priority to US16/021,381 priority patent/US10194141B2/en
Priority to US16/228,888 priority patent/US10595009B2/en
Expired - Fee Related legal-status Critical Current
Adjusted expiration legal-status Critical

Links

Images

Classifications

    • H04N13/0239
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/20Image signal generators
    • H04N13/296Synchronisation thereof; Control thereof
    • H04N13/0296
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/20Image signal generators
    • H04N13/204Image signal generators using stereoscopic image cameras
    • H04N13/239Image signal generators using stereoscopic image cameras using two 2D image sensors having a relative position equal to or related to the interocular distance
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/45Cameras or camera modules comprising electronic image sensors; Control thereof for generating image signals from two or more image sensors being of different type or operating in different modes, e.g. with a CMOS sensor for moving images in combination with a charge-coupled device [CCD] for still images
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/60Control of cameras or camera modules
    • H04N23/667Camera operation mode switching, e.g. between still and video, sport and normal or high- and low-resolution modes
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/70Circuitry for compensating brightness variation in the scene
    • H04N23/73Circuitry for compensating brightness variation in the scene by influencing the exposure time
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N5/00Details of television systems
    • H04N5/04Synchronising
    • H04N5/06Generation of synchronising signals
    • H04N5/067Arrangements or circuits at the transmitter end
    • H04N5/073Arrangements or circuits at the transmitter end for mutually locking plural sources of synchronising signals, e.g. studios or relay stations
    • H04N5/0733Arrangements or circuits at the transmitter end for mutually locking plural sources of synchronising signals, e.g. studios or relay stations for distributing synchronisation pulses to different TV cameras
    • H04N5/2258

Definitions

  • the present disclosure relates to an imaging device and an imaging method, and more particularly, to a technique used upon taking a stereoscopic image ( 3 D image).
  • an imaging device such as a video camera
  • two imaging devices are used. Then, an image for left channel and an image for right channel are individually taken by the respective imaging devices.
  • the two imaging devices are fixed by a coupling mechanism called a rig, and an image is taken so that optical axes of a lens unit attached to each imaging device become parallel to each other.
  • an image signal output from each imaging device is in synchronizing timing.
  • a reference synchronizing signal from a reference synchronizing signal generator is supplied to each imaging device, and synchronizing timing of an image signal (video signal) output from each imaging device agrees with each other.
  • Japanese Patent Laid-Open No. 10-210506 describes an example of a configuration in which a stereoscopic image is taken with the use of two imaging devices.
  • a stereoscopic image to be created will be an image taken at different timing between left channel and right channel. Therefore, when a high speed scene is taken, for example, a rough image is created.
  • a first imaging device and a second imaging device are connected with each other in a communicatable manner, and data of imaging timing and data of an exposure period of the first imaging device are transmitted to the second imaging device.
  • the second imaging device can set imaging timing and an exposure period based on the received data of the imaging timing and data of the exposure period. Therefore, an image is taken with the imaging timing and the exposure period of the first imaging device agreeing with those of the second imaging device.
  • an image can be taken with the imaging timing and the exposure period of one imaging device agreeing with those of the other imaging device by communicating the imaging timing and the exposure period between two imaging devices; therefore, an appropriate image signal as a stereoscopic image signal is obtained.
  • a synchronized imaging can be taken even when a variable speed imaging, in which a frame period for imaging is changed, is performed.
  • FIG. 1 is a diagram illustrating a configuration of an imaging device according to an embodiment of the present disclosure
  • FIG. 2 is a flow chart showing an example of a processing by a master imaging device according to an embodiment of the present disclosure
  • FIG. 3 is a flow chart showing an example of a processing by a slave imaging device according to an embodiment of the present disclosure
  • FIGS. 4A to 4C are explanatory diagrams showing an example of an imaging state (example 1 showing an imaging at a certain state) according to an embodiment of the present disclosure
  • FIGS. 5A to 5C are explanatory diagrams showing an example of an imaging state (example 2 showing an imaging at a certain state) according to an embodiment of the present disclosure
  • FIGS. 6A to 6C are explanatory diagrams showing an example of an imaging state (example 3 showing an imaging at a certain state) according to an embodiment of the present disclosure
  • FIGS. 7A to 7C are explanatory diagrams showing an example of an imaging state (example of a variable imaging) according to an embodiment of the present disclosure
  • FIG. 8 is a diagram illustrating an example of a system that uses an imaging device according to an embodiment of the present disclosure
  • FIGS. 9A to 9E are timing diagrams showing an example of analog shutter pulse according to an embodiment of the present disclosure.
  • FIG. 10 is a flow chart showing an example of a processing when analog shutter pulse is input according to an embodiment of the present disclosure.
  • FIG. 1 Configuration of imaging device ( FIG. 1 ); 2. Example of processing by master imaging device ( FIG. 2 ); 3. Example of processing by slave imaging device ( FIG. 3 ); 4. Example of imaging state ( FIGS. 4, 5, 6, and 7 ); 5. Example when analog shutter pulse is input ( FIGS. 8, 9, and 10 ); and 6. Modification.
  • FIG. 1 A configuration of an imaging device according to an embodiment of the present disclosure (hereinafter referred to as “the embodiment”) will be described with reference to FIG. 1 .
  • FIG. 1 is a diagram illustrating a configuration of an imaging device according to the embodiment.
  • An imaging device 100 has an image sensor and an imaging unit 103 including a circuit for reading out an image signal from the image sensor and the like.
  • the image sensor included in the imaging unit 103 converts image light incident through a lens 201 of a lens unit 200 attached to the imaging device 100 into an electrical image signal.
  • Imaging timing and an imaging period at the imaging unit 103 are controlled by an imaging control unit 113 .
  • An instruction from an operating unit 114 is supplied to the imaging control unit 113 .
  • An operation of the operating unit 114 by a user sets shutter speed and a frame period, and the imaging control unit 113 controls each unit so that an image is taken with the shutter speed and the frame period being set.
  • the setting based on the operation of the operating unit 114 and the like is stored in a memory 115 .
  • a shutter blade 101 is arranged in front of the image sensor included in the imaging unit 103 .
  • the shutter blade 101 is a rotating plate that includes an opening 101 a and a light shielding portion 101 b , and is rotationally driven by a shutter driving unit 102 .
  • a shutter driving unit 102 By rotating the shutter blade 101 , there alternately occurs a state in which the opening 101 a is located in front of the image sensor and image light is incident and a state in which the light shielding portion 101 b is located in front of the image sensor and the image light is shielded. In other words, opening and closing of a shutter are repeated.
  • Rotational driving of the shutter blade 101 by the shutter driving unit 102 is controlled by the imaging control unit 113 .
  • the shutter driving unit 102 supplies a detection signal of a rotation phase of the shutter blade 101 detected by a shutter rotation detection sensor 102 a to the imaging control unit 113 .
  • An image signal output from the imaging unit 103 is supplied to a signal processing unit 104 .
  • the signal processing unit 104 performs various types of signal processings on the supplied image signal and converts the image signal into image data in a prescribed format for output.
  • the image data in the prescribed format obtained at the signal processing unit 104 is supplied to a mixing unit 106 .
  • a memory 105 is connected to the signal processing unit 104 .
  • the signal processing unit 104 performs a temporary storage processing necessary for an image data processing and performs a correction processing on output timing of the image data.
  • the image signal output from the imaging unit 103 may be directly supplied to the mixing unit 106 as the image data for output without performing a format conversion of the image data at the signal processing unit 104 .
  • the mixing unit 106 performs a processing to add synchronizing data generated at an external synchronization generator 107 to the supplied image data.
  • An audio signal input to an audio signal input terminal 192 is delayed at an audio delaying unit 108 to cause timing of the audio signal to agree with that of the image data, and the audio signal is supplied to the mixing unit 106 to be added to the image data.
  • the mixing unit 106 supplies the image data to which the synchronizing data and audio data are added to a video output terminal 191 , and the image data is output from the video output terminal 191 to outside.
  • the image data (video data) to be output from the video output terminal 191 may be a digital signal or an analog signal; however, a digital signal is used for the following explanation.
  • the external synchronization generator 107 generates the synchronizing data, which is output from a synchronizing signal output terminal 193 to outside.
  • the synchronizing data generated at the external synchronization generator 107 is synchronizing data that is in synchronization with the image data output from the video output terminal 191 .
  • Data designating imaging timing and exposure timing is added to the synchronizing data output from the synchronizing signal output terminal 193 .
  • the imaging timing is timing at which imaging is started at the imaging unit 103 .
  • An exposure period is data of a period in which light is exposed by the shutter blade 101 , and for example, is data indicating an open angle of the shutter blade 101 , which is an angle of the opening 101 a . Details on a state of addition of the data designating the imaging timing and the exposure period will be described later.
  • the imaging device 100 includes a synchronizing signal input terminal 194 , and when the synchronizing data is input to the synchronizing signal input terminal 194 , the obtained synchronizing data is supplied to an external synchronization extracting unit 111 and an imaging synchronization extracting unit 112 .
  • the external synchronization extracting unit 111 detects synchronizing timing such as a vertical blanking period and a horizontal blanking period from the supplied synchronizing data. Then, the data indicating the timing detected by the external synchronization extracting unit 111 is supplied to the signal processing unit 104 .
  • the signal processing unit 104 When the data of the synchronizing timing is supplied from the external synchronization extracting unit 111 , the signal processing unit 104 outputs the image data from the video output terminal 191 in synchronization with the timing.
  • the imaging control unit 113 controls the imaging unit 103 so that an image is taken in accordance with the supplied imaging timing.
  • the imaging control unit 113 controls the shutter driving unit 102 so that the shutter blade 101 is opened and closed in association with the data of the exposure period that is supplied.
  • the imaging device 100 includes an analog shutter pulse input terminal 195 .
  • Analog shutter pulse obtained at the input terminal 195 has analog voltage waveform generated at another imaging device or an external shutter pulse generator.
  • the analog shutter pulse obtained at the input terminal 195 is supplied to a pulse generation unit 116 .
  • the pulse generation unit 116 generates driving pulse that is in synchronization with the supplied analog shutter pulse, and supplies the generated driving pulse to the imaging control unit 113 .
  • the analog shutter pulse obtained at the input terminal 195 is supplied to a counter 117 , a count processing in synchronization with the analog shutter pulse is performed, and the count result is supplied to the imaging control unit 113 .
  • the counter 117 performs a count operation for detecting a frame period of the analog shutter pulse and a count operation for detecting an exposure period (open angle of shutter) of the analog shutter pulse. Processings when the analog shutter pulse is input to the input terminal 195 will be described later with reference to FIG. 6 and the subsequent drawings.
  • the two imaging devices 100 shown in FIG. 1 it is possible to take a stereoscopic image. That is, when a stereoscopic image is taken, two imaging devices are fixed by a coupling mechanism (not shown) such as a rig, and an image is taken so that optical axes of a lens unit 200 attached to each imaging device become parallel to each other. An image for left channel is taken by one imaging device and an image for right channel is taken by the other imaging device.
  • the synchronizing signal output terminal 193 of one imaging device and the synchronizing signal input terminal 194 of the other imaging device are connected to each other via a cable.
  • the synchronizing data and the like are transmitted between the two imaging devices via the connected cable to execute a synchronized imaging.
  • the imaging device that outputs the synchronizing data from the synchronizing signal output terminal 193 is assumed to be a master imaging device
  • the imaging device in which the synchronizing data is input from the synchronizing signal input terminal 194 is assumed to be a slave imaging device.
  • the master imaging device is the imaging device that outputs the synchronizing data from the synchronizing signal output terminal 193 .
  • Step S 11 it is determined whether the imaging device is a master imaging device or a slave imaging device.
  • Step S 12 it is determined that the imaging device is a slave imaging device. The processings in the case where the imaging device is a slave imaging device will be described later with reference to FIG. 3 .
  • Step S 11 When it is determined in Step S 11 that the imaging device is a master imaging device, it is determined whether there has been a change in a setting of an imaging state (Step S 13 ). When it is determined that there has been no change in the setting of the imaging state, a processing is performed to add data of current imaging frame frequency, number of delayed frames, and open angle of a shutter to the vertical blanking period of the synchronizing data generated at the external synchronization generator 107 (Step S 14 ).
  • the image data output from the imaging device 100 is digital video data of SDI standard and the synchronizing data output from the synchronizing signal output terminal 193 is also the synchronizing data of the digital video data of SDI standard.
  • each data described above is added to auxiliary data of the vertical blanking period specified by SDI standard.
  • the vertical blanking period of the synchronizing data is data specifying a frame period or a field period of output image data.
  • the imaging frame frequency added to the synchronizing data in the processing of Step S 14 is frame frequency upon taking an image by the imaging unit 103 and not frame frequency of the image data output from the video output terminal 191 .
  • the frame frequency of the image data output from the video output terminal 191 is fixed at certain frequency dependent on an output format.
  • the number of delayed frames indicates the number of delayed frames between taking an image by the imaging unit 103 and output of the image data from the video output terminal 191 .
  • the open angle of the shutter is an angle of the opening 101 a of the shutter blade 101 .
  • the data is data indicating the setting of the current imaging state at the imaging control unit 113 .
  • Step S 14 After a processing for adding the data to the synchronizing data of the vertical blanking period is performed in Step S 14 , it is determined whether it has come to timing to actually take an image by the imaging unit 103 (Step S 15 ), and wait until the timing to actually take an image has come.
  • the timing to actually take an image by the imaging unit 103 is referred to as actual imaging timing.
  • Step S 15 When it is determined in Step S 15 that the actual imaging timing has come, data indicating that the actual imaging timing has come is added to the auxiliary data of the horizontal blanking period of the timing that agrees with the actual imaging timing (Step S 16 ).
  • the horizontal blanking period of the synchronizing data is data specifying a period of a horizontal line of the image data.
  • Step S 16 After the data of the actual imaging timing is added in Step S 16 , it returns to the determination of Step S 13 .
  • Step S 13 When it is determined in Step S 13 that there has been a change in the setting of the imaging state, data of changed imaging frame frequency, number of delayed frames, and open angle of the shutter is added to the vertical blanking period of the synchronizing data generated at the external synchronization generator 107 (Step S 17 ).
  • the change of the setting means a change in a user's operation of the operating unit 114 or a change of the analog shutter pulse input to the analog shutter pulse input terminal 195 .
  • Step S 17 After a processing to add is performed in Step S 17 , it goes to the determination of Step S 15 .
  • the actual imaging timing determined in Step S 15 after the processing of Step S 17 is the actual imaging timing after change.
  • Step S 18 it is determined whether the number of delayed frames added to the synchronizing data has passed, and wait until the corresponding delayed frames have passed. For example, if the number of delayed frame is 1, wait for 1 frame after change.
  • the imaging control unit 113 changes the setting for taking an image for own imaging device to a state indicated by the data added in Step S 17 (Step S 19 ). In other words, the imaging frame frequency and the open angle of the shutter are changed to those after change.
  • the slave imaging device is the imaging device in which the synchronizing data is input to the synchronizing signal input terminal 194 .
  • Step S 31 it is determined whether the imaging device is a master imaging device or a slave imaging device.
  • Step S 32 it goes to the processings shown in the flow chart of FIG. 2 described earlier.
  • Step S 33 it is determined whether auxiliary data of the vertical blanking period contained in the synchronizing data input to the synchronizing signal input terminal 194 has been received.
  • the imaging control unit 113 sets the obtained frame frequency and open angle of the shutter to an image control state of own imaging device.
  • Step S 35 it is determined whether data of the actual imaging timing added to the horizontal blanking period has been received (Step S 35 ), and wait until the data of the actual imaging timing is received.
  • Step S 35 timing to take an image by the image sensor included in the imaging unit 103 is set so as to start taking an image at the timing, and the rotation phase of the shutter blade 101 is made to be a synchronized phase (Step S 36 ).
  • Step S 36 After the imaging timing of the imaging unit 103 and the rotation phase of the shutter blade 101 are set, it returns to the determination of Step S 33 .
  • the processings shown in the flow chart of FIG. 2 are performed in the master imaging device and the processings shown in the flow chart of FIG. 3 are performed in the slave imaging device. Accordingly, an image is taken at the synchronized timing between the master imaging device and the slave imaging device, and the rotation phase of the shutter blade 101 is also synchronized. Therefore, the image data output from the video output terminal 191 of the two imaging devices is the one taken at the same timing, and the rotation phase of the shutter blade 101 and the open angle of the master imaging device agree with those of the slave imaging device, which means that the image has been taken under the same condition.
  • FIGS. 4A to 4C an example in which an image is taken in a state where the number of frames upon taking an image is not changed is shown in FIGS. 4A to 4C .
  • the exposure period shutter open angle
  • FIG. 4A shows the imaging timing and the exposure period at the image sensor included in the imaging unit 103 .
  • FIG. 4B shows a read out period of a signal from the image sensor.
  • FIG. 4C shows the image data output from the video output terminal 191 and the synchronizing data output from the synchronizing signal output terminal 193 .
  • the image data output from the video output terminal 191 and the synchronizing data output from the synchronizing signal output terminal 193 are the data with the same synchronizing timing.
  • FIGS. 4A to 4C show an example of an imaging at 30 P, that is, an example of a progressive imaging in which an image is taken by frame at 30 Hz of frame frequency.
  • the signal is read out from the image sensor at 30 P, which is the same period as the exposure period.
  • the image data output from the video output terminal 191 performs a conversion transmission, converting the image of 1 frame into the image of 2 fields. That is, an odd number field image ODO including 1 ⁇ 2 of a horizontal line and an even number field image EVEN including the remaining 1 ⁇ 2 of the horizontal line are generated from the image of 1 frame. Then, images in each field are sequentially output at 60 Hz of frequency that is two times of 30 Hz of frame frequency. The odd number field image ODO starts to be output when the signal from the image sensor of FIG. 4B has been read out at least for 1 ⁇ 2 frame.
  • the synchronizing data is added to the image data output from the video output terminal 191 .
  • a low level period shown in FIG. 4C is the data of the vertical blanking period.
  • the synchronizing data output from the synchronizing signal output terminal 193 is also the synchronizing data with the same timing.
  • Data D 1 is frame frequency, and shows 30 P in this example.
  • Data D 1 is data designating the frame frequency, but may be data designating a frame period.
  • Data D 2 is the number of the delayed frames, and shows 1 frame delay in this example.
  • Data D 3 is the shutter open angle, and shows 360° in this example. 360° of shutter open angle means that the shutter is completely open.
  • the frame frequency D 1 , the number of the delayed frames D 2 , and the shutter open angle D 3 are added to the synchronizing data of the vertical blanking period of each frame (field) to be transmitted.
  • Data D 4 shown in FIG. 4C is data showing the imaging timing, and as shown in FIGS. 4A and 4B , a processing for reading out the image signal of 1 frame accumulated in the image sensor is started at the imaging timing.
  • the data D 4 of the imaging timing is added to the horizontal blanking period; however, the horizontal blanking period is a period for each 1 horizontal line, and FIG. 4C only shows the timing in which the data is added and does not show the data of the horizontal blanking period.
  • the number of the delayed frame 1 shown as the data D 2 shows the timing at which a state designated by the data D 1 and the data D 3 is set, and shows that the setting is made after 1 frame.
  • FIG. 5 shows an example in which the speed of reading out from the image sensor is set to be 60 Hz, which is two times the speed in the example of FIG. 4 .
  • An imaging at 30 P that is, a progressive imaging at 30 Hz of frame frequency, is performed as shown in FIG. 5A , and reading out of the signal from the image sensor is performed at 60 P (60 Hz), which is half of the exposure period, as shown in FIG. 5B .
  • Segments with diagonal lines in FIG. 5B are periods where reading out is not performed.
  • the image data that has been subjected to field conversion output from the video output terminal 191 sequentially outputs images in each field at 60 Hz of frequency, which is two times of 30 Hz of frame frequency.
  • output of the odd number field image ODO can be started immediately after reading out of the signal from the image sensor shown in FIG. 5B is started. Therefore, as compared with the example of FIG. 4 , output is performed earlier by about 1 field period.
  • the exposure period (shutter open angle) upon taking an image is set to be 180° as compared with the example of FIG. 5 . That is, the 180° period immediately before the timing at which an imaging is indicated by the imaging timing D 4 is an opening period of the shutter, and the shutter is closed for the remaining period.
  • reading out is performed at 60 Hz of frequency at the timing indicated by the imaging timing D 4 .
  • the period in which the shutter is closed and the read out period agree with each other.
  • the image data that has been subjected to field conversion output from the video output terminal 191 sequentially outputs images in each field at 60 Hz of frequency as shown in FIG. 6C .
  • FIGS. 4 to 6 shows a state where there is no change in the imaging state.
  • the state is as shown in FIG. 7 , for example.
  • FIG. 7 shows a state, as an initial state, in which an image is taken at 30 P (30 Hz) of frame frequency with the shutter open angle being 180′. From this state, the frame frequency is changed from 30 P to 34 P, 40 P, 34 P, and then to 30 Hz.
  • the data of the frame frequency D 1 added to the synchronizing data shown in FIG. 7C transmits the setting (change) of the frame frequency described above.
  • the frame delay D 2 is 1 frame
  • the actual exposure period shown in FIG. 7B is set to the corresponding frequency (period) 1 frame after an instruction by the data D 1 added to the synchronizing data of FIG. 7C .
  • the shutter open angle is set to be 180° when an image is taken at 30 P of frame frequency
  • the shutter open angle is set to be 240° when an image is taken at 34 P of frame frequency
  • the shutter open angle is set to be 204° when an image is taken at 40 P of frame frequency.
  • the image data that has been subjected to field conversion is output at 60 Hz of constant frequency. Therefore, the horizontal line to which the data D 4 designating the imaging timing is inserted varies for each imaging frame, and the phase between an imaging and output of the image data also changes in association with the change of the imaging frequency.
  • the master imaging device and the slave imaging device synchronize with each other, and an image can be taken with the shutter phase and the imaging timing of the master imaging device agreeing with those of the slave imaging device.
  • FIG. 8 is a diagram illustrating an example of a system configuration in which analog shutter pulse is used.
  • two imaging devices that is, an imaging device for left channel 100 L and an imaging device for right channel 100 R are used.
  • the imaging devices 100 L and 100 R for each channel are the same as the imaging device 100 shown in FIG. 1 .
  • a synchronizing signal output terminal 193 of the imaging device for left channel 100 L and a synchronizing signal input terminal 194 of the imaging device for right channel 100 R are connected via a cable 92 .
  • the imaging device for left channel 100 L will be master and the imaging device for right channel 100 R will be slave as explained earlier; therefore, an image can be taken with the shutter phase and the imaging timing of the imaging device 100 L agreeing with those of the imaging device 100 R.
  • image data for left channel and image data for right channel are output from cables 91 L and 91 R connected to video output terminals 191 of the two imaging devices 100 L and 100 R.
  • a synchronizing pulse generator 81 is used, and analog shutter pulse is supplied from the synchronizing pulse generator 81 to an analog shutter pulse input terminal 195 of the imaging device for left channel 100 L via a cable 93 .
  • the synchronizing pulse generator 81 generates analog shutter pulse that is in synchronization with a connected apparatus.
  • a lighting device 82 is connected to the synchronizing pulse generator 81 via a cable 94
  • a film camera 83 is connected to the synchronizing pulse generator 81 via a cable 95 .
  • the synchronizing pulse generator 81 generates analog shutter pulse that is in synchronization with a lighting period of light of the lighting device 82 and a filming period of a movie film at the film camera 83 .
  • an image is taken by the two imaging devices 100 L and 1008 at timing in association with the lighting period of the light of the lighting device 82 and the filming period at the film camera 83 .
  • FIG. 9 is a diagram showing an example of analog shutter pulse.
  • the analog shutter pulse shown in FIG. 9A has pulse waveform in which changes of high level and low level are repeated by 1 frame period, and the high level period indicates that the shutter is open and the low level period indicates that the shutter is closed.
  • the analog shutter pulse shown in FIG. 9B is an example in which high level and low level are inverted with respect to the analog shutter pulse shown in FIG. 9A , and the low level period indicates that the shutter is open and the high level period indicates that the shutter is closed.
  • Such analog shutter pulse is supplied from the analog shutter pulse input terminal 195 of the imaging device 100 to the pulse generation unit 116 ( FIG. 1 ); accordingly, driving pulse shown in FIG. 9C is generated.
  • the driving pulse shown in FIG. 9C is pulse that rises at timing where the shutter is opened from the closed state by the analog shutter pulse.
  • the counter 117 performs a count processing to detect 1 frame period, which is a period between a fall of the analog shutter pulse and the next fall of the analog shutter pulse, as shown in FIG. 9D . Further, the period where the analog shutter pulse is being risen is counted for each frame inside the imaging control unit 113 as shown in FIG. 9E . These count processings are counts with a horizontal period of the image data generated in the imaging device being a unit, for example.
  • Such driving pulse and the result of the count are supplied from the pulse generation unit 116 to the imaging control unit 113 ; accordingly, the imaging timing and the exposure timing set by the imaging control unit 113 become timing indicated by the analog shutter pulse. That is, an image is taken in a state where the timing in which the opening 101 a of the shutter blade 101 is located in front of the image sensor agrees with the timing of open indicated by the analog shutter pulse shown in FIG. 9A or FIG. 9B .
  • the data indicating the imaging timing or the exposure period added to the synchronizing data at the external synchronization generator 107 also becomes data indicating corresponding timing or period. Therefore, an image is taken in synchronization with the analog shutter pulse by the master imaging device 100 L to which the analog shutter pulse is input and the slave imaging device 100 R.
  • the analog shutter pulse is counted for the frame period at the pulse generation unit 116 in the master imaging device 100 L.
  • the result of the count is notified from the pulse generation unit 116 to the imaging control unit 113 , and it is determined whether the analog shutter pulse input by the imaging control unit 113 is adequate based on the notification.
  • the flow chart of FIG. 10 shows an example of a processing at the imaging control unit 113 when the analog shutter pulse is input.
  • Step S 41 When the analog shutter pulse is input, it is determined whether the current imaging mode of the imaging device is a variable mode or a fixed mode (Step S 41 ). If it is a fixed mode, a notice is indicated on a view finder or the like (Step S 42 ).
  • This imaging mode is preliminarily set by an operation and the like of the operating unit 114 of the imaging device 100 or the like, and the timing of imaging or exposure is not changed by pulse from outside in the fixed mode.
  • Step S 43 it is determined whether the input analog shutter pulse is positive pulse or negative pulse (Step S 43 ). Whether the input analog shutter pulse is positive pulse or negative pulse may be automatically determined by the phase of the input pulse, but which pulse to be input may be set by an operation of the operating unit 114 or the like.
  • the positive pulse is the pulse shown in FIG. 9A , for example, and the negative pulse is the pulse shown in FIG. 9B , for example.
  • Step S 43 When it is determined in Step S 43 that the input analog shutter pulse is the positive pulse, the driving pulse shown in FIG. 9C is generated at the pulse generation unit 116 with the phase maintained. When it is determined that the input analog shutter pulse is the negative pulse, the driving pulse is generated after the pulse phase is inverted (Step S 44 ). In addition, a count processing based on the analog shutter pulse is performed at the counter 117 (Step S 45 ).
  • the imaging control unit 113 determines the frame period indicated by the analog shutter pulse based on the result of the count of the frame period (Step S 46 ) and determines the shutter open angle indicated by the analog shutter pulse based on the result of the count of the shutter open period (Step S 47 ). As input of the analog shutter pulse continues, processings of Steps S 45 , S 46 , and S 47 are repeated.
  • the imaging control unit 113 determines whether the frame period and the shutter open angle have values within the appropriate range that can be set by the imaging device 100 . For example, the imaging control unit 113 determines whether the frame period is the period that can be set by the imaging device 100 and determines whether the designated shutter open angle can be set within 1 frame period. Based on the respective determinations, control is carried out so that an image is taken in a state where the setting is possible, and if the setting is not possible, it is switched to an imaging at the timing generated inside the imaging device.
  • the shutter pulse which has analog voltage waveform
  • the shutter pulse may slightly vary due to various causes.
  • the last setting is maintained.
  • it is continuously detected several times that the setting is above the range that can be set according to the result of the count of the shutter pulse it is preferable that it is switched to an imaging at the timing generated inside the imaging device.
  • Imaging control based on input of the analog shutter pulse makes it possible to carry out an imaging and an exposure at timing in association with the lighting period of the light of the lighting device 82 and the filming period at the film camera 83 as in the system configuration shown in FIG. 8 .
  • setting regarding an imaging can be changed by an operation of the operating unit 114 included in the imaging device 100 .
  • the setting may be changed by an instruction from a remote controller by connecting the remote controller to the master imaging device 100 , for example.
  • the data of imaging timing, the exposure period of the shutter, and the frame delay is added; however, only one or two of these may be added.
  • the imaging device including the shutter mechanism that has the shutter blade is shown as an example.
  • a control of an electronic shutter function may be carried out by a similar processing when the image sensor included in the imaging unit has the electronic shutter function.
  • the electronic shutter function is provided, a period in which a pixel arranged in the image sensor receives a signal will be the exposure period.
  • the analog shutter pulse is generated in the synchronizing pulse generator 81 located outside and synchronized with the external apparatus.
  • the synchronization may be achieved by similar analog shutter pulse transmission.
  • the synchronization can be achieved by a configuration in which the analog shutter pulse is generated in the imaging device 100 shown in FIG. 1 and the generated analog shutter pulse is output.
  • present technology may also be configured as below.
  • An imaging device comprising:
  • an imaging unit converting an image light incident through a lens to an electrical image signal in a set imaging timing and exposure period
  • a signal processing unit processing the image signal output from the imaging unit to obtain an image data for output
  • an imaging control unit controlling the imaging timing at the imaging unit and transmitting data of the imaging timing and data of the exposure period at the imaging unit to the other imaging device through the terminal portion for synchronization.
  • the imaging control unit adds data of a frame period or a frame frequency at which an image is taken to data output from the terminal portion for synchronization and takes an image at the imaging unit at the added frame period or frame frequency.
  • the imaging device according to (1) or (2),
  • the imaging control unit adds data of a delayed period until the imaging timing is applied to the data of the imaging timing transmitted from the terminal portion for synchronization and performs a setting of the imaging timing at the imaging unit after waiting for a period indicated by data of the period.
  • the imaging device according to any one of (1) to (3),
  • the data output from the terminal portion for synchronization is synchronizing data that is in synchronization with the image data output from the output portion and the data of the imaging timing and the data of the exposure period are added to the synchronizing data.
  • the imaging device according to any one of (1) to (4),
  • the imaging device according to any one of (1) to (5) comprising:
  • a shutter mechanism unit arranged in an optical path between the lens and the imaging unit
  • the data of the exposure period is data indicated by an open angle of the shutter mechanism unit.
  • the imaging device according to any one of (1) to (6) comprising:
  • the imaging control unit sets the imaging timing and the exposure period at the imaging unit based on the analog shutter pulse input to the input terminal for the analog shutter pulse.
  • the imaging control unit sets the imaging timing and the exposure period at the imaging unit based on received data when the data of the imaging timing and the data of the exposure period are received via the terminal portion for synchronization.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Human Computer Interaction (AREA)
  • Studio Devices (AREA)
  • Stereoscopic And Panoramic Photography (AREA)
  • Testing, Inspecting, Measuring Of Stereoscopic Televisions And Televisions (AREA)
  • Stroboscope Apparatuses (AREA)
US13/486,456 2011-06-07 2012-06-01 Imaging device and imaging method Expired - Fee Related US9338436B2 (en)

Priority Applications (3)

Application Number Priority Date Filing Date Title
US15/076,893 US10045009B2 (en) 2011-06-07 2016-03-22 Imaging device and imaging control method with adjustable frame frequency
US16/021,381 US10194141B2 (en) 2011-06-07 2018-06-28 Imaging device and imaging method
US16/228,888 US10595009B2 (en) 2011-06-07 2018-12-21 Imaging device and imaging method

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
JP2011-127414 2011-06-07
JP2011127414A JP5821300B2 (ja) 2011-06-07 2011-06-07 撮像装置及び撮像方法

Related Child Applications (1)

Application Number Title Priority Date Filing Date
US15/076,893 Continuation US10045009B2 (en) 2011-06-07 2016-03-22 Imaging device and imaging control method with adjustable frame frequency

Publications (2)

Publication Number Publication Date
US20120314101A1 US20120314101A1 (en) 2012-12-13
US9338436B2 true US9338436B2 (en) 2016-05-10

Family

ID=46466091

Family Applications (4)

Application Number Title Priority Date Filing Date
US13/486,456 Expired - Fee Related US9338436B2 (en) 2011-06-07 2012-06-01 Imaging device and imaging method
US15/076,893 Active 2032-06-19 US10045009B2 (en) 2011-06-07 2016-03-22 Imaging device and imaging control method with adjustable frame frequency
US16/021,381 Active US10194141B2 (en) 2011-06-07 2018-06-28 Imaging device and imaging method
US16/228,888 Active US10595009B2 (en) 2011-06-07 2018-12-21 Imaging device and imaging method

Family Applications After (3)

Application Number Title Priority Date Filing Date
US15/076,893 Active 2032-06-19 US10045009B2 (en) 2011-06-07 2016-03-22 Imaging device and imaging control method with adjustable frame frequency
US16/021,381 Active US10194141B2 (en) 2011-06-07 2018-06-28 Imaging device and imaging method
US16/228,888 Active US10595009B2 (en) 2011-06-07 2018-12-21 Imaging device and imaging method

Country Status (4)

Country Link
US (4) US9338436B2 (enrdf_load_stackoverflow)
EP (1) EP2533542B1 (enrdf_load_stackoverflow)
JP (1) JP5821300B2 (enrdf_load_stackoverflow)
CN (1) CN102821248A (enrdf_load_stackoverflow)

Families Citing this family (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP6056497B2 (ja) * 2013-01-18 2017-01-11 株式会社ソシオネクスト 画像処理装置、画像処理方法及びプログラム
US20170214903A1 (en) * 2014-03-13 2017-07-27 Sony Corporation Imaging apparatus, imaging system, and control method for imaging apparatus
CN104506779B (zh) * 2014-12-24 2018-11-13 浙江宇视科技有限公司 一种红绿灯色彩矫正方法和摄像设备
US10551154B2 (en) * 2017-01-20 2020-02-04 Vista Outdoor Operations Llc Rifle cartridge with improved bullet upset and separation
CN110326285B (zh) 2017-03-02 2021-06-25 索尼半导体解决方案公司 图像传感器和控制系统
WO2019010626A1 (zh) * 2017-07-11 2019-01-17 深圳市柔宇科技有限公司 拍摄方法和终端

Citations (37)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH10210506A (ja) 1997-01-22 1998-08-07 Sony Corp 3次元画像情報入力装置および3次元画像情報入出力装置
JP2000341719A (ja) 1999-05-25 2000-12-08 Mitsubishi Electric Corp ステレオカメラ
US20020118958A1 (en) * 2001-02-26 2002-08-29 Matsushita Electric Industrial Co., Ltd. Recording system, video camera device and video image recording method
US20020135682A1 (en) * 2001-02-13 2002-09-26 Hiroto Oka Image pickup system
US6493034B1 (en) 1999-08-23 2002-12-10 Elbex Video Ltd. Method and apparatus for remote adjustment of a time delay circuit of externally synchronized video transmitters
US20030011683A1 (en) * 2001-07-13 2003-01-16 Fumitomo Yamasaki Digital camera
US20030133018A1 (en) * 2002-01-16 2003-07-17 Ted Ziemkowski System for near-simultaneous capture of multiple camera images
US20030160870A1 (en) * 2002-02-22 2003-08-28 Ziemkowski Theodore B. Share link between image capturing devices
US6774935B1 (en) * 1998-10-26 2004-08-10 Minolta Co., Ltd. Digital camera
US20040183915A1 (en) * 2002-08-28 2004-09-23 Yukita Gotohda Method, device, and program for controlling imaging device
JP2005039707A (ja) 2003-07-18 2005-02-10 Sony Corp 撮像装置と同期信号発生装置
US20050077450A1 (en) 2003-10-09 2005-04-14 Baer Richard L. CMOS stereo imaging system & method
US20060065915A1 (en) * 2004-09-27 2006-03-30 Kazunobu Kuwazawa Sold-state imaging devices
US20060158526A1 (en) * 2004-12-21 2006-07-20 Kotaro Kashiwa Image editing apparatus, image pickup apparatus, image editing method, and program
US20060261282A1 (en) 2005-05-23 2006-11-23 Takatsugu Nakajima Imaging system, imaging controller, and method and program for vertical synchronization
US20070223901A1 (en) * 2006-03-27 2007-09-27 Fujifilm Corporation Photographic system, photographic device and photographic method
US20070283031A1 (en) * 2006-06-02 2007-12-06 Canon Kabushiki Kaisha Image pickup apparatus having communication function, method for controlling the same, and computer-readable storage medium
US20080232780A1 (en) * 2007-03-23 2008-09-25 Fujifilm Corporation Imaging system and imaging apparatus
US20090015689A1 (en) 2007-07-09 2009-01-15 Jin Murayama Multi-eye image pickup apparatus and adjusting method
US7511737B2 (en) * 2004-06-30 2009-03-31 Scenera Technologies, Llc Synchronized multi-perspective pictures
US20090135262A1 (en) * 2006-07-27 2009-05-28 Nikon Corporation External device, electronic flash device, and camera
JP2010028484A (ja) 2008-07-18 2010-02-04 Sony Corp 撮像装置
US20100134651A1 (en) * 2008-11-28 2010-06-03 Samsung Digital Imaging Co., Ltd. Photographing apparatus and method
US20100157020A1 (en) * 2008-12-22 2010-06-24 Electronics And Telecommunications Research Institute Multiple camera controlling and image storing apparatus for synchronized multiple image acquisition and method thereof
US20100289914A1 (en) * 2009-05-12 2010-11-18 Canon Kabushiki Kaisha Imaging apparatus and imaging method
JP2010277033A (ja) 2009-06-01 2010-12-09 Sony Corp 撮像装置と振れ補正方法
US20110205380A1 (en) * 2010-02-19 2011-08-25 Canon Kabushiki Kaisha Image sensing apparatus, communication apparatus, and control method of these apparatuses
US20110242342A1 (en) * 2010-04-05 2011-10-06 Qualcomm Incorporated Combining data from multiple image sensors
US20110242370A1 (en) * 2010-03-31 2011-10-06 Kazuo Endo Camera system, video processing apparatus, and camera apparatus
US20110267269A1 (en) * 2010-05-03 2011-11-03 Microsoft Corporation Heterogeneous image sensor synchronization
US20120113230A1 (en) * 2010-11-04 2012-05-10 Samsung Electronics Co., Ltd. Digital photographing apparatus and method of controlling the same
US20120154609A1 (en) * 2010-12-15 2012-06-21 Sony Corporation Image recording device, image recording method, and program
US20120162511A1 (en) * 2010-12-22 2012-06-28 Texas Instruments Incorporated Multi-sensor video frame synchronization apparatus and methods
US20120242805A1 (en) * 2011-03-24 2012-09-27 Tyou Syun Imaging device, synchronization control method, reproduction device, and stereoscopic video imaging system
US20120257022A1 (en) * 2011-04-07 2012-10-11 Hidekazu Suto Imaging apparatus and imaging method
US20120300124A1 (en) * 2011-05-26 2012-11-29 Shigeyuki Yamashita Signal transmission apparatus, signal transmission method, signal reception apparatus, signal reception method, and signal transmission system
US20130021450A1 (en) * 2011-07-22 2013-01-24 Yasuo Yoshizawa Stereoscopic imaging system, recording control method, stereoscopic image reproduction system, and reproduction control method

Family Cites Families (27)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP4350936B2 (ja) * 2002-09-30 2009-10-28 富士フイルム株式会社 固体撮像素子の信号読み出し方法
JP2004274219A (ja) * 2003-03-06 2004-09-30 Pioneer Electronic Corp 映像信号のフレームレート変換装置
JP3937332B2 (ja) * 2003-03-11 2007-06-27 ソニー株式会社 撮影システム
JP4458864B2 (ja) * 2004-01-30 2010-04-28 キヤノン株式会社 撮像装置及びその制御方法及びプログラム及び記憶媒体
JP4107269B2 (ja) * 2004-02-23 2008-06-25 ソニー株式会社 固体撮像装置
EP1727360B1 (en) * 2004-03-17 2013-08-28 Fujitsu Semiconductor Limited Automatic gain control circuit
JP5055686B2 (ja) * 2004-05-13 2012-10-24 ソニー株式会社 撮像システム、撮像装置、撮像方法
JP4215266B2 (ja) * 2006-10-30 2009-01-28 パナソニック株式会社 画像生成装置および画像生成方法
JP2008263262A (ja) * 2007-04-10 2008-10-30 Olympus Corp 撮像装置
WO2008128205A1 (en) * 2007-04-13 2008-10-23 Presler Ari M Digital cinema camera system for recording, editing and visualizing images
US7787943B2 (en) * 2007-07-25 2010-08-31 Mcdonough Daniel K Heart rate monitor for swimmers
US8587706B2 (en) * 2008-01-30 2013-11-19 Gentex Corporation Imaging device
JP2009147730A (ja) * 2007-12-14 2009-07-02 Fujifilm Corp 動画生成装置、動画撮影装置、動画生成方法及びプログラム
JP5336936B2 (ja) * 2009-06-08 2013-11-06 オリンパス株式会社 撮像装置及び顕微鏡システム
JP5282689B2 (ja) * 2009-07-23 2013-09-04 ソニー株式会社 撮像装置及び撮像方法
JP5342969B2 (ja) * 2009-09-10 2013-11-13 富士フイルム株式会社 撮像装置及び撮像方法
JP5482154B2 (ja) * 2009-12-02 2014-04-23 セイコーエプソン株式会社 撮像装置、撮像方法および撮像プログラム
JP5432799B2 (ja) * 2010-03-30 2014-03-05 オリンパスイメージング株式会社 撮像装置、撮像システム、撮像方法
JP5533292B2 (ja) * 2010-06-08 2014-06-25 ソニー株式会社 固体撮像素子およびその駆動方法、カメラシステム
CN102742278B (zh) * 2010-07-08 2015-08-12 松下电器产业株式会社 摄像装置
JP2012029223A (ja) * 2010-07-27 2012-02-09 Sanyo Electric Co Ltd イメージセンサおよび撮像装置
US8755785B2 (en) * 2011-03-31 2014-06-17 Intel Corporation Collaborative image control
JP5163728B2 (ja) * 2010-10-13 2013-03-13 セイコーエプソン株式会社 タイミングジェネレーター、撮影装置、ドットクロック出力方法
KR101781533B1 (ko) * 2010-12-23 2017-09-27 삼성디스플레이 주식회사 영상 촬영 장치 및 이의 영상 촬영 방법
EP2682799B1 (en) * 2011-03-02 2015-12-09 Panasonic Corporation Imaging device, semiconductor integrated circuit, and imaging method
JP6354221B2 (ja) * 2014-03-12 2018-07-11 株式会社リコー 撮像装置及び電子機器
US9749532B1 (en) * 2014-08-12 2017-08-29 Amazon Technologies, Inc. Pixel readout of a charge coupled device having a variable aperture

Patent Citations (40)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH10210506A (ja) 1997-01-22 1998-08-07 Sony Corp 3次元画像情報入力装置および3次元画像情報入出力装置
US6774935B1 (en) * 1998-10-26 2004-08-10 Minolta Co., Ltd. Digital camera
JP2000341719A (ja) 1999-05-25 2000-12-08 Mitsubishi Electric Corp ステレオカメラ
US6493034B1 (en) 1999-08-23 2002-12-10 Elbex Video Ltd. Method and apparatus for remote adjustment of a time delay circuit of externally synchronized video transmitters
US20020135682A1 (en) * 2001-02-13 2002-09-26 Hiroto Oka Image pickup system
US20020118958A1 (en) * 2001-02-26 2002-08-29 Matsushita Electric Industrial Co., Ltd. Recording system, video camera device and video image recording method
US20030011683A1 (en) * 2001-07-13 2003-01-16 Fumitomo Yamasaki Digital camera
US20030133018A1 (en) * 2002-01-16 2003-07-17 Ted Ziemkowski System for near-simultaneous capture of multiple camera images
US7046292B2 (en) * 2002-01-16 2006-05-16 Hewlett-Packard Development Company, L.P. System for near-simultaneous capture of multiple camera images
US20030160870A1 (en) * 2002-02-22 2003-08-28 Ziemkowski Theodore B. Share link between image capturing devices
US20040183915A1 (en) * 2002-08-28 2004-09-23 Yukita Gotohda Method, device, and program for controlling imaging device
JP2005039707A (ja) 2003-07-18 2005-02-10 Sony Corp 撮像装置と同期信号発生装置
US20090303337A1 (en) * 2003-07-18 2009-12-10 Katsumi Kaneko Image pick-up device and synchronization-signal-generating device
US20050077450A1 (en) 2003-10-09 2005-04-14 Baer Richard L. CMOS stereo imaging system & method
US7511737B2 (en) * 2004-06-30 2009-03-31 Scenera Technologies, Llc Synchronized multi-perspective pictures
US20060065915A1 (en) * 2004-09-27 2006-03-30 Kazunobu Kuwazawa Sold-state imaging devices
US20060158526A1 (en) * 2004-12-21 2006-07-20 Kotaro Kashiwa Image editing apparatus, image pickup apparatus, image editing method, and program
US20060261282A1 (en) 2005-05-23 2006-11-23 Takatsugu Nakajima Imaging system, imaging controller, and method and program for vertical synchronization
US20070223901A1 (en) * 2006-03-27 2007-09-27 Fujifilm Corporation Photographic system, photographic device and photographic method
US20070283031A1 (en) * 2006-06-02 2007-12-06 Canon Kabushiki Kaisha Image pickup apparatus having communication function, method for controlling the same, and computer-readable storage medium
US20090135262A1 (en) * 2006-07-27 2009-05-28 Nikon Corporation External device, electronic flash device, and camera
US20080232780A1 (en) * 2007-03-23 2008-09-25 Fujifilm Corporation Imaging system and imaging apparatus
US20090015689A1 (en) 2007-07-09 2009-01-15 Jin Murayama Multi-eye image pickup apparatus and adjusting method
JP2010028484A (ja) 2008-07-18 2010-02-04 Sony Corp 撮像装置
US20100134651A1 (en) * 2008-11-28 2010-06-03 Samsung Digital Imaging Co., Ltd. Photographing apparatus and method
US20100157020A1 (en) * 2008-12-22 2010-06-24 Electronics And Telecommunications Research Institute Multiple camera controlling and image storing apparatus for synchronized multiple image acquisition and method thereof
US20100289914A1 (en) * 2009-05-12 2010-11-18 Canon Kabushiki Kaisha Imaging apparatus and imaging method
US8456535B2 (en) * 2009-05-12 2013-06-04 Canon Kabushiki Kaisha Imaging apparatus and imaging method
JP2010277033A (ja) 2009-06-01 2010-12-09 Sony Corp 撮像装置と振れ補正方法
US20110205380A1 (en) * 2010-02-19 2011-08-25 Canon Kabushiki Kaisha Image sensing apparatus, communication apparatus, and control method of these apparatuses
US20110242370A1 (en) * 2010-03-31 2011-10-06 Kazuo Endo Camera system, video processing apparatus, and camera apparatus
US20110242342A1 (en) * 2010-04-05 2011-10-06 Qualcomm Incorporated Combining data from multiple image sensors
US20110267269A1 (en) * 2010-05-03 2011-11-03 Microsoft Corporation Heterogeneous image sensor synchronization
US20120113230A1 (en) * 2010-11-04 2012-05-10 Samsung Electronics Co., Ltd. Digital photographing apparatus and method of controlling the same
US20120154609A1 (en) * 2010-12-15 2012-06-21 Sony Corporation Image recording device, image recording method, and program
US20120162511A1 (en) * 2010-12-22 2012-06-28 Texas Instruments Incorporated Multi-sensor video frame synchronization apparatus and methods
US20120242805A1 (en) * 2011-03-24 2012-09-27 Tyou Syun Imaging device, synchronization control method, reproduction device, and stereoscopic video imaging system
US20120257022A1 (en) * 2011-04-07 2012-10-11 Hidekazu Suto Imaging apparatus and imaging method
US20120300124A1 (en) * 2011-05-26 2012-11-29 Shigeyuki Yamashita Signal transmission apparatus, signal transmission method, signal reception apparatus, signal reception method, and signal transmission system
US20130021450A1 (en) * 2011-07-22 2013-01-24 Yasuo Yoshizawa Stereoscopic imaging system, recording control method, stereoscopic image reproduction system, and reproduction control method

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
Office Action issued in corresponding Japanese application No. 2011-127414 dated Feb. 10, 2015.

Also Published As

Publication number Publication date
US20180324411A1 (en) 2018-11-08
US10045009B2 (en) 2018-08-07
US10595009B2 (en) 2020-03-17
EP2533542A2 (en) 2012-12-12
US20160227198A1 (en) 2016-08-04
JP5821300B2 (ja) 2015-11-24
JP2012255830A (ja) 2012-12-27
US20120314101A1 (en) 2012-12-13
CN102821248A (zh) 2012-12-12
US20190132580A1 (en) 2019-05-02
US10194141B2 (en) 2019-01-29
EP2533542A3 (en) 2015-07-01
EP2533542B1 (en) 2020-06-03

Similar Documents

Publication Publication Date Title
US10595009B2 (en) Imaging device and imaging method
TWI327438B (en) Timing signal processing apparatus for controlling driving of image-capturing element, and camera
US8633998B2 (en) Imaging apparatus and display apparatus
US9992446B2 (en) Image transmission system and image transmission method
JP2005039707A (ja) 撮像装置と同期信号発生装置
JP5103913B2 (ja) 撮像装置及び映像信号発生装置
JP5970748B2 (ja) 動画像撮影システム及び同期制御方法
US20190149702A1 (en) Imaging apparatus
JP2005311699A (ja) 撮像装置、および表示制御方法
GB2620471A (en) Control apparatus and control method of image capture system
JP2014146850A (ja) 撮像装置およびその制御方法
US20130058618A1 (en) Imaging device and imaging method
JP5273895B2 (ja) 撮像装置及びレンズ装置
JP2015108674A (ja) 撮像装置、レンズ装置、撮像装置の制御方法、及びレンズ装置の制御方法
JP4458132B2 (ja) 撮像装置
US20180152612A1 (en) Imaging device, control method thereof, and recording medium
US10212313B2 (en) Data processing device
US20240364852A1 (en) Image processing system, image processing apparatus, control method of image processing apparatus, and storage medium
US9307157B2 (en) Lens apparatus and imaging apparatus detachable therefrom
JP2018017756A (ja) 制御装置、撮像装置、レンズ装置、制御方法、プログラム、および、記憶媒体
Ye et al. Camera Capture and Frame-Rate Synchronization in a Multi-Camera System
JP2013165313A (ja) カメラ制御装置
JP2011234080A (ja) 映像表示制御装置及び映像監視装置
JP2001352488A (ja) デジタルカメラ及びデジタルカメラ調整システム
JP2006352768A (ja) 撮像モジュール及びデータ受信装置

Legal Events

Date Code Title Description
AS Assignment

Owner name: SONY CORPORATION, JAPAN

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:OOBA, YUUJI;REEL/FRAME:028304/0255

Effective date: 20120406

ZAAA Notice of allowance and fees due

Free format text: ORIGINAL CODE: NOA

ZAAB Notice of allowance mailed

Free format text: ORIGINAL CODE: MN/=.

STCF Information on status: patent grant

Free format text: PATENTED CASE

MAFP Maintenance fee payment

Free format text: PAYMENT OF MAINTENANCE FEE, 4TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1551); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

Year of fee payment: 4

FEPP Fee payment procedure

Free format text: MAINTENANCE FEE REMINDER MAILED (ORIGINAL EVENT CODE: REM.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

LAPS Lapse for failure to pay maintenance fees

Free format text: PATENT EXPIRED FOR FAILURE TO PAY MAINTENANCE FEES (ORIGINAL EVENT CODE: EXP.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

STCH Information on status: patent discontinuation

Free format text: PATENT EXPIRED DUE TO NONPAYMENT OF MAINTENANCE FEES UNDER 37 CFR 1.362

FP Lapsed due to failure to pay maintenance fee

Effective date: 20240510